| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433 | # SPDX-License-Identifier: AGPL-3.0-or-later# pylint: disable=global-statement# pylint: disable=missing-module-docstring, missing-class-docstringfrom __future__ import annotationsimport typingimport atexitimport asyncioimport ipaddressfrom itertools import cyclefrom typing import Dictimport httpxfrom searx import logger, sxng_debugfrom searx.extended_types import SXNG_Responsefrom .client import new_client, get_loop, AsyncHTTPTransportNoHttpfrom .raise_for_httperror import raise_for_httperrorlogger = logger.getChild('network')DEFAULT_NAME = '__DEFAULT__'NETWORKS: Dict[str, 'Network'] = {}# requests compatibility when reading proxy settings from settings.ymlPROXY_PATTERN_MAPPING = {    'http': 'http://',    'https': 'https://',    'socks4': 'socks4://',    'socks5': 'socks5://',    'socks5h': 'socks5h://',    'http:': 'http://',    'https:': 'https://',    'socks4:': 'socks4://',    'socks5:': 'socks5://',    'socks5h:': 'socks5h://',}ADDRESS_MAPPING = {'ipv4': '0.0.0.0', 'ipv6': '::'}class Network:    __slots__ = (        'enable_http',        'verify',        'enable_http2',        'max_connections',        'max_keepalive_connections',        'keepalive_expiry',        'local_addresses',        'proxies',        'using_tor_proxy',        'max_redirects',        'retries',        'retry_on_http_error',        '_local_addresses_cycle',        '_proxies_cycle',        '_clients',        '_logger',    )    _TOR_CHECK_RESULT = {}    def __init__(        # pylint: disable=too-many-arguments        self,        enable_http=True,        verify=True,        enable_http2=False,        max_connections=None,        max_keepalive_connections=None,        keepalive_expiry=None,        proxies=None,        using_tor_proxy=False,        local_addresses=None,        retries=0,        retry_on_http_error=None,        max_redirects=30,        logger_name=None,    ):        self.enable_http = enable_http        self.verify = verify        self.enable_http2 = enable_http2        self.max_connections = max_connections        self.max_keepalive_connections = max_keepalive_connections        self.keepalive_expiry = keepalive_expiry        self.proxies = proxies        self.using_tor_proxy = using_tor_proxy        self.local_addresses = local_addresses        self.retries = retries        self.retry_on_http_error = retry_on_http_error        self.max_redirects = max_redirects        self._local_addresses_cycle = self.get_ipaddress_cycle()        self._proxies_cycle = self.get_proxy_cycles()        self._clients = {}        self._logger = logger.getChild(logger_name) if logger_name else logger        self.check_parameters()    def check_parameters(self):        for address in self.iter_ipaddresses():            if '/' in address:                ipaddress.ip_network(address, False)            else:                ipaddress.ip_address(address)        if self.proxies is not None and not isinstance(self.proxies, (str, dict)):            raise ValueError('proxies type has to be str, dict or None')    def iter_ipaddresses(self):        local_addresses = self.local_addresses        if not local_addresses:            return        if isinstance(local_addresses, str):            local_addresses = [local_addresses]        yield from local_addresses    def get_ipaddress_cycle(self):        while True:            count = 0            for address in self.iter_ipaddresses():                if '/' in address:                    for a in ipaddress.ip_network(address, False).hosts():                        yield str(a)                        count += 1                else:                    a = ipaddress.ip_address(address)                    yield str(a)                    count += 1            if count == 0:                yield None    def iter_proxies(self):        if not self.proxies:            return        # https://www.python-httpx.org/compatibility/#proxy-keys        if isinstance(self.proxies, str):            yield 'all://', [self.proxies]        else:            for pattern, proxy_url in self.proxies.items():                pattern = PROXY_PATTERN_MAPPING.get(pattern, pattern)                if isinstance(proxy_url, str):                    proxy_url = [proxy_url]                yield pattern, proxy_url    def get_proxy_cycles(self):        proxy_settings = {}        for pattern, proxy_urls in self.iter_proxies():            proxy_settings[pattern] = cycle(proxy_urls)        while True:            # pylint: disable=stop-iteration-return            yield tuple((pattern, next(proxy_url_cycle)) for pattern, proxy_url_cycle in proxy_settings.items())    async def log_response(self, response: httpx.Response):        request = response.request        status = f"{response.status_code} {response.reason_phrase}"        response_line = f"{response.http_version} {status}"        content_type = response.headers.get("Content-Type")        content_type = f' ({content_type})' if content_type else ''        self._logger.debug(f'HTTP Request: {request.method} {request.url} "{response_line}"{content_type}')    @staticmethod    async def check_tor_proxy(client: httpx.AsyncClient, proxies) -> bool:        if proxies in Network._TOR_CHECK_RESULT:            return Network._TOR_CHECK_RESULT[proxies]        result = True        # ignore client._transport because it is not used with all://        for transport in client._mounts.values():  # pylint: disable=protected-access            if isinstance(transport, AsyncHTTPTransportNoHttp):                continue            if getattr(transport, "_pool") and getattr(                transport._pool, "_rdns", False  # pylint: disable=protected-access            ):                continue            return False        response = await client.get("https://check.torproject.org/api/ip", timeout=60)        if not response.json()["IsTor"]:            result = False        Network._TOR_CHECK_RESULT[proxies] = result        return result    async def get_client(self, verify=None, max_redirects=None) -> httpx.AsyncClient:        verify = self.verify if verify is None else verify        max_redirects = self.max_redirects if max_redirects is None else max_redirects        local_address = next(self._local_addresses_cycle)        proxies = next(self._proxies_cycle)  # is a tuple so it can be part of the key        key = (verify, max_redirects, local_address, proxies)        hook_log_response = self.log_response if sxng_debug else None        if key not in self._clients or self._clients[key].is_closed:            client = new_client(                self.enable_http,                verify,                self.enable_http2,                self.max_connections,                self.max_keepalive_connections,                self.keepalive_expiry,                dict(proxies),                local_address,                0,                max_redirects,                hook_log_response,            )            if self.using_tor_proxy and not await self.check_tor_proxy(client, proxies):                await client.aclose()                raise httpx.ProxyError('Network configuration problem: not using Tor')            self._clients[key] = client        return self._clients[key]    async def aclose(self):        async def close_client(client):            try:                await client.aclose()            except httpx.HTTPError:                pass        await asyncio.gather(*[close_client(client) for client in self._clients.values()], return_exceptions=False)    @staticmethod    def extract_kwargs_clients(kwargs):        kwargs_clients = {}        if 'verify' in kwargs:            kwargs_clients['verify'] = kwargs.pop('verify')        if 'max_redirects' in kwargs:            kwargs_clients['max_redirects'] = kwargs.pop('max_redirects')        if 'allow_redirects' in kwargs:            # see https://github.com/encode/httpx/pull/1808            kwargs['follow_redirects'] = kwargs.pop('allow_redirects')        return kwargs_clients    @staticmethod    def extract_do_raise_for_httperror(kwargs):        do_raise_for_httperror = True        if 'raise_for_httperror' in kwargs:            do_raise_for_httperror = kwargs['raise_for_httperror']            del kwargs['raise_for_httperror']        return do_raise_for_httperror    def patch_response(self, response, do_raise_for_httperror) -> SXNG_Response:        if isinstance(response, httpx.Response):            response = typing.cast(SXNG_Response, response)            # requests compatibility (response is not streamed)            # see also https://www.python-httpx.org/compatibility/#checking-for-4xx5xx-responses            response.ok = not response.is_error            # raise an exception            if do_raise_for_httperror:                try:                    raise_for_httperror(response)                except:                    self._logger.warning(f"HTTP Request failed: {response.request.method} {response.request.url}")                    raise        return response    def is_valid_response(self, response):        # pylint: disable=too-many-boolean-expressions        if (            (self.retry_on_http_error is True and 400 <= response.status_code <= 599)            or (isinstance(self.retry_on_http_error, list) and response.status_code in self.retry_on_http_error)            or (isinstance(self.retry_on_http_error, int) and response.status_code == self.retry_on_http_error)        ):            return False        return True    async def call_client(self, stream, method, url, **kwargs) -> SXNG_Response:        retries = self.retries        was_disconnected = False        do_raise_for_httperror = Network.extract_do_raise_for_httperror(kwargs)        kwargs_clients = Network.extract_kwargs_clients(kwargs)        while retries >= 0:  # pragma: no cover            client = await self.get_client(**kwargs_clients)            cookies = kwargs.pop("cookies", None)            client.cookies = httpx.Cookies(cookies)            try:                if stream:                    response = client.stream(method, url, **kwargs)                else:                    response = await client.request(method, url, **kwargs)                if self.is_valid_response(response) or retries <= 0:                    return self.patch_response(response, do_raise_for_httperror)            except httpx.RemoteProtocolError as e:                if not was_disconnected:                    # the server has closed the connection:                    # try again without decreasing the retries variable & with a new HTTP client                    was_disconnected = True                    await client.aclose()                    self._logger.warning('httpx.RemoteProtocolError: the server has disconnected, retrying')                    continue                if retries <= 0:                    raise e            except (httpx.RequestError, httpx.HTTPStatusError) as e:                if retries <= 0:                    raise e            retries -= 1    async def request(self, method, url, **kwargs):        return await self.call_client(False, method, url, **kwargs)    async def stream(self, method, url, **kwargs):        return await self.call_client(True, method, url, **kwargs)    @classmethod    async def aclose_all(cls):        await asyncio.gather(*[network.aclose() for network in NETWORKS.values()], return_exceptions=False)def get_network(name=None):    return NETWORKS.get(name or DEFAULT_NAME)def check_network_configuration():    async def check():        exception_count = 0        for network in NETWORKS.values():            if network.using_tor_proxy:                try:                    await network.get_client()                except Exception:  # pylint: disable=broad-except                    network._logger.exception('Error')  # pylint: disable=protected-access                    exception_count += 1        return exception_count    future = asyncio.run_coroutine_threadsafe(check(), get_loop())    exception_count = future.result()    if exception_count > 0:        raise RuntimeError("Invalid network configuration")def initialize(settings_engines=None, settings_outgoing=None):    # pylint: disable=import-outside-toplevel)    from searx.engines import engines    from searx import settings    # pylint: enable=import-outside-toplevel)    settings_engines = settings_engines or settings['engines']    settings_outgoing = settings_outgoing or settings['outgoing']    # default parameters for AsyncHTTPTransport    # see https://github.com/encode/httpx/blob/e05a5372eb6172287458b37447c30f650047e1b8/httpx/_transports/default.py#L108-L121  # pylint: disable=line-too-long    default_params = {        'enable_http': False,        'verify': settings_outgoing['verify'],        'enable_http2': settings_outgoing['enable_http2'],        'max_connections': settings_outgoing['pool_connections'],        'max_keepalive_connections': settings_outgoing['pool_maxsize'],        'keepalive_expiry': settings_outgoing['keepalive_expiry'],        'local_addresses': settings_outgoing['source_ips'],        'using_tor_proxy': settings_outgoing['using_tor_proxy'],        'proxies': settings_outgoing['proxies'],        'max_redirects': settings_outgoing['max_redirects'],        'retries': settings_outgoing['retries'],        'retry_on_http_error': None,    }    def new_network(params, logger_name=None):        nonlocal default_params        result = {}        result.update(default_params)        result.update(params)        if logger_name:            result['logger_name'] = logger_name        return Network(**result)    def iter_networks():        nonlocal settings_engines        for engine_spec in settings_engines:            engine_name = engine_spec['name']            engine = engines.get(engine_name)            if engine is None:                continue            network = getattr(engine, 'network', None)            yield engine_name, engine, network    if NETWORKS:        done()    NETWORKS.clear()    NETWORKS[DEFAULT_NAME] = new_network({}, logger_name='default')    NETWORKS['ipv4'] = new_network({'local_addresses': '0.0.0.0'}, logger_name='ipv4')    NETWORKS['ipv6'] = new_network({'local_addresses': '::'}, logger_name='ipv6')    # define networks from outgoing.networks    for network_name, network in settings_outgoing['networks'].items():        NETWORKS[network_name] = new_network(network, logger_name=network_name)    # define networks from engines.[i].network (except references)    for engine_name, engine, network in iter_networks():        if network is None:            network = {}            for attribute_name, attribute_value in default_params.items():                if hasattr(engine, attribute_name):                    network[attribute_name] = getattr(engine, attribute_name)                else:                    network[attribute_name] = attribute_value            NETWORKS[engine_name] = new_network(network, logger_name=engine_name)        elif isinstance(network, dict):            NETWORKS[engine_name] = new_network(network, logger_name=engine_name)    # define networks from engines.[i].network (references)    for engine_name, engine, network in iter_networks():        if isinstance(network, str):            NETWORKS[engine_name] = NETWORKS[network]    # the /image_proxy endpoint has a dedicated network.    # same parameters than the default network, but HTTP/2 is disabled.    # It decreases the CPU load average, and the total time is more or less the same    if 'image_proxy' not in NETWORKS:        image_proxy_params = default_params.copy()        image_proxy_params['enable_http2'] = False        NETWORKS['image_proxy'] = new_network(image_proxy_params, logger_name='image_proxy')@atexit.registerdef done():    """Close all HTTP client    Avoid a warning at exit    See https://github.com/encode/httpx/pull/2026    Note: since Network.aclose has to be async, it is not possible to call this method on Network.__del__    So Network.aclose is called here using atexit.register    """    try:        loop = get_loop()        if loop:            future = asyncio.run_coroutine_threadsafe(Network.aclose_all(), loop)            # wait 3 seconds to close the HTTP clients            future.result(3)    finally:        NETWORKS.clear()NETWORKS[DEFAULT_NAME] = Network()
 |