| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257 | 
							- # SPDX-License-Identifier: AGPL-3.0-or-later
 
- # lint: pylint
 
- # pylint: disable=missing-module-docstring, global-statement
 
- import asyncio
 
- import logging
 
- import threading
 
- import anyio
 
- import httpcore
 
- import httpx
 
- from httpx_socks import AsyncProxyTransport
 
- from python_socks import (
 
-     parse_proxy_url,
 
-     ProxyConnectionError,
 
-     ProxyTimeoutError,
 
-     ProxyError
 
- )
 
- from searx import logger
 
- # Optional uvloop (support Python 3.6)
 
- try:
 
-     import uvloop
 
- except ImportError:
 
-     pass
 
- else:
 
-     uvloop.install()
 
- logger = logger.getChild('searx.network.client')
 
- LOOP = None
 
- SSLCONTEXTS = {}
 
- TRANSPORT_KWARGS = {
 
-     # use anyio :
 
-     # * https://github.com/encode/httpcore/issues/344
 
-     # * https://github.com/encode/httpx/discussions/1511
 
-     'backend': 'anyio',
 
-     'trust_env': False,
 
- }
 
- # pylint: disable=protected-access
 
- async def close_connections_for_url(
 
-     connection_pool: httpcore.AsyncConnectionPool, url: httpcore._utils.URL
 
- ):
 
-     origin = httpcore._utils.url_to_origin(url)
 
-     logger.debug('Drop connections for %r', origin)
 
-     connections_to_close = connection_pool._connections_for_origin(origin)
 
-     for connection in connections_to_close:
 
-         await connection_pool._remove_from_pool(connection)
 
-         try:
 
-             await connection.aclose()
 
-         except httpx.NetworkError as e:
 
-             logger.warning('Error closing an existing connection', exc_info=e)
 
- # pylint: enable=protected-access
 
- def get_sslcontexts(proxy_url=None, cert=None, verify=True, trust_env=True, http2=False):
 
-     key = (proxy_url, cert, verify, trust_env, http2)
 
-     if key not in SSLCONTEXTS:
 
-         SSLCONTEXTS[key] = httpx.create_ssl_context(cert, verify, trust_env, http2)
 
-     return SSLCONTEXTS[key]
 
- class AsyncHTTPTransportNoHttp(httpx.AsyncHTTPTransport):
 
-     """Block HTTP request"""
 
-     async def handle_async_request(
 
-         self, method, url, headers=None, stream=None, extensions=None
 
-     ):
 
-         raise httpx.UnsupportedProtocol('HTTP protocol is disabled')
 
- class AsyncProxyTransportFixed(AsyncProxyTransport):
 
-     """Fix httpx_socks.AsyncProxyTransport
 
-     Map python_socks exceptions to httpx.ProxyError / httpx.ConnectError
 
-     Map socket.gaierror to httpx.ConnectError
 
-     Note: AsyncProxyTransport inherit from AsyncConnectionPool
 
-     """
 
-     async def handle_async_request(
 
-         self, method, url, headers=None, stream=None, extensions=None
 
-     ):
 
-         retry = 2
 
-         while retry > 0:
 
-             retry -= 1
 
-             try:
 
-                 return await super().handle_async_request(
 
-                     method, url, headers=headers, stream=stream, extensions=extensions
 
-                 )
 
-             except (ProxyConnectionError, ProxyTimeoutError, ProxyError) as e:
 
-                 raise httpx.ProxyError from e
 
-             except OSError as e:
 
-                 # socket.gaierror when DNS resolution fails
 
-                 raise httpx.ConnectError from e
 
-             except httpx.NetworkError as e:
 
-                 # httpx.WriteError on HTTP/2 connection leaves a new opened stream
 
-                 # then each new request creates a new stream and raise the same WriteError
 
-                 await close_connections_for_url(self, url)
 
-                 raise e
 
-             except anyio.ClosedResourceError as e:
 
-                 await close_connections_for_url(self, url)
 
-                 raise httpx.CloseError from e
 
-             except httpx.RemoteProtocolError as e:
 
-                 # in case of httpx.RemoteProtocolError: Server disconnected
 
-                 await close_connections_for_url(self, url)
 
-                 logger.warning('httpx.RemoteProtocolError: retry', exc_info=e)
 
-                 # retry
 
- class AsyncHTTPTransportFixed(httpx.AsyncHTTPTransport):
 
-     """Fix httpx.AsyncHTTPTransport"""
 
-     async def handle_async_request(
 
-         self, method, url, headers=None, stream=None, extensions=None
 
-     ):
 
-         retry = 2
 
-         while retry > 0:
 
-             retry -= 1
 
-             try:
 
-                 return await super().handle_async_request(
 
-                     method, url, headers=headers, stream=stream, extensions=extensions
 
-                 )
 
-             except OSError as e:
 
-                 # socket.gaierror when DNS resolution fails
 
-                 raise httpx.ConnectError from e
 
-             except httpx.NetworkError as e:
 
-                 # httpx.WriteError on HTTP/2 connection leaves a new opened stream
 
-                 # then each new request creates a new stream and raise the same WriteError
 
-                 await close_connections_for_url(self._pool, url)
 
-                 raise e
 
-             except anyio.ClosedResourceError as e:
 
-                 await close_connections_for_url(self._pool, url)
 
-                 raise httpx.CloseError from e
 
-             except httpx.RemoteProtocolError as e:
 
-                 # in case of httpx.RemoteProtocolError: Server disconnected
 
-                 await close_connections_for_url(self._pool, url)
 
-                 logger.warning('httpx.RemoteProtocolError: retry', exc_info=e)
 
-                 # retry
 
- def get_transport_for_socks_proxy(verify, http2, local_address, proxy_url, limit, retries):
 
-     # support socks5h (requests compatibility):
 
-     # https://requests.readthedocs.io/en/master/user/advanced/#socks
 
-     # socks5://   hostname is resolved on client side
 
-     # socks5h://  hostname is resolved on proxy side
 
-     rdns = False
 
-     socks5h = 'socks5h://'
 
-     if proxy_url.startswith(socks5h):
 
-         proxy_url = 'socks5://' + proxy_url[len(socks5h):]
 
-         rdns = True
 
-     proxy_type, proxy_host, proxy_port, proxy_username, proxy_password = parse_proxy_url(proxy_url)
 
-     verify = get_sslcontexts(proxy_url, None, True, False, http2) if verify is True else verify
 
-     return AsyncProxyTransportFixed(
 
-         proxy_type=proxy_type, proxy_host=proxy_host, proxy_port=proxy_port,
 
-         username=proxy_username, password=proxy_password,
 
-         rdns=rdns,
 
-         loop=get_loop(),
 
-         verify=verify,
 
-         http2=http2,
 
-         local_address=local_address,
 
-         max_connections=limit.max_connections,
 
-         max_keepalive_connections=limit.max_keepalive_connections,
 
-         keepalive_expiry=limit.keepalive_expiry,
 
-         retries=retries,
 
-         **TRANSPORT_KWARGS
 
-     )
 
- def get_transport(verify, http2, local_address, proxy_url, limit, retries):
 
-     verify = get_sslcontexts(None, None, True, False, http2) if verify is True else verify
 
-     return AsyncHTTPTransportFixed(
 
-         # pylint: disable=protected-access
 
-         verify=verify,
 
-         http2=http2,
 
-         local_address=local_address,
 
-         proxy=httpx._config.Proxy(proxy_url) if proxy_url else None,
 
-         limits=limit,
 
-         retries=retries,
 
-         **TRANSPORT_KWARGS
 
-     )
 
- def new_client(
 
-         # pylint: disable=too-many-arguments
 
-         enable_http, verify, enable_http2,
 
-         max_connections, max_keepalive_connections, keepalive_expiry,
 
-         proxies, local_address, retries, max_redirects, hook_log_response  ):
 
-     limit = httpx.Limits(
 
-         max_connections=max_connections,
 
-         max_keepalive_connections=max_keepalive_connections,
 
-         keepalive_expiry=keepalive_expiry
 
-     )
 
-     # See https://www.python-httpx.org/advanced/#routing
 
-     mounts = {}
 
-     for pattern, proxy_url in proxies.items():
 
-         if not enable_http and pattern.startswith('http://'):
 
-             continue
 
-         if (proxy_url.startswith('socks4://')
 
-            or proxy_url.startswith('socks5://')
 
-             or proxy_url.startswith('socks5h://')
 
-         ):
 
-             mounts[pattern] = get_transport_for_socks_proxy(
 
-                 verify, enable_http2, local_address, proxy_url, limit, retries
 
-             )
 
-         else:
 
-             mounts[pattern] = get_transport(
 
-                 verify, enable_http2, local_address, proxy_url, limit, retries
 
-             )
 
-     if not enable_http:
 
-         mounts['http://'] = AsyncHTTPTransportNoHttp()
 
-     transport = get_transport(verify, enable_http2, local_address, None, limit, retries)
 
-     event_hooks = None
 
-     if hook_log_response:
 
-         event_hooks = {'response': [ hook_log_response ]}
 
-     return httpx.AsyncClient(
 
-         transport=transport,
 
-         mounts=mounts,
 
-         max_redirects=max_redirects,
 
-         event_hooks=event_hooks,
 
-     )
 
- def get_loop():
 
-     return LOOP
 
- def init():
 
-     # log
 
-     for logger_name in ('hpack.hpack', 'hpack.table', 'httpx._client'):
 
-         logging.getLogger(logger_name).setLevel(logging.WARNING)
 
-     # loop
 
-     def loop_thread():
 
-         global LOOP
 
-         LOOP = asyncio.new_event_loop()
 
-         LOOP.run_forever()
 
-     thread = threading.Thread(
 
-         target=loop_thread,
 
-         name='asyncio_loop',
 
-         daemon=True,
 
-     )
 
-     thread.start()
 
- init()
 
 
  |