Merge pull request #596 from dalf/upgrade-httpx

Upgrade httpx
This commit is contained in:
Alexandre Flament 2022-01-05 19:13:48 +01:00 committed by GitHub
commit 9004e84b6e
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 48 additions and 112 deletions

View File

@ -7,9 +7,9 @@ lxml==4.7.1
pygments==2.11.0 pygments==2.11.0
python-dateutil==2.8.2 python-dateutil==2.8.2
pyyaml==6.0 pyyaml==6.0
httpx[http2]==0.19.0 httpx[http2]==0.21.2
Brotli==1.0.9 Brotli==1.0.9
uvloop==0.16.0 uvloop==0.16.0
httpx-socks[asyncio]==0.4.1 httpx-socks[asyncio]==0.7.2
langdetect==1.0.9 langdetect==1.0.9
setproctitle==1.2.2 setproctitle==1.2.2

View File

@ -6,8 +6,6 @@ import asyncio
import logging import logging
import threading import threading
import anyio
import httpcore
import httpx import httpx
from httpx_socks import AsyncProxyTransport from httpx_socks import AsyncProxyTransport
from python_socks import parse_proxy_url, ProxyConnectionError, ProxyTimeoutError, ProxyError from python_socks import parse_proxy_url, ProxyConnectionError, ProxyTimeoutError, ProxyError
@ -27,31 +25,10 @@ logger = logger.getChild('searx.network.client')
LOOP = None LOOP = None
SSLCONTEXTS = {} SSLCONTEXTS = {}
TRANSPORT_KWARGS = { TRANSPORT_KWARGS = {
# use anyio :
# * https://github.com/encode/httpcore/issues/344
# * https://github.com/encode/httpx/discussions/1511
'backend': 'anyio',
'trust_env': False, 'trust_env': False,
} }
# pylint: disable=protected-access
async def close_connections_for_url(connection_pool: httpcore.AsyncConnectionPool, url: httpcore._utils.URL):
origin = httpcore._utils.url_to_origin(url)
logger.debug('Drop connections for %r', origin)
connections_to_close = connection_pool._connections_for_origin(origin)
for connection in connections_to_close:
await connection_pool._remove_from_pool(connection)
try:
await connection.aclose()
except httpx.NetworkError as e:
logger.warning('Error closing an existing connection', exc_info=e)
# pylint: enable=protected-access
def get_sslcontexts(proxy_url=None, cert=None, verify=True, trust_env=True, http2=False): def get_sslcontexts(proxy_url=None, cert=None, verify=True, trust_env=True, http2=False):
key = (proxy_url, cert, verify, trust_env, http2) key = (proxy_url, cert, verify, trust_env, http2)
if key not in SSLCONTEXTS: if key not in SSLCONTEXTS:
@ -62,75 +39,25 @@ def get_sslcontexts(proxy_url=None, cert=None, verify=True, trust_env=True, http
class AsyncHTTPTransportNoHttp(httpx.AsyncHTTPTransport): class AsyncHTTPTransportNoHttp(httpx.AsyncHTTPTransport):
"""Block HTTP request""" """Block HTTP request"""
async def handle_async_request(self, method, url, headers=None, stream=None, extensions=None): async def handle_async_request(self, request):
raise httpx.UnsupportedProtocol('HTTP protocol is disabled') raise httpx.UnsupportedProtocol('HTTP protocol is disabled')
class AsyncProxyTransportFixed(AsyncProxyTransport): class AsyncProxyTransportFixed(AsyncProxyTransport):
"""Fix httpx_socks.AsyncProxyTransport """Fix httpx_socks.AsyncProxyTransport
Map python_socks exceptions to httpx.ProxyError / httpx.ConnectError Map python_socks exceptions to httpx.ProxyError exceptions
Map socket.gaierror to httpx.ConnectError
Note: AsyncProxyTransport inherit from AsyncConnectionPool
""" """
async def handle_async_request(self, method, url, headers=None, stream=None, extensions=None): async def handle_async_request(self, request):
retry = 2
while retry > 0:
retry -= 1
try: try:
return await super().handle_async_request( return await super().handle_async_request(request)
method, url, headers=headers, stream=stream, extensions=extensions except ProxyConnectionError as e:
) raise httpx.ProxyError("ProxyConnectionError: " + e.strerror, request=request) from e
except (ProxyConnectionError, ProxyTimeoutError, ProxyError) as e: except ProxyTimeoutError as e:
raise httpx.ProxyError from e raise httpx.ProxyError("ProxyTimeoutError: " + e.args[0], request=request) from e
except OSError as e: except ProxyError as e:
# socket.gaierror when DNS resolution fails raise httpx.ProxyError("ProxyError: " + e.args[0], request=request) from e
raise httpx.ConnectError from e
except httpx.NetworkError as e:
# httpx.WriteError on HTTP/2 connection leaves a new opened stream
# then each new request creates a new stream and raise the same WriteError
await close_connections_for_url(self, url)
raise e
except anyio.ClosedResourceError as e:
await close_connections_for_url(self, url)
raise httpx.CloseError from e
except httpx.RemoteProtocolError as e:
# in case of httpx.RemoteProtocolError: Server disconnected
await close_connections_for_url(self, url)
logger.warning('httpx.RemoteProtocolError: retry', exc_info=e)
# retry
class AsyncHTTPTransportFixed(httpx.AsyncHTTPTransport):
"""Fix httpx.AsyncHTTPTransport"""
async def handle_async_request(self, method, url, headers=None, stream=None, extensions=None):
retry = 2
while retry > 0:
retry -= 1
try:
return await super().handle_async_request(
method, url, headers=headers, stream=stream, extensions=extensions
)
except OSError as e:
# socket.gaierror when DNS resolution fails
raise httpx.ConnectError from e
except httpx.NetworkError as e:
# httpx.WriteError on HTTP/2 connection leaves a new opened stream
# then each new request creates a new stream and raise the same WriteError
await close_connections_for_url(self._pool, url)
raise e
except anyio.ClosedResourceError as e:
await close_connections_for_url(self._pool, url)
raise httpx.CloseError from e
except httpx.RemoteProtocolError as e:
# in case of httpx.RemoteProtocolError: Server disconnected
await close_connections_for_url(self._pool, url)
logger.warning('httpx.RemoteProtocolError: retry', exc_info=e)
# retry
def get_transport_for_socks_proxy(verify, http2, local_address, proxy_url, limit, retries): def get_transport_for_socks_proxy(verify, http2, local_address, proxy_url, limit, retries):
@ -157,9 +84,7 @@ def get_transport_for_socks_proxy(verify, http2, local_address, proxy_url, limit
verify=verify, verify=verify,
http2=http2, http2=http2,
local_address=local_address, local_address=local_address,
max_connections=limit.max_connections, limits=limit,
max_keepalive_connections=limit.max_keepalive_connections,
keepalive_expiry=limit.keepalive_expiry,
retries=retries, retries=retries,
**TRANSPORT_KWARGS, **TRANSPORT_KWARGS,
) )
@ -167,13 +92,13 @@ def get_transport_for_socks_proxy(verify, http2, local_address, proxy_url, limit
def get_transport(verify, http2, local_address, proxy_url, limit, retries): def get_transport(verify, http2, local_address, proxy_url, limit, retries):
verify = get_sslcontexts(None, None, True, False, http2) if verify is True else verify verify = get_sslcontexts(None, None, True, False, http2) if verify is True else verify
return AsyncHTTPTransportFixed( return httpx.AsyncHTTPTransport(
# pylint: disable=protected-access # pylint: disable=protected-access
verify=verify, verify=verify,
http2=http2, http2=http2,
local_address=local_address,
proxy=httpx._config.Proxy(proxy_url) if proxy_url else None,
limits=limit, limits=limit,
proxy=httpx._config.Proxy(proxy_url) if proxy_url else None,
local_address=local_address,
retries=retries, retries=retries,
**TRANSPORT_KWARGS, **TRANSPORT_KWARGS,
) )

View File

@ -213,15 +213,18 @@ class Network:
await asyncio.gather(*[close_client(client) for client in self._clients.values()], return_exceptions=False) await asyncio.gather(*[close_client(client) for client in self._clients.values()], return_exceptions=False)
@staticmethod @staticmethod
def get_kwargs_clients(kwargs): def extract_kwargs_clients(kwargs):
kwargs_clients = {} kwargs_clients = {}
if 'verify' in kwargs: if 'verify' in kwargs:
kwargs_clients['verify'] = kwargs.pop('verify') kwargs_clients['verify'] = kwargs.pop('verify')
if 'max_redirects' in kwargs: if 'max_redirects' in kwargs:
kwargs_clients['max_redirects'] = kwargs.pop('max_redirects') kwargs_clients['max_redirects'] = kwargs.pop('max_redirects')
if 'allow_redirects' in kwargs:
# see https://github.com/encode/httpx/pull/1808
kwargs['follow_redirects'] = kwargs.pop('allow_redirects')
return kwargs_clients return kwargs_clients
def is_valid_respones(self, response): def is_valid_response(self, response):
# pylint: disable=too-many-boolean-expressions # pylint: disable=too-many-boolean-expressions
if ( if (
(self.retry_on_http_error is True and 400 <= response.status_code <= 599) (self.retry_on_http_error is True and 400 <= response.status_code <= 599)
@ -231,33 +234,39 @@ class Network:
return False return False
return True return True
async def request(self, method, url, **kwargs): async def call_client(self, stream, method, url, **kwargs):
retries = self.retries retries = self.retries
was_disconnected = False
kwargs_clients = Network.extract_kwargs_clients(kwargs)
while retries >= 0: # pragma: no cover while retries >= 0: # pragma: no cover
kwargs_clients = Network.get_kwargs_clients(kwargs)
client = await self.get_client(**kwargs_clients) client = await self.get_client(**kwargs_clients)
try: try:
if stream:
response = client.stream(method, url, **kwargs)
else:
response = await client.request(method, url, **kwargs) response = await client.request(method, url, **kwargs)
if self.is_valid_respones(response) or retries <= 0: if self.is_valid_response(response) or retries <= 0:
return response return response
except httpx.RemoteProtocolError as e:
if not was_disconnected:
# the server has closed the connection:
# try again without decreasing the retries variable & with a new HTTP client
was_disconnected = True
await client.aclose()
self._logger.warning('httpx.RemoteProtocolError: the server has disconnected, retrying')
continue
if retries <= 0:
raise e
except (httpx.RequestError, httpx.HTTPStatusError) as e: except (httpx.RequestError, httpx.HTTPStatusError) as e:
if retries <= 0: if retries <= 0:
raise e raise e
retries -= 1 retries -= 1
async def request(self, method, url, **kwargs):
return await self.call_client(False, method, url, **kwargs)
async def stream(self, method, url, **kwargs): async def stream(self, method, url, **kwargs):
retries = self.retries return await self.call_client(True, method, url, **kwargs)
while retries >= 0: # pragma: no cover
kwargs_clients = Network.get_kwargs_clients(kwargs)
client = await self.get_client(**kwargs_clients)
try:
response = client.stream(method, url, **kwargs)
if self.is_valid_respones(response) or retries <= 0:
return response
except (httpx.RequestError, httpx.HTTPStatusError) as e:
if retries <= 0:
raise e
retries -= 1
@classmethod @classmethod
async def aclose_all(cls): async def aclose_all(cls):

View File

@ -76,13 +76,15 @@ class TestNetwork(SearxTestCase):
'verify': True, 'verify': True,
'max_redirects': 5, 'max_redirects': 5,
'timeout': 2, 'timeout': 2,
'allow_redirects': True,
} }
kwargs_client = Network.get_kwargs_clients(kwargs) kwargs_client = Network.extract_kwargs_clients(kwargs)
self.assertEqual(len(kwargs_client), 2) self.assertEqual(len(kwargs_client), 2)
self.assertEqual(len(kwargs), 1) self.assertEqual(len(kwargs), 2)
self.assertEqual(kwargs['timeout'], 2) self.assertEqual(kwargs['timeout'], 2)
self.assertEqual(kwargs['follow_redirects'], True)
self.assertTrue(kwargs_client['verify']) self.assertTrue(kwargs_client['verify'])
self.assertEqual(kwargs_client['max_redirects'], 5) self.assertEqual(kwargs_client['max_redirects'], 5)