forked from zaclys/searxng
d14994dc73
settings.yml: * outgoing.networks: * can contains network definition * propertiers: enable_http, verify, http2, max_connections, max_keepalive_connections, keepalive_expiry, local_addresses, support_ipv4, support_ipv6, proxies, max_redirects, retries * retries: 0 by default, number of times searx retries to send the HTTP request (using different IP & proxy each time) * local_addresses can be "192.168.0.1/24" (it supports IPv6) * support_ipv4 & support_ipv6: both True by default see https://github.com/searx/searx/pull/1034 * each engine can define a "network" section: * either a full network description * either reference an existing network * all HTTP requests of engine use the same HTTP configuration (it was not the case before, see proxy configuration in master)
138 lines
3.8 KiB
Python
138 lines
3.8 KiB
Python
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
"""
|
|
DuckDuckGo (Web)
|
|
"""
|
|
|
|
from lxml.html import fromstring
|
|
from json import loads
|
|
from searx.utils import extract_text, match_language, eval_xpath, dict_subset
|
|
from searx.network import get
|
|
|
|
# about
|
|
about = {
|
|
"website": 'https://duckduckgo.com/',
|
|
"wikidata_id": 'Q12805',
|
|
"official_api_documentation": 'https://duckduckgo.com/api',
|
|
"use_official_api": False,
|
|
"require_api_key": False,
|
|
"results": 'HTML',
|
|
}
|
|
|
|
# engine dependent config
|
|
categories = ['general']
|
|
paging = False
|
|
supported_languages_url = 'https://duckduckgo.com/util/u172.js'
|
|
time_range_support = True
|
|
|
|
language_aliases = {
|
|
'ar-SA': 'ar-XA',
|
|
'es-419': 'es-XL',
|
|
'ja': 'jp-JP',
|
|
'ko': 'kr-KR',
|
|
'sl-SI': 'sl-SL',
|
|
'zh-TW': 'tzh-TW',
|
|
'zh-HK': 'tzh-HK'
|
|
}
|
|
|
|
# search-url
|
|
url = 'https://html.duckduckgo.com/html'
|
|
url_ping = 'https://duckduckgo.com/t/sl_h'
|
|
time_range_dict = {'day': 'd',
|
|
'week': 'w',
|
|
'month': 'm',
|
|
'year': 'y'}
|
|
|
|
# specific xpath variables
|
|
result_xpath = '//div[@class="result results_links results_links_deep web-result "]' # noqa
|
|
url_xpath = './/a[@class="result__a"]/@href'
|
|
title_xpath = './/a[@class="result__a"]'
|
|
content_xpath = './/a[@class="result__snippet"]'
|
|
correction_xpath = '//div[@id="did_you_mean"]//a'
|
|
|
|
|
|
# match query's language to a region code that duckduckgo will accept
|
|
def get_region_code(lang, lang_list=None):
|
|
if lang == 'all':
|
|
return None
|
|
|
|
lang_code = match_language(lang, lang_list or [], language_aliases, 'wt-WT')
|
|
lang_parts = lang_code.split('-')
|
|
|
|
# country code goes first
|
|
return lang_parts[1].lower() + '-' + lang_parts[0].lower()
|
|
|
|
|
|
def request(query, params):
|
|
if params['time_range'] is not None and params['time_range'] not in time_range_dict:
|
|
return params
|
|
|
|
params['url'] = url
|
|
params['method'] = 'POST'
|
|
params['data']['q'] = query
|
|
params['data']['b'] = ''
|
|
|
|
region_code = get_region_code(params['language'], supported_languages)
|
|
if region_code:
|
|
params['data']['kl'] = region_code
|
|
params['cookies']['kl'] = region_code
|
|
|
|
if params['time_range'] in time_range_dict:
|
|
params['data']['df'] = time_range_dict[params['time_range']]
|
|
|
|
params['allow_redirects'] = False
|
|
return params
|
|
|
|
|
|
# get response from search-request
|
|
def response(resp):
|
|
if resp.status_code == 303:
|
|
return []
|
|
|
|
# ping
|
|
headers_ping = dict_subset(resp.request.headers, ['User-Agent', 'Accept-Encoding', 'Accept', 'Cookie'])
|
|
get(url_ping, headers=headers_ping)
|
|
|
|
# parse the response
|
|
results = []
|
|
doc = fromstring(resp.text)
|
|
for i, r in enumerate(eval_xpath(doc, result_xpath)):
|
|
if i >= 30:
|
|
break
|
|
try:
|
|
res_url = eval_xpath(r, url_xpath)[-1]
|
|
except:
|
|
continue
|
|
|
|
if not res_url:
|
|
continue
|
|
|
|
title = extract_text(eval_xpath(r, title_xpath))
|
|
content = extract_text(eval_xpath(r, content_xpath))
|
|
|
|
# append result
|
|
results.append({'title': title,
|
|
'content': content,
|
|
'url': res_url})
|
|
|
|
# parse correction
|
|
for correction in eval_xpath(doc, correction_xpath):
|
|
# append correction
|
|
results.append({'correction': extract_text(correction)})
|
|
|
|
# return results
|
|
return results
|
|
|
|
|
|
# get supported languages from their site
|
|
def _fetch_supported_languages(resp):
|
|
|
|
# response is a js file with regions as an embedded object
|
|
response_page = resp.text
|
|
response_page = response_page[response_page.find('regions:{') + 8:]
|
|
response_page = response_page[:response_page.find('}') + 1]
|
|
|
|
regions_json = loads(response_page)
|
|
supported_languages = map((lambda x: x[3:] + '-' + x[:2].upper()), regions_json.keys())
|
|
|
|
return list(supported_languages)
|