forked from zaclys/searxng
Merge branch 'master' into patch-1
This commit is contained in:
commit
9012ad498c
|
@ -35,9 +35,12 @@ site_url = 'https://duckduckgo.com/?{query}&iar=images&iax=1&ia=images'
|
|||
|
||||
# run query in site to get vqd number needed for requesting images
|
||||
# TODO: find a way to get this number without an extra request (is it a hash of the query?)
|
||||
def get_vqd(query):
|
||||
res = get(site_url.format(query=urlencode({'q': query})))
|
||||
def get_vqd(query, headers):
|
||||
query_url = site_url.format(query=urlencode({'q': query}))
|
||||
res = get(query_url, headers=headers)
|
||||
content = res.text
|
||||
if content.find('vqd=\'') == -1:
|
||||
raise Exception('Request failed')
|
||||
vqd = content[content.find('vqd=\'') + 5:]
|
||||
vqd = vqd[:vqd.find('\'')]
|
||||
return vqd
|
||||
|
@ -47,7 +50,7 @@ def get_vqd(query):
|
|||
def request(query, params):
|
||||
# to avoid running actual external requests when testing
|
||||
if 'is_test' not in params:
|
||||
vqd = get_vqd(query)
|
||||
vqd = get_vqd(query, params['headers'])
|
||||
else:
|
||||
vqd = '12345'
|
||||
|
||||
|
@ -74,7 +77,7 @@ def response(resp):
|
|||
try:
|
||||
res_json = loads(content)
|
||||
except:
|
||||
return []
|
||||
raise Exception('Cannot parse results')
|
||||
|
||||
# parse results
|
||||
for result in res_json['results']:
|
||||
|
|
|
@ -11,7 +11,6 @@
|
|||
"""
|
||||
|
||||
from datetime import date, timedelta
|
||||
from json import loads
|
||||
from lxml import html
|
||||
from searx.url_utils import urlencode, urlparse, parse_qs
|
||||
|
||||
|
@ -39,7 +38,6 @@ time_range_dict = {'day': 'd',
|
|||
# do search-request
|
||||
def request(query, params):
|
||||
search_options = {
|
||||
'ijn': params['pageno'] - 1,
|
||||
'start': (params['pageno'] - 1) * number_of_results
|
||||
}
|
||||
|
||||
|
@ -53,7 +51,7 @@ def request(query, params):
|
|||
search_options['tbs'] = time_range_custom_attr.format(start=start, end=end)
|
||||
|
||||
if safesearch and params['safesearch']:
|
||||
search_options['safe'] = 'on'
|
||||
search_options['safe'] = 'active'
|
||||
|
||||
params['url'] = search_url.format(query=urlencode({'q': query}),
|
||||
search_options=urlencode(search_options))
|
||||
|
@ -63,24 +61,30 @@ def request(query, params):
|
|||
|
||||
# get response from search-request
|
||||
def response(resp):
|
||||
results = []
|
||||
|
||||
dom = html.fromstring(resp.text)
|
||||
|
||||
# parse results
|
||||
for img in dom.xpath('//a'):
|
||||
r = {
|
||||
'title': u' '.join(img.xpath('.//div[class="rg_ilmbg"]//text()')),
|
||||
results = []
|
||||
for element in dom.xpath('//div[@id="search"] //td'):
|
||||
link = element.xpath('./a')[0]
|
||||
|
||||
google_url = urlparse(link.xpath('.//@href')[0])
|
||||
query = parse_qs(google_url.query)
|
||||
source_url = next(iter(query.get('q', [])), None)
|
||||
|
||||
title_parts = element.xpath('./cite//following-sibling::*/text()')
|
||||
title_parts.extend(element.xpath('./cite//following-sibling::text()')[:-1])
|
||||
|
||||
result = {
|
||||
'title': ''.join(title_parts),
|
||||
'content': '',
|
||||
'template': 'images.html',
|
||||
'url': source_url,
|
||||
'img_src': source_url,
|
||||
'thumbnail_src': next(iter(link.xpath('.//img //@src')), None)
|
||||
}
|
||||
url = urlparse(img.xpath('.//@href')[0])
|
||||
query = parse_qs(url.query)
|
||||
r['url'] = query['imgrefurl'][0]
|
||||
r['img_src'] = query['imgurl'][0]
|
||||
r['thumbnail_src'] = r['img_src']
|
||||
# append result
|
||||
results.append(r)
|
||||
|
||||
# return results
|
||||
if not source_url or not result['thumbnail_src']:
|
||||
continue
|
||||
|
||||
results.append(result)
|
||||
return results
|
||||
|
|
|
@ -41,7 +41,7 @@ class TestDuckduckgoImagesEngine(SearxTestCase):
|
|||
self.assertRaises(AttributeError, duckduckgo_images.response, '[]')
|
||||
|
||||
response = mock.Mock(text='If this error persists, please let us know: ops@duckduckgo.com')
|
||||
self.assertEqual(duckduckgo_images.response(response), [])
|
||||
self.assertRaises(Exception, duckduckgo_images.response, response)
|
||||
|
||||
json = u"""
|
||||
{
|
||||
|
|
Loading…
Reference in New Issue