mirror of
https://github.com/searxng/searxng
synced 2024-01-01 19:24:07 +01:00

We have been using a static type checker (pyright) for a long time, but its check was not yet a prerequisite for passing the quality gate. It was checked in the CI, but the error messages were only logged. As is always the case in life, with checks that you have to do but which have no consequences; you neglect them :-) We didn't activate the checks back then because we (even today) have too much monkey patching in our code (not only in the engines, httpx and others objects are also affected). We want to replace monkey patching with clear interfaces for a long time, the basis for this is increased typing and we can only achieve this if we make type checking an integral part of the quality gate. This PR activates the type check; in order to pass the check, a few typings were corrected in the code, but most type inconsistencies were deactivated via inline comments. This was particularly necessary in places where the code uses properties that stick to the objects (monkey patching). The sticking of properties only happens in a few places, but the access to these properties extends over the entire code, which is why there are many `# type: ignore` markers in the code ... which we will hopefully be able to remove again successively in the future. Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
135 lines
4.1 KiB
Python
135 lines
4.1 KiB
Python
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
"""Flickr (Images)
|
|
|
|
"""
|
|
|
|
import json
|
|
from time import time
|
|
import re
|
|
from urllib.parse import urlencode
|
|
from searx.utils import ecma_unescape, html_to_text
|
|
|
|
|
|
about = {
|
|
"website": 'https://www.flickr.com',
|
|
"wikidata_id": 'Q103204',
|
|
"official_api_documentation": 'https://secure.flickr.com/services/api/flickr.photos.search.html',
|
|
"use_official_api": False,
|
|
"require_api_key": False,
|
|
"results": 'HTML',
|
|
}
|
|
|
|
# engine dependent config
|
|
categories = ['images']
|
|
paging = True
|
|
time_range_support = True
|
|
safesearch = False
|
|
|
|
time_range_dict = {
|
|
'day': 60 * 60 * 24,
|
|
'week': 60 * 60 * 24 * 7,
|
|
'month': 60 * 60 * 24 * 7 * 4,
|
|
'year': 60 * 60 * 24 * 7 * 52,
|
|
}
|
|
image_sizes = ('o', 'k', 'h', 'b', 'c', 'z', 'm', 'n', 't', 'q', 's')
|
|
|
|
search_url = 'https://www.flickr.com/search?{query}&page={page}'
|
|
time_range_url = '&min_upload_date={start}&max_upload_date={end}'
|
|
photo_url = 'https://www.flickr.com/photos/{userid}/{photoid}'
|
|
modelexport_re = re.compile(r"^\s*modelExport:\s*({.*}),$", re.M)
|
|
|
|
|
|
def build_flickr_url(user_id, photo_id):
|
|
return photo_url.format(userid=user_id, photoid=photo_id)
|
|
|
|
|
|
def _get_time_range_url(time_range):
|
|
if time_range in time_range_dict:
|
|
return time_range_url.format(start=time(), end=str(int(time()) - time_range_dict[time_range]))
|
|
return ''
|
|
|
|
|
|
def request(query, params):
|
|
params['url'] = search_url.format(query=urlencode({'text': query}), page=params['pageno']) + _get_time_range_url(
|
|
params['time_range']
|
|
)
|
|
return params
|
|
|
|
|
|
def response(resp): # pylint: disable=too-many-branches
|
|
results = []
|
|
|
|
matches = modelexport_re.search(resp.text)
|
|
if matches is None:
|
|
return results
|
|
|
|
match = matches.group(1)
|
|
model_export = json.loads(match)
|
|
|
|
if 'legend' not in model_export:
|
|
return results
|
|
legend = model_export['legend']
|
|
|
|
# handle empty page
|
|
if not legend or not legend[0]:
|
|
return results
|
|
|
|
for x, index in enumerate(legend):
|
|
if len(index) != 8:
|
|
logger.debug("skip legend enty %s : %s", x, index)
|
|
continue
|
|
|
|
photo = model_export['main'][index[0]][int(index[1])][index[2]][index[3]][index[4]][index[5]][int(index[6])][
|
|
index[7]
|
|
]
|
|
author = ecma_unescape(photo.get('realname', ''))
|
|
source = ecma_unescape(photo.get('username', ''))
|
|
if source:
|
|
source += ' @ Flickr'
|
|
title = ecma_unescape(photo.get('title', ''))
|
|
content = html_to_text(ecma_unescape(photo.get('description', '')))
|
|
img_src = None
|
|
|
|
# From the biggest to the lowest format
|
|
size_data = None
|
|
for image_size in image_sizes:
|
|
if image_size in photo['sizes']['data']:
|
|
size_data = photo['sizes']['data'][image_size]['data']
|
|
break
|
|
|
|
if not size_data:
|
|
logger.debug('cannot find valid image size: {0}'.format(repr(photo['sizes']['data'])))
|
|
continue
|
|
|
|
img_src = size_data['url']
|
|
resolution = f"{size_data['width']} x {size_data['height']}"
|
|
|
|
# For a bigger thumbnail, keep only the url_z, not the url_n
|
|
if 'n' in photo['sizes']['data']:
|
|
thumbnail_src = photo['sizes']['data']['n']['data']['url']
|
|
elif 'z' in photo['sizes']['data']:
|
|
thumbnail_src = photo['sizes']['data']['z']['data']['url']
|
|
else:
|
|
thumbnail_src = img_src
|
|
|
|
if 'ownerNsid' not in photo:
|
|
# should not happen, disowned photo? Show it anyway
|
|
url = img_src
|
|
else:
|
|
url = build_flickr_url(photo['ownerNsid'], photo['id'])
|
|
|
|
result = {
|
|
'url': url,
|
|
'img_src': img_src,
|
|
'thumbnail_src': thumbnail_src,
|
|
'source': source,
|
|
'resolution': resolution,
|
|
'template': 'images.html',
|
|
}
|
|
result['author'] = author.encode(errors='ignore').decode()
|
|
result['source'] = source.encode(errors='ignore').decode()
|
|
result['title'] = title.encode(errors='ignore').decode()
|
|
result['content'] = content.encode(errors='ignore').decode()
|
|
results.append(result)
|
|
|
|
return results
|