mirror of
https://github.com/searxng/searxng
synced 2024-01-01 18:24:07 +00:00
d6a2d4f969
Slightly modified merge of commit [1cb1d3ac] from searx [PR 2543]: This adds Docker Hub .. as a search engine .. the engine's favicon was downloaded from the Docker Hub website with wget and converted to a PNG with ImageMagick .. It supports the parsing of URLs, titles, content, published dates, and thumbnails of Docker images. [1cb1d3ac] https://github.com/searx/searx/pull/2543/commits/1cb1d3ac [PR 2543] https://github.com/searx/searx/pull/2543 Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
66 lines
1.9 KiB
Python
66 lines
1.9 KiB
Python
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
# lint: pylint
|
|
# pylint: disable=missing-function-docstring
|
|
"""Docker Hub (IT)
|
|
|
|
"""
|
|
|
|
from json import loads
|
|
from urllib.parse import urlencode
|
|
from dateutil import parser
|
|
|
|
about = {
|
|
"website": 'https://hub.docker.com',
|
|
"wikidata_id": 'Q100769064',
|
|
"official_api_documentation": 'https://docs.docker.com/registry/spec/api/',
|
|
"use_official_api": True,
|
|
"require_api_key": False,
|
|
"results": 'JSON',
|
|
}
|
|
|
|
categories = ['it'] # optional
|
|
paging = True
|
|
|
|
base_url = "https://hub.docker.com/"
|
|
search_url = base_url + "api/content/v1/products/search?{query}&type=image&page_size=25"
|
|
|
|
def request(query, params):
|
|
|
|
params['url'] = search_url.format(query=urlencode(dict(q=query, page=params["pageno"])))
|
|
params["headers"]["Search-Version"] = "v3"
|
|
|
|
return params
|
|
|
|
def response(resp):
|
|
'''post-response callback
|
|
resp: requests response object
|
|
'''
|
|
results = []
|
|
body = loads(resp.text)
|
|
|
|
# Make sure `summaries` isn't `null`
|
|
search_res = body.get("summaries")
|
|
if search_res:
|
|
for item in search_res:
|
|
result = {}
|
|
|
|
# Make sure correct URL is set
|
|
filter_type = item.get("filter_type")
|
|
is_official = filter_type in ["store", "official"]
|
|
|
|
if is_official:
|
|
result["url"] = base_url + "_/" + item.get('slug', "")
|
|
else:
|
|
result["url"] = base_url + "r/" + item.get('slug', "")
|
|
result["title"] = item.get("name")
|
|
result["content"] = item.get("short_description")
|
|
result["publishedDate"] = parser.parse(
|
|
item.get("updated_at") or item.get("created_at")
|
|
)
|
|
result["thumbnail"] = (
|
|
item["logo_url"].get("large") or item["logo_url"].get("small")
|
|
)
|
|
results.append(result)
|
|
|
|
return results
|