forked from zaclys/searxng
52023e3d6e
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
117 lines
3.5 KiB
Python
117 lines
3.5 KiB
Python
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
# lint: pylint
|
|
"""CORE (science)
|
|
|
|
"""
|
|
|
|
from datetime import datetime
|
|
from urllib.parse import urlencode
|
|
|
|
from searx.exceptions import SearxEngineAPIException
|
|
|
|
about = {
|
|
"website": 'https://core.ac.uk',
|
|
"wikidata_id": 'Q22661180',
|
|
"official_api_documentation": 'https://core.ac.uk/documentation/api/',
|
|
"use_official_api": True,
|
|
"require_api_key": True,
|
|
"results": 'JSON',
|
|
}
|
|
|
|
categories = ['science', 'scientific publications']
|
|
paging = True
|
|
nb_per_page = 10
|
|
|
|
api_key = 'unset'
|
|
|
|
base_url = 'https://core.ac.uk:443/api-v2/search/'
|
|
search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
|
|
|
|
|
|
def request(query, params):
|
|
|
|
if api_key == 'unset':
|
|
raise SearxEngineAPIException('missing CORE API key')
|
|
|
|
search_path = search_string.format(
|
|
query=urlencode({'q': query}),
|
|
nb_per_page=nb_per_page,
|
|
page=params['pageno'],
|
|
apikey=api_key,
|
|
)
|
|
params['url'] = base_url + search_path
|
|
|
|
return params
|
|
|
|
|
|
def response(resp):
|
|
results = []
|
|
json_data = resp.json()
|
|
|
|
for result in json_data['data']:
|
|
source = result['_source']
|
|
url = None
|
|
if source.get('urls'):
|
|
url = source['urls'][0].replace('http://', 'https://', 1)
|
|
|
|
if url is None and source.get('doi'):
|
|
# use the DOI reference
|
|
url = 'https://doi.org/' + source['doi']
|
|
|
|
if url is None and source.get('downloadUrl'):
|
|
# use the downloadUrl
|
|
url = source['downloadUrl']
|
|
|
|
if url is None and source.get('identifiers'):
|
|
# try to find an ark id, see
|
|
# https://www.wikidata.org/wiki/Property:P8091
|
|
# and https://en.wikipedia.org/wiki/Archival_Resource_Key
|
|
arkids = [
|
|
identifier[5:] # 5 is the length of "ark:/"
|
|
for identifier in source.get('identifiers')
|
|
if isinstance(identifier, str) and identifier.startswith('ark:/')
|
|
]
|
|
if len(arkids) > 0:
|
|
url = 'https://n2t.net/' + arkids[0]
|
|
|
|
if url is None:
|
|
continue
|
|
|
|
publishedDate = None
|
|
time = source['publishedDate'] or source['depositedDate']
|
|
if time:
|
|
publishedDate = datetime.fromtimestamp(time / 1000)
|
|
|
|
# sometimes the 'title' is None / filter None values
|
|
journals = [j['title'] for j in (source.get('journals') or []) if j['title']]
|
|
|
|
publisher = source['publisher']
|
|
if publisher:
|
|
publisher = source['publisher'].strip("'")
|
|
|
|
results.append(
|
|
{
|
|
'template': 'paper.html',
|
|
'title': source['title'],
|
|
'url': url,
|
|
'content': source['description'] or '',
|
|
# 'comments': '',
|
|
'tags': source['topics'],
|
|
'publishedDate': publishedDate,
|
|
'type': (source['types'] or [None])[0],
|
|
'authors': source['authors'],
|
|
'editor': ', '.join(source['contributors'] or []),
|
|
'publisher': publisher,
|
|
'journal': ', '.join(journals),
|
|
# 'volume': '',
|
|
# 'pages' : '',
|
|
# 'number': '',
|
|
'doi': source['doi'],
|
|
'issn': [x for x in [source.get('issn')] if x],
|
|
'isbn': [x for x in [source.get('isbn')] if x], # exists in the rawRecordXml
|
|
'pdf_url': source.get('repositoryDocument', {}).get('pdfOrigin'),
|
|
}
|
|
)
|
|
|
|
return results
|