# SPDX-License-Identifier: AGPL-3.0-or-later # lint: pylint """CORE (science) """ from datetime import datetime from urllib.parse import urlencode from searx.exceptions import SearxEngineAPIException about = { "website": 'https://core.ac.uk', "wikidata_id": 'Q22661180', "official_api_documentation": 'https://core.ac.uk/documentation/api/', "use_official_api": True, "require_api_key": True, "results": 'JSON', } categories = ['science', 'scientific publications'] paging = True nb_per_page = 10 api_key = 'unset' base_url = 'https://core.ac.uk:443/api-v2/search/' search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}' def request(query, params): if api_key == 'unset': raise SearxEngineAPIException('missing CORE API key') search_path = search_string.format( query=urlencode({'q': query}), nb_per_page=nb_per_page, page=params['pageno'], apikey=api_key, ) params['url'] = base_url + search_path return params def response(resp): results = [] json_data = resp.json() for result in json_data['data']: source = result['_source'] url = None if source.get('urls'): url = source['urls'][0].replace('http://', 'https://', 1) if url is None and source.get('doi'): # use the DOI reference url = 'https://doi.org/' + source['doi'] if url is None and source.get('downloadUrl'): # use the downloadUrl url = source['downloadUrl'] if url is None and source.get('identifiers'): # try to find an ark id, see # https://www.wikidata.org/wiki/Property:P8091 # and https://en.wikipedia.org/wiki/Archival_Resource_Key arkids = [ identifier[5:] # 5 is the length of "ark:/" for identifier in source.get('identifiers') if isinstance(identifier, str) and identifier.startswith('ark:/') ] if len(arkids) > 0: url = 'https://n2t.net/' + arkids[0] if url is None: continue publishedDate = None time = source['publishedDate'] or source['depositedDate'] if time: publishedDate = datetime.fromtimestamp(time / 1000) # sometimes the 'title' is None / filter None values journals = [j['title'] for j in (source.get('journals') or []) if j['title']] publisher = source['publisher'] if publisher: publisher = source['publisher'].strip("'") results.append( { 'template': 'paper.html', 'title': source['title'], 'url': url, 'content': source['description'] or '', # 'comments': '', 'tags': source['topics'], 'publishedDate': publishedDate, 'type': (source['types'] or [None])[0], 'authors': source['authors'], 'editor': ', '.join(source['contributors'] or []), 'publisher': publisher, 'journal': ', '.join(journals), # 'volume': '', # 'pages' : '', # 'number': '', 'doi': source['doi'], 'issn': [x for x in [source.get('issn')] if x], 'isbn': [x for x in [source.get('isbn')] if x], # exists in the rawRecordXml 'pdf_url': source.get('repositoryDocument', {}).get('pdfOrigin'), } ) return results