| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115 | 
							- # SPDX-License-Identifier: AGPL-3.0-or-later
 
- """CORE (science)
 
- """
 
- from datetime import datetime
 
- from urllib.parse import urlencode
 
- from searx.exceptions import SearxEngineAPIException
 
- about = {
 
-     "website": 'https://core.ac.uk',
 
-     "wikidata_id": 'Q22661180',
 
-     "official_api_documentation": 'https://core.ac.uk/documentation/api/',
 
-     "use_official_api": True,
 
-     "require_api_key": True,
 
-     "results": 'JSON',
 
- }
 
- categories = ['science', 'scientific publications']
 
- paging = True
 
- nb_per_page = 10
 
- api_key = 'unset'
 
- base_url = 'https://core.ac.uk:443/api-v2/search/'
 
- search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
 
- def request(query, params):
 
-     if api_key == 'unset':
 
-         raise SearxEngineAPIException('missing CORE API key')
 
-     search_path = search_string.format(
 
-         query=urlencode({'q': query}),
 
-         nb_per_page=nb_per_page,
 
-         page=params['pageno'],
 
-         apikey=api_key,
 
-     )
 
-     params['url'] = base_url + search_path
 
-     return params
 
- def response(resp):
 
-     results = []
 
-     json_data = resp.json()
 
-     for result in json_data['data']:
 
-         source = result['_source']
 
-         url = None
 
-         if source.get('urls'):
 
-             url = source['urls'][0].replace('http://', 'https://', 1)
 
-         if url is None and source.get('doi'):
 
-             # use the DOI reference
 
-             url = 'https://doi.org/' + source['doi']
 
-         if url is None and source.get('downloadUrl'):
 
-             # use the downloadUrl
 
-             url = source['downloadUrl']
 
-         if url is None and source.get('identifiers'):
 
-             # try to find an ark id, see
 
-             # https://www.wikidata.org/wiki/Property:P8091
 
-             # and https://en.wikipedia.org/wiki/Archival_Resource_Key
 
-             arkids = [
 
-                 identifier[5:]  # 5 is the length of "ark:/"
 
-                 for identifier in source.get('identifiers')
 
-                 if isinstance(identifier, str) and identifier.startswith('ark:/')
 
-             ]
 
-             if len(arkids) > 0:
 
-                 url = 'https://n2t.net/' + arkids[0]
 
-         if url is None:
 
-             continue
 
-         publishedDate = None
 
-         time = source['publishedDate'] or source['depositedDate']
 
-         if time:
 
-             publishedDate = datetime.fromtimestamp(time / 1000)
 
-         # sometimes the 'title' is None / filter None values
 
-         journals = [j['title'] for j in (source.get('journals') or []) if j['title']]
 
-         publisher = source['publisher']
 
-         if publisher:
 
-             publisher = source['publisher'].strip("'")
 
-         results.append(
 
-             {
 
-                 'template': 'paper.html',
 
-                 'title': source['title'],
 
-                 'url': url,
 
-                 'content': source['description'] or '',
 
-                 # 'comments': '',
 
-                 'tags': source['topics'],
 
-                 'publishedDate': publishedDate,
 
-                 'type': (source['types'] or [None])[0],
 
-                 'authors': source['authors'],
 
-                 'editor': ', '.join(source['contributors'] or []),
 
-                 'publisher': publisher,
 
-                 'journal': ', '.join(journals),
 
-                 # 'volume': '',
 
-                 # 'pages' : '',
 
-                 # 'number': '',
 
-                 'doi': source['doi'],
 
-                 'issn': [x for x in [source.get('issn')] if x],
 
-                 'isbn': [x for x in [source.get('isbn')] if x],  # exists in the rawRecordXml
 
-                 'pdf_url': source.get('repositoryDocument', {}).get('pdfOrigin'),
 
-             }
 
-         )
 
-     return results
 
 
  |