Browse Source

[fix] engine: core.ac.uk implement API v3 / v2 is no longer supported

Tan Yong Sheng 1 month ago
parent
commit
40feede51e
3 changed files with 111 additions and 62 deletions
  1. 13 0
      docs/dev/engines/online/core.rst
  2. 97 61
      searx/engines/core.py
  3. 1 1
      searx/engines/mariadb_server.py

+ 13 - 0
docs/dev/engines/online/core.rst

@@ -0,0 +1,13 @@
+.. _core engine:
+
+====
+CORE
+====
+
+.. contents::
+   :depth: 2
+   :local:
+   :backlinks: entry
+
+.. automodule:: searx.engines.core
+   :members:

+ 97 - 61
searx/engines/core.py

@@ -1,7 +1,33 @@
 # SPDX-License-Identifier: AGPL-3.0-or-later
-"""CORE (science)
+"""CORE_ (COnnecting REpositories) provides a comprehensive bibliographic
+database of the world’s scholarly literature, collecting and indexing
+research from repositories and journals.
+
+.. _CORE: https://core.ac.uk/about
+
+.. _core engine config:
+
+Configuration
+=============
+
+The engine has the following additional settings:
+
+- :py:obj:`api_key`
+
+.. code:: yaml
+
+  - name: core.ac.uk
+    engine: core
+    categories: science
+    shortcut: cor
+    api_key: "..."
+    timeout: 5
+
+Implementations
+===============
 
 """
+# pylint: disable=too-many-branches
 
 from datetime import datetime
 from urllib.parse import urlencode
@@ -11,34 +37,36 @@ from searx.exceptions import SearxEngineAPIException
 about = {
     "website": 'https://core.ac.uk',
     "wikidata_id": 'Q22661180',
-    "official_api_documentation": 'https://core.ac.uk/documentation/api/',
+    "official_api_documentation": 'https://api.core.ac.uk/docs/v3',
     "use_official_api": True,
     "require_api_key": True,
     "results": 'JSON',
 }
 
+api_key = 'unset'
+"""For an API key register at https://core.ac.uk/services/api and insert
+the API key in the engine :ref:`core engine config`."""
+
 categories = ['science', 'scientific publications']
 paging = True
 nb_per_page = 10
-
-api_key = 'unset'
-
-base_url = 'https://core.ac.uk:443/api-v2/search/'
-search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
+base_url = 'https://api.core.ac.uk/v3/search/works/'
 
 
 def request(query, params):
-
     if api_key == 'unset':
         raise SearxEngineAPIException('missing CORE API key')
 
-    search_path = search_string.format(
-        query=urlencode({'q': query}),
-        nb_per_page=nb_per_page,
-        page=params['pageno'],
-        apikey=api_key,
-    )
-    params['url'] = base_url + search_path
+    # API v3 uses different parameters
+    search_params = {
+        'q': query,
+        'offset': (params['pageno'] - 1) * nb_per_page,
+        'limit': nb_per_page,
+        'sort': 'relevance',
+    }
+
+    params['url'] = base_url + '?' + urlencode(search_params)
+    params['headers'] = {'Authorization': f'Bearer {api_key}'}
 
     return params
 
@@ -47,68 +75,76 @@ def response(resp):
     results = []
     json_data = resp.json()
 
-    for result in json_data['data']:
-        source = result['_source']
+    for result in json_data.get('results', []):
+        # Get title
+        if not result.get('title'):
+            continue
+
+        # Get URL - try different options
         url = None
-        if source.get('urls'):
-            url = source['urls'][0].replace('http://', 'https://', 1)
 
-        if url is None and source.get('doi'):
+        # Try DOI first
+        doi = result.get('doi')
+        if doi:
+            url = f'https://doi.org/{doi}'
+
+        if url is None and result.get('doi'):
             # use the DOI reference
-            url = 'https://doi.org/' + source['doi']
-
-        if url is None and source.get('downloadUrl'):
-            # use the downloadUrl
-            url = source['downloadUrl']
-
-        if url is None and source.get('identifiers'):
-            # try to find an ark id, see
-            # https://www.wikidata.org/wiki/Property:P8091
-            # and https://en.wikipedia.org/wiki/Archival_Resource_Key
-            arkids = [
-                identifier[5:]  # 5 is the length of "ark:/"
-                for identifier in source.get('identifiers')
-                if isinstance(identifier, str) and identifier.startswith('ark:/')
-            ]
-            if len(arkids) > 0:
-                url = 'https://n2t.net/' + arkids[0]
-
-        if url is None:
+            url = 'https://doi.org/' + str(result['doi'])
+        elif result.get('id'):
+            url = 'https://core.ac.uk/works/' + str(result['id'])
+        elif result.get('downloadUrl'):
+            url = result['downloadUrl']
+        elif result.get('sourceFulltextUrls'):
+            url = result['sourceFulltextUrls']
+        else:
             continue
 
-        publishedDate = None
-        time = source['publishedDate'] or source['depositedDate']
-        if time:
-            publishedDate = datetime.fromtimestamp(time / 1000)
+        # Published date
+        published_date = None
 
-        # sometimes the 'title' is None / filter None values
-        journals = [j['title'] for j in (source.get('journals') or []) if j['title']]
+        raw_date = result.get('publishedDate') or result.get('depositedDate')
+        if raw_date:
+            try:
+                published_date = datetime.fromisoformat(result['publishedDate'].replace('Z', '+00:00'))
+            except (ValueError, AttributeError):
+                pass
 
-        publisher = source['publisher']
+        # Handle journals
+        journals = []
+        if result.get('journals'):
+            journals = [j.get('title') for j in result['journals'] if j.get('title')]
+
+        # Handle publisher
+        publisher = result.get('publisher', '').strip("'")
         if publisher:
-            publisher = source['publisher'].strip("'")
+            publisher = publisher.strip("'")
+
+        # Handle authors
+        authors = set()
+        for i in result.get('authors', []):
+            name = i.get("name")
+            if name:
+                authors.add(name)
 
         results.append(
             {
                 'template': 'paper.html',
-                'title': source['title'],
+                'title': result.get('title'),
                 'url': url,
-                'content': source['description'] or '',
+                'content': result.get('fullText', '') or '',
                 # 'comments': '',
-                'tags': source['topics'],
-                'publishedDate': publishedDate,
-                'type': (source['types'] or [None])[0],
-                'authors': source['authors'],
-                'editor': ', '.join(source['contributors'] or []),
+                'tags': result.get('fieldOfStudy', []),
+                'publishedDate': published_date,
+                'type': result.get('documentType', '') or '',
+                'authors': authors,
+                'editor': ', '.join(result.get('contributors', [])),
                 'publisher': publisher,
                 'journal': ', '.join(journals),
-                # 'volume': '',
-                # 'pages' : '',
-                # 'number': '',
-                'doi': source['doi'],
-                'issn': [x for x in [source.get('issn')] if x],
-                'isbn': [x for x in [source.get('isbn')] if x],  # exists in the rawRecordXml
-                'pdf_url': source.get('repositoryDocument', {}).get('pdfOrigin'),
+                'doi': result.get('doi'),
+                # 'issn' : ''
+                # 'isbn' : ''
+                'pdf_url': result.get('downloadUrl', {}) or result.get("sourceFulltextUrls", {}),
             }
         )
 

+ 1 - 1
searx/engines/mariadb_server.py

@@ -29,7 +29,7 @@ Implementations
 from typing import TYPE_CHECKING
 
 try:
-    import mariadb
+    import mariadb  # pyright: ignore [reportMissingImports]
 except ImportError:
     # import error is ignored because the admin has to install mysql manually to use
     # the engine