Browse Source

[mod] engine ask.com - parse JS result to JSON

Parse the result list from ask.com given in the variable named
window.MESON.initialState::

    <script nonce="..">
        window.MESON = window.MESON || {};
        window.MESON.initialState = {"siteConfig": ...
          ...}};
        window.MESON.loadedLang = "en";
    </script>

The result list is in field::

    json_resp['search']['webResults']['results']

Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
Markus Heiser 1 year ago
parent
commit
76845ea42c
1 changed files with 34 additions and 16 deletions
  1. 34 16
      searx/engines/ask.py

+ 34 - 16
searx/engines/ask.py

@@ -3,8 +3,9 @@
 """Ask.com"""
 
 from urllib.parse import urlencode
-import re
+import dateutil
 from lxml import html
+from searx import utils
 
 # Metadata
 about = {
@@ -37,20 +38,37 @@ def request(query, params):
 
 def response(resp):
 
-    text = html.fromstring(resp.text).text_content()
-    urls_match = re.findall(r'"url":"(.*?)"', text)
-    titles_match = re.findall(r'"title":"(.*?)"', text)[3:]
-    content_match = re.findall(r'"abstract":"(.*?)"', text)
-
-    results = [
-        {
-            "url": url,
-            "title": title,
-            "content": content,
-        }
-        for url, title, content in zip(urls_match, titles_match, content_match)
-        if "&qo=relatedSearchNarrow" not in url
-        # Related searches shouldn't be in the search results: www.ask.com/web&q=related
-    ]
+    start_tag = 'window.MESON.initialState = {'
+    end_tag = '}};'
+
+    dom = html.fromstring(resp.text)
+    script = utils.eval_xpath_getindex(dom, '//script', 0, default=None).text
+
+    pos = script.index(start_tag) + len(start_tag) - 1
+    script = script[pos:]
+    pos = script.index(end_tag) + len(end_tag) - 1
+    script = script[:pos]
+
+    json_resp = utils.js_variable_to_python(script)
+
+    results = []
+
+    for item in json_resp['search']['webResults']['results']:
+
+        pubdate_original = item.get('pubdate_original')
+        if pubdate_original:
+            pubdate_original = dateutil.parser.parse(pubdate_original)
+        metadata = [item.get(field) for field in ['category_l1', 'catsy'] if item.get(field)]
+
+        results.append(
+            {
+                "url": item['url'],
+                "title": item['title'],
+                "content": item['abstract'],
+                "publishedDate": pubdate_original,
+                # "img_src": item.get('image_url') or None, # these are not thumbs / to large
+                "metadata": ' | '.join(metadata),
+            }
+        )
 
     return results