| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176 | # SPDX-License-Identifier: AGPL-3.0-or-later# lint: pylint"""This module implements functions needed for the autocompleter."""from json import loadsfrom urllib.parse import urlencodefrom lxml import etreefrom httpx import HTTPErrorfrom searx import settingsfrom searx.data import ENGINES_LANGUAGESfrom searx.network import get as http_getfrom searx.exceptions import SearxEngineResponseException# a fetch_supported_languages() for XPath engines isn't available right now# _brave = ENGINES_LANGUAGES['brave'].keys()def get(*args, **kwargs):    if 'timeout' not in kwargs:        kwargs['timeout'] = settings['outgoing']['request_timeout']    kwargs['raise_for_httperror'] = True    return http_get(*args, **kwargs)def brave(query, _lang):    # brave search autocompleter    url = 'https://search.brave.com/api/suggest?'    url += urlencode({'q': query})    country = 'all'    # if lang in _brave:    #    country = lang    kwargs = {'cookies': {'country': country}}    resp = get(url, **kwargs)    results = []    if resp.ok:        data = resp.json()        for item in data[1]:            results.append(item)    return resultsdef dbpedia(query, _lang):    # dbpedia autocompleter, no HTTPS    autocomplete_url = 'https://lookup.dbpedia.org/api/search.asmx/KeywordSearch?'    response = get(autocomplete_url + urlencode(dict(QueryString=query)))    results = []    if response.ok:        dom = etree.fromstring(response.content)        results = dom.xpath('//Result/Label//text()')    return resultsdef duckduckgo(query, _lang):    # duckduckgo autocompleter    url = 'https://ac.duckduckgo.com/ac/?{0}&type=list'    resp = loads(get(url.format(urlencode(dict(q=query)))).text)    if len(resp) > 1:        return resp[1]    return []def google(query, lang):    # google autocompleter    autocomplete_url = 'https://suggestqueries.google.com/complete/search?client=toolbar&'    response = get(autocomplete_url + urlencode(dict(hl=lang, q=query)))    results = []    if response.ok:        dom = etree.fromstring(response.text)        results = dom.xpath('//suggestion/@data')    return resultsdef seznam(query, _lang):    # seznam search autocompleter    url = 'https://suggest.seznam.cz/fulltext/cs?{query}'    resp = get(        url.format(            query=urlencode(                {'phrase': query, 'cursorPosition': len(query), 'format': 'json-2', 'highlight': '1', 'count': '6'}            )        )    )    if not resp.ok:        return []    data = resp.json()    return [        ''.join([part.get('text', '') for part in item.get('text', [])])        for item in data.get('result', [])        if item.get('itemType', None) == 'ItemType.TEXT'    ]def startpage(query, lang):    # startpage autocompleter    lui = ENGINES_LANGUAGES['startpage'].get(lang, 'english')    url = 'https://startpage.com/suggestions?{query}'    resp = get(url.format(query=urlencode({'q': query, 'segment': 'startpage.udog', 'lui': lui})))    data = resp.json()    return [e['text'] for e in data.get('suggestions', []) if 'text' in e]def swisscows(query, _lang):    # swisscows autocompleter    url = 'https://swisscows.ch/api/suggest?{query}&itemsCount=5'    resp = loads(get(url.format(query=urlencode({'query': query}))).text)    return respdef qwant(query, lang):    # qwant autocompleter (additional parameter : lang=en_en&count=xxx )    url = 'https://api.qwant.com/api/suggest?{query}'    resp = get(url.format(query=urlencode({'q': query, 'lang': lang})))    results = []    if resp.ok:        data = loads(resp.text)        if data['status'] == 'success':            for item in data['data']['items']:                results.append(item['value'])    return resultsdef wikipedia(query, lang):    # wikipedia autocompleter    url = 'https://' + lang + '.wikipedia.org/w/api.php?action=opensearch&{0}&limit=10&namespace=0&format=json'    resp = loads(get(url.format(urlencode(dict(search=query)))).text)    if len(resp) > 1:        return resp[1]    return []backends = {    'dbpedia': dbpedia,    'duckduckgo': duckduckgo,    'google': google,    'seznam': seznam,    'startpage': startpage,    'swisscows': swisscows,    'qwant': qwant,    'wikipedia': wikipedia,    'brave': brave,}def search_autocomplete(backend_name, query, lang):    backend = backends.get(backend_name)    if backend is None:        return []    try:        return backend(query, lang)    except (HTTPError, SearxEngineResponseException):        return []
 |