| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586 | 
							- # SPDX-License-Identifier: AGPL-3.0-or-later
 
- """
 
-  Doku Wiki
 
- """
 
- from urllib.parse import urlencode
 
- from lxml.html import fromstring
 
- from searx.utils import extract_text, eval_xpath
 
- # about
 
- about = {
 
-     "website": 'https://www.dokuwiki.org/',
 
-     "wikidata_id": 'Q851864',
 
-     "official_api_documentation": 'https://www.dokuwiki.org/devel:xmlrpc',
 
-     "use_official_api": False,
 
-     "require_api_key": False,
 
-     "results": 'HTML',
 
- }
 
- # engine dependent config
 
- categories = ['general']  # TODO , 'images', 'music', 'videos', 'files'
 
- paging = False
 
- number_of_results = 5
 
- # search-url
 
- # Doku is OpenSearch compatible
 
- base_url = 'http://localhost:8090'
 
- search_url = (
 
-     # fmt: off
 
-     '/?do=search'
 
-     '&{query}'
 
-     # fmt: on
 
- )
 
- # TODO  '&startRecord={offset}'
 
- # TODO  '&maximumRecords={limit}'
 
- # do search-request
 
- def request(query, params):
 
-     params['url'] = base_url + search_url.format(query=urlencode({'id': query}))
 
-     return params
 
- # get response from search-request
 
- def response(resp):
 
-     results = []
 
-     doc = fromstring(resp.text)
 
-     # parse results
 
-     # Quickhits
 
-     for r in eval_xpath(doc, '//div[@class="search_quickresult"]/ul/li'):
 
-         try:
 
-             res_url = eval_xpath(r, './/a[@class="wikilink1"]/@href')[-1]
 
-         except:
 
-             continue
 
-         if not res_url:
 
-             continue
 
-         title = extract_text(eval_xpath(r, './/a[@class="wikilink1"]/@title'))
 
-         # append result
 
-         results.append({'title': title, 'content': "", 'url': base_url + res_url})
 
-     # Search results
 
-     for r in eval_xpath(doc, '//dl[@class="search_results"]/*'):
 
-         try:
 
-             if r.tag == "dt":
 
-                 res_url = eval_xpath(r, './/a[@class="wikilink1"]/@href')[-1]
 
-                 title = extract_text(eval_xpath(r, './/a[@class="wikilink1"]/@title'))
 
-             elif r.tag == "dd":
 
-                 content = extract_text(eval_xpath(r, '.'))
 
-                 # append result
 
-                 results.append({'title': title, 'content': content, 'url': base_url + res_url})
 
-         except:
 
-             continue
 
-         if not res_url:
 
-             continue
 
-     # return results
 
-     return results
 
 
  |