| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455 | 
							- # SPDX-License-Identifier: AGPL-3.0-or-later
 
- """lib.rs (packages)"""
 
- from urllib.parse import quote_plus
 
- from lxml import html
 
- from searx.utils import eval_xpath, eval_xpath_list, extract_text
 
- about = {
 
-     'website': 'https://lib.rs',
 
-     'wikidata_id': 'Q113486010',
 
-     'use_official_api': False,
 
-     'require_api_key': False,
 
-     'results': "HTML",
 
- }
 
- categories = ["it", "packages"]
 
- base_url = 'https://lib.rs'
 
- results_xpath = '/html/body/main/div/ol/li/a'
 
- url_xpath = './@href'
 
- title_xpath = './div[@class="h"]/h4'
 
- content_xpath = './div[@class="h"]/p'
 
- version_xpath = './div[@class="meta"]/span[contains(@class, "version")]'
 
- download_count_xpath = './div[@class="meta"]/span[@class="downloads"]'
 
- tags_xpath = './div[@class="meta"]/span[contains(@class, "k")]/text()'
 
- def request(query, params):
 
-     params['url'] = f"{base_url}/search?q={quote_plus(query)}"
 
-     return params
 
- def response(resp):
 
-     results = []
 
-     doc = html.fromstring(resp.text)
 
-     for result in eval_xpath_list(doc, results_xpath):
 
-         package_name = extract_text(eval_xpath(result, title_xpath))
 
-         results.append(
 
-             {
 
-                 'template': 'packages.html',
 
-                 'title': package_name,
 
-                 'url': base_url + extract_text(eval_xpath(result, url_xpath)),  # type: ignore
 
-                 'content': extract_text(eval_xpath(result, content_xpath)),
 
-                 'package_name': package_name,
 
-                 'version': extract_text(eval_xpath(result, version_xpath)),
 
-                 'popularity': extract_text(eval_xpath(result, download_count_xpath)),
 
-                 'tags': eval_xpath_list(result, tags_xpath),
 
-             }
 
-         )
 
-     return results
 
 
  |