# SPDX-License-Identifier: AGPL-3.0-or-later
"""Ollama model search engine for searxng"""

from urllib.parse import urlencode
from datetime import datetime
from lxml import html

from searx.utils import eval_xpath_list, eval_xpath_getindex, eval_xpath, extract_text
from searx.result_types import EngineResults

about = {
    "website": "https://ollama.com",
    "wikidata_id": "Q124636097",
    "use_official_api": False,
    "require_api_key": False,
    "results": "HTML",
}

categories = ["it", "repos"]

base_url = "https://ollama.com"

results_xpath = '//li[@x-test-model]'
title_xpath = './/span[@x-test-search-response-title]/text()'
content_xpath = './/p[@class="max-w-lg break-words text-neutral-800 text-md"]/text()'
url_xpath = './a/@href'
publish_date_xpath = './/span[contains(@class, "flex items-center")]/@title'


def request(query, params):
    query_params = {"q": query}

    params['url'] = f"{base_url}/search?{urlencode(query_params)}"
    return params


def response(resp) -> EngineResults:
    res = EngineResults()

    dom = html.fromstring(resp.text)

    for item in eval_xpath_list(dom, results_xpath):
        published_date = None
        try:
            published_date = datetime.strptime(
                extract_text(eval_xpath(item, publish_date_xpath)), "%b %d, %Y %I:%M %p %Z"
            )
        except ValueError:
            pass

        res.add(
            res.types.MainResult(
                title=extract_text(eval_xpath(item, title_xpath)),
                content=extract_text(eval_xpath(item, content_xpath)),
                url=f"{base_url}{eval_xpath_getindex(item, url_xpath, 0)}",
                publishedDate=published_date,
            )
        )

    return res