123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223 |
- # SPDX-License-Identifier: AGPL-3.0-or-later
- """ChinaSo_, a search engine for the chinese language area.
- .. attention::
- ChinaSo engine does not return real URL, the links from these search
- engines violate the privacy of the users!!
- We try to find a solution for this problem, please follow `issue #4694`_.
- As long as the problem has not been resolved, these engines are
- not active in a standard setup (``inactive: true``).
- .. _ChinaSo: https://www.chinaso.com/
- .. _issue #4694: https://github.com/searxng/searxng/issues/4694
- Configuration
- =============
- The engine has the following additional settings:
- - :py:obj:`chinaso_category` (:py:obj:`ChinasoCategoryType`)
- - :py:obj:`chinaso_news_source` (:py:obj:`ChinasoNewsSourceType`)
- In the example below, all three ChinaSO engines are using the :ref:`network
- <engine network>` from the ``chinaso news`` engine.
- .. code:: yaml
- - name: chinaso news
- engine: chinaso
- shortcut: chinaso
- categories: [news]
- chinaso_category: news
- chinaso_news_source: all
- - name: chinaso images
- engine: chinaso
- network: chinaso news
- shortcut: chinasoi
- categories: [images]
- chinaso_category: images
- - name: chinaso videos
- engine: chinaso
- network: chinaso news
- shortcut: chinasov
- categories: [videos]
- chinaso_category: videos
- Implementations
- ===============
- """
- import typing
- from urllib.parse import urlencode
- from datetime import datetime
- from searx.exceptions import SearxEngineAPIException
- from searx.utils import html_to_text
- about = {
- "website": "https://www.chinaso.com/",
- "wikidata_id": "Q10846064",
- "use_official_api": False,
- "require_api_key": False,
- "results": "JSON",
- "language": "zh",
- }
- paging = True
- time_range_support = True
- results_per_page = 10
- categories = []
- ChinasoCategoryType = typing.Literal['news', 'videos', 'images']
- """ChinaSo supports news, videos, images search.
- - ``news``: search for news
- - ``videos``: search for videos
- - ``images``: search for images
- In the category ``news`` you can additionally filter by option
- :py:obj:`chinaso_news_source`.
- """
- chinaso_category = 'news'
- """Configure ChinaSo category (:py:obj:`ChinasoCategoryType`)."""
- ChinasoNewsSourceType = typing.Literal['CENTRAL', 'LOCAL', 'BUSINESS', 'EPAPER', 'all']
- """Filtering ChinaSo-News results by source:
- - ``CENTRAL``: central publication
- - ``LOCAL``: local publication
- - ``BUSINESS``: business publication
- - ``EPAPER``: E-Paper
- - ``all``: all sources
- """
- chinaso_news_source: ChinasoNewsSourceType = 'all'
- """Configure ChinaSo-News type (:py:obj:`ChinasoNewsSourceType`)."""
- time_range_dict = {'day': '24h', 'week': '1w', 'month': '1m', 'year': '1y'}
- base_url = "https://www.chinaso.com"
- def init(_):
- if chinaso_category not in ('news', 'videos', 'images'):
- raise ValueError(f"Unsupported category: {chinaso_category}")
- if chinaso_category == 'news' and chinaso_news_source not in typing.get_args(ChinasoNewsSourceType):
- raise ValueError(f"Unsupported news source: {chinaso_news_source}")
- def request(query, params):
- query_params = {"q": query}
- if time_range_dict.get(params['time_range']):
- query_params["stime"] = time_range_dict[params['time_range']]
- query_params["etime"] = 'now'
- category_config = {
- 'news': {'endpoint': '/v5/general/v1/web/search', 'params': {'pn': params["pageno"], 'ps': results_per_page}},
- 'images': {
- 'endpoint': '/v5/general/v1/search/image',
- 'params': {'start_index': (params["pageno"] - 1) * results_per_page, 'rn': results_per_page},
- },
- 'videos': {
- 'endpoint': '/v5/general/v1/search/video',
- 'params': {'start_index': (params["pageno"] - 1) * results_per_page, 'rn': results_per_page},
- },
- }
- if chinaso_news_source != 'all':
- if chinaso_news_source == 'EPAPER':
- category_config['news']['params']["type"] = 'EPAPER'
- else:
- category_config['news']['params']["cate"] = chinaso_news_source
- query_params.update(category_config[chinaso_category]['params'])
- params["url"] = f"{base_url}{category_config[chinaso_category]['endpoint']}?{urlencode(query_params)}"
- return params
- def response(resp):
- try:
- data = resp.json()
- except Exception as e:
- raise SearxEngineAPIException(f"Invalid response: {e}") from e
- parsers = {'news': parse_news, 'images': parse_images, 'videos': parse_videos}
- return parsers[chinaso_category](data)
- def parse_news(data):
- results = []
- if not data.get("data", {}).get("data"):
- raise SearxEngineAPIException("Invalid response")
- for entry in data["data"]["data"]:
- published_date = None
- if entry.get("timestamp"):
- try:
- published_date = datetime.fromtimestamp(int(entry["timestamp"]))
- except (ValueError, TypeError):
- pass
- results.append(
- {
- 'title': html_to_text(entry["title"]),
- 'url': entry["url"],
- 'content': html_to_text(entry["snippet"]),
- 'publishedDate': published_date,
- }
- )
- return results
- def parse_images(data):
- results = []
- if not data.get("data", {}).get("arrRes"):
- raise SearxEngineAPIException("Invalid response")
- for entry in data["data"]["arrRes"]:
- results.append(
- {
- 'url': entry["web_url"],
- 'title': html_to_text(entry["title"]),
- 'content': html_to_text(entry["ImageInfo"]),
- 'template': 'images.html',
- 'img_src': entry["url"].replace("http://", "https://"),
- 'thumbnail_src': entry["largeimage"].replace("http://", "https://"),
- }
- )
- return results
- def parse_videos(data):
- results = []
- if not data.get("data", {}).get("arrRes"):
- raise SearxEngineAPIException("Invalid response")
- for entry in data["data"]["arrRes"]:
- published_date = None
- if entry.get("VideoPubDate"):
- try:
- published_date = datetime.fromtimestamp(int(entry["VideoPubDate"]))
- except (ValueError, TypeError):
- pass
- results.append(
- {
- 'url': entry["url"],
- 'title': html_to_text(entry["raw_title"]),
- 'template': 'videos.html',
- 'publishedDate': published_date,
- 'thumbnail': entry["image_src"].replace("http://", "https://"),
- }
- )
- return results
|