peertube.py 2.7 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192
  1. # SPDX-License-Identifier: AGPL-3.0-or-later
  2. """
  3. peertube (Videos)
  4. """
  5. from json import loads
  6. from datetime import datetime
  7. from urllib.parse import urlencode
  8. from searx.utils import html_to_text
  9. # about
  10. about = {
  11. "website": 'https://joinpeertube.org',
  12. "wikidata_id": 'Q50938515',
  13. "official_api_documentation": 'https://docs.joinpeertube.org/api-rest-reference.html',
  14. "use_official_api": True,
  15. "require_api_key": False,
  16. "results": 'JSON',
  17. }
  18. # engine dependent config
  19. categories = ["videos"]
  20. paging = True
  21. base_url = "https://peer.tube"
  22. supported_languages_url = (
  23. 'https://framagit.org/framasoft/peertube/search-index/-/raw/master/client/src/views/Search.vue'
  24. )
  25. # do search-request
  26. def request(query, params):
  27. sanitized_url = base_url.rstrip("/")
  28. pageno = (params["pageno"] - 1) * 15
  29. search_url = sanitized_url + "/api/v1/search/videos/?pageno={pageno}&{query}"
  30. query_dict = {"search": query}
  31. language = params["language"].split("-")[0]
  32. if "all" != language and language in supported_languages:
  33. query_dict["languageOneOf"] = language
  34. params["url"] = search_url.format(query=urlencode(query_dict), pageno=pageno)
  35. return params
  36. def _get_offset_from_pageno(pageno):
  37. return (pageno - 1) * 15 + 1
  38. # get response from search-request
  39. def response(resp):
  40. sanitized_url = base_url.rstrip("/")
  41. results = []
  42. search_res = loads(resp.text)
  43. # return empty array if there are no results
  44. if "data" not in search_res:
  45. return []
  46. # parse results
  47. for res in search_res["data"]:
  48. title = res["name"]
  49. url = sanitized_url + "/videos/watch/" + res["uuid"]
  50. description = res["description"]
  51. if description:
  52. content = html_to_text(res["description"])
  53. else:
  54. content = ""
  55. thumbnail = sanitized_url + res["thumbnailPath"]
  56. publishedDate = datetime.strptime(res["publishedAt"], "%Y-%m-%dT%H:%M:%S.%fZ")
  57. results.append(
  58. {
  59. "template": "videos.html",
  60. "url": url,
  61. "title": title,
  62. "content": content,
  63. "publishedDate": publishedDate,
  64. "iframe_src": sanitized_url + res["embedPath"],
  65. "thumbnail": thumbnail,
  66. }
  67. )
  68. # return results
  69. return results
  70. def _fetch_supported_languages(resp):
  71. import re
  72. # https://docs.python.org/3/howto/regex.html#greedy-versus-non-greedy
  73. videolanguages = re.search(r"videoLanguages \(\)[^\n]+(.*?)\]", resp.text, re.DOTALL)
  74. peertube_languages = [m.group(1) for m in re.finditer(r"\{ id: '([a-z]+)', label:", videolanguages.group(1))]
  75. return peertube_languages