peertube.py 2.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899
  1. # SPDX-License-Identifier: AGPL-3.0-or-later
  2. """
  3. peertube (Videos)
  4. """
  5. from json import loads
  6. from datetime import datetime
  7. from urllib.parse import urlencode
  8. from searx.utils import html_to_text
  9. # about
  10. about = {
  11. "website": 'https://joinpeertube.org',
  12. "wikidata_id": 'Q50938515',
  13. "official_api_documentation": 'https://docs.joinpeertube.org/api-rest-reference.html',
  14. "use_official_api": True,
  15. "require_api_key": False,
  16. "results": 'JSON',
  17. }
  18. # engine dependent config
  19. categories = ["videos"]
  20. paging = True
  21. base_url = "https://peer.tube"
  22. supported_languages_url = (
  23. 'https://framagit.org/framasoft/peertube/search-index/-/raw/master/client/src/views/Search.vue'
  24. )
  25. # do search-request
  26. def request(query, params):
  27. sanitized_url = base_url.rstrip("/")
  28. pageno = (params["pageno"] - 1) * 15
  29. search_url = sanitized_url + "/api/v1/search/videos/?pageno={pageno}&{query}"
  30. query_dict = {"search": query}
  31. language = params["language"].split("-")[0]
  32. if "all" != language and language in supported_languages:
  33. query_dict["languageOneOf"] = language
  34. params["url"] = search_url.format(query=urlencode(query_dict), pageno=pageno)
  35. return params
  36. def _get_offset_from_pageno(pageno):
  37. return (pageno - 1) * 15 + 1
  38. # get response from search-request
  39. def response(resp):
  40. sanitized_url = base_url.rstrip("/")
  41. results = []
  42. search_res = loads(resp.text)
  43. embedded_url = (
  44. '<iframe width="560" height="315" sandbox="allow-same-origin allow-scripts allow-popups" '
  45. + 'src="'
  46. + sanitized_url
  47. + '{embed_path}" frameborder="0" allowfullscreen></iframe>'
  48. )
  49. # return empty array if there are no results
  50. if "data" not in search_res:
  51. return []
  52. # parse results
  53. for res in search_res["data"]:
  54. title = res["name"]
  55. url = sanitized_url + "/videos/watch/" + res["uuid"]
  56. description = res["description"]
  57. if description:
  58. content = html_to_text(res["description"])
  59. else:
  60. content = ""
  61. thumbnail = sanitized_url + res["thumbnailPath"]
  62. publishedDate = datetime.strptime(res["publishedAt"], "%Y-%m-%dT%H:%M:%S.%fZ")
  63. embedded = embedded_url.format(embed_path=res["embedPath"])
  64. results.append(
  65. {
  66. "template": "videos.html",
  67. "url": url,
  68. "title": title,
  69. "content": content,
  70. "publishedDate": publishedDate,
  71. "embedded": embedded,
  72. "thumbnail": thumbnail,
  73. }
  74. )
  75. # return results
  76. return results
  77. def _fetch_supported_languages(resp):
  78. import re
  79. # https://docs.python.org/3/howto/regex.html#greedy-versus-non-greedy
  80. videolanguages = re.search(r"videoLanguages \(\)[^\n]+(.*?)\]", resp.text, re.DOTALL)
  81. peertube_languages = [m.group(1) for m in re.finditer(r"\{ id: '([a-z]+)', label:", videolanguages.group(1))]
  82. return peertube_languages