peertube.py 3.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102
  1. # SPDX-License-Identifier: AGPL-3.0-or-later
  2. """
  3. peertube (Videos)
  4. """
  5. from json import loads
  6. from datetime import datetime
  7. from urllib.parse import urlencode
  8. from searx.utils import html_to_text
  9. # about
  10. about = {
  11. "website": 'https://joinpeertube.org',
  12. "wikidata_id": 'Q50938515',
  13. "official_api_documentation": 'https://docs.joinpeertube.org/api-rest-reference.html',
  14. "use_official_api": True,
  15. "require_api_key": False,
  16. "results": 'JSON',
  17. }
  18. # engine dependent config
  19. categories = ["videos"]
  20. paging = True
  21. base_url = "https://peer.tube"
  22. supported_languages_url = (
  23. 'https://framagit.org/framasoft/peertube/search-index/-/raw/master/client/src/views/Search.vue'
  24. )
  25. # do search-request
  26. def request(query, params):
  27. sanitized_url = base_url.rstrip("/")
  28. pageno = (params["pageno"] - 1) * 15
  29. search_url = sanitized_url + "/api/v1/search/videos/?pageno={pageno}&{query}"
  30. query_dict = {"search": query}
  31. language = params["language"].split("-")[0]
  32. # pylint: disable=undefined-variable
  33. if "all" != language and language in supported_languages:
  34. query_dict["languageOneOf"] = language
  35. params["url"] = search_url.format(
  36. query=urlencode(query_dict), pageno=pageno
  37. )
  38. return params
  39. def _get_offset_from_pageno(pageno):
  40. return (pageno - 1) * 15 + 1
  41. # get response from search-request
  42. def response(resp):
  43. sanitized_url = base_url.rstrip("/")
  44. results = []
  45. search_res = loads(resp.text)
  46. embedded_url = (
  47. '<iframe width="560" height="315" sandbox="allow-same-origin allow-scripts allow-popups" '
  48. + 'src="'
  49. + sanitized_url
  50. + '{embed_path}" frameborder="0" allowfullscreen></iframe>'
  51. )
  52. # return empty array if there are no results
  53. if "data" not in search_res:
  54. return []
  55. # parse results
  56. for res in search_res["data"]:
  57. title = res["name"]
  58. url = sanitized_url + "/videos/watch/" + res["uuid"]
  59. description = res["description"]
  60. if description:
  61. content = html_to_text(res["description"])
  62. else:
  63. content = ""
  64. thumbnail = sanitized_url + res["thumbnailPath"]
  65. publishedDate = datetime.strptime(res["publishedAt"], "%Y-%m-%dT%H:%M:%S.%fZ")
  66. embedded = embedded_url.format(embed_path=res["embedPath"])
  67. results.append(
  68. {
  69. "template": "videos.html",
  70. "url": url,
  71. "title": title,
  72. "content": content,
  73. "publishedDate": publishedDate,
  74. "embedded": embedded,
  75. "thumbnail": thumbnail,
  76. }
  77. )
  78. # return results
  79. return results
  80. def _fetch_supported_languages(resp):
  81. import re
  82. # https://docs.python.org/3/howto/regex.html#greedy-versus-non-greedy
  83. videolanguages = re.search(r"videoLanguages \(\)[^\n]+(.*?)\]", resp.text, re.DOTALL)
  84. peertube_languages = [m.group(1) for m in re.finditer(r"\{ id: '([a-z]+)', label:", videolanguages.group(1))]
  85. return peertube_languages