| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159 | 
							- """
 
-  Yahoo (Web)
 
-  @website     https://search.yahoo.com/web
 
-  @provide-api yes (https://developer.yahoo.com/boss/search/),
 
-               $0.80/1000 queries
 
-  @using-api   no (because pricing)
 
-  @results     HTML (using search portal)
 
-  @stable      no (HTML can change)
 
-  @parse       url, title, content, suggestion
 
- """
 
- from urllib.parse import unquote, urlencode
 
- from lxml import html
 
- from searx.utils import extract_text, extract_url, match_language, eval_xpath
 
- # engine dependent config
 
- categories = ['general']
 
- paging = True
 
- language_support = True
 
- time_range_support = True
 
- # search-url
 
- base_url = 'https://search.yahoo.com/'
 
- search_url = 'search?{query}&b={offset}&fl=1&vl=lang_{lang}'
 
- search_url_with_time = 'search?{query}&b={offset}&fl=1&vl=lang_{lang}&age={age}&btf={btf}&fr2=time'
 
- supported_languages_url = 'https://search.yahoo.com/web/advanced'
 
- # specific xpath variables
 
- results_xpath = "//div[contains(concat(' ', normalize-space(@class), ' '), ' Sr ')]"
 
- url_xpath = './/h3/a/@href'
 
- title_xpath = './/h3/a'
 
- content_xpath = './/div[contains(@class, "compText")]'
 
- suggestion_xpath = "//div[contains(concat(' ', normalize-space(@class), ' '), ' AlsoTry ')]//a"
 
- time_range_dict = {'day': ['1d', 'd'],
 
-                    'week': ['1w', 'w'],
 
-                    'month': ['1m', 'm']}
 
- language_aliases = {'zh-CN': 'zh-CHS', 'zh-TW': 'zh-CHT', 'zh-HK': 'zh-CHT'}
 
- # remove yahoo-specific tracking-url
 
- def parse_url(url_string):
 
-     endings = ['/RS', '/RK']
 
-     endpositions = []
 
-     start = url_string.find('http', url_string.find('/RU=') + 1)
 
-     for ending in endings:
 
-         endpos = url_string.rfind(ending)
 
-         if endpos > -1:
 
-             endpositions.append(endpos)
 
-     if start == 0 or len(endpositions) == 0:
 
-         return url_string
 
-     else:
 
-         end = min(endpositions)
 
-         return unquote(url_string[start:end])
 
- def _get_url(query, offset, language, time_range):
 
-     if time_range in time_range_dict:
 
-         return base_url + search_url_with_time.format(offset=offset,
 
-                                                       query=urlencode({'p': query}),
 
-                                                       lang=language,
 
-                                                       age=time_range_dict[time_range][0],
 
-                                                       btf=time_range_dict[time_range][1])
 
-     return base_url + search_url.format(offset=offset,
 
-                                         query=urlencode({'p': query}),
 
-                                         lang=language)
 
- def _get_language(params):
 
-     if params['language'] == 'all':
 
-         return 'en'
 
-     language = match_language(params['language'], supported_languages, language_aliases)
 
-     if language not in language_aliases.values():
 
-         language = language.split('-')[0]
 
-     language = language.replace('-', '_').lower()
 
-     return language
 
- # do search-request
 
- def request(query, params):
 
-     if params['time_range'] and params['time_range'] not in time_range_dict:
 
-         return params
 
-     offset = (params['pageno'] - 1) * 10 + 1
 
-     language = _get_language(params)
 
-     params['url'] = _get_url(query, offset, language, params['time_range'])
 
-     # TODO required?
 
-     params['cookies']['sB'] = 'fl=1&vl=lang_{lang}&sh=1&rw=new&v=1'\
 
-         .format(lang=language)
 
-     return params
 
- # get response from search-request
 
- def response(resp):
 
-     results = []
 
-     dom = html.fromstring(resp.text)
 
-     try:
 
-         results_num = int(eval_xpath(dom, '//div[@class="compPagination"]/span[last()]/text()')[0]
 
-                           .split()[0].replace(',', ''))
 
-         results.append({'number_of_results': results_num})
 
-     except:
 
-         pass
 
-     # parse results
 
-     for result in eval_xpath(dom, results_xpath):
 
-         try:
 
-             url = parse_url(extract_url(eval_xpath(result, url_xpath), search_url))
 
-             title = extract_text(eval_xpath(result, title_xpath)[0])
 
-         except:
 
-             continue
 
-         content = extract_text(eval_xpath(result, content_xpath)[0])
 
-         # append result
 
-         results.append({'url': url,
 
-                         'title': title,
 
-                         'content': content})
 
-     # if no suggestion found, return results
 
-     suggestions = eval_xpath(dom, suggestion_xpath)
 
-     if not suggestions:
 
-         return results
 
-     # parse suggestion
 
-     for suggestion in suggestions:
 
-         # append suggestion
 
-         results.append({'suggestion': extract_text(suggestion)})
 
-     # return results
 
-     return results
 
- # get supported languages from their site
 
- def _fetch_supported_languages(resp):
 
-     supported_languages = []
 
-     dom = html.fromstring(resp.text)
 
-     options = eval_xpath(dom, '//div[@id="yschlang"]/span/label/input')
 
-     for option in options:
 
-         code_parts = eval_xpath(option, './@value')[0][5:].split('_')
 
-         if len(code_parts) == 2:
 
-             code = code_parts[0] + '-' + code_parts[1].upper()
 
-         else:
 
-             code = code_parts[0]
 
-         supported_languages.append(code)
 
-     return supported_languages
 
 
  |