Browse Source

[enh] 1st version of filecrop engine, to discover how searx works

pw3t 11 years ago
parent
commit
b71bddad06
2 changed files with 76 additions and 0 deletions
  1. 3 0
      .gitignore
  2. 73 0
      searx/engines/filecrop.py

+ 3 - 0
.gitignore

@@ -1,2 +1,5 @@
 env
 engines.cfg
+
+*.pyc
+*/*.pyc

+ 73 - 0
searx/engines/filecrop.py

@@ -0,0 +1,73 @@
+from json import loads
+from urllib import urlencode
+from searx.utils import html_to_text
+from HTMLParser import HTMLParser
+
+url = 'http://www.filecrop.com/'
+search_url = url + '/search.php?w={query}&size_i=0&size_f=100000000&engine_r=1&engine_d=1&engine_e=1&engine_4=1&engine_m=1'
+
+class FilecropResultParser(HTMLParser):
+    def __init__(self):
+        HTMLParser.__init__(self)
+        self.__start_processing = False
+        
+        self.results = []
+        self.result = {}
+
+        self.tr_counter = 0
+        self.data_counter = 0
+
+    def handle_starttag(self, tag, attrs):
+
+        if tag == 'tr':
+            if ('bgcolor', '#edeff5') in attrs or ('bgcolor', '#ffffff') in attrs:
+                self.__start_processing = True
+                
+        if not self.__start_processing:
+            return
+
+        if tag == 'label':
+            self.result['title'] = [attr[1] for attr in attrs if attr[0] == 'title'][0]
+        elif tag == 'a' and ('rel', 'nofollow') in attrs and ('class', 'sourcelink') in attrs:
+            if 'content' in self.result:
+                self.result['content'] += [attr[1] for attr in attrs if attr[0] == 'title'][0]
+            else:
+                self.result['content'] = [attr[1] for attr in attrs if attr[0] == 'title'][0]
+            self.result['content'] += ' '
+        elif tag == 'a':
+            self.result['url'] = url + [attr[1] for attr in attrs if attr[0] == 'href'][0]
+
+    def handle_endtag(self, tag):
+        if self.__start_processing is False:
+            return
+
+        if tag == 'tr':
+            self.tr_counter += 1
+
+            if self.tr_counter == 2:
+                self.__start_processing = False
+                self.tr_counter = 0
+                self.data_counter = 0
+                self.results.append(self.result)
+                self.result = {}
+                                
+    def handle_data(self, data):
+        if not self.__start_processing:
+            return
+
+        if 'content' in self.result:
+            self.result['content'] += data + ' '
+        else:
+            self.result['content'] = data + ' '
+        
+        self.data_counter += 1
+
+def request(query, params):
+    params['url'] = search_url.format(query=urlencode({'q': query}))
+    return params
+
+def response(resp):
+    parser = FilecropResultParser()
+    parser.feed(resp.text)
+
+    return parser.results