summaryrefslogtreecommitdiff
path: root/searx/engines/filecrop.py
blob: b2adff235efad5c46abc5556a0289fc9be2858ae (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
from json import loads
from urllib import urlencode
from searx.utils import html_to_text
from HTMLParser import HTMLParser

url = 'http://www.filecrop.com/'
search_url = url + '/search.php?w={query}&size_i=0&size_f=100000000&engine_r=1&engine_d=1&engine_e=1&engine_4=1&engine_m=1'

class FilecropResultParser(HTMLParser):
    def __init__(self):
        HTMLParser.__init__(self)
        self.__start_processing = False
        
        self.results = []
        self.result = {}

        self.tr_counter = 0
        self.data_counter = 0

    def handle_starttag(self, tag, attrs):

        if tag == 'tr':
            if ('bgcolor', '#edeff5') in attrs or ('bgcolor', '#ffffff') in attrs:
                self.__start_processing = True
                
        if not self.__start_processing:
            return

        if tag == 'label':
            self.result['title'] = [attr[1] for attr in attrs if attr[0] == 'title'][0]
        elif tag == 'a' and ('rel', 'nofollow') in attrs and ('class', 'sourcelink') in attrs:
            if 'content' in self.result:
                self.result['content'] += [attr[1] for attr in attrs if attr[0] == 'title'][0]
            else:
                self.result['content'] = [attr[1] for attr in attrs if attr[0] == 'title'][0]
            self.result['content'] += ' '
        elif tag == 'a':
            self.result['url'] = url + [attr[1] for attr in attrs if attr[0] == 'href'][0]

    def handle_endtag(self, tag):
        if self.__start_processing is False:
            return

        if tag == 'tr':
            self.tr_counter += 1

            if self.tr_counter == 2:
                self.__start_processing = False
                self.tr_counter = 0
                self.data_counter = 0
                self.results.append(self.result)
                self.result = {}
                                
    def handle_data(self, data):
        if not self.__start_processing:
            return

        if 'content' in self.result:
            self.result['content'] += data + ' '
        else:
            self.result['content'] = data + ' '
        
        self.data_counter += 1

def request(query, params):
    params['url'] = search_url.format(query=urlencode({'q': query}))
    return params

def response(resp):
    parser = FilecropResultParser()
    parser.feed(resp.text)

    return parser.results