libremiami-search/searx/engines/flickr.py

49 lines
2.1 KiB
Python

#!/usr/bin/env python
from urllib import urlencode
from json import loads
#from urlparse import urljoin
categories = ['images']
# url = 'https://secure.flickr.com/'
# search_url = url+'search/?{query}&page={page}'
# results_xpath = '//div[@id="thumbnails"]//a[@class="rapidnofollow photo-click" and @data-track="photo-click"]' # noqa
paging = True
# text=[query]
# TODO clean "extras"
search_url = 'https://api.flickr.com/services/rest?extras=can_addmeta%2Ccan_comment%2Ccan_download%2Ccan_share%2Ccontact%2Ccount_comments%2Ccount_faves%2Ccount_notes%2Cdate_taken%2Cdate_upload%2Cdescription%2Cicon_urls_deep%2Cisfavorite%2Cispro%2Clicense%2Cmedia%2Cneeds_interstitial%2Cowner_name%2Cowner_datecreate%2Cpath_alias%2Crealname%2Csafety_level%2Csecret_k%2Csecret_h%2Curl_c%2Curl_h%2Curl_k%2Curl_l%2Curl_m%2Curl_n%2Curl_o%2Curl_q%2Curl_s%2Curl_sq%2Curl_t%2Curl_z%2Cvisibility&per_page=50&page={page}&{query}&sort=relevance&method=flickr.photos.search&api_key=ad11b34c341305471e3c410a02e671d0&format=json' # noqa
def request(query, params):
params['url'] = search_url.format(query=urlencode({'text': query}),
page=params['pageno'])
#params['url'] = search_url.format(query=urlencode({'q': query}),
# page=params['pageno'])
return params
def response(resp):
results = []
images = loads(resp.text[14:-1])["photos"]["photo"]
for i in images:
results.append({'url': i['url_s'],
'title': i['title'],
'img_src': i['url_s'],
'template': 'images.html'})
#dom = html.fromstring(resp.text)
#for result in dom.xpath(results_xpath):
# href = urljoin(url, result.attrib.get('href'))
# img = result.xpath('.//img')[0]
# title = img.attrib.get('alt', '')
# img_src = img.attrib.get('data-defer-src')
# if not img_src:
# continue
# results.append({'url': href,
# 'title': title,
# 'img_src': img_src,
# 'template': 'images.html'})
return results