forked from zaclys/searxng
		
	Merge branch 'master' into engines/unsplash
This commit is contained in:
		
						commit
						4a127b19de
					
				
					 5 changed files with 27242 additions and 135 deletions
				
			
		|  | @ -32,6 +32,7 @@ RUN echo "@commuedge http://nl.alpinelinux.org/alpine/edge/community" >> /etc/ap | |||
|     openssl-dev \ | ||||
|     ca-certificates \ | ||||
|     tini@commuedge \ | ||||
|  && pip install --upgrade pip \ | ||||
|  && pip install --no-cache -r requirements.txt \ | ||||
|  && apk del \ | ||||
|     build-base \ | ||||
|  |  | |||
										
											
												File diff suppressed because one or more lines are too long
											
										
									
								
							|  | @ -1,115 +0,0 @@ | |||
| """ | ||||
| FindX (General, Images, Videos) | ||||
| 
 | ||||
| @website     https://www.findx.com | ||||
| @provide-api no | ||||
| @using-api   no | ||||
| @results     HTML | ||||
| @stable      no | ||||
| @parse       url, title, content, embedded, img_src, thumbnail_src | ||||
| """ | ||||
| 
 | ||||
| from dateutil import parser | ||||
| from json import loads | ||||
| import re | ||||
| 
 | ||||
| from lxml import html | ||||
| 
 | ||||
| from searx import logger | ||||
| from searx.engines.xpath import extract_text | ||||
| from searx.engines.youtube_noapi import base_youtube_url, embedded_url | ||||
| from searx.url_utils import urlencode | ||||
| 
 | ||||
| 
 | ||||
| paging = True | ||||
| results_xpath = '//script[@id="initial-state"]' | ||||
| search_url = 'https://www.findx.com/{category}?{q}' | ||||
| type_map = { | ||||
|     'none': 'web', | ||||
|     'general': 'web', | ||||
|     'images': 'images', | ||||
|     'videos': 'videos', | ||||
| } | ||||
| 
 | ||||
| 
 | ||||
| def request(query, params): | ||||
|     params['url'] = search_url.format( | ||||
|         category=type_map[params['category']], | ||||
|         q=urlencode({ | ||||
|             'q': query, | ||||
|             'page': params['pageno'] | ||||
|         }) | ||||
|     ) | ||||
|     return params | ||||
| 
 | ||||
| 
 | ||||
| def response(resp): | ||||
|     dom = html.fromstring(resp.text) | ||||
|     results_raw_json = dom.xpath(results_xpath) | ||||
|     results_json = loads(extract_text(results_raw_json)) | ||||
| 
 | ||||
|     if len(results_json['web']['results']) > 0: | ||||
|         return _general_results(results_json['web']['results']['webSearch']['results']) | ||||
| 
 | ||||
|     if len(results_json['images']['results']) > 0: | ||||
|         return _images_results(results_json['images']['results']) | ||||
| 
 | ||||
|     if len(results_json['video']['results']) > 0: | ||||
|         return _videos_results(results_json['video']['results']) | ||||
| 
 | ||||
|     return [] | ||||
| 
 | ||||
| 
 | ||||
| def _general_results(general_results): | ||||
|     results = [] | ||||
|     for result in general_results: | ||||
|         results.append({ | ||||
|             'url': result['url'], | ||||
|             'title': result['title'], | ||||
|             'content': result['sum'], | ||||
|         }) | ||||
|     return results | ||||
| 
 | ||||
| 
 | ||||
| def _images_results(image_results): | ||||
|     results = [] | ||||
|     for result in image_results: | ||||
|         results.append({ | ||||
|             'url': result['sourceURL'], | ||||
|             'title': result['title'], | ||||
|             'content': result['source'], | ||||
|             'thumbnail_src': _extract_url(result['assets']['thumb']['url']), | ||||
|             'img_src': _extract_url(result['assets']['file']['url']), | ||||
|             'template': 'images.html', | ||||
|         }) | ||||
|     return results | ||||
| 
 | ||||
| 
 | ||||
| def _videos_results(video_results): | ||||
|     results = [] | ||||
|     for result in video_results: | ||||
|         if not result['kind'].startswith('youtube'): | ||||
|             logger.warn('Unknown video kind in findx: {}'.format(result['kind'])) | ||||
|             continue | ||||
| 
 | ||||
|         description = result['snippet']['description'] | ||||
|         if len(description) > 300: | ||||
|             description = description[:300] + '...' | ||||
| 
 | ||||
|         results.append({ | ||||
|             'url': base_youtube_url + result['id'], | ||||
|             'title': result['snippet']['title'], | ||||
|             'content': description, | ||||
|             'thumbnail': _extract_url(result['snippet']['thumbnails']['default']['url']), | ||||
|             'publishedDate': parser.parse(result['snippet']['publishedAt']), | ||||
|             'embedded': embedded_url.format(videoid=result['id']), | ||||
|             'template': 'videos.html', | ||||
|         }) | ||||
|     return results | ||||
| 
 | ||||
| 
 | ||||
| def _extract_url(url): | ||||
|     matching = re.search('(/https?://[^)]+)', url) | ||||
|     if matching: | ||||
|         return matching.group(0)[1:] | ||||
|     return '' | ||||
|  | @ -218,24 +218,6 @@ engines: | |||
|     shortcut : fd | ||||
|     disabled : True | ||||
| 
 | ||||
|   - name : findx | ||||
|     engine : findx | ||||
|     shortcut : fx | ||||
|     categories : general | ||||
|     disabled : True | ||||
| 
 | ||||
|   - name : findx images | ||||
|     engine : findx | ||||
|     shortcut : fxi | ||||
|     categories : images | ||||
|     disabled : True | ||||
| 
 | ||||
|   - name : findx videos | ||||
|     engine : findx | ||||
|     shortcut : fxv | ||||
|     categories : videos | ||||
|     disabled : True | ||||
| 
 | ||||
|   - name : flickr | ||||
|     categories : images | ||||
|     shortcut : fl | ||||
|  | @ -728,6 +710,19 @@ engines: | |||
|     shortcut : du | ||||
|     disabled : True | ||||
| 
 | ||||
|   - name : seznam | ||||
|     shortcut: szn | ||||
|     engine: xpath | ||||
|     paging : True | ||||
|     search_url : https://search.seznam.cz/?q={query}&count=10&from={pageno} | ||||
|     results_xpath: //div[@class="Page-content"]//div[@class="Result "] | ||||
|     url_xpath : ./h3/a/@href | ||||
|     title_xpath : ./h3 | ||||
|     content_xpath : .//p[@class="Result-description"] | ||||
|     first_page_num : 0 | ||||
|     page_size : 10 | ||||
|     disabled : True | ||||
| 
 | ||||
| #  - name : yacy | ||||
| #    engine : yacy | ||||
| #    shortcut : ya | ||||
|  |  | |||
|  | @ -27,12 +27,14 @@ def fetch_supported_languages(): | |||
|         if hasattr(engines[engine_name], 'fetch_supported_languages'): | ||||
|             try: | ||||
|                 engines_languages[engine_name] = engines[engine_name].fetch_supported_languages() | ||||
|                 if type(engines_languages[engine_name]) == list: | ||||
|                     engines_languages[engine_name] = sorted(engines_languages[engine_name]) | ||||
|             except Exception as e: | ||||
|                 print(e) | ||||
| 
 | ||||
|     # write json file | ||||
|     with io.open(engines_languages_file, "w", encoding="utf-8") as f: | ||||
|         dump(engines_languages, f, ensure_ascii=False) | ||||
|         dump(engines_languages, f, ensure_ascii=False, indent=4, separators=(',', ': ')) | ||||
| 
 | ||||
|     return engines_languages | ||||
| 
 | ||||
|  |  | |||
		Loading…
	
	Add table
		
		Reference in a new issue
	
	 d-tux
						d-tux