#!/usr/bin/env python # SPDX-License-Identifier: AGPL-3.0-or-later # lint: pylint """This script fetches engine data from engines `engine_data_url`` and updates: - :py:obj:`write_languages_file` updates :origin:`searx/languages.py` - :py:obj:`fetch_engine_data` updates :origin:`searx/data/engines_datas.json` This script is triggered by CI in job :origin:`updateData <.github/workflows/data-update.yml>`. """ # pylint: disable=invalid-name from unicodedata import lookup from pprint import pformat from babel import Locale, UnknownLocaleError from babel.languages import get_global from babel.core import parse_locale from searx import settings from searx import network from searx.engines import ( load_engines, engines, EngineData, EngineDataDict, ) from searx.locales import LANGUAGES_FILE from searx.utils import gen_useragent def fetch_engine_data(): """Fetch :class:`EngineData` for each engine and persist JSON in file. The script checks all engines about a function:: def _fetch_engine_data(resp, engine_data): ... and a variable named ``engine_data_url``. The HTTP GET response of ``engine_data_url`` is passed to the ``_fetch_engine_data`` function including a instance of :py:obj:`searx.engines.EngineData`. .. hint:: This implementation is backward compatible and supports the (depricated) ``_fetch_supported_languages`` interface. On the long term the depricated implementations in the engines will be replaced by ``_fetch_engine_data``.""" network.set_timeout_for_thread(10.0) engine_data_dict = EngineDataDict() names = list(engines) names.sort() # The headers has been moved here from commit 9b6ffed06: Some engines (at # least bing and startpage) return a different result list of supported # languages depending on the IP location where the HTTP request comes from. # The IP based results (from bing) can be avoided by setting a # 'Accept-Language' in the HTTP request. headers = { 'User-Agent': gen_useragent(), 'Accept-Language': "en-US,en;q=0.5", # bing needs to set the English language } for engine_name in names: engine = engines[engine_name] fetch_data = getattr(engine, '_fetch_engine_data', None) # depricated: _fetch_supported_languages fetch_languages = getattr(engine, '_fetch_supported_languages', None) if fetch_data is not None: # data_type = 'engine_data' engine_data = EngineData() resp = network.get(engine.engine_data_url, headers=headers) fetch_data(resp, engine_data) print("%s: %s languages" % (engine_name, len(engine_data.languages))) print("%s: %s regions" % (engine_name, len(engine_data.regions))) elif fetch_languages is not None: # depricated: data_type = 'supported_languages' print("%s: using deepricated _fetch_supported_languages" % engine_name) resp = network.get(engine.supported_languages_url, headers=headers) engine_languages = fetch_languages(resp) if isinstance(engine_languages, list): engine_languages.sort() print( "%s: fetched language %s containing %s items" % (engine_name, engine_languages.__class__.__name__, len(engine_languages)) ) engine_data = EngineData(data_type='supported_languages') engine_data.languages = engine_languages else: continue engine_data_dict[engine_name] = engine_data print("fetched properties from %s engines" % len(engine_data_dict)) print("write json file: %s" % (engine_data_dict.ENGINE_DATA_FILE)) engine_data_dict.save_data() return engine_data_dict # Get babel Locale object from lang_code if possible. def get_locale(lang_code): try: locale = Locale.parse(lang_code, sep='-') return locale except (UnknownLocaleError, ValueError): return None lang2emoji = { 'ha': '\U0001F1F3\U0001F1EA', # Hausa / Niger 'bs': '\U0001F1E7\U0001F1E6', # Bosnian / Bosnia & Herzegovina 'jp': '\U0001F1EF\U0001F1F5', # Japanese 'ua': '\U0001F1FA\U0001F1E6', # Ukrainian 'he': '\U0001F1EE\U0001F1F7', # Hebrew } def get_unicode_flag(lang_code): """Determine a unicode flag (emoji) that fits to the ``lang_code``""" emoji = lang2emoji.get(lang_code.lower()) if emoji: return emoji if len(lang_code) == 2: return '\U0001F310' language = territory = script = variant = '' try: language, territory, script, variant = parse_locale(lang_code, '-') except ValueError as exc: print(exc) # https://en.wikipedia.org/wiki/ISO_3166-1_alpha-2 if not territory: # https://www.unicode.org/emoji/charts/emoji-list.html#country-flag emoji = lang2emoji.get(language) if not emoji: print( "%s --> language: %s / territory: %s / script: %s / variant: %s" % (lang_code, language, territory, script, variant) ) return emoji emoji = lang2emoji.get(territory.lower()) if emoji: return emoji try: c1 = lookup('REGIONAL INDICATOR SYMBOL LETTER ' + territory[0]) c2 = lookup('REGIONAL INDICATOR SYMBOL LETTER ' + territory[1]) # print("%s --> territory: %s --> %s%s" %(lang_code, territory, c1, c2 )) except KeyError as exc: print("%s --> territory: %s --> %s" % (lang_code, territory, exc)) return None return c1 + c2 def get_territory_name(lang_code): country_name = None locale = get_locale(lang_code) try: if locale is not None: country_name = locale.get_territory_name() except FileNotFoundError as exc: print("ERROR: %s --> %s" % (locale, exc)) return country_name def join_language_lists(engine_data_dict): """Join all languages of the engines into one list. The returned language list contains language codes (``zh``) and region codes (``zh-TW``). The codes can be parsed by babel:: babel.Locale.parse(language_list[n]) """ # pylint: disable=too-many-branches language_list = {} for engine_name in engine_data_dict: engine = engines[engine_name] engine_data = engine_data_dict[engine_name] if engine_data.data_type == 'engine_data': # items of type 'engine_data' do have regions & languages, the list # of engine_codes should contain both. engine_codes = engine_data.regions engine_codes.update(engine_data.languages) engine_codes = engine_codes.keys() elif engine_data.data_type == 'supported_languages': # depricated engine_languages = engine_data.languages if isinstance(engine_languages, dict): engine_languages = engine_languages.keys() language_aliases_values = getattr(engine, 'language_aliases', {}).values() engine_codes = [] for lang_code in engine_languages: if lang_code in language_aliases_values: lang_code = next(lc for lc, alias in engine.language_aliases.items() if lang_code == alias) engine_codes.append(lang_code) else: raise TypeError('unknown type of engine data: %s' % engine_data.data_type) for lang_code in engine_codes: locale = get_locale(lang_code) # ensure that lang_code uses standard language and country codes if locale and locale.territory: lang_code = "{lang}-{country}".format(lang=locale.language, country=locale.territory) short_code = lang_code.split('-')[0] # add language without country if not in list if short_code not in language_list: if locale: # get language's data from babel's Locale object language_name = locale.get_language_name().title() english_name = locale.english_name.split(' (')[0] elif short_code in engine_data_dict['wikipedia'].languages: # get language's data from wikipedia if not known by babel language_name = engine_data_dict['wikipedia'].languages[short_code]['name'] english_name = engine_data_dict['wikipedia'].languages[short_code]['english_name'] else: language_name = None english_name = None # add language to list language_list[short_code] = { 'name': language_name, 'english_name': english_name, 'counter': set(), 'countries': {}, } # add language with country if not in list if lang_code != short_code and lang_code not in language_list[short_code]['countries']: country_name = '' if locale: # get country name from babel's Locale object try: country_name = locale.get_territory_name() except FileNotFoundError as exc: print("ERROR: %s --> %s" % (locale, exc)) locale = None language_list[short_code]['countries'][lang_code] = { 'country_name': country_name, 'counter': set(), } # count engine for both language_country combination and language alone language_list[short_code]['counter'].add(engine_name) if lang_code != short_code: language_list[short_code]['countries'][lang_code]['counter'].add(engine_name) return language_list def filter_language_list(all_languages): """Filter language list so it only includes the most supported languages and countries. """ min_engines_per_lang = 12 min_engines_per_country = 7 # pylint: disable=consider-using-dict-items, consider-iterating-dictionary main_engines = [ engine_name for engine_name in engines.keys() if 'general' in engines[engine_name].categories and hasattr(engines[engine_name], 'supported_languages') and engines[engine_name].supported_languages and not engines[engine_name].disabled ] # filter list to include only languages supported by most engines or all default general engines filtered_languages = { code: lang for code, lang in all_languages.items() if ( len(lang['counter']) >= min_engines_per_lang or all(main_engine in lang['counter'] for main_engine in main_engines) ) } def _copy_lang_data(lang, country_name=None): new_dict = {} new_dict['name'] = all_languages[lang]['name'] new_dict['english_name'] = all_languages[lang]['english_name'] if country_name: new_dict['country_name'] = country_name return new_dict # for each language get country codes supported by most engines or at least one country code filtered_languages_with_countries = {} for lang, lang_data in filtered_languages.items(): countries = lang_data['countries'] filtered_countries = {} # get language's country codes with enough supported engines for lang_country, country_data in countries.items(): if len(country_data['counter']) >= min_engines_per_country: filtered_countries[lang_country] = _copy_lang_data(lang, country_data['country_name']) # add language without countries too if there's more than one country to choose from if len(filtered_countries) > 1: filtered_countries[lang] = _copy_lang_data(lang, None) elif len(filtered_countries) == 1: lang_country = next(iter(filtered_countries)) # if no country has enough engines try to get most likely country code from babel if not filtered_countries: lang_country = None subtags = get_global('likely_subtags').get(lang) if subtags: country_code = subtags.split('_')[-1] if len(country_code) == 2: lang_country = "{lang}-{country}".format(lang=lang, country=country_code) if lang_country: filtered_countries[lang_country] = _copy_lang_data(lang, None) else: filtered_countries[lang] = _copy_lang_data(lang, None) filtered_languages_with_countries.update(filtered_countries) return filtered_languages_with_countries class UnicodeEscape(str): """Escape unicode string in :py:obj:`pprint.pformat`""" def __repr__(self): return "'" + "".join([chr(c) for c in self.encode('unicode-escape')]) + "'" # Write languages.py. def write_languages_file(languages): """Generates :origin:`searx/languages.py`.""" file_headers = ( "# -*- coding: utf-8 -*-", "# list of language codes", "# this file is generated automatically by:", "#", "# ./manage pyenv.cmd searxng_extra/update/update_languages.py", "language_codes = (\n", ) language_codes = [] for code in sorted(languages): name = languages[code]['name'] if name is None: print("ERROR: languages['%s'] --> %s" % (code, languages[code])) continue flag = get_unicode_flag(code) or '' item = ( code, languages[code]['name'].split(' (')[0], get_territory_name(code) or '', languages[code].get('english_name') or '', UnicodeEscape(flag), ) language_codes.append(item) language_codes = tuple(language_codes) with open(LANGUAGES_FILE, 'w', encoding='utf-8') as new_file: file_content = "{file_headers} {language_codes},\n)\n".format( # fmt: off file_headers = '\n'.join(file_headers), language_codes = pformat(language_codes, indent=4)[1:-1] # fmt: on ) new_file.write(file_content) new_file.close() if __name__ == "__main__": load_engines(settings['engines']) _engine_data_dict = fetch_engine_data() _all_languages = join_language_lists(_engine_data_dict) _filtered_languages = filter_language_list(_all_languages) write_languages_file(_filtered_languages)