[mod] add documentation about searx.utils

This module is a toolbox for the engines.
Is should be documented.

In addition, searx/utils.py is checked by pylint.
This commit is contained in:
Alexandre Flament 2022-01-29 11:16:28 +01:00
parent f240a67bd7
commit 0eacc46ee3
3 changed files with 104 additions and 100 deletions

8
docs/src/searx.utils.rst Normal file
View file

@ -0,0 +1,8 @@
.. _searx.utils:
=================================
Utility functions for the engines
=================================
.. automodule:: searx.utils
:members:

View file

@ -1,7 +1,13 @@
# -*- coding: utf-8 -*- # SPDX-License-Identifier: AGPL-3.0-or-later
# lint: pylint
# pyright: basic
"""Utility functions for the engines
"""
import re import re
import importlib import importlib
from typing import Optional, Union, Any, Set, List, Dict, MutableMapping, Tuple, Callable
from numbers import Number from numbers import Number
from os.path import splitext, join from os.path import splitext, join
from random import choice from random import choice
@ -23,42 +29,56 @@ from searx import logger
logger = logger.getChild('utils') logger = logger.getChild('utils')
blocked_tags = ('script', 'style') XPathSpecType = Union[str, XPath]
ecma_unescape4_re = re.compile(r'%u([0-9a-fA-F]{4})', re.UNICODE) _BLOCKED_TAGS = ('script', 'style')
ecma_unescape2_re = re.compile(r'%([0-9a-fA-F]{2})', re.UNICODE)
xpath_cache = dict() _ECMA_UNESCAPE4_RE = re.compile(r'%u([0-9a-fA-F]{4})', re.UNICODE)
lang_to_lc_cache = dict() _ECMA_UNESCAPE2_RE = re.compile(r'%([0-9a-fA-F]{2})', re.UNICODE)
_STORAGE_UNIT_VALUE: Dict[str, int] = {
'TB': 1024 * 1024 * 1024 * 1024,
'GB': 1024 * 1024 * 1024,
'MB': 1024 * 1024,
'TiB': 1000 * 1000 * 1000 * 1000,
'MiB': 1000 * 1000,
'KiB': 1000,
}
_XPATH_CACHE = {}
_LANG_TO_LC_CACHE = {}
class NotSetClass: class _NotSetClass: # pylint: disable=too-few-public-methods
pass """Internal class for this module, do not create instance of this class.
Replace the None value, allow explicitly pass None as a function argument"""
NOTSET = NotSetClass() _NOTSET = _NotSetClass()
def searx_useragent(): def searx_useragent() -> str:
"""Return the searx User Agent""" """Return the searx User Agent"""
return 'searx/{searx_version} {suffix}'.format( return 'searx/{searx_version} {suffix}'.format(
searx_version=VERSION_TAG, suffix=settings['outgoing']['useragent_suffix'] searx_version=VERSION_TAG, suffix=settings['outgoing']['useragent_suffix']
).strip() ).strip()
def gen_useragent(os=None): def gen_useragent(os_string: str = None) -> str:
"""Return a random browser User Agent """Return a random browser User Agent
See searx/data/useragents.json See searx/data/useragents.json
""" """
return str(USER_AGENTS['ua'].format(os=os or choice(USER_AGENTS['os']), version=choice(USER_AGENTS['versions']))) return USER_AGENTS['ua'].format(os=os_string or choice(USER_AGENTS['os']), version=choice(USER_AGENTS['versions']))
class HTMLTextExtractorException(Exception): class _HTMLTextExtractorException(Exception):
pass """Internal exception raised when the HTML is invalid"""
class HTMLTextExtractor(HTMLParser): # pylint: disable=W0223 # (see https://bugs.python.org/issue31844) class _HTMLTextExtractor(HTMLParser): # pylint: disable=W0223 # (see https://bugs.python.org/issue31844)
"""Internal class to extract text from HTML"""
def __init__(self): def __init__(self):
HTMLParser.__init__(self) HTMLParser.__init__(self)
self.result = [] self.result = []
@ -72,12 +92,12 @@ class HTMLTextExtractor(HTMLParser): # pylint: disable=W0223 # (see https://bu
return return
if tag != self.tags[-1]: if tag != self.tags[-1]:
raise HTMLTextExtractorException() raise _HTMLTextExtractorException()
self.tags.pop() self.tags.pop()
def is_valid_tag(self): def is_valid_tag(self):
return not self.tags or self.tags[-1] not in blocked_tags return not self.tags or self.tags[-1] not in _BLOCKED_TAGS
def handle_data(self, data): def handle_data(self, data):
if not self.is_valid_tag(): if not self.is_valid_tag():
@ -104,7 +124,7 @@ class HTMLTextExtractor(HTMLParser): # pylint: disable=W0223 # (see https://bu
return ''.join(self.result).strip() return ''.join(self.result).strip()
def html_to_text(html_str): def html_to_text(html_str: str) -> str:
"""Extract text from a HTML string """Extract text from a HTML string
Args: Args:
@ -122,15 +142,15 @@ def html_to_text(html_str):
""" """
html_str = html_str.replace('\n', ' ') html_str = html_str.replace('\n', ' ')
html_str = ' '.join(html_str.split()) html_str = ' '.join(html_str.split())
s = HTMLTextExtractor() s = _HTMLTextExtractor()
try: try:
s.feed(html_str) s.feed(html_str)
except HTMLTextExtractorException: except _HTMLTextExtractorException:
logger.debug("HTMLTextExtractor: invalid HTML\n%s", html_str) logger.debug("HTMLTextExtractor: invalid HTML\n%s", html_str)
return s.get_text() return s.get_text()
def extract_text(xpath_results, allow_none=False): def extract_text(xpath_results, allow_none: bool = False):
"""Extract text from a lxml result """Extract text from a lxml result
* if xpath_results is list, extract the text from each result and concat the list * if xpath_results is list, extract the text from each result and concat the list
@ -142,24 +162,23 @@ def extract_text(xpath_results, allow_none=False):
# it's list of result : concat everything using recursive call # it's list of result : concat everything using recursive call
result = '' result = ''
for e in xpath_results: for e in xpath_results:
result = result + extract_text(e) result = result + (extract_text(e) or '')
return result.strip() return result.strip()
elif isinstance(xpath_results, ElementBase): if isinstance(xpath_results, ElementBase):
# it's a element # it's a element
text = html.tostring(xpath_results, encoding='unicode', method='text', with_tail=False) text: str = html.tostring(xpath_results, encoding='unicode', method='text', with_tail=False)
text = text.strip().replace('\n', ' ') text = text.strip().replace('\n', ' ')
return ' '.join(text.split()) return ' '.join(text.split())
elif isinstance(xpath_results, (_ElementStringResult, _ElementUnicodeResult, str, Number, bool)): if isinstance(xpath_results, (_ElementStringResult, _ElementUnicodeResult, str, Number, bool)):
return str(xpath_results) return str(xpath_results)
elif xpath_results is None and allow_none: if xpath_results is None and allow_none:
return None return None
elif xpath_results is None and not allow_none: if xpath_results is None and not allow_none:
raise ValueError('extract_text(None, allow_none=False)') raise ValueError('extract_text(None, allow_none=False)')
else: raise ValueError('unsupported type')
raise ValueError('unsupported type')
def normalize_url(url, base_url): def normalize_url(url: str, base_url: str) -> str:
"""Normalize URL: add protocol, join URL with base_url, add trailing slash if there is no path """Normalize URL: add protocol, join URL with base_url, add trailing slash if there is no path
Args: Args:
@ -209,7 +228,7 @@ def normalize_url(url, base_url):
return url return url
def extract_url(xpath_results, base_url): def extract_url(xpath_results, base_url) -> str:
"""Extract and normalize URL from lxml Element """Extract and normalize URL from lxml Element
Args: Args:
@ -248,7 +267,7 @@ def extract_url(xpath_results, base_url):
return normalize_url(url, base_url) return normalize_url(url, base_url)
def dict_subset(d, properties): def dict_subset(dictionnary: MutableMapping, properties: Set[str]) -> Dict:
"""Extract a subset of a dict """Extract a subset of a dict
Examples: Examples:
@ -257,10 +276,10 @@ def dict_subset(d, properties):
>>> >> dict_subset({'A': 'a', 'B': 'b', 'C': 'c'}, ['A', 'D']) >>> >> dict_subset({'A': 'a', 'B': 'b', 'C': 'c'}, ['A', 'D'])
{'A': 'a'} {'A': 'a'}
""" """
return {k: d[k] for k in properties if k in d} return {k: dictionnary[k] for k in properties if k in dictionnary}
def get_torrent_size(filesize, filesize_multiplier): def get_torrent_size(filesize: str, filesize_multiplier: str) -> Optional[int]:
""" """
Args: Args:
@ -277,39 +296,20 @@ def get_torrent_size(filesize, filesize_multiplier):
3140000 3140000
""" """
try: try:
filesize = float(filesize) multiplier = _STORAGE_UNIT_VALUE.get(filesize_multiplier, 1)
return int(float(filesize) * multiplier)
if filesize_multiplier == 'TB':
filesize = int(filesize * 1024 * 1024 * 1024 * 1024)
elif filesize_multiplier == 'GB':
filesize = int(filesize * 1024 * 1024 * 1024)
elif filesize_multiplier == 'MB':
filesize = int(filesize * 1024 * 1024)
elif filesize_multiplier == 'KB':
filesize = int(filesize * 1024)
elif filesize_multiplier == 'TiB':
filesize = int(filesize * 1000 * 1000 * 1000 * 1000)
elif filesize_multiplier == 'GiB':
filesize = int(filesize * 1000 * 1000 * 1000)
elif filesize_multiplier == 'MiB':
filesize = int(filesize * 1000 * 1000)
elif filesize_multiplier == 'KiB':
filesize = int(filesize * 1000)
except ValueError: except ValueError:
filesize = None return None
return filesize
def convert_str_to_int(number_str): def convert_str_to_int(number_str: str) -> int:
"""Convert number_str to int or 0 if number_str is not a number.""" """Convert number_str to int or 0 if number_str is not a number."""
if number_str.isdigit(): if number_str.isdigit():
return int(number_str) return int(number_str)
else: return 0
return 0
def int_or_zero(num): def int_or_zero(num: Union[List[str], str]) -> int:
"""Convert num to int or 0. num can be either a str or a list. """Convert num to int or 0. num can be either a str or a list.
If num is a list, the first element is converted to int (or return 0 if the list is empty). If num is a list, the first element is converted to int (or return 0 if the list is empty).
If num is a str, see convert_str_to_int If num is a str, see convert_str_to_int
@ -321,12 +321,12 @@ def int_or_zero(num):
return convert_str_to_int(num) return convert_str_to_int(num)
def is_valid_lang(lang): def is_valid_lang(lang) -> Optional[Tuple[bool, str, str]]:
"""Return language code and name if lang describe a language. """Return language code and name if lang describe a language.
Examples: Examples:
>>> is_valid_lang('zz') >>> is_valid_lang('zz')
False None
>>> is_valid_lang('uk') >>> is_valid_lang('uk')
(True, 'uk', 'ukrainian') (True, 'uk', 'ukrainian')
>>> is_valid_lang(b'uk') >>> is_valid_lang(b'uk')
@ -346,28 +346,27 @@ def is_valid_lang(lang):
for l in language_codes: for l in language_codes:
if l[0][:2] == lang: if l[0][:2] == lang:
return (True, l[0][:2], l[3].lower()) return (True, l[0][:2], l[3].lower())
return False return None
else: for l in language_codes:
for l in language_codes: if l[1].lower() == lang or l[3].lower() == lang:
if l[1].lower() == lang or l[3].lower() == lang: return (True, l[0][:2], l[3].lower())
return (True, l[0][:2], l[3].lower()) return None
return False
def _get_lang_to_lc_dict(lang_list): def _get_lang_to_lc_dict(lang_list: List[str]) -> Dict[str, str]:
key = str(lang_list) key = str(lang_list)
value = lang_to_lc_cache.get(key, None) value = _LANG_TO_LC_CACHE.get(key, None)
if value is None: if value is None:
value = dict() value = {}
for lc in lang_list: for lang in lang_list:
value.setdefault(lc.split('-')[0], lc) value.setdefault(lang.split('-')[0], lang)
lang_to_lc_cache[key] = value _LANG_TO_LC_CACHE[key] = value
return value return value
# babel's get_global contains all sorts of miscellaneous locale and territory related data # babel's get_global contains all sorts of miscellaneous locale and territory related data
# see get_global in: https://github.com/python-babel/babel/blob/master/babel/core.py # see get_global in: https://github.com/python-babel/babel/blob/master/babel/core.py
def _get_from_babel(lang_code, key): def _get_from_babel(lang_code, key: str):
match = get_global(key).get(lang_code.replace('-', '_')) match = get_global(key).get(lang_code.replace('-', '_'))
# for some keys, such as territory_aliases, match may be a list # for some keys, such as territory_aliases, match may be a list
if isinstance(match, str): if isinstance(match, str):
@ -375,7 +374,7 @@ def _get_from_babel(lang_code, key):
return match return match
def _match_language(lang_code, lang_list=[], custom_aliases={}): # pylint: disable=W0102 def _match_language(lang_code, lang_list=[], custom_aliases={}) -> Optional[str]: # pylint: disable=W0102
"""auxiliary function to match lang_code in lang_list""" """auxiliary function to match lang_code in lang_list"""
# replace language code with a custom alias if necessary # replace language code with a custom alias if necessary
if lang_code in custom_aliases: if lang_code in custom_aliases:
@ -400,7 +399,7 @@ def _match_language(lang_code, lang_list=[], custom_aliases={}): # pylint: disa
return _get_lang_to_lc_dict(lang_list).get(lang_code, None) return _get_lang_to_lc_dict(lang_list).get(lang_code, None)
def match_language(locale_code, lang_list=[], custom_aliases={}, fallback='en-US'): # pylint: disable=W0102 def match_language(locale_code, lang_list=[], custom_aliases={}, fallback='en-US') -> str: # pylint: disable=W0102
"""get the language code from lang_list that best matches locale_code""" """get the language code from lang_list that best matches locale_code"""
# try to get language from given locale_code # try to get language from given locale_code
language = _match_language(locale_code, lang_list, custom_aliases) language = _match_language(locale_code, lang_list, custom_aliases)
@ -438,7 +437,7 @@ def match_language(locale_code, lang_list=[], custom_aliases={}, fallback='en-US
return language or fallback return language or fallback
def load_module(filename, module_dir): def load_module(filename: str, module_dir: str):
modname = splitext(filename)[0] modname = splitext(filename)[0]
filepath = join(module_dir, filename) filepath = join(module_dir, filename)
# and https://docs.python.org/3/library/importlib.html#importing-a-source-file-directly # and https://docs.python.org/3/library/importlib.html#importing-a-source-file-directly
@ -448,19 +447,16 @@ def load_module(filename, module_dir):
return module return module
def to_string(obj): def to_string(obj: Any) -> str:
"""Convert obj to its string representation.""" """Convert obj to its string representation."""
if isinstance(obj, str): if isinstance(obj, str):
return obj return obj
if isinstance(obj, Number):
return str(obj)
if hasattr(obj, '__str__'): if hasattr(obj, '__str__'):
return obj.__str__() return obj.__str__()
if hasattr(obj, '__repr__'): return repr(obj)
return obj.__repr__()
def ecma_unescape(s): def ecma_unescape(string: str) -> str:
"""Python implementation of the unescape javascript function """Python implementation of the unescape javascript function
https://www.ecma-international.org/ecma-262/6.0/#sec-unescape-string https://www.ecma-international.org/ecma-262/6.0/#sec-unescape-string
@ -475,23 +471,23 @@ def ecma_unescape(s):
'ó' 'ó'
""" """
# "%u5409" becomes "吉" # "%u5409" becomes "吉"
s = ecma_unescape4_re.sub(lambda e: chr(int(e.group(1), 16)), s) string = _ECMA_UNESCAPE4_RE.sub(lambda e: chr(int(e.group(1), 16)), string)
# "%20" becomes " ", "%F3" becomes "ó" # "%20" becomes " ", "%F3" becomes "ó"
s = ecma_unescape2_re.sub(lambda e: chr(int(e.group(1), 16)), s) string = _ECMA_UNESCAPE2_RE.sub(lambda e: chr(int(e.group(1), 16)), string)
return s return string
def get_string_replaces_function(replaces): def get_string_replaces_function(replaces: Dict[str, str]) -> Callable:
rep = {re.escape(k): v for k, v in replaces.items()} rep = {re.escape(k): v for k, v in replaces.items()}
pattern = re.compile("|".join(rep.keys())) pattern = re.compile("|".join(rep.keys()))
def f(text): def func(text):
return pattern.sub(lambda m: rep[re.escape(m.group(0))], text) return pattern.sub(lambda m: rep[re.escape(m.group(0))], text)
return f return func
def get_engine_from_settings(name): def get_engine_from_settings(name: str) -> Dict:
"""Return engine configuration from settings.yml of a given engine name""" """Return engine configuration from settings.yml of a given engine name"""
if 'engines' not in settings: if 'engines' not in settings:
@ -506,7 +502,7 @@ def get_engine_from_settings(name):
return {} return {}
def get_xpath(xpath_spec): def get_xpath(xpath_spec: XPathSpecType) -> XPath:
"""Return cached compiled XPath """Return cached compiled XPath
There is no thread lock. There is no thread lock.
@ -523,13 +519,13 @@ def get_xpath(xpath_spec):
* SearxXPathSyntaxException: Raise when there is a syntax error in the XPath * SearxXPathSyntaxException: Raise when there is a syntax error in the XPath
""" """
if isinstance(xpath_spec, str): if isinstance(xpath_spec, str):
result = xpath_cache.get(xpath_spec, None) result = _XPATH_CACHE.get(xpath_spec, None)
if result is None: if result is None:
try: try:
result = XPath(xpath_spec) result = XPath(xpath_spec)
except XPathSyntaxError as e: except XPathSyntaxError as e:
raise SearxXPathSyntaxException(xpath_spec, str(e.msg)) from e raise SearxXPathSyntaxException(xpath_spec, str(e.msg)) from e
xpath_cache[xpath_spec] = result _XPATH_CACHE[xpath_spec] = result
return result return result
if isinstance(xpath_spec, XPath): if isinstance(xpath_spec, XPath):
@ -538,7 +534,7 @@ def get_xpath(xpath_spec):
raise TypeError('xpath_spec must be either a str or a lxml.etree.XPath') raise TypeError('xpath_spec must be either a str or a lxml.etree.XPath')
def eval_xpath(element, xpath_spec): def eval_xpath(element: ElementBase, xpath_spec: XPathSpecType):
"""Equivalent of element.xpath(xpath_str) but compile xpath_str once for all. """Equivalent of element.xpath(xpath_str) but compile xpath_str once for all.
See https://lxml.de/xpathxslt.html#xpath-return-values See https://lxml.de/xpathxslt.html#xpath-return-values
@ -562,7 +558,7 @@ def eval_xpath(element, xpath_spec):
raise SearxEngineXPathException(xpath_spec, arg) from e raise SearxEngineXPathException(xpath_spec, arg) from e
def eval_xpath_list(element, xpath_spec, min_len=None): def eval_xpath_list(element: ElementBase, xpath_spec: XPathSpecType, min_len: int = None):
"""Same as eval_xpath, check if the result is a list """Same as eval_xpath, check if the result is a list
Args: Args:
@ -586,7 +582,7 @@ def eval_xpath_list(element, xpath_spec, min_len=None):
return result return result
def eval_xpath_getindex(elements, xpath_spec, index, default=NOTSET): def eval_xpath_getindex(elements: ElementBase, xpath_spec: XPathSpecType, index: int, default=_NOTSET):
"""Call eval_xpath_list then get one element using the index parameter. """Call eval_xpath_list then get one element using the index parameter.
If the index does not exist, either aise an exception is default is not set, If the index does not exist, either aise an exception is default is not set,
other return the default value (can be None). other return the default value (can be None).
@ -606,9 +602,9 @@ def eval_xpath_getindex(elements, xpath_spec, index, default=NOTSET):
* result (bool, float, list, str): Results. * result (bool, float, list, str): Results.
""" """
result = eval_xpath_list(elements, xpath_spec) result = eval_xpath_list(elements, xpath_spec)
if index >= -len(result) and index < len(result): if -len(result) <= index < len(result):
return result[index] return result[index]
if default == NOTSET: if default == _NOTSET:
# raise an SearxEngineXPathException instead of IndexError # raise an SearxEngineXPathException instead of IndexError
# to record xpath_spec # to record xpath_spec
raise SearxEngineXPathException(xpath_spec, 'index ' + str(index) + ' not found') raise SearxEngineXPathException(xpath_spec, 'index ' + str(index) + ' not found')

View file

@ -128,7 +128,7 @@ class TestUtils(SearxTestCase):
class TestHTMLTextExtractor(SearxTestCase): class TestHTMLTextExtractor(SearxTestCase):
def setUp(self): def setUp(self):
self.html_text_extractor = utils.HTMLTextExtractor() self.html_text_extractor = utils._HTMLTextExtractor()
def test__init__(self): def test__init__(self):
self.assertEqual(self.html_text_extractor.result, []) self.assertEqual(self.html_text_extractor.result, [])
@ -149,7 +149,7 @@ class TestHTMLTextExtractor(SearxTestCase):
def test_invalid_html(self): def test_invalid_html(self):
text = '<p><b>Lorem ipsum</i>dolor sit amet</p>' text = '<p><b>Lorem ipsum</i>dolor sit amet</p>'
with self.assertRaises(utils.HTMLTextExtractorException): with self.assertRaises(utils._HTMLTextExtractorException):
self.html_text_extractor.feed(text) self.html_text_extractor.feed(text)