Source code for darc.db

# -*- coding: utf-8 -*-
"""Link Database
===================

The :mod:`darc` project utilises `Redis`_ based database
to provide tele-process communication.

.. note::

   In its first implementation, the :mod:`darc` project used
   :class:`~multiprocessing.Queue` to support such communication.
   However, as noticed when runtime, the :class:`~multiprocessing.Queue`
   object will be much affected by the lack of memory.

There will be three databases, all following the save naming
convension with ``queue_`` prefix:

* the hostname database -- ``queue_hostname``
* the :mod:`requests` database -- ``queue_requests``
* the :mod:`selenium` database -- ``queue_selenium``

For ``queue_hostname``, it is a `Redis`_ **set** data type;
and for ``queue_requests`` and ``queue_selenium``, they
are both `Redis`_ **sorted set** data type.

.. _Redis: https://redis.io/

"""

import math
import os
import pickle
import pprint
import shutil
import sys
import time
import warnings

import redis.lock as redis_lock
import stem.util.term

import darc.typing as typing
from darc._compat import nullcontext
from darc.const import CHECK
from darc.const import REDIS as redis
from darc.const import TIME_CACHE, VERBOSE
from darc.error import LockWarning, RedisCommandFailed, render_error
from darc.link import Link
from darc.parse import _check

# use lock?
REDIS_LOCK = bool(int(os.getenv('DARC_REDIS_LOCK', '0')))

# Redis retry interval
REDIS_RETRY = float(os.getenv('DARC_REDIS_RETRY', '10'))
if not math.isfinite(REDIS_RETRY):
    REDIS_RETRY = None

# lock blocking timeout
LOCK_TIMEOUT = float(os.getenv('DARC_LOCK_TIMEOUT', '10'))
if not math.isfinite(LOCK_TIMEOUT):
    LOCK_TIMEOUT = None

# bulk size
BULK_SIZE = int(os.getenv('DARC_BULK_SIZE', '100'))

# max pool
MAX_POOL = float(os.getenv('DARC_MAX_POOL', '100'))
if math.isfinite(MAX_POOL):
    MAX_POOL = math.floor(MAX_POOL)


[docs]def redis_command(command: str, *args, **kwargs) -> typing.Any: """Wrapper function for Redis command. Args: command: Command name. *args: Arbitrary arguments for the Redis command. Keyword Args: **kwargs: Arbitrary keyword arguments for the Redis command. Return: Values returned from the Redis command. Warns: RedisCommandFailed: Warns at each round when the command failed. See Also: Between each retry, the function sleeps for :data:`~darc.db.REDIS_RETRY` second(s) if such value is **NOT** :data:`None`. """ _args = None _kwargs = None method = getattr(redis, command) while True: try: value = method(*args, **kwargs) except Exception as error: if _args is None: _args = ', '.join(map(repr, args)) if _kwargs is None: _kwargs = ', '.join(f'{k}={v!r}' for k, v in kwargs.items()) warning = warnings.formatwarning(error, RedisCommandFailed, __file__, 85, f'value = redis.{command}({_args}, {_kwargs})') print(render_error(warning, stem.util.term.Color.YELLOW), end='', file=sys.stderr) # pylint: disable=no-member if REDIS_RETRY is not None: time.sleep(REDIS_RETRY) continue break return value
[docs]def get_lock(name: str, timeout: typing.Optional[float] = None, sleep: float = 0.1, blocking_timeout: typing.Optional[float] = None, lock_class: typing.Optional[redis_lock.Lock] = None, thread_local: bool = True) -> typing.Union[redis_lock.Lock, nullcontext]: """Get a lock for Redis operations. Args: name: Lock name. timeout: Maximum life for the lock. sleep: Amount of time to sleep per loop iteration when the lock is in blocking mode and another client is currently holding the lock. blocking_timeout: Maximum amount of time in seconds to spend trying to acquire the lock. lock_class: Lock implementation. thread_local: Whether the lock token is placed in thread-local storage. Returns: Return a new :class:`redis.lock.Lock` object using key ``name`` that mimics the behavior of :class:`threading.Lock`. Seel Also: If :data:`~darc.db.REDIS_LOCK` is :data:`False`, returns a :class:`contextlib.nullcontext` instead. """ if REDIS_LOCK: return redis_command('lock', name, timeout, sleep, blocking_timeout, lock_class, thread_local) return nullcontext()
[docs]def have_hostname(link: Link) -> bool: """Check if current link is a new host. The function checks the ``queue_hostname`` database. Args: link: Link to check against. Returns: If such link is a new host. """ with get_lock('lock_queue_hostname'): code = redis_command('sadd', 'queue_hostname', link.host) flag = not bool(code) # 1 - success; 0 - failed return flag
[docs]def drop_hostname(link: Link): """Remove link from the hostname database. The function updates the ``queue_hostname`` database. Args: link: Link to be removed. """ with get_lock('lock_queue_hostname'): redis_command('zrem', 'queue_hostname', link.host)
[docs]def drop_requests(link: Link): """Remove link from the :mod:`requests` database. The function updates the ``queue_requests`` database. Args: link: Link to be removed. """ with get_lock('lock_queue_requests'): redis_command('zrem', 'queue_requests', pickle.dumps(link))
[docs]def drop_selenium(link: Link): """Remove link from the :mod:`selenium` database. The function updates the ``queue_selenium`` database. Args: link: Link to be removed. """ with get_lock('lock_queue_selenium'): redis_command('zrem', 'queue_selenium', pickle.dumps(link))
[docs]def save_requests(entries: typing.List[Link], single: bool = False, score=None, nx=False, xx=False): """Save link to the :mod:`requests` database. The function updates the ``queue_requests`` database. Args: entries: Links to be added to the :mod:`requests` database. It can be either a :obj:`list` of links, or a single link string (if ``single`` set as :data:`True`). single: Indicate if ``entries`` is a :obj:`list` of links or a single link string. score: Score to for the Redis sorted set. nx: Forces ``ZADD`` to only create new elements and not to update scores for elements that already exist. xx: Forces ``ZADD`` to only update scores of elements that already exist. New elements will not be added. When ``entries`` is a list of :class:`~darc.link.Link` instances, we tries to perform *bulk* update to easy the memory consumption. The *bulk* size is defined by :data:`~darc.db.BULK_SIZE`. Notes: The ``entries`` will be dumped through :mod:`pickle` so that :mod:`darc` do not need to parse them again. """ if not entries: return if score is None: score = time.time() if not single: for i in range(0, len(entries), BULK_SIZE): mapping = { pickle.dumps(link): score for link in entries[i:i + BULK_SIZE] } with get_lock('lock_queue_requests'): redis_command('zadd', 'queue_requests', mapping, nx=nx, xx=xx) return mapping = { pickle.dumps(entries): score, } with get_lock('lock_queue_requests'): redis_command('zadd', 'queue_requests', mapping, nx=nx, xx=xx)
[docs]def save_selenium(entries: typing.List[Link], single: bool = False, score=None, nx=False, xx=False): """Save link to the :mod:`selenium` database. The function updates the ``queue_selenium`` database. Args: entries: Links to be added to the :mod:`selenium` database. It can be either an *iterable* of links, or a single link string (if ``single`` set as :data:`True`). single: Indicate if ``entries`` is an *iterable* of links or a single link string. score: Score to for the Redis sorted set. nx: Forces ``ZADD`` to only create new elements and not to update scores for elements that already exist. xx: Forces ``ZADD`` to only update scores of elements that already exist. New elements will not be added. When ``entries`` is a list of :class:`~darc.link.Link` instances, we tries to perform *bulk* update to easy the memory consumption. The *bulk* size is defined by :data:`~darc.db.BULK_SIZE`. Notes: The ``entries`` will be dumped through :mod:`pickle` so that :mod:`darc` do not need to parse them again. """ if not entries: return if score is None: score = time.time() if not single: for i in range(0, len(entries), BULK_SIZE): mapping = { pickle.dumps(link): score for link in entries[i:i + BULK_SIZE] } with get_lock('lock_queue_selenium'): redis_command('zadd', 'queue_selenium', mapping, nx=nx, xx=xx) return mapping = { pickle.dumps(entries): score, } with get_lock('lock_queue_selenium'): redis_command('zadd', 'queue_selenium', mapping, nx=nx, xx=xx)
[docs]def load_requests(check: bool = CHECK) -> typing.List[Link]: """Load link from the :mod:`requests` database. The function reads the ``queue_requests`` database. Args: check: If perform checks on loaded links, default to :data:`~darc.const.CHECK`. Returns: List of loaded links from the :mod:`requests` database. Note: At runtime, the function will load links with maximum number at :data:`~darc.db.MAX_POOL` to limit the memory usage. """ now = time.time() if TIME_CACHE is None: max_score = now else: sec_delta = TIME_CACHE.total_seconds() max_score = now - sec_delta try: with get_lock('lock_queue_requests', blocking_timeout=LOCK_TIMEOUT): link_pool = [pickle.loads(link) for link in redis_command('zrangebyscore', 'queue_requests', min=0, max=max_score, start=0, num=MAX_POOL)] if TIME_CACHE is not None: new_score = now + sec_delta save_requests(link_pool, score=new_score) # force update records except redis_lock.LockError: warning = warnings.formatwarning(f'[REQUESTS] Failed to acquire Redis lock after {LOCK_TIMEOUT} second(s)', LockWarning, __file__, 250, "get_lock('lock_queue_requests')") print(render_error(warning, stem.util.term.Color.YELLOW), end='', file=sys.stderr) # pylint: disable=no-member link_pool = list() if check: link_pool = _check(link_pool) if VERBOSE: print(stem.util.term.format('-*- [REQUESTS] LINK POOL -*-', stem.util.term.Color.MAGENTA)) # pylint: disable=no-member print(render_error(pprint.pformat(sorted(link.url for link in link_pool)), stem.util.term.Color.MAGENTA)) # pylint: disable=no-member print(stem.util.term.format('-' * shutil.get_terminal_size().columns, stem.util.term.Color.MAGENTA)) # pylint: disable=no-member return link_pool
[docs]def load_selenium(check: bool = CHECK) -> typing.List[Link]: """Load link from the :mod:`selenium` database. The function reads the ``queue_selenium`` database. Args: check: If perform checks on loaded links, default to :data:`~darc.const.CHECK`. Returns: List of loaded links from the :mod:`selenium` database. Note: At runtime, the function will load links with maximum number at :data:`~darc.db.MAX_POOL` to limit the memory usage. """ now = time.time() if TIME_CACHE is None: max_score = now else: sec_delta = TIME_CACHE.total_seconds() max_score = now - sec_delta try: with get_lock('lock_queue_selenium', blocking_timeout=LOCK_TIMEOUT): link_pool = [pickle.loads(link) for link in redis_command('zrangebyscore', 'queue_selenium', min=0, max=max_score, start=0, num=MAX_POOL)] if TIME_CACHE is not None: new_score = now + sec_delta save_selenium(link_pool, score=new_score) # force update records except redis_lock.LockError: warning = warnings.formatwarning(f'[SELENIUM] Failed to acquire Redis lock after {LOCK_TIMEOUT} second(s)', LockWarning, __file__, 299, "get_lock('lock_queue_selenium')") print(render_error(warning, stem.util.term.Color.YELLOW), end='', file=sys.stderr) # pylint: disable=no-member link_pool = list() if check: link_pool = _check(link_pool) if VERBOSE: print(stem.util.term.format('-*- [SELENIUM] LINK POOL -*-', stem.util.term.Color.MAGENTA)) # pylint: disable=no-member print(render_error(pprint.pformat(sorted(link.url for link in link_pool)), stem.util.term.Color.MAGENTA)) # pylint: disable=no-member print(stem.util.term.format('-' * shutil.get_terminal_size().columns, stem.util.term.Color.MAGENTA)) # pylint: disable=no-member return link_pool