|
""" |
|
Utilities for declaring and retrieving active model layers using a shared DHT. |
|
""" |
|
from __future__ import annotations |
|
|
|
from functools import partial |
|
from typing import Dict, List, Optional, Sequence, Union |
|
|
|
from hivemind.dht import DHT, DHTNode, DHTValue |
|
from hivemind.moe.client.remote_expert_worker import RemoteExpertWorker |
|
from hivemind.p2p import P2P, PeerID |
|
from hivemind.utils import DHTExpiration, MPFuture, get_dht_time, get_logger, use_hivemind_log_handler |
|
|
|
import src |
|
from src.data_structures import CHAIN_DELIMITER, UID_DELIMITER, ModuleUID, RemoteModuleInfo |
|
|
|
use_hivemind_log_handler("in_root_logger") |
|
logger = get_logger(__file__) |
|
|
|
|
|
def declare_active_modules( |
|
dht: DHT, |
|
uids: Sequence[ModuleUID], |
|
expiration_time: DHTExpiration, |
|
throughput: Optional[float] = None, |
|
wait: bool = True, |
|
) -> Union[Dict[ModuleUID, bool], MPFuture[Dict[ModuleUID, bool]]]: |
|
""" |
|
Declare that your node serves the specified modules; update timestamps if declared previously |
|
|
|
:param uids: a list of module ids to declare |
|
:param wait: if True, awaits for declaration to finish, otherwise runs in background |
|
:param throughput: optionally specify your performance in terms of compute throughput |
|
:param expiration_time: declated modules will be visible for this many seconds |
|
:returns: if wait, returns store status for every key (True = store succeeded, False = store rejected) |
|
""" |
|
if isinstance(uids, str): |
|
uids = [uids] |
|
if not isinstance(uids, list): |
|
uids = list(uids) |
|
for uid in uids: |
|
assert isinstance(uid, ModuleUID) and UID_DELIMITER in uid and CHAIN_DELIMITER not in uid |
|
return dht.run_coroutine( |
|
partial(_declare_active_modules, uids=uids, expiration_time=expiration_time, throughput=throughput), |
|
return_future=not wait, |
|
) |
|
|
|
|
|
async def _declare_active_modules( |
|
dht: DHT, |
|
node: DHTNode, |
|
uids: List[ModuleUID], |
|
expiration_time: DHTExpiration, |
|
throughput: Optional[float] = None, |
|
) -> Dict[ModuleUID, bool]: |
|
num_workers = len(uids) if dht.num_workers is None else min(len(uids), dht.num_workers) |
|
return await node.store_many( |
|
keys=uids, |
|
subkeys=[dht.peer_id.to_base58()] * len(uids), |
|
values=[throughput] * len(uids), |
|
expiration_time=expiration_time, |
|
num_workers=num_workers, |
|
) |
|
|
|
|
|
def get_remote_module( |
|
dht: DHT, |
|
uid_or_uids: Union[ModuleUID, List[ModuleUID]], |
|
expiration_time: Optional[DHTExpiration] = None, |
|
return_future: bool = False, |
|
) -> Union[List[Optional[src.RemoteTransformerBlock]], MPFuture[List[Optional[src.RemoteTransformerBlock]]]]: |
|
""" |
|
:param uid_or_uids: find one or more modules with these ids from across the DHT |
|
:param expiration_time: if specified, return modules that expire no sooner than this (based on get_dht_time) |
|
:param return_future: if False (default), return when finished. Otherwise return MPFuture and run in background. |
|
:returns: a list of [RemoteTransformerBlock if found else None] |
|
""" |
|
single_uid = isinstance(uid_or_uids, ModuleUID) |
|
uids = [uid_or_uids] if single_uid else uid_or_uids |
|
infos = dht.run_coroutine( |
|
partial(_get_remote_module_infos, uids=uids, expiration_time=expiration_time), return_future |
|
) |
|
|
|
if return_future: |
|
|
|
async def _unpack(infos_future: MPFuture, dht: DHT): |
|
p2p = await dht.replicate_p2p() |
|
modules = _create_remote_modules_from_infos(await infos_future, p2p) |
|
return modules[0] if single_uid else modules |
|
|
|
return RemoteExpertWorker.run_coroutine(_unpack(infos, dht), return_future) |
|
p2p = RemoteExpertWorker.run_coroutine(dht.replicate_p2p()) |
|
modules = _create_remote_modules_from_infos(infos, p2p) |
|
return modules[0] if single_uid else modules |
|
|
|
|
|
async def _get_remote_module_infos( |
|
dht: DHT, node: DHTNode, uids: List[ModuleUID], expiration_time: Optional[DHTExpiration] |
|
) -> List[Optional[RemoteModuleInfo]]: |
|
if expiration_time is None: |
|
expiration_time = get_dht_time() |
|
num_workers = len(uids) if dht.num_workers is None else min(len(uids), dht.num_workers) |
|
found: Dict[ModuleUID, DHTValue] = await node.get_many(uids, expiration_time, num_workers=num_workers) |
|
|
|
modules: List[Optional[RemoteModuleInfo]] = [None] * len(uids) |
|
for i, uid in enumerate(uids): |
|
metadata = found[uid] |
|
if metadata is None or not isinstance(metadata.value, dict): |
|
if metadata is not None: |
|
logger.error(f"Incorrect metadata for {uid}: {metadata}") |
|
continue |
|
valid_entries = set() |
|
for maybe_peer_id, _unused_value in metadata.value.items(): |
|
try: |
|
valid_entries.add(PeerID.from_base58(maybe_peer_id)) |
|
except: |
|
logger.error(f"Incorrect peer entry for {uid}: {maybe_peer_id}") |
|
if valid_entries: |
|
modules[i] = RemoteModuleInfo(uid, valid_entries) |
|
return modules |
|
|
|
|
|
def _create_remote_modules_from_infos( |
|
infos: Sequence[Optional[RemoteModuleInfo]], p2p: P2P |
|
) -> List[Optional[src.RemoteTransformerBlock]]: |
|
modules: List[Optional[src.RemoteTransformerBlock]] = [] |
|
for info in infos: |
|
if info is not None: |
|
modules.append(src.RemoteTransformerBlock(info, p2p)) |
|
else: |
|
modules.append(None) |
|
return modules |
|
|