Source code for one.registration

"""Session creation and datasets registration.

The RegistrationClient provides an high-level API for creating experimentation sessions on Alyx
and registering associated datasets.

Summary of methods
------------------
create_new_session - Create a new local session folder and optionally create session record on Alyx
create_sessions - Create sessions and register files for folder containing a given flag file
register_session - Create a session on Alyx from local path and register any ALF datasets present
register_files - Register a list of files to their respective sessions on Alyx
"""
import pathlib
from uuid import UUID
from pathlib import Path, PurePosixPath
import datetime
import logging
import itertools
from collections import defaultdict
from fnmatch import fnmatch
import shutil

import requests.exceptions

from iblutil.io import hashfile
from iblutil.util import Bunch, ensure_list

import one.alf.io as alfio
from one.alf.path import ALFPath, session_path_parts, ensure_alf_path, folder_parts
from one.alf.spec import is_valid
import one.alf.exceptions as alferr
from one.api import ONE
from one.webclient import no_cache

_logger = logging.getLogger(__name__)


[docs] def get_dataset_type(filename, dtypes): """Get the dataset type from a given filename. A dataset type is matched one of two ways: 1. the filename matches the dataset type filename_pattern; 2. if filename_pattern is empty, the filename object.attribute matches the dataset type name. Parameters ---------- filename : str, pathlib.Path The filename or filepath. dtypes : iterable An iterable of dataset type objects with the attributes ('name', 'filename_pattern'). Returns ------- The matching dataset type object for filename. Raises ------ ValueError filename doesn't match any of the dataset types filename matches multiple dataset types """ dataset_types = [] filename = ensure_alf_path(filename) for dt in dtypes: if not dt.filename_pattern.strip(): # If the filename pattern is null, check whether the filename object.attribute matches # the dataset type name. if is_valid(filename.name): obj_attr = '.'.join(filename.dataset_name_parts[1:3]) else: # will match name against filename sans extension obj_attr = filename.stem if dt.name == obj_attr: dataset_types.append(dt) # Check whether pattern matches filename elif fnmatch(filename.name.lower(), dt.filename_pattern.lower()): dataset_types.append(dt) n = len(dataset_types) if n == 0: raise ValueError(f'No dataset type found for filename "{filename.name}"') elif n >= 2: raise ValueError('Multiple matching dataset types found for filename ' f'"{filename.name}": \n{", ".join(map(str, dataset_types))}') return dataset_types[0]
[docs] class RegistrationClient: """ Object that keeps the ONE instance and provides method to create sessions and register data. """ def __init__(self, one=None): self.one = one if not one: self.one = ONE(cache_rest=None) elif one.alyx.cache_mode == 'GET': _logger.warning('AlyxClient REST cache active; ' 'this may cause issues with registration.') self.dtypes = list(map(Bunch, self.one.alyx.rest('dataset-types', 'list'))) self.registration_patterns = [ dt['filename_pattern'] for dt in self.dtypes if dt['filename_pattern']] self.file_extensions = [df['file_extension'] for df in self.one.alyx.rest('data-formats', 'list', no_cache=True)]
[docs] def create_sessions(self, root_data_folder, glob_pattern='**/create_me.flag', register_files=False, dry=False): """ Create sessions looking recursively for flag files. Parameters ---------- root_data_folder : str, pathlib.Path Folder to look for sessions. glob_pattern : str Register valid sessions that contain this pattern. register_files : bool If true, register all valid datasets within the session folder. dry : bool If true returns list of sessions without creating them on Alyx. Returns ------- list of pathlib.Paths Newly created session paths. list of dicts Alyx session records. """ flag_files = list(Path(root_data_folder).glob(glob_pattern)) records = [] for flag_file in flag_files: if dry: records.append(print(flag_file)) continue session_path = ALFPath(flag_file.parent) _logger.info('creating session for ' + str(session_path)) # providing a false flag stops the registration after session creation session_info, _ = self.register_session(session_path, file_list=register_files) records.append(session_info) flag_file.unlink() return [ALFPath(ff.parent) for ff in flag_files], records
[docs] def create_new_session(self, subject, session_root=None, date=None, register=True, **kwargs): """Create a new local session folder and optionally create session record on Alyx. Parameters ---------- subject : str The subject name. Must exist on Alyx. session_root : str, pathlib.Path The root folder in which to create the subject/date/number folder. Defaults to ONE cache directory. date : datetime.datetime, datetime.date, str An optional date for the session. If None the current time is used. register : bool If true, create session record on Alyx database. kwargs Optional arguments for RegistrationClient.register_session. Returns ------- pathlib.Path New local session path. uuid.UUID The experiment UUID if register is True. Examples -------- Create a local session only >>> session_path, _ = RegistrationClient().create_new_session('Ian', register=False) Register a session on Alyx in a specific location >>> session_path, eid = RegistrationClient().create_new_session('Sy', '/data/lab/Subjects') Create a session for a given date >>> session_path, eid = RegistrationClient().create_new_session('Ian', date='2020-01-01') """ assert not self.one.offline, 'ONE must be in online mode' date = self.ensure_ISO8601(date) # Format, validate # Ensure subject exists on Alyx self.assert_exists(subject, 'subjects') session_root = Path(session_root or self.one.alyx.cache_dir) / subject / date[:10] session_path = session_root / alfio.next_num_folder(session_root) session_path.mkdir(exist_ok=True, parents=True) # Ensure folder exists on disk if register: session_info, _ = self.register_session(session_path, **kwargs) eid = UUID(session_info['url'][-36:]) else: eid = None return session_path, eid
[docs] def find_files(self, session_path): """ Returns a generator of file names that match one of the dataset type patterns in Alyx. Parameters ---------- session_path : str, pathlib.Path The session path to search. Yields ------- pathlib.Path File paths that match the dataset type patterns in Alyx. """ session_path = ALFPath(session_path) for p in session_path.iter_datasets(recursive=True): if any(p.name.endswith(ext) for ext in self.file_extensions): try: get_dataset_type(p, self.dtypes) yield p except ValueError as ex: _logger.debug('%s', ex.args[0])
[docs] def assert_exists(self, member, endpoint): """Raise an error if a given member doesn't exist on Alyx database. Parameters ---------- member : str, uuid.UUID, list The member ID(s) to verify endpoint: str The endpoint at which to look it up Examples -------- >>> client.assert_exists('ALK_036', 'subjects') >>> client.assert_exists('user_45', 'users') >>> client.assert_exists('local_server', 'repositories') Raises ------- one.alf.exceptions.AlyxSubjectNotFound Subject does not exist on Alyx one.alf.exceptions.ALFError Member does not exist on Alyx requests.exceptions.HTTPError Failed to connect to Alyx database or endpoint not found Returns ------- dict, list of dict The endpoint data if member exists. """ if isinstance(member, (str, UUID)): try: return self.one.alyx.rest(endpoint, 'read', id=str(member), no_cache=True) except requests.exceptions.HTTPError as ex: if ex.response.status_code != 404: raise ex elif endpoint == 'subjects': raise alferr.AlyxSubjectNotFound(member) else: raise alferr.ALFError(f'Member "{member}" doesn\'t exist in {endpoint}') else: return [self.assert_exists(x, endpoint) for x in member]
[docs] @staticmethod def ensure_ISO8601(date) -> str: """Ensure provided date is ISO 8601 compliant Parameters ---------- date : str, None, datetime.date, datetime.datetime An optional date to convert to ISO string. If None, the current datetime is used. Returns ------- str The datetime as an ISO 8601 string """ date = date or datetime.datetime.now() # If None get current time if isinstance(date, str): # FIXME support timezone aware strings, e.g. '2023-03-09T17:08:12.4465024+00:00' date = datetime.datetime.fromisoformat(date) # Validate by parsing elif type(date) is datetime.date: date = datetime.datetime.fromordinal(date.toordinal()) return datetime.datetime.isoformat(date)
[docs] def register_session(self, ses_path, users=None, file_list=True, **kwargs): """ Register session in Alyx. NB: If providing a lab or start_time kwarg, they must match the lab (if there is one) and date of the session path. Parameters ---------- ses_path : str, pathlib.Path The local session path users : str, list The user(s) to attribute to the session file_list : bool, list An optional list of file paths to register. If True, all valid files within the session folder are registered. If False, no files are registered location : str The optional location within the lab where the experiment takes place procedures : str, list An optional list of procedures, e.g. 'Behavior training/tasks' n_correct_trials : int The number of correct trials (optional) n_trials : int The total number of completed trials (optional) json : dict, str Optional JSON data projects: str, list The project(s) to which the experiment belongs (optional) type : str The experiment type, e.g. 'Experiment', 'Base' task_protocol : str The task protocol (optional) lab : str The name of the lab where the session took place. If None the lab name will be taken from the path. If no lab name is found in the path (i.e. no <lab>/Subjects) the default lab on Alyx will be used. start_time : str, datetime.datetime The precise start time of the session. The date must match the date in the session path. end_time : str, datetime.datetime The precise end time of the session. Returns ------- dict An Alyx session record list, None Alyx file records (or None if file_list is False) Raises ------ AssertionError Subject does not exist on Alyx or provided start_time does not match date in session path. ValueError The provided lab name does not match the one found in the session path or start_time/end_time is not a valid ISO date time. requests.HTTPError A 400 status code means the submitted data was incorrect (e.g. task_protocol was an int instead of a str); A 500 status code means there was a server error. ConnectionError Failed to connect to Alyx, most likely due to a bad internet connection. """ ses_path = ALFPath(ses_path) details = session_path_parts(ses_path.as_posix(), as_dict=True, assert_valid=True) # query alyx endpoints for subject, error if not found self.assert_exists(details['subject'], 'subjects') # look for a session from the same subject, same number on the same day with no_cache(self.one.alyx): session_id, session = self.one.search(subject=details['subject'], date_range=details['date'], number=details['number'], details=True, query_type='remote') users = ensure_list(users or self.one.alyx.user) self.assert_exists(users, 'users') # if nothing found create a new session in Alyx ses_ = {'subject': details['subject'], 'users': users, 'type': 'Experiment', 'number': details['number']} if kwargs.get('end_time', False): ses_['end_time'] = self.ensure_ISO8601(kwargs.pop('end_time')) start_time = self.ensure_ISO8601(kwargs.pop('start_time', details['date'])) assert start_time[:10] == details['date'], 'start_time doesn\'t match session path' if kwargs.get('procedures', False): ses_['procedures'] = ensure_list(kwargs.pop('procedures')) if kwargs.get('projects', False): ses_['projects'] = ensure_list(kwargs.pop('projects')) assert ('subject', 'number') not in kwargs if 'lab' not in kwargs and details['lab']: kwargs.update({'lab': details['lab']}) elif details['lab'] and kwargs.get('lab', details['lab']) != details['lab']: names = (kwargs['lab'], details['lab']) raise ValueError('lab kwarg "%s" does not match lab name in path ("%s")' % names) ses_.update(kwargs) if not session: # Create from scratch ses_['start_time'] = start_time session = self.one.alyx.rest('sessions', 'create', data=ses_) else: # Update existing if start_time: ses_['start_time'] = self.ensure_ISO8601(start_time) session = self.one.alyx.rest('sessions', 'update', id=session_id[0], data=ses_) _logger.info(session['url'] + ' ') # at this point the session has been created. If create only, exit if not file_list: return session, None recs = self.register_files(self.find_files(ses_path) if file_list is True else file_list) if recs: # Update local session data after registering files session['data_dataset_session_related'] = ensure_list(recs) return session, recs
[docs] def prepare_files(self, file_list, versions=None): """ Validates file list for registration and splits files into a list of files per session path. Parameters ---------- file_list : list, str, pathlib.Path A filepath (or list thereof) of ALF datasets to register to Alyx. versions : str, list of str Optional version tags. Returns ------- list of dicts A dict containing a list of files for each session. list of dicts A dict containing a list of versions for each session. list A list of files converted to paths. bool A boolean indicating if input was a single file. """ F = defaultdict(list) # empty map whose keys will be session paths V = defaultdict(list) # empty map for versions if single_file := isinstance(file_list, (str, pathlib.Path)): file_list = [file_list] file_list = list(map(ALFPath, file_list)) # Ensure list of path objects if versions is None or isinstance(versions, str): versions = itertools.repeat(versions) else: versions = itertools.cycle(versions) # Filter valid files and sort by session for fn, ver in zip(file_list, versions): session_path = fn.session_path() if not session_path: _logger.debug(f'{fn}: Invalid session path') continue if fn.suffix not in self.file_extensions: _logger.debug(f'{fn}: No matching extension "{fn.suffix}" in database') continue try: get_dataset_type(fn, self.dtypes) except ValueError as ex: _logger.debug('%s', ex.args[0]) continue F[session_path].append(fn.relative_to(session_path)) V[session_path].append(ver) return F, V, file_list, single_file
[docs] def check_protected_files(self, file_list, created_by=None): """ Check whether a set of files associated to a session are protected Parameters ---------- file_list : list, str, pathlib.Path A filepath (or list thereof) of ALF datasets to register to Alyx. created_by : str Name of Alyx user (defaults to whoever is logged in to ONE instance). Returns ------- list of dicts, dict A status for each session whether any of the files specified are protected datasets or not.If none of the datasets are protected, a response with status 200 is returned, if any of the files are protected a response with status 403 is returned. """ # Validate files and rearrange into list per session F, _, _, single_file = self.prepare_files(file_list) # For each unique session, make a separate POST request records = [] for session_path, files in F.items(): # this is the generic relative path: subject/yyyy-mm-dd/NNN details = session_path_parts(session_path.as_posix(), as_dict=True, assert_valid=True) rel_path = PurePosixPath(details['subject'], details['date'], details['number']) r_ = {'created_by': created_by or self.one.alyx.user, 'path': rel_path.as_posix(), 'filenames': [x.as_posix() for x in files] } records.append(self.one.alyx.get('/check-protected', data=r_, clobber=True)) return records[0] if single_file else records
[docs] def register_files(self, file_list, versions=None, default=True, created_by=None, server_only=False, repository=None, exists=True, dry=False, max_md5_size=None, **kwargs): """ Registers a set of files belonging to a session only on the server. Parameters ---------- file_list : list, str, pathlib.Path A filepath (or list thereof) of ALF datasets to register to Alyx. versions : str, list of str Optional version tags. default : bool Whether to set as default revision (defaults to True). created_by : str Name of Alyx user (defaults to whoever is logged in to ONE instance). server_only : bool Will only create file records in the 'online' repositories and skips local repositories repository : str Name of the repository in Alyx to register to. exists : bool Whether the files exist on the repository (defaults to True). dry : bool When true returns POST data for registration endpoint without submitting the data. max_md5_size : int Maximum file in bytes to compute md5 sum (always compute if None). exists : bool Whether files exist in the repository. May be set to False when registering files before copying to the repository. kwargs Extra arguments directly passed as REST request data to /register-files endpoint. Returns ------- list of dicts, dict A list of newly created Alyx dataset records or the registration data if dry. If a single file is passed in, a single dict is returned. Notes ----- - The registered files may be automatically moved to new revision folders if they are protected on Alyx, therefore it's important to check the relative paths of the output. - Protected datasets are not checked in dry mode. - In most circumstances a new revision will be added automatically, however if this fails a 403 HTTP status may be returned. Raises ------ requests.exceptions.HTTPError Submitted data not valid (400 status code) Server side database error (500 status code) Revision protected (403 status code) """ F, V, file_list, single_file = self.prepare_files(file_list, versions=versions) # For each unique session, make a separate POST request records = [None] * (len(F) if dry else len(file_list)) # If dry return data per session for session_path, files in F.items(): # this is the generic relative path: subject/yyyy-mm-dd/NNN details = session_path_parts(session_path.as_posix(), as_dict=True, assert_valid=True) rel_path = PurePosixPath(details['subject'], details['date'], details['number']) file_sizes = [session_path.joinpath(fn).stat().st_size for fn in files] # computing the md5 can be very long, so this is an option to skip if the file is # bigger than a certain threshold md5s = [hashfile.md5(session_path.joinpath(fn)) if (max_md5_size is None or sz < max_md5_size) else None for fn, sz in zip(files, file_sizes)] _logger.info('Registering ' + str(files)) r_ = {'created_by': created_by or self.one.alyx.user, 'path': rel_path.as_posix(), 'filenames': [x.as_posix() for x in files], 'hashes': md5s, 'filesizes': file_sizes, 'name': repository, 'exists': exists, 'server_only': server_only, 'default': default, 'versions': V[session_path], 'check_protected': True, **kwargs } # Add optional field if details['lab'] and 'labs' not in kwargs: r_['labs'] = details['lab'] # If dry, store POST data, otherwise store resulting file records if dry: records[list(F).index(session_path)] = r_ continue try: response = self.one.alyx.post('/register-file', data=r_) # Ensure we keep the order of the output records: the files missing will remain # as None type for f, r in zip(files, response): records[file_list.index(session_path / f)] = r except requests.exceptions.HTTPError as err: # 403 response when datasets already registered and protected by tags err_message = err.response.json() if not (err_message.get('status_code') == 403 and err_message.get('error') == 'One or more datasets is protected'): raise err # Some other error occurred; re-raise response = err_message['details'] today_revision = datetime.datetime.today().strftime('%Y-%m-%d') new_file_list = [] for fl, res in zip(files, response): (name, prot_info), = res.items() # Dataset has not yet been registered if not prot_info: new_file_list.append(fl) continue # Check to see if the file path already has a revision in it file_revision = folder_parts(rel_path / fl, as_dict=True)['revision'] # Find existing protected revisions existing_revisions = [k for pr in prot_info for k, v in pr.items() if v] if file_revision: # If the revision explicitly defined by the user doesn't exist or # is not protected, register as is if file_revision not in existing_revisions: revision_path = fl.parent else: # Find the next sub-revision that isn't protected new_revision = self._next_revision(file_revision, existing_revisions) revision_path = fl.parent.parent.joinpath(f'#{new_revision}#') if revision_path != fl.parent: session_path.joinpath(revision_path).mkdir(exist_ok=True) _logger.info('Moving %s -> %s', fl, revision_path.joinpath(fl.name)) shutil.move(session_path / fl, session_path / revision_path / fl.name) new_file_list.append(revision_path.joinpath(fl.name)) continue # The file wasn't in a revision folder but is protected fl_path = fl.parent assert name == fl_path.joinpath(fl.name).as_posix() # Find info about the latest revision # N.B on django side prot_info is sorted by latest revisions first (latest_revision, protected), = prot_info[0].items() # If the latest revision is the original and it is unprotected # no need for revision e.g {'clusters.amp.npy': [{'': False}]} if latest_revision == '' and not protected: # Use original path revision_path = fl_path # If there already is a revision but it is unprotected, # move into this revision folder e.g # {'clusters.amp.npy': # [{'2022-10-31': False}, {'2022-05-31': True}, {'': True}]} elif not protected: # Check that the latest_revision has the date naming convention we expect # i.e. 'YYYY-MM-DD' try: _ = datetime.datetime.strptime(latest_revision[:10], '%Y-%m-%d') revision_path = fl_path.joinpath(f'#{latest_revision}#') # If it doesn't it probably has been made manually so we don't want to # overwrite this and instead use today's date except ValueError: # NB: It's possible that today's date revision is also protected but is # not the most recent revision. In this case it's safer to let fail. revision_path = fl_path.joinpath(f'#{today_revision}#') # If protected and the latest protected revision is from today we need to make # a sub-revision elif protected and today_revision in latest_revision: if latest_revision == today_revision: # iterate from appending 'a' new_revision = self._next_revision(today_revision, existing_revisions) else: # assume the revision is date + character, e.g. '2020-01-01c' alpha = latest_revision[-1] # iterate from this character new_revision = self._next_revision( today_revision, existing_revisions, alpha) revision_path = fl_path.joinpath(f'#{new_revision}#') # Otherwise cases move into revision from today # e.g {'clusters.amp.npy': [{'': True}]} # e.g {'clusters.amp.npy': [{'2022-10-31': True}, {'': True}]} else: revision_path = fl_path.joinpath(f'#{today_revision}#') # Only move for the cases where a revision folder has been made if revision_path != fl_path: session_path.joinpath(revision_path).mkdir(exist_ok=True) _logger.info('Moving %s -> %s', fl, revision_path.joinpath(fl.name)) shutil.move(session_path / fl, session_path / revision_path / fl.name) new_file_list.append(revision_path.joinpath(fl.name)) assert len(new_file_list) == len(files) r_['filenames'] = [p.as_posix() for p in new_file_list] r_['filesizes'] = [session_path.joinpath(p).stat().st_size for p in new_file_list] r_['check_protected'] = False # Speed things up by ignoring server-side checks response = self.one.alyx.post('/register-file', data=r_) for f, r in zip(files, response): # Populate records list in correct order records[file_list.index(session_path / f)] = r files = new_file_list # Log file names _logger.info(f'ALYX REGISTERED DATA {"!DRY!" if dry else ""}: {rel_path}') for p in files: _logger.info(f'ALYX REGISTERED DATA: {p}') return records[0] if single_file else records
@staticmethod def _next_revision(revision: str, reserved: list = None, alpha: str = 'a') -> str: """ Return the next logical revision that is not already in the provided list. Revisions will increment by appending a letter to a date or other identifier. Parameters ---------- revision : str The revision on which to base the new revision. reserved : list of str A list of reserved (i.e. already existing) revision strings. alpha : str The starting character as an integer, defaults to 'a'. Returns ------- str The next logical revision string that's not in the reserved list. Examples -------- >>> RegistrationClient._next_revision('2020-01-01') '2020-01-01a' >>> RegistrationClient._next_revision('2020-01-01', ['2020-01-01a', '2020-01-01b']) '2020-01-01c' >>> RegistrationClient._next_revision('2020-01-01', ['2020-01-01a', '2020-01-01b']) '2020-01-01c' """ if len(alpha) != 1: raise TypeError( f'`alpha` must be a character; received a string of length {len(alpha)}' ) i = ord(alpha) new_revision = revision + chr(i) while new_revision in (reserved or []): i += 1 new_revision = revision + chr(i) return new_revision
[docs] def register_water_administration(self, subject, volume, **kwargs): """ Register a water administration to Alyx for a given subject Parameters ---------- subject : str A subject nickname that exists on Alyx volume : float The total volume administrated in ml date_time : str, datetime.datetime, datetime.date The time of administration. If None, the current time is used. water_type : str A water type that exists in Alyx; default is 'Water' user : str The user who administrated the water. Currently logged-in user is the default. session : str, UUID, pathlib.Path, dict An optional experiment ID to associate adlib : bool If true, indicates that the subject was given water ad libitum Returns ------- dict A water administration record Raises ------ one.alf.exceptions.AlyxSubjectNotFound Subject does not exist on Alyx one.alf.exceptions.ALFError User does not exist on Alyx ValueError date_time is not a valid ISO date time or session ID is not valid requests.exceptions.HTTPError Failed to connect to database, or submitted data not valid (500) """ # Ensure subject exists self.assert_exists(subject, 'subjects') # Ensure user(s) exist user = kwargs.pop('user', self.one.alyx.user) self.assert_exists(user, 'users') # Ensure volume not zero if volume == 0: raise ValueError('Water volume must be greater than zero') # Post water admin wa_ = { 'subject': subject, 'date_time': self.ensure_ISO8601(kwargs.pop('date_time', None)), 'water_administered': float(f'{volume:.4g}'), # Round to 4 s.f. 'water_type': kwargs.pop('water_type', 'Water'), 'user': user, 'adlib': kwargs.pop('adlib', False) } # Ensure session is valid; convert to eid if kwargs.get('session', False): wa_['session'] = self.one.to_eid(kwargs.pop('session')) if not wa_['session']: raise ValueError('Failed to parse session ID') return self.one.alyx.rest('water-administrations', 'create', data=wa_)
[docs] def register_weight(self, subject, weight, date_time=None, user=None): """ Register a subject weight to Alyx. Parameters ---------- subject : str A subject nickname that exists on Alyx. weight : float The subject weight in grams. date_time : str, datetime.datetime, datetime.date The time of weighing. If None, the current time is used. user : str The user who performed the weighing. Currently logged-in user is the default. Returns ------- dict An Alyx weight record Raises ------ one.alf.exceptions.AlyxSubjectNotFound Subject does not exist on Alyx one.alf.exceptions.ALFError User does not exist on Alyx ValueError date_time is not a valid ISO date time or weight < 1e-4 requests.exceptions.HTTPError Failed to connect to database, or submitted data not valid (500) """ # Ensure subject exists self.assert_exists(subject, 'subjects') # Ensure user(s) exist user = user or self.one.alyx.user self.assert_exists(user, 'users') # Ensure weight not zero if weight == 0: raise ValueError('Water volume must be greater than 0') # Post water admin wei_ = {'subject': subject, 'date_time': self.ensure_ISO8601(date_time), 'weight': float(f'{weight:.4g}'), # Round to 4 s.f. 'user': user} return self.one.alyx.rest('weighings', 'create', data=wei_)