"""
Core datastructures used in the validphys data model.
"""
import dataclasses
import enum
import functools
import inspect
import json
import logging
from pathlib import Path
import re
import numpy as np
from nnpdf_data.theorydbutils import fetch_theory
from reportengine import namespaces
from reportengine.baseexceptions import AsInputError
from reportengine.compat import yaml
# TODO: There is a bit of a circular dependency between filters.py and this.
# Maybe move the cuts logic to its own module?
from validphys import filters, lhaindex
from validphys.commondataparser import get_plot_kinlabels, load_commondata, peek_commondata_metadata
from validphys.fkparser import load_fktable, parse_cfactor
from validphys.hyperoptplot import HyperoptTrial
from validphys.lhapdfset import LHAPDFSet
from validphys.tableloader import parse_exp_mat
from validphys.utils import experiments_to_dataset_inputs
log = logging.getLogger(__name__)
[docs]
class TupleComp:
[docs]
@classmethod
def argnames(cls):
return list(inspect.signature(cls.__init__).parameters.keys())[1:]
def __init__(self, *args, **kwargs):
self.comp_tuple = (*args, *kwargs.values())
def __eq__(self, other):
return self.comp_tuple == other.comp_tuple
def __hash__(self):
return hash(self.comp_tuple)
def __repr__(self):
argvals = ', '.join('%s=%r' % vals for vals in zip(self.argnames(), self.comp_tuple))
return f'{self.__class__.__qualname__}({argvals})'
[docs]
class PDFDoesNotExist(Exception):
pass
class _PDFSETS:
"""Convenient way to access installed PDFS, by e.g. tab completing
in ipython."""
def __getattr__(self, attr):
if lhaindex.isinstalled(attr):
return PDF(attr)
raise AttributeError()
def __dir__(self):
return lhaindex.expand_local_names('*')
PDFSETS = _PDFSETS()
# TODO: make the PDF into a dataclass instead of a TupleComp
# https://github.com/NNPDF/nnpdf/issues/408
[docs]
class PDF(TupleComp):
"""Base validphys PDF providing high level access to metadata.
Statistical estimators which depends on the PDF type (MC, Hessian...)
are exposed as a :py:class:`Stats` object through the :py:attr:`stats_class` attribute
The LHAPDF metadata can directly be accessed through the :py:attr:`info` attribute
Examples
--------
>>> from validphys.api import API
>>> from validphys.convolution import predictions
>>> args = {"dataset_input":{"dataset": "ATLASTTBARTOT"}, "theoryid":162, "use_cuts":"internal"}
>>> ds = API.dataset(**args)
>>> pdf = API.pdf(pdf="NNPDF40_nnlo_as_01180")
>>> preds = predictions(ds, pdf)
>>> preds.shape
(3, 100)
"""
def __init__(self, name, boundary=None):
if boundary is not None:
raise ValueError(
"The keyword argument boundary is a transitional argument for tuplecomp"
)
self.name = name
self._plotname = name
self._info = None
self._stats_class = None
# Boundary conditions:
self.unpolarized_bc = None
super().__init__(name, boundary)
@property
def label(self):
return self._plotname
@property
def is_polarized(self):
"""Returns ``True`` if the PDF has a boundary condition associated to it.
At the moment LHAPDF provides no mechanism to know whether a PDF is polarized."""
return self.unpolarized_bc is not None
@label.setter
def label(self, label):
self._plotname = label
@property
def stats_class(self):
"""Return the stats calculator for this error type"""
if self._stats_class is None:
try:
klass = STAT_TYPES[self.error_type]
except KeyError:
raise NotImplementedError(f"No Stats class for error type {self.error_type}")
if self.error_conf_level is not None:
klass = functools.partial(klass, rescale_factor=self._rescale_factor())
self._stats_class = klass
return self._stats_class
@property
def infopath(self):
return Path(lhaindex.infofilename(self.name))
@property
def info(self):
"""Information contained in the LHAPDF .info file"""
if self._info is None:
try:
self._info = lhaindex.parse_info(self.name)
except OSError as e:
raise PDFDoesNotExist(self.name) from e
return self._info
@property
def q_min(self):
"""Minimum Q as given by the LHAPDF .info file"""
return self.info["QMin"]
@property
def error_type(self):
"""Error type as defined in the LHAPDF .info file"""
try:
return self.info["ErrorType"]
except KeyError as e:
# If the error type is not defined _but_ the PDF only has one member:
if self.info.get("NumMembers") == 1:
return "replicas"
raise e
@property
def alphas_mz(self):
"""Alpha_s(M_Z) as defined in the LHAPDF .info file"""
return self.info["AlphaS_MZ"]
@property
def alphas_vals(self):
"""List of alpha_s(Q) at various Q for interpolation based alphas.
Values as defined in the LHAPDF .info file"""
return self.info["AlphaS_Vals"]
@property
def error_conf_level(self):
"""Error confidence level as defined in the LHAPDF .info file
if no number is given in the LHAPDF .info file defaults to 68%
"""
key_name = "ErrorConfLevel"
# Check possible misconfigured info file
if self.error_type == "replicas":
if key_name in self.info:
raise ValueError(
f"Attribute at {self.infopath} 'ErrorConfLevel' doesn't "
"make sense with 'replicas' error type"
)
return None
return self.info.get(key_name, 68)
@property
def isinstalled(self):
try:
return self.infopath.is_file()
except FileNotFoundError:
return False
def _rescale_factor(self):
"""Compute the rescale factor for the stats class"""
# This is imported here for performance reasons.
import scipy.stats
val = scipy.stats.norm.isf((1 - 0.01 * self.error_conf_level) / 2)
if np.isnan(val):
raise ValueError(f"Invalid 'ErrorConfLevel' for PDF {self}: {val}")
return val
[docs]
@functools.lru_cache(maxsize=16)
def load(self):
return LHAPDFSet(self.name, self.error_type)
[docs]
@functools.lru_cache(maxsize=2)
def load_t0(self):
"""Load the PDF as a t0 set"""
return LHAPDFSet(self.name, "t0")
def __str__(self):
return self.label
def __len__(self):
return self.info["NumMembers"]
[docs]
def get_members(self):
"""Return the number of members selected in ``pdf.load().grid_values``"""
return len(self)
[docs]
def register_boundary(self, unpolarized_bc=None):
"""Register other PDFs as boundary conditions of this PDF"""
if unpolarized_bc is not None:
self.unpolarized_bc = unpolarized_bc
# Update `comp_tuple` so that the pseudo-dataclass still works
self.comp_tuple = (self.name, unpolarized_bc.name)
[docs]
def make_only_cv(self):
return PDFcv(self.name)
[docs]
class PDFcv(PDF):
"""An add-on for the PDF class that makes only the central value available"""
[docs]
def load(self):
return self.load_t0()
def __len__(self):
return 1
[docs]
class CommonDataSpec(TupleComp):
"""Holds all the information necessary to load a commondata file and provides
methods to easily access them
Arguments
---------
name: str
name of the commondata
metadata: ObservableMetaData
instance of ObservableMetaData holding all information about the dataset
legacy: bool
whether this is an old or new format metadata file
The ``datafile``, ``sysfile`` and `plotfiles`` arguments are deprecated
and only to be used with ``legacy=True``
"""
def __init__(self, name, metadata, legacy=False, datafile=None, sysfile=None, plotfiles=None):
self.legacy = legacy
self._metadata = metadata
# Some checks
if legacy:
# TODO: this will start raising an error soon
if datafile is None or sysfile is None or plotfiles is None:
raise ValueError(
"Legacy CommonDataSpec need datafile, sysfile and plotfiles arguments"
)
else:
if sysfile is not None:
raise ValueError("New CommonDataSpec don't need sysfile input")
if plotfiles is not None:
raise ValueError("New CommonDataSpec don't need plotfile input")
self.datafile = datafile
self.sysfile = sysfile
if legacy:
self.plotfiles = tuple(plotfiles)
super().__init__(datafile, sysfile, self.plotfiles)
else:
self.plotfiles = False
super().__init__(name, self.metadata)
[docs]
def with_modified_data(self, central_data_file, uncertainties_file=None):
"""Returns a copy of this instance with a new data file in the metadata"""
if self.legacy:
return self.__class__(
self.name,
self.metadata,
legacy=True,
datafile=central_data_file,
sysfile=self.sysfile,
plotfiles=self.plotfiles,
)
modified_args = {"data_central": central_data_file}
if uncertainties_file is not None:
modified_args["data_uncertainties"] = [uncertainties_file]
new_metadata = dataclasses.replace(self.metadata, **modified_args)
return self.__class__(self.name, new_metadata)
@property
def name(self):
return self.metadata.name
@functools.cached_property
def nsys(self):
if self.legacy:
return self.metadata.nsys
else:
cd = self.load()
return cd.nsys
@property
def ndata(self):
return self.metadata.ndata
@property
def process_type(self):
return self.metadata.process_type
@property
def metadata(self):
if self.legacy:
self._metadata = peek_commondata_metadata(self.datafile)
return self._metadata
@functools.cached_property
def legacy_names(self):
return self.load().legacy_names
@property
def theory_metadata(self):
if self.legacy:
return None
return self.metadata.theory
def __str__(self):
return self.name
def __iter__(self):
return iter((self.datafile, self.sysfile, self.plotfiles))
[docs]
def load(self):
"""
load a validphys.core.CommonDataSpec to validphys.core.CommonData
"""
return load_commondata(self)
@property
def plot_kinlabels(self):
if self.legacy:
return get_plot_kinlabels(self)
else:
return self.metadata.kinlabels
[docs]
class CutsPolicy(enum.Enum):
INTERNAL = "internal"
NOCUTS = "nocuts"
FROMFIT = "fromfit"
FROM_CUT_INTERSECTION_NAMESPACE = "fromintersection"
FROM_SIMILAR_PREDICTIONS_NAMESPACE = "fromsimilarpredictions"
[docs]
class Cuts(TupleComp):
def __init__(self, commondata, path):
"""Represents a file containing cuts for a given dataset"""
name = commondata.name
self.name = name
self.path = path
self._mask = None
if path is None:
log.debug("No filter found for %s, all points allowed", name)
self._mask = np.arange(commondata.ndata)
self._legacy_mask = None
super().__init__(name, path)
[docs]
def load(self):
if self._mask is not None:
return self._mask
log.debug("Loading cuts for %s", self.name)
return np.atleast_1d(np.loadtxt(self.path, dtype=int))
[docs]
class InternalCutsWrapper(TupleComp):
def __init__(self, commondata, rules):
self.rules = rules if rules else tuple()
self.commondata = commondata
super().__init__(commondata, tuple(self.rules))
[docs]
def load(self):
return np.atleast_1d(
np.asarray(filters.get_cuts_for_dataset(self.commondata, self.rules), dtype=int)
)
[docs]
class MatchedCuts(TupleComp):
def __init__(self, othercuts, ndata):
self.othercuts = tuple(othercuts)
self.ndata = ndata
super().__init__(self.othercuts, self.ndata)
[docs]
def load(self):
loaded = [c.load() for c in self.othercuts if c]
if loaded:
return functools.reduce(np.intersect1d, loaded)
self._full = True
return np.arange(self.ndata)
[docs]
class SimilarCuts(TupleComp):
def __init__(self, inputs, threshold):
if len(inputs) != 2:
raise ValueError("Expecting two input tuples")
firstcuts, secondcuts = inputs[0][0].cuts, inputs[1][0].cuts
if firstcuts != secondcuts:
raise ValueError("Expecting cuts to be the same for all datasets")
self.inputs = inputs
self.threshold = threshold
super().__init__(self.inputs, self.threshold)
[docs]
@functools.lru_cache
def load(self):
# TODO: Update this when a suitable interace becomes available
from validphys.convolution import central_predictions
from validphys.covmats import covmat_from_systematics
first, second = self.inputs
first_ds = first[0]
exp_err = np.sqrt(
np.diag(
covmat_from_systematics(
load_commondata(first_ds.commondata).with_cuts(first_ds.cuts),
first_ds, # DataSetSpec has weight attr
use_weights_in_covmat=False, # Don't weight covmat
)
)
)
# Compute matched predictions
delta = np.abs((central_predictions(*first) - central_predictions(*second)).squeeze(axis=1))
ratio = delta / exp_err
passed = ratio < self.threshold
return passed[passed].index
[docs]
def cut_mask(cuts):
"""Return an objects that will act as the cuts when applied as a slice"""
if cuts is None:
return slice(None)
return cuts.load()
[docs]
class DataSetSpec(TupleComp):
def __init__(
self, *, name, commondata, fkspecs, thspec, cuts, frac=1, op=None, weight=1, rules=()
):
self.name = name
self.commondata = commondata
self.rules = rules
if isinstance(fkspecs, FKTableSpec):
fkspecs = (fkspecs,)
fkspecs = tuple(fkspecs)
self.fkspecs = fkspecs
self.thspec = thspec
self.cuts = cuts
self.frac = frac
# If OP is None, check whether the commondata is setting an operation
# TODO: eventually the operation will _always_ be set from the commondata, but for legacy
# compatibility it will be also controllable as an input argument
if op is None:
if commondata.theory_metadata is None:
op = 'NULL'
else:
op = commondata.theory_metadata.operation
self.op = op
self.weight = weight
super().__init__(name, commondata, fkspecs, thspec, cuts, frac, op, weight, rules)
[docs]
@functools.lru_cache
def load_commondata(self):
"""Strips the commondata loading from `load`"""
cd = self.commondata.load()
if self.cuts is not None:
loaded_cuts = self.cuts.load()
if not (hasattr(loaded_cuts, '_full') and loaded_cuts._full):
intmask = [int(ele) for ele in loaded_cuts]
cd = cd.with_cuts(intmask)
return cd
[docs]
def to_unweighted(self):
"""Return a copy of the dataset with the weight set to one."""
return self.__class__(
name=self.name,
commondata=self.commondata,
fkspecs=self.fkspecs,
thspec=self.thspec,
cuts=self.cuts,
frac=self.frac,
op=self.op,
weight=1,
)
def __str__(self):
return self.name
[docs]
class FKTableSpec(TupleComp):
"""
Each FKTable is formed by a number of sub-fktables to be concatenated
each of which having its own path.
Therefore the ``fkpath`` variable is a list of paths.
Before the pineappl implementation, FKTable were already pre-concatenated.
The Legacy interface therefore relies on fkpath being just a string or path instead
The metadata of the FKTable for the given dataset is stored as an attribute to this function.
This is transitional, eventually it will be held by the associated CommonData in the new format.
"""
def __init__(self, fkpath, cfactors, metadata=None):
self.cfactors = cfactors if cfactors is not None else []
self.legacy = False
# NOTE: The legacy interface is currently used by fkparser to decide
# whether to read an FKTable using the old parser or the pineappl parser
# this attribute (and the difference between both) might be removed in future
# releases of NNPDF so please don't write code that relies on it
if not isinstance(fkpath, (tuple, list)):
self.legacy = True
else:
# Make it into a tuple only for the new format
fkpath = tuple(fkpath)
self.fkpath = fkpath
self.metadata = metadata
# For non-legacy theory, add the metadata since it defines how the theory is to be loaded
# and thus, it should also define the hash of the class
if not self.legacy:
super().__init__(fkpath, cfactors, self.metadata)
else:
super().__init__(fkpath, cfactors)
[docs]
def load_cfactors(self):
"""Each of the sub-fktables that form the complete FKTable can have several cfactors
applied to it. This function uses ``parse_cfactor`` to make them into CFactorData
"""
if self.legacy:
raise NotImplementedError("cfactor loading from spec not implemented for old theories")
return [[parse_cfactor(c.open("rb")) for c in cfacs] for cfacs in self.cfactors]
[docs]
@functools.lru_cache
def load_with_cuts(self, cuts):
"""Load the fktable and apply cuts immediately. Returns a FKTableData"""
return load_fktable(self).with_cuts(cuts)
[docs]
class LagrangeSetSpec(DataSetSpec):
"""Extends DataSetSpec to work around the particularities of the positivity, integrability
and other Lagrange Multiplier datasets.
"""
def __init__(self, name, commondataspec, fkspec, maxlambda, thspec, rules):
cuts = InternalCutsWrapper(commondataspec, rules)
self.maxlambda = maxlambda
super().__init__(
name=name,
commondata=commondataspec,
fkspecs=fkspec,
thspec=thspec,
cuts=cuts,
rules=rules,
)
[docs]
def to_unweighted(self):
log.warning(
"Trying to unweight %s, %s are always unweighted", self.__class__.__name__, self.name
)
return self
[docs]
class PositivitySetSpec(LagrangeSetSpec):
pass
[docs]
class IntegrabilitySetSpec(LagrangeSetSpec):
pass
# We allow to expand the experiment as a list of datasets
[docs]
class DataGroupSpec(TupleComp, namespaces.NSList):
def __init__(self, name, datasets, dsinputs=None):
# This needs to be hashable
datasets = tuple(datasets)
# TODO: Find a better way for interactive usage.
if dsinputs is not None:
dsinputs = tuple(dsinputs)
self.name = name
self.datasets = datasets
self.dsinputs = dsinputs
# TODO: Add dsinputs to comp tuple?
super().__init__(name, datasets)
# TODO: Can we do better cooperative inherece trick than this?
namespaces.NSList.__init__(self, dsinputs, nskey='dataset_input')
[docs]
@functools.lru_cache(maxsize=32)
def load_commondata(self):
return [d.load_commondata() for d in self.datasets]
[docs]
def load_commondata_instance(self):
"""
Given Experiment load list of validphys.coredata.CommonData
objects with cuts already applied
"""
commodata_list = []
for dataset in self.datasets:
cd = dataset.commondata.load()
if dataset.cuts is None:
commodata_list.append(cd)
else:
commodata_list.append(cd.with_cuts(dataset.cuts.load()))
return commodata_list
@property
def thspec(self):
# TODO: Is this good enough? Should we explicitly pass the theory
return self.datasets[0].thspec
def __str__(self):
return self.name
# Need this so that it doesn't try to iterte over itself.
@property
def as_markdown(self):
return str(self)
[docs]
def to_unweighted(self):
"""Return a copy of the group with the weights for all experiments set
to one. Note that the results cannot be used as a namespace."""
return self.__class__(
name=self.name, datasets=[ds.to_unweighted() for ds in self.datasets], dsinputs=None
)
[docs]
class FitSpec(TupleComp):
def __init__(self, name, path):
self.name = name
self.path = path
self.label = name
super().__init__(name, path)
def __iter__(self):
yield self.name
yield self.path
def __str__(self):
return self.label
__slots__ = ('label', 'name', 'path')
[docs]
class HyperscanSpec(FitSpec):
"""The hyperscan spec is just a special case of FitSpec"""
def __init__(self, name, path):
super().__init__(name, path)
self._tries_files = None
@property
def tries_files(self):
"""Return a dictionary with all tries.json files mapped to their replica number"""
if self._tries_files is None:
re_idx = re.compile(r"(?<=replica_)\d+$")
get_idx = lambda x: int(re_idx.findall(x.as_posix())[-1])
all_rep = map(get_idx, self.path.glob("nnfit/replica_*"))
# Now loop over all replicas and save them when they include a tries.json file
tries = {}
for idx in sorted(all_rep):
test_path = self.path / f"nnfit/replica_{idx}/tries.json"
if test_path.exists():
tries[idx] = test_path
self._tries_files = tries
return self._tries_files
[docs]
def get_all_trials(self, base_params=None):
"""Read all trials from all tries files.
If there are original runcard-based parameters, a reference to them can be passed
to the trials so that a full hyperparameter dictionary can be defined
Each hyperopt trial object will also have a reference to all trials in its own file
"""
all_trials = []
for trial_file in self.tries_files.values():
with open(trial_file) as tf:
run_trials = []
for trial in json.load(tf):
trial = HyperoptTrial(trial, base_params=base_params, linked_trials=run_trials)
run_trials.append(trial)
all_trials += run_trials
return all_trials
[docs]
def sample_trials(self, n=None, base_params=None, sigma=4.0):
"""Parse all trials in the hyperscan object
and then return an array of ``n`` trials read from the ``tries.json`` files
and sampled according to their reward.
If ``n`` is ``None``, no sapling is performed and all trials are returned
Returns
-------
Dictionary on the form {parameters: list of trials}
"""
all_trials_raw = self.get_all_trials(base_params=base_params)
# Drop all failing trials
all_trials = list(filter(lambda i: i.reward, all_trials_raw))
if n is None:
return all_trials
if len(all_trials) < n:
log.warning("Asked for %d trials, only %d valid trials found", n, len(all_trials))
# Compute weights proportionally to the reward (goes from 0 (worst) to 1 (best, loss=1))
rewards = np.array([i.weighted_reward for i in all_trials])
weight_raw = np.exp(sigma * rewards**2)
total = np.sum(weight_raw)
weights = weight_raw / total
return np.random.choice(all_trials, replace=False, size=n, p=weights)
[docs]
@dataclasses.dataclass
class TheoryIDSpec:
id: int
path: Path
dbpath: Path
def __iter__(self):
yield self.id
yield self.path
[docs]
def get_description(self):
return fetch_theory(self.dbpath, self.id)
def __repr__(self):
return f"{self.__class__.__name__}(id={self.id}, path={self.path!r})"
def __str__(self):
return f"Theory {self.id}"
def __hash__(self):
return hash(self.path.as_posix())
[docs]
def is_pineappl(self):
"""Check whether this theory is a pineappl-based theory
Assume yes unless a compound directory is found
"""
if (self.path / "compound").is_dir():
return False
return True
[docs]
class ThCovMatSpec:
def __init__(self, path):
self.path = path
# maxsize relatively low here, expect single experiments so one load per dataspec
[docs]
@functools.lru_cache(maxsize=8)
def load(self):
return parse_exp_mat(self.path)
def __str__(self):
return str(self.path)
[docs]
class Stats:
"""Class holding statistical information about the objects used in validphys.
This object can be a PDF or any function of a PDF (such as hadronic observable).
By convention, member 0 corresponds to the central value of the PDF.
Accordingly, the method ``central_value`` will return the result held for member 0.
Note that this is equal to the mean of the ``error_members`` only for the PDF itself
and linear functions of the PDF (such as DIS-type observable).
If you want to obtain the average of the error members you can do:
``np.mean(stats_instance.error_members, axis=0)``
"""
def __init__(self, data):
"""`data `should be N_pdf*N_bins"""
self.data = np.atleast_2d(data)
[docs]
def central_value(self):
return self.data[0]
[docs]
def error_members(self):
return self.data[1:]
[docs]
def std_error(self):
raise NotImplementedError()
[docs]
def moment(self, order):
raise NotImplementedError()
[docs]
def sample_values(self, size):
raise NotImplementedError()
[docs]
def std_interval(self, nsigma):
raise NotImplementedError()
[docs]
def errorbar68(self):
raise NotImplementedError()
[docs]
def errorbarstd(self):
return (self.central_value() - self.std_error(), self.central_value() + self.std_error())
# TODO...
...
[docs]
class MCStats(Stats):
"""Result obtained from a Monte Carlo sample"""
[docs]
def std_error(self):
return np.std(self.error_members(), axis=0)
[docs]
def moment(self, order):
return np.mean(np.power(self.error_members() - self.central_value(), order), axis=0)
[docs]
def errorbar68(self):
# Use nanpercentile here because we can have e.g. 0/0==nan normalization somewhere
down = np.nanpercentile(self.error_members(), 15.87, axis=0)
up = np.nanpercentile(self.error_members(), 84.13, axis=0)
return down, up
[docs]
def sample_values(self, size):
return np.random.choice(self, size=size)
[docs]
class SymmHessianStats(Stats):
"""Compute stats in the 'symetric' hessian format: The first index (0)
is the
central value. The rest of the indexes are results for each eigenvector.
A 'rescale_factor is allowed in case the eigenvector confidence interval
is not 68%'."""
def __init__(self, data, rescale_factor=1):
super().__init__(data)
self.rescale_factor = rescale_factor
[docs]
def errorbar68(self):
return self.errorbarstd()
[docs]
def std_error(self):
data = self.data
diffsq = (data[0] - data[1:]) ** 2
return np.sqrt(diffsq.sum(axis=0)) / self.rescale_factor
[docs]
def moment(self, order):
data = self.data
return np.sum(np.power((data[0] - data[1:]) / self.rescale_factor, order), axis=0)
[docs]
class HessianStats(SymmHessianStats):
"""Compute stats in the 'assymetric' hessian format: The first index (0)
is the
central value. The odd indexes are the
results for lower eigenvectors and the
even are the upper eigenvectors.A 'rescale_factor is allowed in
case the eigenvector confidence interval
is not 68%'."""
[docs]
def std_error(self):
data = self.data
diffsq = (data[1::2] - data[2::2]) ** 2
return np.sqrt(diffsq.sum(axis=0)) / self.rescale_factor / 2
[docs]
def moment(self, order):
data = self.data
return np.sum(np.power((data[1::2] - data[2::2]) / self.rescale_factor / 2, order), axis=0)
STAT_TYPES = dict(symmhessian=SymmHessianStats, hessian=HessianStats, replicas=MCStats)
[docs]
class Filter:
def __init__(self, indexes, label, **kwargs):
self.indexes = indexes
self.label = label
self.kwargs = kwargs
[docs]
def as_pair(self):
return self.label, self.indexes
def __str__(self):
return f'{self.label}: {self.indexes}'