2023-11-09 14:20:41 +01:00
|
|
|
|
import itertools
|
2021-01-15 18:32:32 +01:00
|
|
|
|
from copy import deepcopy
|
2021-01-19 18:26:40 +01:00
|
|
|
|
from typing import Union
|
2021-01-06 14:58:29 +01:00
|
|
|
|
import numpy as np
|
2021-01-15 18:32:32 +01:00
|
|
|
|
from sklearn.linear_model import LogisticRegression
|
2021-05-10 13:36:35 +02:00
|
|
|
|
from sklearn.metrics import f1_score, make_scorer, accuracy_score
|
2021-01-15 18:32:32 +01:00
|
|
|
|
from sklearn.model_selection import GridSearchCV, cross_val_predict
|
2021-05-10 13:36:35 +02:00
|
|
|
|
from tqdm import tqdm
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
import quapy as qp
|
2021-01-19 18:26:40 +01:00
|
|
|
|
from quapy import functional as F
|
2021-05-10 13:36:35 +02:00
|
|
|
|
from quapy.data import LabelledCollection
|
2021-01-15 18:32:32 +01:00
|
|
|
|
from quapy.model_selection import GridSearchQ
|
2023-11-09 14:20:41 +01:00
|
|
|
|
from quapy.method.base import BaseQuantifier, BinaryQuantifier
|
2023-11-15 10:55:13 +01:00
|
|
|
|
from quapy.method.aggregative import CC, ACC, PACC, HDy, EMQ, AggregativeQuantifier
|
2021-05-10 13:36:35 +02:00
|
|
|
|
|
|
|
|
|
try:
|
2024-01-25 14:33:41 +01:00
|
|
|
|
from . import _neural
|
2021-05-10 13:36:35 +02:00
|
|
|
|
except ModuleNotFoundError:
|
2024-01-25 14:33:41 +01:00
|
|
|
|
_neural = None
|
2023-11-09 14:20:41 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2024-01-25 14:33:41 +01:00
|
|
|
|
if _neural:
|
|
|
|
|
QuaNet = _neural.QuaNetTrainer
|
2021-05-10 13:36:35 +02:00
|
|
|
|
else:
|
|
|
|
|
QuaNet = "QuaNet is not available due to missing torch package"
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
|
2023-11-15 10:55:13 +01:00
|
|
|
|
class MedianEstimator2(BinaryQuantifier):
|
2023-11-09 14:20:41 +01:00
|
|
|
|
"""
|
|
|
|
|
This method is a meta-quantifier that returns, as the estimated class prevalence values, the median of the
|
|
|
|
|
estimation returned by differently (hyper)parameterized base quantifiers.
|
|
|
|
|
The median of unit-vectors is only guaranteed to be a unit-vector for n=2 dimensions,
|
|
|
|
|
i.e., in cases of binary quantification.
|
|
|
|
|
|
|
|
|
|
:param base_quantifier: the base, binary quantifier
|
|
|
|
|
:param random_state: a seed to be set before fitting any base quantifier (default None)
|
|
|
|
|
:param param_grid: the grid or parameters towards which the median will be computed
|
|
|
|
|
:param n_jobs: number of parllel workes
|
|
|
|
|
"""
|
|
|
|
|
def __init__(self, base_quantifier: BinaryQuantifier, param_grid: dict, random_state=None, n_jobs=None):
|
|
|
|
|
self.base_quantifier = base_quantifier
|
|
|
|
|
self.param_grid = param_grid
|
|
|
|
|
self.random_state = random_state
|
|
|
|
|
self.n_jobs = qp._get_njobs(n_jobs)
|
|
|
|
|
|
|
|
|
|
def get_params(self, deep=True):
|
|
|
|
|
return self.base_quantifier.get_params(deep)
|
|
|
|
|
|
|
|
|
|
def set_params(self, **params):
|
|
|
|
|
self.base_quantifier.set_params(**params)
|
|
|
|
|
|
|
|
|
|
def _delayed_fit(self, args):
|
|
|
|
|
with qp.util.temp_seed(self.random_state):
|
|
|
|
|
params, training = args
|
|
|
|
|
model = deepcopy(self.base_quantifier)
|
|
|
|
|
model.set_params(**params)
|
|
|
|
|
model.fit(training)
|
|
|
|
|
return model
|
|
|
|
|
|
|
|
|
|
def fit(self, training: LabelledCollection):
|
|
|
|
|
self._check_binary(training, self.__class__.__name__)
|
2023-11-15 10:55:13 +01:00
|
|
|
|
|
|
|
|
|
configs = qp.model_selection.expand_grid(self.param_grid)
|
2023-11-09 14:20:41 +01:00
|
|
|
|
self.models = qp.util.parallel(
|
|
|
|
|
self._delayed_fit,
|
2023-11-15 10:55:13 +01:00
|
|
|
|
((params, training) for params in configs),
|
2023-11-09 14:20:41 +01:00
|
|
|
|
seed=qp.environ.get('_R_SEED', None),
|
|
|
|
|
n_jobs=self.n_jobs
|
|
|
|
|
)
|
|
|
|
|
return self
|
|
|
|
|
|
|
|
|
|
def _delayed_predict(self, args):
|
|
|
|
|
model, instances = args
|
|
|
|
|
return model.quantify(instances)
|
|
|
|
|
|
|
|
|
|
def quantify(self, instances):
|
|
|
|
|
prev_preds = qp.util.parallel(
|
|
|
|
|
self._delayed_predict,
|
|
|
|
|
((model, instances) for model in self.models),
|
|
|
|
|
seed=qp.environ.get('_R_SEED', None),
|
|
|
|
|
n_jobs=self.n_jobs
|
|
|
|
|
)
|
|
|
|
|
prev_preds = np.asarray(prev_preds)
|
|
|
|
|
return np.median(prev_preds, axis=0)
|
|
|
|
|
|
|
|
|
|
|
2023-11-15 10:55:13 +01:00
|
|
|
|
class MedianEstimator(BinaryQuantifier):
|
|
|
|
|
"""
|
|
|
|
|
This method is a meta-quantifier that returns, as the estimated class prevalence values, the median of the
|
|
|
|
|
estimation returned by differently (hyper)parameterized base quantifiers.
|
|
|
|
|
The median of unit-vectors is only guaranteed to be a unit-vector for n=2 dimensions,
|
|
|
|
|
i.e., in cases of binary quantification.
|
|
|
|
|
|
|
|
|
|
:param base_quantifier: the base, binary quantifier
|
|
|
|
|
:param random_state: a seed to be set before fitting any base quantifier (default None)
|
|
|
|
|
:param param_grid: the grid or parameters towards which the median will be computed
|
|
|
|
|
:param n_jobs: number of parllel workes
|
|
|
|
|
"""
|
|
|
|
|
def __init__(self, base_quantifier: BinaryQuantifier, param_grid: dict, random_state=None, n_jobs=None):
|
|
|
|
|
self.base_quantifier = base_quantifier
|
|
|
|
|
self.param_grid = param_grid
|
|
|
|
|
self.random_state = random_state
|
|
|
|
|
self.n_jobs = qp._get_njobs(n_jobs)
|
|
|
|
|
|
|
|
|
|
def get_params(self, deep=True):
|
|
|
|
|
return self.base_quantifier.get_params(deep)
|
|
|
|
|
|
|
|
|
|
def set_params(self, **params):
|
|
|
|
|
self.base_quantifier.set_params(**params)
|
|
|
|
|
|
|
|
|
|
def _delayed_fit(self, args):
|
|
|
|
|
with qp.util.temp_seed(self.random_state):
|
|
|
|
|
params, training = args
|
|
|
|
|
model = deepcopy(self.base_quantifier)
|
|
|
|
|
model.set_params(**params)
|
|
|
|
|
model.fit(training)
|
|
|
|
|
return model
|
|
|
|
|
|
|
|
|
|
def _delayed_fit_classifier(self, args):
|
|
|
|
|
with qp.util.temp_seed(self.random_state):
|
|
|
|
|
cls_params, training = args
|
|
|
|
|
model = deepcopy(self.base_quantifier)
|
|
|
|
|
model.set_params(**cls_params)
|
|
|
|
|
predictions = model.classifier_fit_predict(training, predict_on=model.val_split)
|
|
|
|
|
return (model, predictions)
|
|
|
|
|
|
|
|
|
|
def _delayed_fit_aggregation(self, args):
|
|
|
|
|
with qp.util.temp_seed(self.random_state):
|
|
|
|
|
((model, predictions), q_params), training = args
|
|
|
|
|
model = deepcopy(model)
|
|
|
|
|
model.set_params(**q_params)
|
|
|
|
|
model.aggregation_fit(predictions, training)
|
|
|
|
|
return model
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def fit(self, training: LabelledCollection):
|
|
|
|
|
self._check_binary(training, self.__class__.__name__)
|
|
|
|
|
|
|
|
|
|
if isinstance(self.base_quantifier, AggregativeQuantifier):
|
|
|
|
|
cls_configs, q_configs = qp.model_selection.group_params(self.param_grid)
|
|
|
|
|
|
|
|
|
|
if len(cls_configs) > 1:
|
|
|
|
|
models_preds = qp.util.parallel(
|
|
|
|
|
self._delayed_fit_classifier,
|
|
|
|
|
((params, training) for params in cls_configs),
|
|
|
|
|
seed=qp.environ.get('_R_SEED', None),
|
|
|
|
|
n_jobs=self.n_jobs,
|
|
|
|
|
asarray=False
|
|
|
|
|
)
|
|
|
|
|
else:
|
|
|
|
|
model = self.base_quantifier
|
|
|
|
|
model.set_params(**cls_configs[0])
|
|
|
|
|
predictions = model.classifier_fit_predict(training, predict_on=model.val_split)
|
|
|
|
|
models_preds = [(model, predictions)]
|
|
|
|
|
|
|
|
|
|
self.models = qp.util.parallel(
|
|
|
|
|
self._delayed_fit_aggregation,
|
|
|
|
|
((setup, training) for setup in itertools.product(models_preds, q_configs)),
|
|
|
|
|
seed=qp.environ.get('_R_SEED', None),
|
|
|
|
|
n_jobs=self.n_jobs,
|
|
|
|
|
asarray=False
|
|
|
|
|
)
|
|
|
|
|
else:
|
|
|
|
|
configs = qp.model_selection.expand_grid(self.param_grid)
|
|
|
|
|
self.models = qp.util.parallel(
|
|
|
|
|
self._delayed_fit,
|
|
|
|
|
((params, training) for params in configs),
|
|
|
|
|
seed=qp.environ.get('_R_SEED', None),
|
|
|
|
|
n_jobs=self.n_jobs,
|
|
|
|
|
asarray=False
|
|
|
|
|
)
|
|
|
|
|
return self
|
|
|
|
|
|
|
|
|
|
def _delayed_predict(self, args):
|
|
|
|
|
model, instances = args
|
|
|
|
|
return model.quantify(instances)
|
|
|
|
|
|
|
|
|
|
def quantify(self, instances):
|
|
|
|
|
prev_preds = qp.util.parallel(
|
|
|
|
|
self._delayed_predict,
|
|
|
|
|
((model, instances) for model in self.models),
|
|
|
|
|
seed=qp.environ.get('_R_SEED', None),
|
|
|
|
|
n_jobs=self.n_jobs,
|
|
|
|
|
asarray=False
|
|
|
|
|
)
|
|
|
|
|
prev_preds = np.asarray(prev_preds)
|
|
|
|
|
return np.median(prev_preds, axis=0)
|
|
|
|
|
|
2023-11-09 14:20:41 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
class Ensemble(BaseQuantifier):
|
|
|
|
|
VALID_POLICIES = {'ave', 'ptr', 'ds'} | qp.error.QUANTIFICATION_ERROR_NAMES
|
|
|
|
|
|
|
|
|
|
"""
|
2021-12-15 15:27:43 +01:00
|
|
|
|
Implementation of the Ensemble methods for quantification described by
|
|
|
|
|
`Pérez-Gállego et al., 2017 <https://www.sciencedirect.com/science/article/pii/S1566253516300628>`_
|
2021-01-06 14:58:29 +01:00
|
|
|
|
and
|
2021-12-15 15:27:43 +01:00
|
|
|
|
`Pérez-Gállego et al., 2019 <https://www.sciencedirect.com/science/article/pii/S1566253517303652>`_.
|
|
|
|
|
The policies implemented include:
|
|
|
|
|
|
|
|
|
|
- Average (`policy='ave'`): computes class prevalence estimates as the average of the estimates
|
|
|
|
|
returned by the base quantifiers.
|
|
|
|
|
- Training Prevalence (`policy='ptr'`): applies a dynamic selection to the ensemble’s members by retaining only
|
|
|
|
|
those members such that the class prevalence values in the samples they use as training set are closest to
|
|
|
|
|
preliminary class prevalence estimates computed as the average of the estimates of all the members. The final
|
|
|
|
|
estimate is recomputed by considering only the selected members.
|
|
|
|
|
- Distribution Similarity (`policy='ds'`): performs a dynamic selection of base members by retaining
|
|
|
|
|
the members trained on samples whose distribution of posterior probabilities is closest, in terms of the
|
|
|
|
|
Hellinger Distance, to the distribution of posterior probabilities in the test sample
|
|
|
|
|
- Accuracy (`policy='<valid error name>'`): performs a static selection of the ensemble members by
|
|
|
|
|
retaining those that minimize a quantification error measure, which is passed as an argument.
|
|
|
|
|
|
|
|
|
|
Example:
|
|
|
|
|
|
|
|
|
|
>>> model = Ensemble(quantifier=ACC(LogisticRegression()), size=30, policy='ave', n_jobs=-1)
|
|
|
|
|
|
|
|
|
|
:param quantifier: base quantification member of the ensemble
|
|
|
|
|
:param size: number of members
|
|
|
|
|
:param red_size: number of members to retain after selection (depending on the policy)
|
|
|
|
|
:param min_pos: minimum number of positive instances to consider a sample as valid
|
|
|
|
|
:param policy: the selection policy; available policies include: `ave` (default), `ptr`, `ds`, and accuracy
|
|
|
|
|
(which is instantiated via a valid error name, e.g., `mae`)
|
|
|
|
|
:param max_sample_size: maximum number of instances to consider in the samples (set to None
|
|
|
|
|
to indicate no limit, default)
|
|
|
|
|
:param val_split: a float in range (0,1) indicating the proportion of data to be used as a stratified held-out
|
|
|
|
|
validation split, or a :class:`quapy.data.base.LabelledCollection` (the split itself).
|
|
|
|
|
:param n_jobs: number of parallel workers (default 1)
|
|
|
|
|
:param verbose: set to True (default is False) to get some information in standard output
|
2021-01-06 14:58:29 +01:00
|
|
|
|
"""
|
|
|
|
|
|
2021-01-22 18:01:51 +01:00
|
|
|
|
def __init__(self,
|
|
|
|
|
quantifier: BaseQuantifier,
|
|
|
|
|
size=50,
|
|
|
|
|
red_size=25,
|
2021-01-28 18:22:43 +01:00
|
|
|
|
min_pos=5,
|
2021-01-22 18:01:51 +01:00
|
|
|
|
policy='ave',
|
|
|
|
|
max_sample_size=None,
|
2021-12-15 15:27:43 +01:00
|
|
|
|
val_split:Union[qp.data.LabelledCollection, float]=None,
|
2022-06-14 09:35:39 +02:00
|
|
|
|
n_jobs=None,
|
2021-01-22 18:01:51 +01:00
|
|
|
|
verbose=False):
|
2021-01-22 09:58:12 +01:00
|
|
|
|
assert policy in Ensemble.VALID_POLICIES, \
|
|
|
|
|
f'unknown policy={policy}; valid are {Ensemble.VALID_POLICIES}'
|
|
|
|
|
assert max_sample_size is None or max_sample_size > 0, \
|
2021-01-22 18:01:51 +01:00
|
|
|
|
'wrong value for max_sample_size; set it to a positive number or None'
|
2021-01-06 14:58:29 +01:00
|
|
|
|
self.base_quantifier = quantifier
|
|
|
|
|
self.size = size
|
|
|
|
|
self.min_pos = min_pos
|
|
|
|
|
self.red_size = red_size
|
|
|
|
|
self.policy = policy
|
2021-01-22 18:01:51 +01:00
|
|
|
|
self.val_split = val_split
|
2023-02-08 19:06:53 +01:00
|
|
|
|
self.n_jobs = qp._get_njobs(n_jobs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
self.post_proba_fn = None
|
2021-01-19 18:26:40 +01:00
|
|
|
|
self.verbose = verbose
|
2021-01-22 09:58:12 +01:00
|
|
|
|
self.max_sample_size = max_sample_size
|
2021-01-19 18:26:40 +01:00
|
|
|
|
|
2021-12-15 15:27:43 +01:00
|
|
|
|
def _sout(self, msg):
|
2021-01-19 18:26:40 +01:00
|
|
|
|
if self.verbose:
|
|
|
|
|
print('[Ensemble]' + msg)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2021-05-10 13:36:35 +02:00
|
|
|
|
def fit(self, data: qp.data.LabelledCollection, val_split: Union[qp.data.LabelledCollection, float] = None):
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-05-10 13:36:35 +02:00
|
|
|
|
if self.policy == 'ds' and not data.binary:
|
2021-01-06 14:58:29 +01:00
|
|
|
|
raise ValueError(f'ds policy is only defined for binary quantification, but this dataset is not binary')
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-01-22 18:01:51 +01:00
|
|
|
|
if val_split is None:
|
|
|
|
|
val_split = self.val_split
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
# randomly chooses the prevalences for each member of the ensemble (preventing classes with less than
|
|
|
|
|
# min_pos positive examples)
|
2021-01-25 08:54:10 +01:00
|
|
|
|
sample_size = len(data) if self.max_sample_size is None else min(self.max_sample_size, len(data))
|
|
|
|
|
prevs = [_draw_simplex(ndim=data.n_classes, min_val=self.min_pos / sample_size) for _ in range(self.size)]
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
posteriors = None
|
|
|
|
|
if self.policy == 'ds':
|
|
|
|
|
# precompute the training posterior probabilities
|
2021-12-15 15:27:43 +01:00
|
|
|
|
posteriors, self.post_proba_fn = self._ds_policy_get_posteriors(data)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
is_static_policy = (self.policy in qp.error.QUANTIFICATION_ERROR_NAMES)
|
2021-01-25 08:54:10 +01:00
|
|
|
|
|
2021-01-27 09:54:41 +01:00
|
|
|
|
args = (
|
|
|
|
|
(self.base_quantifier, data, val_split, prev, posteriors, is_static_policy, self.verbose, sample_size)
|
|
|
|
|
for prev in prevs
|
2021-01-06 14:58:29 +01:00
|
|
|
|
)
|
2021-01-27 09:54:41 +01:00
|
|
|
|
self.ensemble = qp.util.parallel(
|
|
|
|
|
_delayed_new_instance,
|
2021-01-28 18:22:43 +01:00
|
|
|
|
tqdm(args, desc='fitting ensamble', total=self.size) if self.verbose else args,
|
2024-01-29 09:43:29 +01:00
|
|
|
|
asarray=False,
|
2021-01-27 09:54:41 +01:00
|
|
|
|
n_jobs=self.n_jobs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
# static selection policy (the name of a quantification-oriented error function to minimize)
|
|
|
|
|
if self.policy in qp.error.QUANTIFICATION_ERROR_NAMES:
|
2021-12-15 15:27:43 +01:00
|
|
|
|
self._accuracy_policy(error_name=self.policy)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2021-12-15 15:27:43 +01:00
|
|
|
|
self._sout('Fit [Done]')
|
2021-01-06 14:58:29 +01:00
|
|
|
|
return self
|
|
|
|
|
|
|
|
|
|
def quantify(self, instances):
|
2021-01-27 09:54:41 +01:00
|
|
|
|
predictions = np.asarray(
|
|
|
|
|
qp.util.parallel(_delayed_quantify, ((Qi, instances) for Qi in self.ensemble), n_jobs=self.n_jobs)
|
|
|
|
|
)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
if self.policy == 'ptr':
|
2021-12-15 15:27:43 +01:00
|
|
|
|
predictions = self._ptr_policy(predictions)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
elif self.policy == 'ds':
|
2021-12-15 15:27:43 +01:00
|
|
|
|
predictions = self._ds_policy(predictions, instances)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
predictions = np.mean(predictions, axis=0)
|
|
|
|
|
return F.normalize_prevalence(predictions)
|
|
|
|
|
|
|
|
|
|
def set_params(self, **parameters):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
This function should not be used within :class:`quapy.model_selection.GridSearchQ` (is here for compatibility
|
|
|
|
|
with the abstract class).
|
|
|
|
|
Instead, use `Ensemble(GridSearchQ(q),...)`, with `q` a Quantifier (recommended), or
|
2023-01-27 18:13:23 +01:00
|
|
|
|
`Ensemble(Q(GridSearchCV(l)))` with `Q` a quantifier class that has a classifier `l` optimized for
|
2023-02-08 19:06:53 +01:00
|
|
|
|
classification (not recommended).
|
2021-12-15 15:27:43 +01:00
|
|
|
|
|
|
|
|
|
:param parameters: dictionary
|
|
|
|
|
:return: raises an Exception
|
|
|
|
|
"""
|
2021-01-06 14:58:29 +01:00
|
|
|
|
raise NotImplementedError(f'{self.__class__.__name__} should not be used within GridSearchQ; '
|
2021-01-19 18:26:40 +01:00
|
|
|
|
f'instead, use Ensemble(GridSearchQ(q),...), with q a Quantifier (recommended), '
|
2023-01-27 18:13:23 +01:00
|
|
|
|
f'or Ensemble(Q(GridSearchCV(l))) with Q a quantifier class that has a classifier '
|
2021-01-19 18:26:40 +01:00
|
|
|
|
f'l optimized for classification (not recommended).')
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
def get_params(self, deep=True):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
This function should not be used within :class:`quapy.model_selection.GridSearchQ` (is here for compatibility
|
|
|
|
|
with the abstract class).
|
|
|
|
|
Instead, use `Ensemble(GridSearchQ(q),...)`, with `q` a Quantifier (recommended), or
|
2023-01-27 18:13:23 +01:00
|
|
|
|
`Ensemble(Q(GridSearchCV(l)))` with `Q` a quantifier class that has a classifier `l` optimized for
|
2023-02-08 19:06:53 +01:00
|
|
|
|
classification (not recommended).
|
2021-12-15 15:27:43 +01:00
|
|
|
|
|
2023-02-08 19:06:53 +01:00
|
|
|
|
:param deep: for compatibility with scikit-learn
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:return: raises an Exception
|
|
|
|
|
"""
|
2023-02-08 19:06:53 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
2021-12-15 15:27:43 +01:00
|
|
|
|
def _accuracy_policy(self, error_name):
|
2021-01-06 14:58:29 +01:00
|
|
|
|
"""
|
|
|
|
|
Selects the red_size best performant quantifiers in a static way (i.e., dropping all non-selected instances).
|
|
|
|
|
For each model in the ensemble, the performance is measured in terms of _error_name_ on the quantification of
|
|
|
|
|
the samples used for training the rest of the models in the ensemble.
|
|
|
|
|
"""
|
2023-02-14 19:15:59 +01:00
|
|
|
|
from quapy.evaluation import evaluate_on_samples
|
2021-01-27 09:54:41 +01:00
|
|
|
|
error = qp.error.from_name(error_name)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
tests = [m[3] for m in self.ensemble]
|
|
|
|
|
scores = []
|
|
|
|
|
for i, model in enumerate(self.ensemble):
|
2023-02-13 19:27:48 +01:00
|
|
|
|
scores.append(evaluate_on_samples(model[0], tests[:i] + tests[i + 1:], error))
|
2021-01-06 14:58:29 +01:00
|
|
|
|
order = np.argsort(scores)
|
|
|
|
|
|
2021-01-28 18:22:43 +01:00
|
|
|
|
self.ensemble = _select_k(self.ensemble, order, k=self.red_size)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2021-12-15 15:27:43 +01:00
|
|
|
|
def _ptr_policy(self, predictions):
|
2021-01-06 14:58:29 +01:00
|
|
|
|
"""
|
|
|
|
|
Selects the predictions made by models that have been trained on samples with a prevalence that is most similar
|
|
|
|
|
to a first approximation of the test prevalence as made by all models in the ensemble.
|
|
|
|
|
"""
|
|
|
|
|
test_prev_estim = predictions.mean(axis=0)
|
|
|
|
|
tr_prevs = [m[1] for m in self.ensemble]
|
|
|
|
|
ptr_differences = [qp.error.mse(ptr_i, test_prev_estim) for ptr_i in tr_prevs]
|
|
|
|
|
order = np.argsort(ptr_differences)
|
2021-01-28 18:22:43 +01:00
|
|
|
|
return _select_k(predictions, order, k=self.red_size)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2021-12-15 15:27:43 +01:00
|
|
|
|
def _ds_policy_get_posteriors(self, data: LabelledCollection):
|
2021-01-06 14:58:29 +01:00
|
|
|
|
"""
|
2024-01-29 09:43:29 +01:00
|
|
|
|
In the original article, there are some aspects regarding this method that are not mentioned. The paper says
|
2021-01-06 14:58:29 +01:00
|
|
|
|
that the distribution of posterior probabilities from training and test examples is compared by means of the
|
|
|
|
|
Hellinger Distance. However, how these posterior probabilities are generated is not specified. In the article,
|
|
|
|
|
a Logistic Regressor (LR) is used as the classifier device and that could be used for this purpose. However, in
|
2021-01-22 09:58:12 +01:00
|
|
|
|
general, a Quantifier is not necessarily an instance of Aggreggative Probabilistic Quantifiers, and so, that the
|
2021-01-06 14:58:29 +01:00
|
|
|
|
quantifier builds on top of a probabilistic classifier cannot be given for granted. Additionally, it would not
|
2024-01-29 09:43:29 +01:00
|
|
|
|
be correct to generate the posterior probabilities for training instances that have concurred in training the
|
2021-01-06 14:58:29 +01:00
|
|
|
|
classifier that generates them.
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
This function thus generates the posterior probabilities for all training documents in a cross-validation way,
|
2024-01-29 09:43:29 +01:00
|
|
|
|
using LR with hyperparameters that have previously been optimized via grid search in 5FCV.
|
|
|
|
|
|
|
|
|
|
:param data: a LabelledCollection
|
|
|
|
|
:return: (P,f,) where P is an ndarray containing the posterior probabilities of the training data, generated via
|
|
|
|
|
cross-validation and using an optimized LR, and the function to be used in order to generate posterior
|
|
|
|
|
probabilities for test instances.
|
2021-01-06 14:58:29 +01:00
|
|
|
|
"""
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
X, y = data.Xy
|
|
|
|
|
lr_base = LogisticRegression(class_weight='balanced', max_iter=1000)
|
|
|
|
|
|
2024-01-29 09:43:29 +01:00
|
|
|
|
param_grid = {'C': np.logspace(-4, 4, 9)}
|
|
|
|
|
optim = GridSearchCV(lr_base, param_grid=param_grid, cv=5, n_jobs=self.n_jobs, refit=True).fit(X, y)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2024-01-29 09:43:29 +01:00
|
|
|
|
posteriors = cross_val_predict(optim.best_estimator_, X, y, cv=5, n_jobs=self.n_jobs, method='predict_proba')
|
2021-01-06 14:58:29 +01:00
|
|
|
|
posteriors_generator = optim.best_estimator_.predict_proba
|
|
|
|
|
|
|
|
|
|
return posteriors, posteriors_generator
|
|
|
|
|
|
2021-12-15 15:27:43 +01:00
|
|
|
|
def _ds_policy(self, predictions, test):
|
2021-01-06 14:58:29 +01:00
|
|
|
|
test_posteriors = self.post_proba_fn(test)
|
|
|
|
|
test_distribution = get_probability_distribution(test_posteriors)
|
|
|
|
|
tr_distributions = [m[2] for m in self.ensemble]
|
|
|
|
|
dist = [F.HellingerDistance(tr_dist_i, test_distribution) for tr_dist_i in tr_distributions]
|
|
|
|
|
order = np.argsort(dist)
|
2021-01-28 18:22:43 +01:00
|
|
|
|
return _select_k(predictions, order, k=self.red_size)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2021-01-18 16:52:19 +01:00
|
|
|
|
@property
|
|
|
|
|
def aggregative(self):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Indicates that the quantifier is not aggregative.
|
|
|
|
|
|
|
|
|
|
:return: False
|
|
|
|
|
"""
|
2021-01-19 18:26:40 +01:00
|
|
|
|
return False
|
2021-01-18 16:52:19 +01:00
|
|
|
|
|
|
|
|
|
@property
|
|
|
|
|
def probabilistic(self):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Indicates that the quantifier is not probabilistic.
|
|
|
|
|
|
|
|
|
|
:return: False
|
|
|
|
|
"""
|
2021-01-19 18:26:40 +01:00
|
|
|
|
return False
|
2021-01-18 16:52:19 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
def get_probability_distribution(posterior_probabilities, bins=8):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Gets a histogram out of the posterior probabilities (only for the binary case).
|
|
|
|
|
|
|
|
|
|
:param posterior_probabilities: array-like of shape `(n_instances, 2,)`
|
|
|
|
|
:param bins: integer
|
|
|
|
|
:return: `np.ndarray` with the relative frequencies for each bin (for the positive class only)
|
|
|
|
|
"""
|
2021-05-10 13:36:35 +02:00
|
|
|
|
assert posterior_probabilities.shape[1] == 2, 'the posterior probabilities do not seem to be for a binary problem'
|
|
|
|
|
posterior_probabilities = posterior_probabilities[:, 1] # take the positive posteriors only
|
2021-01-06 14:58:29 +01:00
|
|
|
|
distribution, _ = np.histogram(posterior_probabilities, bins=bins, range=(0, 1), density=True)
|
|
|
|
|
return distribution
|
|
|
|
|
|
|
|
|
|
|
2021-01-28 18:22:43 +01:00
|
|
|
|
def _select_k(elements, order, k):
|
2021-01-06 14:58:29 +01:00
|
|
|
|
return [elements[idx] for idx in order[:k]]
|
|
|
|
|
|
|
|
|
|
|
2021-01-27 09:54:41 +01:00
|
|
|
|
def _delayed_new_instance(args):
|
|
|
|
|
base_quantifier, data, val_split, prev, posteriors, keep_samples, verbose, sample_size = args
|
2021-01-19 18:26:40 +01:00
|
|
|
|
if verbose:
|
2021-01-22 09:58:12 +01:00
|
|
|
|
print(f'\tfit-start for prev {F.strprev(prev)}, sample_size={sample_size}')
|
2021-01-06 14:58:29 +01:00
|
|
|
|
model = deepcopy(base_quantifier)
|
2021-01-22 18:01:51 +01:00
|
|
|
|
|
|
|
|
|
if val_split is not None:
|
|
|
|
|
if isinstance(val_split, float):
|
|
|
|
|
assert 0 < val_split < 1, 'val_split should be in (0,1)'
|
2021-05-10 13:36:35 +02:00
|
|
|
|
data, val_split = data.split_stratified(train_prop=1 - val_split)
|
2021-01-22 18:01:51 +01:00
|
|
|
|
|
2021-01-22 09:58:12 +01:00
|
|
|
|
sample_index = data.sampling_index(sample_size, *prev)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
sample = data.sampling_from_index(sample_index)
|
2021-01-22 18:01:51 +01:00
|
|
|
|
|
|
|
|
|
if val_split is not None:
|
2021-01-19 18:26:40 +01:00
|
|
|
|
model.fit(sample, val_split=val_split)
|
2021-01-22 18:01:51 +01:00
|
|
|
|
else:
|
|
|
|
|
model.fit(sample)
|
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
tr_prevalence = sample.prevalence()
|
|
|
|
|
tr_distribution = get_probability_distribution(posteriors[sample_index]) if (posteriors is not None) else None
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-01-19 18:26:40 +01:00
|
|
|
|
if verbose:
|
|
|
|
|
print(f'\t\--fit-ended for prev {F.strprev(prev)}')
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
return (model, tr_prevalence, tr_distribution, sample if keep_samples else None)
|
|
|
|
|
|
|
|
|
|
|
2021-01-27 09:54:41 +01:00
|
|
|
|
def _delayed_quantify(args):
|
|
|
|
|
quantifier, instances = args
|
2021-01-06 14:58:29 +01:00
|
|
|
|
return quantifier[0].quantify(instances)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def _draw_simplex(ndim, min_val, max_trials=100):
|
|
|
|
|
"""
|
2024-01-29 09:43:29 +01:00
|
|
|
|
Returns a uniform sampling from the ndim-dimensional simplex but guarantees that all dimensions
|
2021-01-06 14:58:29 +01:00
|
|
|
|
are >= min_class_prev (for min_val>0, this makes the sampling not truly uniform)
|
2024-01-29 09:43:29 +01:00
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
:param ndim: number of dimensions of the simplex
|
|
|
|
|
:param min_val: minimum class prevalence allowed. If less than 1/ndim a ValueError will be throw since
|
|
|
|
|
there is no possible solution.
|
|
|
|
|
:return: a sample from the ndim-dimensional simplex that is uniform in S(ndim)-R where S(ndim) is the simplex
|
|
|
|
|
and R is the simplex subset containing dimensions lower than min_val
|
|
|
|
|
"""
|
2021-05-10 13:36:35 +02:00
|
|
|
|
if min_val >= 1 / ndim:
|
2021-01-06 14:58:29 +01:00
|
|
|
|
raise ValueError(f'no sample can be draw from the {ndim}-dimensional simplex so that '
|
|
|
|
|
f'all its values are >={min_val} (try with a larger value for min_pos)')
|
|
|
|
|
trials = 0
|
|
|
|
|
while True:
|
|
|
|
|
u = F.uniform_simplex_sampling(ndim)
|
|
|
|
|
if all(u >= min_val):
|
|
|
|
|
return u
|
|
|
|
|
trials += 1
|
|
|
|
|
if trials >= max_trials:
|
|
|
|
|
raise ValueError(f'it looks like finding a random simplex with all its dimensions being'
|
|
|
|
|
f'>= {min_val} is unlikely (it failed after {max_trials} trials)')
|
|
|
|
|
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def _instantiate_ensemble(classifier, base_quantifier_class, param_grid, optim, param_model_sel, **kwargs):
|
2021-01-06 14:58:29 +01:00
|
|
|
|
if optim is None:
|
2023-01-27 18:13:23 +01:00
|
|
|
|
base_quantifier = base_quantifier_class(classifier)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
elif optim in qp.error.CLASSIFICATION_ERROR:
|
2021-02-17 18:05:22 +01:00
|
|
|
|
if optim == qp.error.f1e:
|
|
|
|
|
scoring = make_scorer(f1_score)
|
|
|
|
|
elif optim == qp.error.acce:
|
|
|
|
|
scoring = make_scorer(accuracy_score)
|
2023-01-27 18:13:23 +01:00
|
|
|
|
classifier = GridSearchCV(classifier, param_grid, scoring=scoring)
|
|
|
|
|
base_quantifier = base_quantifier_class(classifier)
|
2021-01-27 09:54:41 +01:00
|
|
|
|
else:
|
2023-01-27 18:13:23 +01:00
|
|
|
|
base_quantifier = GridSearchQ(base_quantifier_class(classifier),
|
2021-01-06 14:58:29 +01:00
|
|
|
|
param_grid=param_grid,
|
2021-01-19 18:26:40 +01:00
|
|
|
|
**param_model_sel,
|
2021-01-06 14:58:29 +01:00
|
|
|
|
error=optim)
|
|
|
|
|
|
|
|
|
|
return Ensemble(base_quantifier, **kwargs)
|
|
|
|
|
|
|
|
|
|
|
2021-01-19 18:26:40 +01:00
|
|
|
|
def _check_error(error):
|
|
|
|
|
if error is None:
|
|
|
|
|
return None
|
|
|
|
|
if error in qp.error.QUANTIFICATION_ERROR or error in qp.error.CLASSIFICATION_ERROR:
|
|
|
|
|
return error
|
|
|
|
|
elif isinstance(error, str):
|
2021-01-27 09:54:41 +01:00
|
|
|
|
return qp.error.from_name(error)
|
2021-01-19 18:26:40 +01:00
|
|
|
|
else:
|
|
|
|
|
raise ValueError(f'unexpected error type; must either be a callable function or a str representing\n'
|
|
|
|
|
f'the name of an error function in {qp.error.ERROR_NAMES}')
|
|
|
|
|
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def ensembleFactory(classifier, base_quantifier_class, param_grid=None, optim=None, param_model_sel: dict = None,
|
2021-05-10 13:36:35 +02:00
|
|
|
|
**kwargs):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Ensemble factory. Provides a unified interface for instantiating ensembles that can be optimized (via model
|
|
|
|
|
selection for quantification) for a given evaluation metric using :class:`quapy.model_selection.GridSearchQ`.
|
|
|
|
|
If the evaluation metric is classification-oriented
|
|
|
|
|
(instead of quantification-oriented), then the optimization will be carried out via sklearn's
|
|
|
|
|
`GridSearchCV <https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.GridSearchCV.html>`_.
|
|
|
|
|
|
2021-12-15 15:46:15 +01:00
|
|
|
|
Example to instantiate an :class:`Ensemble` based on :class:`quapy.method.aggregative.PACC`
|
|
|
|
|
in which the base members are optimized for :meth:`quapy.error.mae` via
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`. The ensemble follows the policy `Accuracy` based
|
|
|
|
|
on :meth:`quapy.error.mae` (the same measure being optimized),
|
|
|
|
|
meaning that a static selection of members of the ensemble is made based on their performance
|
|
|
|
|
in terms of this error.
|
|
|
|
|
|
|
|
|
|
>>> param_grid = {
|
|
|
|
|
>>> 'C': np.logspace(-3,3,7),
|
|
|
|
|
>>> 'class_weight': ['balanced', None]
|
|
|
|
|
>>> }
|
|
|
|
|
>>> param_mod_sel = {
|
|
|
|
|
>>> 'sample_size': 500,
|
|
|
|
|
>>> 'protocol': 'app'
|
|
|
|
|
>>> }
|
|
|
|
|
>>> common={
|
|
|
|
|
>>> 'max_sample_size': 1000,
|
|
|
|
|
>>> 'n_jobs': -1,
|
|
|
|
|
>>> 'param_grid': param_grid,
|
|
|
|
|
>>> 'param_mod_sel': param_mod_sel,
|
|
|
|
|
>>> }
|
|
|
|
|
>>>
|
|
|
|
|
>>> ensembleFactory(LogisticRegression(), PACC, optim='mae', policy='mae', **common)
|
2021-12-15 15:27:43 +01:00
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
:param classifier: sklearn's Estimator that generates a classifier
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:param base_quantifier_class: a class of quantifiers
|
|
|
|
|
:param param_grid: a dictionary with the grid of parameters to optimize for
|
|
|
|
|
:param optim: a valid quantification or classification error, or a string name of it
|
|
|
|
|
:param param_model_sel: a dictionary containing any keyworded argument to pass to
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`
|
|
|
|
|
:param kwargs: kwargs for the class :class:`Ensemble`
|
|
|
|
|
:return: an instance of :class:`Ensemble`
|
|
|
|
|
"""
|
2021-05-10 13:36:35 +02:00
|
|
|
|
if optim is not None:
|
|
|
|
|
if param_grid is None:
|
|
|
|
|
raise ValueError(f'param_grid is None but optim was requested.')
|
|
|
|
|
if param_model_sel is None:
|
|
|
|
|
raise ValueError(f'param_model_sel is None but optim was requested.')
|
|
|
|
|
error = _check_error(optim)
|
2023-01-27 18:13:23 +01:00
|
|
|
|
return _instantiate_ensemble(classifier, base_quantifier_class, param_grid, error, param_model_sel, **kwargs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def ECC(classifier, param_grid=None, optim=None, param_mod_sel=None, **kwargs):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Implements an ensemble of :class:`quapy.method.aggregative.CC` quantifiers, as used by
|
|
|
|
|
`Pérez-Gállego et al., 2019 <https://www.sciencedirect.com/science/article/pii/S1566253517303652>`_.
|
|
|
|
|
|
2021-12-15 15:46:15 +01:00
|
|
|
|
Equivalent to:
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
>>> ensembleFactory(classifier, CC, param_grid, optim, param_mod_sel, **kwargs)
|
2021-12-15 15:46:15 +01:00
|
|
|
|
|
|
|
|
|
See :meth:`ensembleFactory` for further details.
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
:param classifier: sklearn's Estimator that generates a classifier
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:param param_grid: a dictionary with the grid of parameters to optimize for
|
|
|
|
|
:param optim: a valid quantification or classification error, or a string name of it
|
|
|
|
|
:param param_model_sel: a dictionary containing any keyworded argument to pass to
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`
|
|
|
|
|
:param kwargs: kwargs for the class :class:`Ensemble`
|
|
|
|
|
:return: an instance of :class:`Ensemble`
|
|
|
|
|
"""
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
return ensembleFactory(classifier, CC, param_grid, optim, param_mod_sel, **kwargs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def EACC(classifier, param_grid=None, optim=None, param_mod_sel=None, **kwargs):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Implements an ensemble of :class:`quapy.method.aggregative.ACC` quantifiers, as used by
|
|
|
|
|
`Pérez-Gállego et al., 2019 <https://www.sciencedirect.com/science/article/pii/S1566253517303652>`_.
|
|
|
|
|
|
2021-12-15 15:46:15 +01:00
|
|
|
|
Equivalent to:
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
>>> ensembleFactory(classifier, ACC, param_grid, optim, param_mod_sel, **kwargs)
|
2021-12-15 15:46:15 +01:00
|
|
|
|
|
|
|
|
|
See :meth:`ensembleFactory` for further details.
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
:param classifier: sklearn's Estimator that generates a classifier
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:param param_grid: a dictionary with the grid of parameters to optimize for
|
|
|
|
|
:param optim: a valid quantification or classification error, or a string name of it
|
|
|
|
|
:param param_model_sel: a dictionary containing any keyworded argument to pass to
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`
|
|
|
|
|
:param kwargs: kwargs for the class :class:`Ensemble`
|
|
|
|
|
:return: an instance of :class:`Ensemble`
|
|
|
|
|
"""
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
return ensembleFactory(classifier, ACC, param_grid, optim, param_mod_sel, **kwargs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def EPACC(classifier, param_grid=None, optim=None, param_mod_sel=None, **kwargs):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Implements an ensemble of :class:`quapy.method.aggregative.PACC` quantifiers.
|
|
|
|
|
|
2021-12-15 15:46:15 +01:00
|
|
|
|
Equivalent to:
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
>>> ensembleFactory(classifier, PACC, param_grid, optim, param_mod_sel, **kwargs)
|
2021-12-15 15:46:15 +01:00
|
|
|
|
|
|
|
|
|
See :meth:`ensembleFactory` for further details.
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
:param classifier: sklearn's Estimator that generates a classifier
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:param param_grid: a dictionary with the grid of parameters to optimize for
|
|
|
|
|
:param optim: a valid quantification or classification error, or a string name of it
|
|
|
|
|
:param param_model_sel: a dictionary containing any keyworded argument to pass to
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`
|
|
|
|
|
:param kwargs: kwargs for the class :class:`Ensemble`
|
|
|
|
|
:return: an instance of :class:`Ensemble`
|
|
|
|
|
"""
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
return ensembleFactory(classifier, PACC, param_grid, optim, param_mod_sel, **kwargs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def EHDy(classifier, param_grid=None, optim=None, param_mod_sel=None, **kwargs):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Implements an ensemble of :class:`quapy.method.aggregative.HDy` quantifiers, as used by
|
|
|
|
|
`Pérez-Gállego et al., 2019 <https://www.sciencedirect.com/science/article/pii/S1566253517303652>`_.
|
|
|
|
|
|
2021-12-15 15:46:15 +01:00
|
|
|
|
Equivalent to:
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
>>> ensembleFactory(classifier, HDy, param_grid, optim, param_mod_sel, **kwargs)
|
2021-12-15 15:46:15 +01:00
|
|
|
|
|
|
|
|
|
See :meth:`ensembleFactory` for further details.
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
:param classifier: sklearn's Estimator that generates a classifier
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:param param_grid: a dictionary with the grid of parameters to optimize for
|
|
|
|
|
:param optim: a valid quantification or classification error, or a string name of it
|
|
|
|
|
:param param_model_sel: a dictionary containing any keyworded argument to pass to
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`
|
|
|
|
|
:param kwargs: kwargs for the class :class:`Ensemble`
|
|
|
|
|
:return: an instance of :class:`Ensemble`
|
|
|
|
|
"""
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
return ensembleFactory(classifier, HDy, param_grid, optim, param_mod_sel, **kwargs)
|
2021-01-06 14:58:29 +01:00
|
|
|
|
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
def EEMQ(classifier, param_grid=None, optim=None, param_mod_sel=None, **kwargs):
|
2021-12-15 15:27:43 +01:00
|
|
|
|
"""
|
|
|
|
|
Implements an ensemble of :class:`quapy.method.aggregative.EMQ` quantifiers.
|
|
|
|
|
|
2021-12-15 15:46:15 +01:00
|
|
|
|
Equivalent to:
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
>>> ensembleFactory(classifier, EMQ, param_grid, optim, param_mod_sel, **kwargs)
|
2021-12-15 15:46:15 +01:00
|
|
|
|
|
|
|
|
|
See :meth:`ensembleFactory` for further details.
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
:param classifier: sklearn's Estimator that generates a classifier
|
2021-12-15 15:27:43 +01:00
|
|
|
|
:param param_grid: a dictionary with the grid of parameters to optimize for
|
|
|
|
|
:param optim: a valid quantification or classification error, or a string name of it
|
|
|
|
|
:param param_model_sel: a dictionary containing any keyworded argument to pass to
|
|
|
|
|
:class:`quapy.model_selection.GridSearchQ`
|
|
|
|
|
:param kwargs: kwargs for the class :class:`Ensemble`
|
|
|
|
|
:return: an instance of :class:`Ensemble`
|
|
|
|
|
"""
|
|
|
|
|
|
2023-01-27 18:13:23 +01:00
|
|
|
|
return ensembleFactory(classifier, EMQ, param_grid, optim, param_mod_sel, **kwargs)
|