The key idea of LAL is to train a regressor that predicts the expected error reduction for a candidate sample in a particular learning state.
The regressor is trained on 2D datasets and can score unseen data from real datasets. The method yields strategies that work well on real data from a wide range of domains.
In alipy, LAL will use a pre-extracted data provided by the authors to train the regressor. It will download the data file if no accepted file is found. You can also download 'LAL-iterativetree-simulatedunbalanced-big.npz' and 'LAL-randomtree-simulatedunbalanced-big.npz' from https://github.com/ksenia-konyushkova/LAL. and specify the dir to the file for training.
The implementation is refer to the https://github.com/ksenia-konyushkova/LAL/ directly.
References
----------
[1] Ksenia Konyushkova, and Sznitman Raphael. 2017. Learning Active Learning from Data. In The 31st Conference on Neural Information Processing Systems (NIPS 2017), 4228-4238.
__init__(self, X, y, mode='LAL_iterative', data_path='.', cls_est=50, train_slt=True, **kwargs)
Parameters: |
|
---|
download_data(self)
Download the training data for training the regressor to evaluate unlabeled data.
train_selector_from_file(self, file_path=None, reg_est=2000, reg_depth=40, feat=6)
Train a random forest as the instance selector. Note that, if the parameters of the forest is too high to your computer, it will take a lot of time for training.
Parameters: |
|
---|
select(self, label_index, unlabel_index, batch_size=1, **kwargs)
Select indexes from the unlabel_index for querying.
Parameters: |
|
---|---|
Returns: |
|
Copyright © 2018, alipy developers (BSD 3 License).