| Literature DB >> 33286675 |
Ilona Kulikovskikh1, Tomislav Lipic2, Tomislav Šmuc2.
Abstract
Machines usually employ a guess-and-check strategy to analyze data: they take the data, make a guess, check the answer, adjust it with regard to the correct one if necessary, and try again on a new data set. An active learning environment guarantees better performance while training on less, but carefully chosen, data which reduces the costs of both annotating and analyzing large data sets. This issue becomes even more critical for deep learning applications. Human-like active learning integrates a variety of strategies and instructional models chosen by a teacher to contribute to learners' knowledge, while machine active learning strategies lack versatile tools for shifting the focus of instruction away from knowledge transmission to learners' knowledge construction. We approach this gap by considering an active learning environment in an educational setting. We propose a new strategy that measures the information capacity of data using the information function from the four-parameter logistic item response theory (4PL IRT). We compared the proposed strategy with the most common active learning strategies-Least Confidence and Entropy Sampling. The results of computational experiments showed that the Information Capacity strategy shares similar behavior but provides a more flexible framework for building transparent knowledge models in deep learning.Entities:
Keywords: active learning; deep learning; item information; item response theory; multiple-choice testing; pool-based sampling
Year: 2020 PMID: 33286675 PMCID: PMC7517531 DOI: 10.3390/e22080906
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524