E-resources
-
Chen, Kuilin
01/2022Dissertation
Building models with limited data is one of the key steps towards the application of deep learning models in realistic scenarios. Under the framework of representation learning, we propose several novel algorithms for tackling challenging tasks in building models with limited data, including standard few-shot learning, incremental few-shot learning, and unsupervised few-shot learning. In the first part of this thesis, we propose an approach to learn low-rank representation that generalizes well to a new task using just a few training samples. High-quality representation can be found by averaging the weights of neural networks during the pre-training phase. Our approach achieves strong performance on both few-shot classification and regression benchmarks. We then consider incremental few-shot learning, in which the model incrementally learns new tasks from few-shot samples without forgetting old ones. We propose an approach to harmonize old knowledge preserving and new knowledge adaptation through quantized vectors of the learned representation. Prediction is made in a nonparametric way using similarity to learned reference vectors, which circumvents biased weights in a parametric classification layer during incremental few-shot learning. Experimental results demonstrate that the proposed method outperforms other state-of-the-art methods in incremental learning. In addition, we develop deep Laplacian eigenmaps to learn representation from unlabeled image data for downstream few-shot learning tasks. Our method learns representation by grouping similar images together and can be intuitively interpreted by random walks on augmented training data. The proposed method significantly closes the performance gap between supervised and unsupervised few-shot learning. To get us closer to general unsupervised representation learning across different data types, we present a domain-agnostic self-supervised learning method, which learns representation from unlabeled data without domain-specific data augmentations. The proposed method is adversarial perturbation based latent reconstruction (APLR), which is closely related to multi-dimensional Hirschfeld-Gebelein-Renyi maximal correlation and has theoretical guarantees on the linear probe error. APLR not only outperforms existing domain-agnostic self-supervised learning methods, but also closes the performance gap to domain-specific self-supervised learning methods on various domains, such as tabular data, images, and audio.
Author
![loading ... loading ...](themes/default/img/ajax-loading.gif)
Shelf entry
Permalink
- URL:
Impact factor
Access to the JCR database is permitted only to users from Slovenia. Your current IP address is not on the list of IP addresses with access permission, and authentication with the relevant AAI accout is required.
Year | Impact factor | Edition | Category | Classification | ||||
---|---|---|---|---|---|---|---|---|
JCR | SNIP | JCR | SNIP | JCR | SNIP | JCR | SNIP |
Select the library membership card:
If the library membership card is not in the list,
add a new one.
DRS, in which the journal is indexed
Database name | Field | Year |
---|
Links to authors' personal bibliographies | Links to information on researchers in the SICRIS system |
---|
Source: Personal bibliographies
and: SICRIS
The material is available in full text. If you wish to order the material anyway, click the Continue button.