Augmentation-aware self-supervised learning with conditioned projector

2024
journal article
article
3
dc.abstract.enSelf-supervised learning (SSL) is a powerful technique for learning from unlabeled data. By learning to remain invariant to applied data augmentations, methods such as SimCLR and MoCo can reach quality on par with supervised approaches. However, this invariance may be detrimental for solving downstream tasks that depend on traits affected by augmentations used during pretraining, such as color. In this paper, we propose to foster sensitivity to such characteristics in the representation space by modifying the projector network, a common component of self-supervised architectures. Specifically, we supplement the projector with information about augmentations applied to images. For the projector to take advantage of this auxiliary conditioning when solving the SSL task, the feature extractor learns to preserve the augmentation information in its representations. Our approach, coined Conditional Augmentation-aware Self-supervised Learning (CASSLE), is directly applicable to typical joint-embedding SSL methods regardless of their objective functions. Moreover, it does not require major changes in the network architecture or prior knowledge of downstream tasks. In addition to an analysis of sensitivity towards different data augmentations, we conduct a series of experiments, which show that CASSLE improves over various SSL methods, reaching state-of-the-art performance in multiple downstream tasks.
dc.affiliationSzkoła Doktorska Nauk Ścisłych i Przyrodniczych
dc.affiliationWydział Matematyki i Informatyki : Instytut Informatyki i Matematyki Komputerowej
dc.contributor.authorPrzewięźlikowski, Marcin - 421101
dc.contributor.authorPyla, Mateusz - 477206
dc.contributor.authorZieliński, Bartosz - 106948
dc.contributor.authorTwardowski, Bartłomiej
dc.contributor.authorTabor, Jacek - 132362
dc.contributor.authorŚmieja, Marek - 135996
dc.date.accessioned2025-02-06T06:50:41Z
dc.date.available2025-02-06T06:50:41Z
dc.date.createdat2025-02-04T12:37:02Zen
dc.date.issued2024
dc.date.openaccess0
dc.description.accesstimew momencie opublikowania
dc.description.physical112572
dc.description.versionostateczna wersja wydawcy
dc.description.volume305
dc.identifier.doi10.1016/j.knosys.2024.112572
dc.identifier.eissn1872-7409
dc.identifier.issn0950-7051
dc.identifier.projectHORIZON-CL4-2022-HUMAN-02
dc.identifier.project101120237
dc.identifier.project2023/49/N/ST6/03268
dc.identifier.project2022/45/B/ST6/01117
dc.identifier.project2022/47/B/ST6/03397
dc.identifier.projectRYC2021-032765-I
dc.identifier.urihttps://ruj.uj.edu.pl/handle/item/547199
dc.languageeng
dc.language.containereng
dc.rightsUdzielam licencji. Uznanie autorstwa - Użycie niekomercyjne 4.0 Międzynarodowa
dc.rights.licenceCC-BY-NC
dc.rights.urihttp://creativecommons.org/licenses/by-nc/4.0/legalcode.pl
dc.share.typeinne
dc.source.integratorfalse
dc.subject.enself-supervised learning
dc.subject.enaugmentation-aware
dc.subject.encontrastive learning
dc.subject.enprojector
dc.subject.enconditional models
dc.subtypeArticle
dc.titleAugmentation-aware self-supervised learning with conditioned projector
dc.title.journalKnowledge-Based Systems
dc.typeJournalArticle
dspace.entity.typePublicationen
dc.abstract.en
Self-supervised learning (SSL) is a powerful technique for learning from unlabeled data. By learning to remain invariant to applied data augmentations, methods such as SimCLR and MoCo can reach quality on par with supervised approaches. However, this invariance may be detrimental for solving downstream tasks that depend on traits affected by augmentations used during pretraining, such as color. In this paper, we propose to foster sensitivity to such characteristics in the representation space by modifying the projector network, a common component of self-supervised architectures. Specifically, we supplement the projector with information about augmentations applied to images. For the projector to take advantage of this auxiliary conditioning when solving the SSL task, the feature extractor learns to preserve the augmentation information in its representations. Our approach, coined Conditional Augmentation-aware Self-supervised Learning (CASSLE), is directly applicable to typical joint-embedding SSL methods regardless of their objective functions. Moreover, it does not require major changes in the network architecture or prior knowledge of downstream tasks. In addition to an analysis of sensitivity towards different data augmentations, we conduct a series of experiments, which show that CASSLE improves over various SSL methods, reaching state-of-the-art performance in multiple downstream tasks.
dc.affiliation
Szkoła Doktorska Nauk Ścisłych i Przyrodniczych
dc.affiliation
Wydział Matematyki i Informatyki : Instytut Informatyki i Matematyki Komputerowej
dc.contributor.author
Przewięźlikowski, Marcin - 421101
dc.contributor.author
Pyla, Mateusz - 477206
dc.contributor.author
Zieliński, Bartosz - 106948
dc.contributor.author
Twardowski, Bartłomiej
dc.contributor.author
Tabor, Jacek - 132362
dc.contributor.author
Śmieja, Marek - 135996
dc.date.accessioned
2025-02-06T06:50:41Z
dc.date.available
2025-02-06T06:50:41Z
dc.date.createdaten
2025-02-04T12:37:02Z
dc.date.issued
2024
dc.date.openaccess
0
dc.description.accesstime
w momencie opublikowania
dc.description.physical
112572
dc.description.version
ostateczna wersja wydawcy
dc.description.volume
305
dc.identifier.doi
10.1016/j.knosys.2024.112572
dc.identifier.eissn
1872-7409
dc.identifier.issn
0950-7051
dc.identifier.project
HORIZON-CL4-2022-HUMAN-02
dc.identifier.project
101120237
dc.identifier.project
2023/49/N/ST6/03268
dc.identifier.project
2022/45/B/ST6/01117
dc.identifier.project
2022/47/B/ST6/03397
dc.identifier.project
RYC2021-032765-I
dc.identifier.uri
https://ruj.uj.edu.pl/handle/item/547199
dc.language
eng
dc.language.container
eng
dc.rights
Udzielam licencji. Uznanie autorstwa - Użycie niekomercyjne 4.0 Międzynarodowa
dc.rights.licence
CC-BY-NC
dc.rights.uri
http://creativecommons.org/licenses/by-nc/4.0/legalcode.pl
dc.share.type
inne
dc.source.integrator
false
dc.subject.en
self-supervised learning
dc.subject.en
augmentation-aware
dc.subject.en
contrastive learning
dc.subject.en
projector
dc.subject.en
conditional models
dc.subtype
Article
dc.title
Augmentation-aware self-supervised learning with conditioned projector
dc.title.journal
Knowledge-Based Systems
dc.type
JournalArticle
dspace.entity.typeen
Publication
Affiliations

* The migration of download and view statistics prior to the date of April 8, 2024 is in progress.

Views
22
Views per month
Views per city
Krakow
13
Amsterdam
1
Iwiny
1
Liszki
1
Wroclaw
1
Downloads
przewiezlikowski_et-al_augmentation-aware_self-supervised_learning_2024.pdf
15