DC Field | Value | Language |
---|---|---|
dc.contributor.author | Choi, June Suk | ko |
dc.contributor.author | Lim, Chae-Gyun | ko |
dc.contributor.author | Choi, Ho-Jin | ko |
dc.date.accessioned | 2022-09-05T02:02:20Z | - |
dc.date.available | 2022-09-05T02:02:20Z | - |
dc.date.created | 2022-09-01 | - |
dc.date.created | 2022-09-01 | - |
dc.date.issued | 2022-01 | - |
dc.identifier.citation | IEEE International Conference on Big Data and Smart Computing (BigComp), pp.382 - 384 | - |
dc.identifier.issn | 2375-933X | - |
dc.identifier.uri | http://hdl.handle.net/10203/298312 | - |
dc.description.abstract | Adapter tuning is a parameter-efficient way of transfer learning for natural language processing, where adapter modules are inserted in a pretrained model and only the modules are updated, with the pretrained model fixed. This also allows easier continual learning, as the number of parameter linearly grows with the number of tasks. In this paper, we apply the concept of adapter tuning and further improve the method by sharing parameters among adapters plugged in different encoder layers. We attempt a lot of different configurations of parameter sharing to find the optimal setup for each GLUE task and find the settings where the model is on par with or even outperforms the default adapter tuning with just 1, 2, 3, or 4 adapters in a BERT-base model. In the experiment, we also analyze the training results to find some patterns among the configurations and their meanings in the perspective of improving transfer learning. | - |
dc.language | English | - |
dc.publisher | IEEE | - |
dc.title | Boosting Adapter Transfer Learning via Weak Parameter Sharing | - |
dc.type | Conference | - |
dc.identifier.wosid | 000835722100077 | - |
dc.identifier.scopusid | 2-s2.0-85127613275 | - |
dc.type.rims | CONF | - |
dc.citation.beginningpage | 382 | - |
dc.citation.endingpage | 384 | - |
dc.citation.publicationname | IEEE International Conference on Big Data and Smart Computing (BigComp) | - |
dc.identifier.conferencecountry | KO | - |
dc.identifier.conferencelocation | Daegu | - |
dc.identifier.doi | 10.1109/BigComp54360.2022.00086 | - |
dc.contributor.localauthor | Choi, Ho-Jin | - |
dc.contributor.nonIdAuthor | Choi, June Suk | - |
dc.contributor.nonIdAuthor | Lim, Chae-Gyun | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.