Cited 1 time in
Improved utilization methodology of BERT specialized in text classification
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | So, H. | - |
| dc.contributor.author | Rhee, J. | - |
| dc.date.accessioned | 2023-04-27T19:40:54Z | - |
| dc.date.available | 2023-04-27T19:40:54Z | - |
| dc.date.issued | 2021-03-05 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/5598 | - |
| dc.description.abstract | Recent language models are pre-trained to generate universal word representations. This study proposes a BERT-Triplet model and its utilization methodology to generate word representations specialized for the text classification task. Specifically, we use class information of the data in the pre-training stage of the proposed BERT-Triplet model to closely distribute the embedding vectors of words or sentences with a high probability of being classified into the same class in the vector space, unlike existing language models. The proposed methodology obtains improvement of the classification performance and is expected to be used in various sub-fields of text classification and in language models other than BERT. © 2021 ACM. | - |
| dc.format.extent | 5 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | Association for Computing Machinery | - |
| dc.title | Improved utilization methodology of BERT specialized in text classification | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1145/3471985.3472384 | - |
| dc.identifier.scopusid | 2-s2.0-85117691807 | - |
| dc.identifier.bibliographicCitation | ACM International Conference Proceeding Series, pp 114 - 118 | - |
| dc.citation.title | ACM International Conference Proceeding Series | - |
| dc.citation.startPage | 114 | - |
| dc.citation.endPage | 118 | - |
| dc.type.docType | Conference Paper | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.subject.keywordAuthor | BERT | - |
| dc.subject.keywordAuthor | Text classification | - |
| dc.subject.keywordAuthor | Triplet loss | - |
| dc.subject.keywordAuthor | Word embedding | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
