000 | 00000nam c2200205 c 4500 | |
001 | 000045953578 | |
005 | 20230712091445 | |
007 | ta | |
008 | 180625s2018 ulk bmAC 000c eng | |
040 | ▼a 211009 ▼c 211009 ▼d 211009 | |
085 | ▼a 0510 ▼2 KDCP | |
090 | ▼a 0510 ▼b 6D36 ▼c 1082 | |
100 | 1 | ▼a 이도명 ▼g 李道明 |
245 | 1 1 | ▼a (A) unified approach to word sense representation and disambiguation / ▼d Do-myoung Lee |
260 | ▼a Seoul : ▼b Graduate School, Korea University, ▼c 2018 | |
300 | ▼a iv, 28장 ; ▼c 26 cm | |
500 | ▼a 지도교수: 이상근 | |
502 | 0 | ▼a 학위논문(석사)-- ▼b 고려대학교 대학원, ▼c 컴퓨터·전파통신공학과, ▼d 2018. 8 |
504 | ▼a 참고문헌: 장 25-28 | |
530 | ▼a PDF 파일로도 이용가능; ▼c Requires PDF file reader(application/pdf) | |
653 | ▼a Computational Intelligence ▼a Artificial neural nets ▼a Recurrent neural networks ▼a Natural language processing | |
776 | 0 | ▼t A Unified Approach to Word Sense Representation and Disambiguation ▼w (DCOLL211009)000000081710 |
900 | 1 0 | ▼a Lee, Do-myoung,, ▼e 저 |
900 | 1 0 | ▼a 이상근, ▼g 李尙根, ▼d 1971-, ▼e 지도교수 ▼0 AUTH(211009)153285 |
945 | ▼a KLPA |
전자정보
소장정보
No. | 소장처 | 청구기호 | 등록번호 | 도서상태 | 반납예정일 | 예약 | 서비스 |
---|---|---|---|---|---|---|---|
No. 1 | 소장처 과학도서관/학위논문서고/ | 청구기호 0510 6D36 1082 | 등록번호 123059621 | 도서상태 대출가능 | 반납예정일 | 예약 | 서비스 |
No. 2 | 소장처 과학도서관/학위논문서고/ | 청구기호 0510 6D36 1082 | 등록번호 123059622 | 도서상태 대출가능 | 반납예정일 | 예약 | 서비스 |
컨텐츠정보
초록
The lexical ambiguity of words has been successfully clarified by representing words at a sense level instead of a word level. This is known as word sense representation (WSR). However, WSR models are typically trained in an unsupervised fashion without any guidance from sense inventories. Therefore, the number of sense vectors assigned to a word varies from model to model. This implies that some senses are missed or unnecessarily added. Moreover, to utilize their sense vectors in natural language processing tasks, we must determine which sense of a word to choose. In this paper, we introduce a unified neural model that incorporates WSR into word sense disambiguation (WSD), thereby leveraging the sense inventory. We use bidirectional long short-term memory networks to capture the sequential information of contexts effectively. To overcome the limitation of size with the labeled dataset, we train our model in a semi-supervised fashion to scale up the size of the dataset by leveraging a large-scale unlabeled dataset. We evaluate our proposed model on both WSR and WSD tasks. The experimental results demonstrate that our model outperforms state-of-the-art on WSR task by 0.27\%, while, on WSD task, by 1.4\% in terms of Spearman's correlation and F1-score, respectively.
목차
1 Introduction 1 2 Preliminary 4 2.1 WordNet 4 2.2 Pseudo Label 5 3 Model Architecture 6 4 Training Process 9 4.1 Training with labeled data 9 4.2 Pseudo label prediction with unlabeled data 10 4.3 Training with labeled and pseudo labeled data 10 5 Performance Evaluation 13 5.1 Datasets 13 5.1.1 Labeled data 13 5.1.2 Unlabeled data 14 5.2 Experimental Settings 14 5.3 WSR evaluation 15 5.4 WSD evaluation 16 6 Model Analysis 18 7 Related Work 21 7.1 WSR 21 7.2 WSD 22 7.3 WSR and WSD 22 8 Conclusion 24 Bibliography 25 Acknowledgement 29