A novel embedding approach to learn word vectors by weighting semantic relations: SemSpace


ORHAN U., Tulu C. N.

EXPERT SYSTEMS WITH APPLICATIONS, cilt.180, 2021 (SCI-Expanded) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 180
  • Basım Tarihi: 2021
  • Doi Numarası: 10.1016/j.eswa.2021.115146
  • Dergi Adı: EXPERT SYSTEMS WITH APPLICATIONS
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus, Academic Search Premier, PASCAL, Aerospace Database, Applied Science & Technology Source, Communication Abstracts, Computer & Applied Sciences, INSPEC, Metadex, Public Affairs Index, Civil Engineering Abstracts
  • Anahtar Kelimeler: SemSpace, Embedding, Word vectors, Aligning semantic relations to weights, WordNet, SEARCH
  • Çukurova Üniversitesi Adresli: Evet

Özet

In this study, we propose a novel embedding approach, called as SemSpace, to determine word vectors of synsets and to find the best weights for semantic relations. First, SemSpace finds the optimum weights to the semantic relations in WordNet by aligning them to values produced by human intelligence, and then, determines word vectors of synsets by adjusting euclidean distances among them. Proposed approach requires two inputs; first, a lexical-semantic network such as WordNet, second, a word-level similarity dataset generated by people. In the experiments, we used WordNet 3.0 data for the lexical-semantic network, and three (RG65, WS353, and MEN3K) benchmark testsets to align semantic weights. Using the aligned semantic weights and the determined word vectors, the obtained resultsresults on the benchmark testsets are compared with literature studies. According to the obtained results, it might be concluded that SemSpace is not only successful to find word level semantic similarity values and semantic weights, but also to discover new semantic relations with their semantic levels.