2019 ifKakao 방문기

  • 2019 if Kakao는 29일에서 30일까지 이틀에 걸쳐 진행되었습니다.
  • 29일에는 ‘FE’, 30일에는 ‘데이터’에 대한 내용이 주를 이루었습니다.

Read more

Hierarchical Agglomerative clustering

A. Feature

  • Make hierarchical tree of clusters, where each depth has different number of clusters and the proper depth can be chose depending on total number of clusters.
  • A child node imply one of the meanings parent node indicates in detail.
    • Assuming clustering sentences, Keywords extracted from each cluster(node) can be regarded as hierachchical topic tree from major topic to sub topics by its parent-child relationship.
  • The results can vary greatly depending on which metric you use to calculate the distance.

Read more

FCM, The Fuzzy C-Means clustering algorithm

A. Feature

  • Each data point can belong to more than one cluster (Soft Clustering).
    • The degree of belonging to each centroid is represented by scalar between 0 and 1.
    • It seems proper to cluster which the boundary is ambiguous such as natural language task.
  • Likewise other prototype-based clustering such as K-means family, Fuzzy C-means clustering finds optimal status through repetition.

Read more

Word Movers Distance

TBD

영어 섹션입니다. 버튼 css 추가 추가된 영어 섹션입니다. 처음에는 영어로 나옵니다. get the PDF2 추가된 영어 섹션입니다2.

Read more

A Structured Self-attentive Sentence Embedding

TL;DR:

  • Previous attention needs source vector to conjuncture the relevance with task. In encoder-decoder architecture, an output of encoder is source vector while an output of decoder is target task vector.
  • Self-Attention embeds contextual information - each tokens’ significance in given task - into a matrix rather than a vector, assuming that there can be more than 1 contextual weight vector for 1 sentence, While former Attention embeds contextual information into vector.
  • Self-Attention helps model to pay attention to significant parts of sentence for target task relieving some long-term memorization burden from LSTM, and provides attention matrix for visualization.

Read more

Neural Machine Translation by jointly learning to align and translate

TL;DR:

  • This paper provided clue to solve long term dependency problem and to develop self-attention, transformer, and BERT, the most popular model in 2019.
  • It is undeniable that attention, which supports decoder to search where the relatively significant parts are, is novel approach itself compared to previous one which embed source sentence into one fixed-length vector according to distributional hypothesis.
  • Eventually, attention contributed to broaden model variation of NLP, expanding the existing options that were limited to recurrent network family.

Read more

Pagination


© 2022.03. by bigshane