Home

Awesome

(한국어) 텍스트 마이닝을 위한 튜토리얼

텍스트 마이닝을 공부하기 위한 자료입니다. 언어에 상관없이 적용할 수 있는 자연어처리 / 머신러닝 관련 자료도 포함되지만, 한국어 분석을 위한 자료들도 포함됩니다.

Contents

  1. Python basic
    1. jupyter tutorial
  2. From text to vector (KoNLPy)
    1. n-gram
    2. from text to vector using KoNLPy
  3. Word extraction and tokenization (Korean)
    1. word extractor
    2. unsupervised tokenizer
    3. noun extractor
    4. dictionary based pos tagger
  4. Document classification
    1. Logistic Regression and Lasso regression
    2. SVM (linear, RBF)
    3. k-nearest neighbors classifier
    4. Feed-forward neural network
    5. Decision Tree
    6. Naive Bayes
  5. Sequential labeling
    1. Conditional Random Field
  6. Embedding for representation
    1. Word2Vec / Doc2Vec
    2. GloVe
    3. FastText (word embedding using subword)
    4. FastText (supervised word embedding)
    5. Sparse Coding
    6. Nonnegative Matrix Factorization (NMF) for topic modeling
  7. Embedding for vector visualization
    1. MDS, ISOMAP, Locally Linear Embedding, PCA, Kernel PCA
    2. t-SNE
    3. t-SNE (detailed)
  8. Keyword / Related words analysis
    1. co-occurrence based keyword / related word analysis
  9. Document clustering
    1. k-means is good for document clustering
    2. DBSCAN, hierarchical, GMM, BGMM are not appropriate for document clustering
  10. Finding similar documents (neighbor search)
    1. Random Projection
    2. Locality Sensitive Hashing
    3. Inverted Index
  11. Graph similarity and ranking (centrality)
    1. SimRank & Random Walk with Restart
    2. PageRank, HITS, WordRank, TextRank
    3. kr-wordrank keyword extraction
  12. String similarity
    1. Levenshtein / Cosine / Jaccard distance
  13. Convolutional Neural Network (CNN)
    1. Introduction of CNN
    2. Word-level CNN for sentence classification (Yoon Kim)
    3. Character-level CNN (LeCun)
    4. BOW-CNN
  14. Recurrent Neural Network (RNN)
    1. Introduction of RNN
    2. LSTM, GRU
    3. Deep RNN & ELMo
    4. Sequence to sequence & seq2seq with attention
    5. Skip-thought vector
    6. Attention mechanism for sentence classification
    7. Hierarchical Attention Network (HAN) for document classification
    8. Transformer & BERT
  15. Applications
    1. soyspacing: heuristic Korean space correction
    2. crf-based Korean soace correction
    3. HMM & CRF-based part-of-speech tagger (morphological analyzer)
    4. semantic movie search using IMDB
  16. TBD

Thanks to

자료를 리뷰하고 함께 토론해주는 고마운 동료들이 많습니다. 특히 많은 시간과 정성을 들여 도와주는 태욱에게 고마움을 표합니다.