Ailln Nlp Roadmap Save

๐Ÿ—บ๏ธ ไธ€ไธช่‡ช็„ถ่ฏญ่จ€ๅค„็†็š„ๅญฆไน ่ทฏ็บฟๅ›พ

Project README

Natural Language Processing Roadmap

๐Ÿ—บ๏ธ ไธ€ไธชใ€Œ่‡ช็„ถ่ฏญ่จ€ๅค„็†ใ€็š„ๅญฆไน ่ทฏ็บฟๅ›พใ€‚

โš ๏ธ ๆณจๆ„:

  1. ่ฟ™ไธช้กน็›ฎๅŒ…ๅซไธ€ไธชๅไธบ PCB ็š„ๅฐๅฎž้ชŒ๏ผŒ่ฟ™ไธช็š„ PCB ไธๆ˜ฏๅฐๅˆท็”ต่ทฏๆฟ Printed Circuit Board๏ผŒไนŸไธๆ˜ฏ่ฟ›็จ‹ๆŽงๅˆถๅ— Process Control Block๏ผŒ่€Œๆ˜ฏ Paper Code Blog ็š„็ผฉๅ†™ใ€‚ๆˆ‘่ฎคไธบ ่ฎบๆ–‡ใ€ไปฃ็  ๅ’Œ ๅšๅฎข ่ฟ™ไธ‰ไธชไธœ่ฅฟ๏ผŒๅฏไปฅ่ฎฉๆˆ‘ไปฌๅ…ผ้กพ็†่ฎบๅ’Œๅฎž่ทตๅŒๆ—ถ๏ผŒๅฟซ้€ŸๅœฐๆŽŒๆก็Ÿฅ่ฏ†็‚น๏ผ

  2. ๆฏ็ฏ‡่ฎบๆ–‡ๅŽ้ข็š„ๆ˜Ÿๆ˜Ÿไธชๆ•ฐไปฃ่กจ่ฎบๆ–‡็š„้‡่ฆๆ€ง๏ผˆไธป่ง‚ๆ„่ง๏ผŒไป…ไพ›ๅ‚่€ƒ๏ผ‰ใ€‚

    1. ๐ŸŒŸ: ไธ€่ˆฌ๏ผ›
    2. ๐ŸŒŸ๐ŸŒŸ: ้‡่ฆ๏ผ›
    3. ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ: ้žๅธธ้‡่ฆใ€‚

1 ๅˆ†่ฏ Word Segmentation

่ฏๆ˜ฏ่ƒฝๅคŸ็‹ฌ็ซ‹ๆดปๅŠจ็š„ๆœ€ๅฐ่ฏญ่จ€ๅ•ไฝใ€‚ ๅœจ่‡ช็„ถ่ฏญ่จ€ๅค„็†ไธญ๏ผŒ้€šๅธธ้ƒฝๆ˜ฏไปฅ่ฏไฝœไธบๅŸบๆœฌๅ•ไฝ่ฟ›่กŒๅค„็†็š„ใ€‚็”ฑไบŽ่‹ฑๆ–‡ๆœฌ่บซๅ…ทๆœ‰ๅคฉ็”Ÿ็š„ไผ˜ๅŠฟ๏ผŒไปฅ็ฉบๆ ผๅˆ’ๅˆ†ๆ‰€ๆœ‰่ฏใ€‚่€Œไธญๆ–‡็š„่ฏไธŽ่ฏไน‹้—ดๆฒกๆœ‰ๆ˜Žๆ˜พ็š„ๅˆ†ๅ‰ฒๆ ‡่ฎฐ๏ผŒๆ‰€ไปฅๅœจๅšไธญๆ–‡่ฏญ่จ€ๅค„็†ๅ‰็š„้ฆ–่ฆไปปๅŠก๏ผŒๅฐฑๆ˜ฏๆŠŠ่ฟž็ปญไธญๆ–‡ๅฅๅญๅˆ†ๅ‰ฒๆˆใ€Œ่ฏๅบๅˆ—ใ€ใ€‚่ฟ™ไธชๅˆ†ๅ‰ฒ็š„่ฟ‡็จ‹ๅฐฑๅซๅˆ†่ฏใ€‚ไบ†่งฃๆ›ดๅคš

็ปผ่ฟฐ

  • ๆฑ‰่ฏญๅˆ†่ฏๆŠ€ๆœฏ็ปผ่ฟฐ {Paper} ๐ŸŒŸ
  • ๅ›ฝๅ†…ไธญๆ–‡่‡ชๅŠจๅˆ†่ฏๆŠ€ๆœฏ็ ”็ฉถ็ปผ่ฟฐ {Paper} ๐ŸŒŸ
  • ๆฑ‰่ฏญ่‡ชๅŠจๅˆ†่ฏ็š„็ ”็ฉถ็Žฐ็ŠถไธŽๅ›ฐ้šพ {Paper} ๐ŸŒŸ๐ŸŒŸ
  • ๆฑ‰่ฏญ่‡ชๅŠจๅˆ†่ฏ็ ”็ฉถ่ฏ„่ฟฐ {Paper} ๐ŸŒŸ๐ŸŒŸ
  • ไธญๆ–‡ๅˆ†่ฏๅๅนดๅˆๅ›ž้กพ: 2007-2017 {Paper} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • chinese-word-segmentation {Code}
  • ๆทฑๅบฆๅญฆไน ไธญๆ–‡ๅˆ†่ฏ่ฐƒ็ ” {Blog}

2 ่ฏๅตŒๅ…ฅ Word Embedding

่ฏๅตŒๅ…ฅๅฐฑๆ˜ฏๆ‰พๅˆฐไธ€ไธชๆ˜ ๅฐ„ๆˆ–่€…ๅ‡ฝๆ•ฐ๏ผŒ็”Ÿๆˆๅœจไธ€ไธชๆ–ฐ็š„็ฉบ้—ดไธŠ็š„่กจ็คบ๏ผŒ่ฏฅ่กจ็คบ่ขซ็งฐไธบใ€Œๅ•่ฏ่กจ็คบใ€ใ€‚ไบ†่งฃๆ›ดๅคš

็ปผ่ฟฐ

  • Word Embeddings: A Survey {Paper} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • Visualizing Attention in Transformer-Based Language Representation Models {Paper} ๐ŸŒŸ๐ŸŒŸ
  • PTMs: Pre-trained Models for Natural Language Processing: A Survey {Paper} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • Efficient Transformers: A Survey {Paper} ๐ŸŒŸ๐ŸŒŸ
  • A Survey of Transformers {Paper} ๐ŸŒŸ๐ŸŒŸ
  • Pre-Trained Models: Past, Present and Future {Paper} ๐ŸŒŸ๐ŸŒŸ
  • Pretrained Language Models for Text Generation: A Survey {Paper} ๐ŸŒŸ
  • A Practical Survey on Faster and Lighter Transformers {Paper} ๐ŸŒŸ
  • The NLP Cookbook: Modern Recipes for Transformer based Deep Learning Architectures {Paper} ๐ŸŒŸ๐ŸŒŸ

ๆ ธๅฟƒ

  • NNLM: A Neural Probabilistic Language Model {Paper} {Code} {Blog} ๐ŸŒŸ
  • W2V: Efficient Estimation of Word Representations in Vector Space {Paper} ๐ŸŒŸ๐ŸŒŸ
  • Glove: Global Vectors for Word Representation {Paper} ๐ŸŒŸ๐ŸŒŸ
  • CharCNN: Character-level Convolutional Networks for Text Classification {Paper} {Blog} ๐ŸŒŸ
  • ULMFiT: Universal Language Model Fine-tuning for Text Classification {Paper} ๐ŸŒŸ
  • SiATL: An Embarrassingly Simple Approach for Transfer Learning from Pretrained Language Models {Paper} ๐ŸŒŸ
  • FastText: Bag of Tricks for Efficient Text Classification {Paper} ๐ŸŒŸ๐ŸŒŸ
  • CoVe: Learned in Translation: Contextualized Word Vectors {Paper} ๐ŸŒŸ
  • ELMo: Deep contextualized word representations {Paper} ๐ŸŒŸ๐ŸŒŸ
  • Transformer: Attention is All you Need {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • GPT: Improving Language Understanding by Generative Pre-Training {Paper} ๐ŸŒŸ
  • GPT2: Language Models are Unsupervised Multitask Learners {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • GPT3: Language Models are Few-Shot Learners {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • GPT4: GPT-4 Technical Report {Paper} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • UniLM: Unified Language Model Pre-training for Natural Language Understanding and Generation {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer {Paper} {Code} {Blog} ๐ŸŒŸ
  • ERNIE(Baidu): Enhanced Representation through Knowledge Integration {Paper} {Code} ๐ŸŒŸ
  • ERNIE(Tsinghua): Enhanced Language Representation with Informative Entities {Paper} {Code} ๐ŸŒŸ
  • RoBERTa: A Robustly Optimized BERT Pretraining Approach {Paper} ๐ŸŒŸ
  • ALBERT: A Lite BERT for Self-supervised Learning of Language Representations {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ
  • TinyBERT: Distilling BERT for Natural Language Understanding {Paper} ๐ŸŒŸ๐ŸŒŸ
  • FastFormers: Highly Efficient Transformer Models for Natural Language Understanding {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ

ๅ…ถไป–

  • word2vec Parameter Learning Explained {Paper} ๐ŸŒŸ๐ŸŒŸ
  • Semi-supervised Sequence Learning {Paper} ๐ŸŒŸ๐ŸŒŸ
  • BERT Rediscovers the Classical NLP Pipeline {Paper} ๐ŸŒŸ
  • Pre-trained Languge Model Papers {Blog}
  • HuggingFace Transformers {Code}
  • Fudan FastNLP {Code}

3 ๆ–‡ๆœฌๅˆ†็ฑป Text Classification

็ปผ่ฟฐ

  • A Survey on Text Classification: From Shallow to Deep Learning {Paper} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • Deep Learning Based Text Classification: A Comprehensive Review {Paper} ๐ŸŒŸ๐ŸŒŸ

CNN

  • TextCNN:Convolutional Neural Networks for Sentence Classification {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • Convolutional Neural Networks for Text Categorization: Shallow Word-level vs. Deep Character-level {Paper} ๐ŸŒŸ
  • DPCNN: Deep Pyramid Convolutional Neural Networks for Text Categorization {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ

4 ๅบๅˆ—ๆ ‡ๆณจ Sequence Labeling

็ปผ่ฟฐ

  • Sequence Labeling ็š„ๅ‘ๅฑ•ๅฒ๏ผˆDNNs+CRF๏ผ‰{Blog}

Bi-LSTM + CRF

  • End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF {Paper} ๐ŸŒŸ๐ŸŒŸ

  • pytorch_NER_BiLSTM_CNN_CRF {Code}

  • NN_NER_tensorFlow {Code}

  • End-to-end-Sequence-Labeling-via-Bi-directional-LSTM-CNNs-CRF-Tutorial {Code}

  • Bi-directional LSTM-CNNs-CRF {Code}

ๅ…ถไป–

  • Sequence to Sequence Learning with Neural Networks {Paper} ๐ŸŒŸ
  • Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks {Paper} ๐ŸŒŸ

5 ๅฏน่ฏ็ณป็ปŸ Dialogue Systems

็ปผ่ฟฐ

  • A Survey on Dialogue Systems: Recent Advances and New Frontiers {Paper} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • ๅฐๅ“ฅๅ“ฅ๏ผŒๆฃ€็ดขๅผchatbotไบ†่งฃไธ€ไธ‹๏ผŸ {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • Recent Neural Methods on Slot Filling and Intent Classification for Task-Oriented Dialogue Systems: A Survey {Paper} ๐ŸŒŸ๐ŸŒŸ

Open Domain Dialogue Systems

  • HERD: Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ
  • Adversarial Learning for Neural Dialogue Generation {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ

Task Oriented Dialogue Systems

  • Joint NLU: Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ
  • BERT for Joint Intent Classification and Slot Filling {Paper} ๐ŸŒŸ
  • Sequicity: Simplifying Task-oriented Dialogue Systems with Single Sequence-to-Sequence Architectures {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ
  • Attention with Intention for a Neural Network Conversation Model {Paper} ๐ŸŒŸ
  • REDP: Few-Shot Generalization Across Dialogue Tasks {Paper} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • TEDP: Dialogue Transformers {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ

Conversational Response Selection

  • Multi-view Response Selection for Human-Computer Conversation {Paper} ๐ŸŒŸ๐ŸŒŸ
  • SMN: Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ:
  • DUA: Modeling Multi-turn Conversation with Deep Utterance Aggregation {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • DAM: Multi-Turn Response Selection for Chatbots with Deep Attention Matching Network {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • IMN: Interactive Matching Network for Multi-Turn Response Selection in Retrieval-Based Chatbots {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • Dialogue Transformers {Paper} ๐ŸŒŸ๐ŸŒŸ

6 ไธป้ข˜ๆจกๅž‹ Topic Model

LDA

  • Latent Dirichlet Allocation {Paper} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ

7 ็Ÿฅ่ฏ†ๅ›พ่ฐฑ Knowledge Graph

็ปผ่ฟฐ

  • Towards a Definition of Knowledge Graphs {Paper} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ

8 ๆ็คบๅญฆไน  Prompt Learning

็ปผ่ฟฐ

  • PPP: Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing {Paper} {Blog} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ

9 ๅ›พ็ฅž็ป็ฝ‘็ปœ Graph Neural Network

็ปผ่ฟฐ

  • Graph Neural Networks for Natural Language Processing: A Survey {Paper} ๐ŸŒŸ๐ŸŒŸ

10 ๅฅๅตŒๅ…ฅ Sentence Embedding

ๆ ธๅฟƒ

  • InferSent: Supervised Learning of Universal Sentence Representations from Natural Language Inference Data {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ
  • Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ
  • BERT-flow: On the Sentence Embeddings from Pre-trained Language Models {Paper} {Code} {Blog} ๐ŸŒŸ๐ŸŒŸ
  • SimCSE: Simple Contrastive Learning of Sentence Embeddings {Paper} {Code} ๐ŸŒŸ๐ŸŒŸ๐ŸŒŸ

ๅ‚่€ƒ

Open Source Agenda is not affiliated with "Ailln Nlp Roadmap" Project. README Source: Ailln/nlp-roadmap
Stars
83
Open Issues
0
Last Commit
1 year ago
Repository
License
MIT

Open Source Agenda Badge

Open Source Agenda Rating