SentencePiece is a subword tokenizer and detokenizer for natural language processing. It performs subword segmentation, supporting the byte-pair-encoding (BPE) algorithm and unigram language model, and then converts this text into an id sequence guarantee perfect reproducibility of the normalization and subword segmentation.
Source: SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text ProcessingPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Language Modelling | 98 | 9.84% |
Question Answering | 59 | 5.92% |
Sentence | 48 | 4.82% |
Text Generation | 40 | 4.02% |
Translation | 31 | 3.11% |
Retrieval | 29 | 2.91% |
Machine Translation | 27 | 2.71% |
Natural Language Understanding | 21 | 2.11% |
Sentiment Analysis | 18 | 1.81% |