Description
State-of-the-art NLP models for text classification without annotated data
Summary
- Natural language processing is a very exciting field right now.
- However, models of this size remain impractical for real-world use.
- If we were to use word vectors as our label representations, however, we would need annotated data to learn an alignment between the S-BERT sequence representations and the word2vec label representations.
- In the most recent version of their paper, the authors also discuss an iterative self-training procedure on top of PET which reports an impressive accuracy of $70.7\%$ on Yahoo Answers, which nearly approaches the performance of state-of-the-art supervised classification methods.