文件名称:How NLP Cracked Transfer Learing
文件大小:2.68MB
文件格式:PDF
更新时间:2021-12-24 11:11:23
深度学习 NLP 迁移学习
The year 2018 has been an inflection point for machine learning models handling text (or more accurately, Natural Language Processing or NLP for short). Our conceptual understanding of how best to represent words and sentences in a way that best captures underlying meanings and relationships is rapidly evolving. Moreover, the NLP community has been putting forward incredibly powerful components that you can freely download and use in your own models and pipelines . One of the latest milestones in this development is the release (https://ai.googleblog.com/2018/11/open-sourcing-bertstate- of-art-pre.html) of BERT (https://github.com/google-research/bert), an event described (https://twitter.com/lmthang/status/1050543868041555969) as marking the beginning of a new era in NLP. BERT is a model that broke several records for how well models can handle language-based tasks. Soon after the release of the paper describing the model, the team also open-sourced the code of the model, and made available for download versions of the model that were already pre-trained on massive datasets. This is a momentous development since it enables anyone building a machine learning model involving language processing to use this powerhouse as a readily-available component – saving the time, energy, knowledge, and resources that would have gone to training a language-processing model from scratch. (It’s been referred to as NLP’s ImageNet moment (http://ruder.io/nlp-imagenet/), referencing how years ago similar developments accelerated the development of machine learning in Computer Vision tasks)