Talk:BERT (language model)
This article is rated Start-class on Wikipedia's content assessment scale. It is of interest to the following WikiProjects: | ||||||||||||||||||||||||||||||||||||||||||||||||
|
Title[edit]
This page should probably be moved to BERT (language representation model) rather than language model. A language model has a specific meaning in that it models the joint probability distribution of words, whereas BERT doesn't do that, although it can predict a masked word it can't give you the probability distribution.
This would also be consistent with Wikipedia's own definition of a language model.
Categories:
- Start-Class Google articles
- Unknown-importance Google articles
- WikiProject Google articles
- Start-Class Computer science articles
- Unknown-importance Computer science articles
- WikiProject Computer science articles
- Start-Class Linguistics articles
- Unknown-importance Linguistics articles
- Start-Class applied linguistics articles
- Applied Linguistics Task Force articles
- WikiProject Linguistics articles