Talk:Language model

Non statistical language models
what about non statistical language models, like cfgs? 84.162.237.4 (talk) 20:44, 7 December 2008 (UTC)
 * PCFG can also be used as a language model, and its performance is said to be worse than n-gram, though I doubt it. Took (talk) 00:18, 29 January 2009 (UTC)

Language Models
Isnt the term language models in Information Retrieval used a little differently from the NLP interpretation? — Preceding unsigned comment added by GreenEdu (talk • contribs) 16:16, 1 March 2011 (UTC)

External links modified
Hello fellow Wikipedians,

I have just modified one external link on Language model. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:
 * Added archive https://web.archive.org/web/20120302151523/http://www-speech.sri.com/projects/srilm/ to http://www-speech.sri.com/projects/srilm

When you have finished reviewing my changes, you may follow the instructions on the template below to fix any issues with the URLs.

Cheers.— InternetArchiveBot  (Report bug) 22:05, 16 December 2017 (UTC)

"Neuronal" language models?!
Recent changes in the page have replaced the word "neural" (as in "neural net language models") to "neuronal", saying that the latter is the adjective form of "neuron". While that might be true, the change is completely wrong on several accounts:


 * 1) The generally accepted term is "neural net". Nobody uses "neuronal".
 * 2) The WP page is also titled "Artificial neural network". The change to "neuronal" here is inconsistent with the wording there or elsewhere on WP.
 * 3) Even in biology, where the inspiration comes from, the network is called neural. That it is made up of neurons is a secondary detail.

I do not wish to start an edit war, so I would like to ask the editors to step in and change "neuronal" back to "neural". As I understand, WP aims to be an impartial encyclopedia, and certainly, using the established terms is part of that.

— Preceding unsigned comment added by 176.63.22.138 (talk) 09:51, 27 February 2020 (UTC)

Transformer, and models based on it
Non-RNN attention-based Transformer model, as well as models based on it (e.g. BERT, GPT, GPT-3), are not covered in the article's text. Could anybody cover them accordingly please? Thank you in advace, --Olexa Riznyk (talk) 20:31, 1 November 2020 (UTC)

Unigram models -- why FSA?
The section on unigram models is needlessly complicated: these are simple Bernoulli models, there is no need to bring in Finite State Automata at all. But before removing the unnecessary complexity I'd like to ask if anybody recalls why it was put there in the first place, maybe I'm missing something. — Preceding unsigned comment added by SnoTraveller (talk • contribs) 21:44, 16 March 2022 (UTC)

Erick is the best teacher in njombe
Erick is the best 197.250.225.87 (talk) 15:45, 2 November 2022 (UTC)

Hi
I ned story I’m 94.249.104.173 (talk) 13:25, 14 December 2022 (UTC)

Criticism section is misleading
GPT-2 is not a recurrent neural network, but rather based on Transformer attention based architecture. Would be nice if somebody provided truthfull critical view, because there are plenty of issues in the idea of posing language learning as pure statistical problem. There is real danger that common people will missinterpret the output of such models as it happens with almost every other deep learning architecture. [citation needed] 31.182.202.212 (talk) 21:10, 5 January 2023 (UTC)

Trimming/merging list of language models section
The "notable language models" section currently contains a number of models which are not language models per se, but rather involve a language component (including text-to speech and text-to-image models). I'm removing these, and will probably merge the contents with the table at Large language model, since the list doesn't seem to include any LMs that aren't LLMs.

For posterity, here's a permalink to the section as it existed before I gutted it. It might be useful if someone ever wants to create a list like List of natural language processing models or something. Colin M (talk) 18:24, 9 March 2023 (UTC)