Submitted by Emergency_Apricot_77 t3_zmd6l8 in MachineLearning
Been experimenting with language models a lot lately and wondering if human generated text (i.e. "natural" text) is really supposed to be maximally likely according to language models even after training. For example, has someone checked likelihood of human translated text to likelihood of machine translated text according to a language model like GPT-3 ?
​
Are there any works that do this already ? Does this idea even make sense to begin with ?
breezedeus t1_j0alnlc wrote
Actually, it's really not like that. If our words came out that way, people would know what you were going to say without even having to say it.