Word-level Text Generation from Language Models

dc.contributor.authorPonrudee Netisopakul
dc.contributor.authorUsanisa Taoto
dc.date.accessioned2026-05-08T19:22:42Z
dc.date.issued2021-11-6
dc.description.abstractThis research constructs and evaluates text generation models created from three different language models, n-gram, a Continuous Bag of Words (CBOW) and gated recurrent unit (GRU), using two training corpora, Berkeley Restaurant (Berkeley) and Alice's Adventures in Wonderland (Alice), and evaluated using two evaluation metrics; perplexity measure and count of grammar errors. The mean perplexities of all three models are comparable for each corpus, the N-gram model produces slightly lower values of perplexity. As for the number of grammatical errors in the Alice corpus, all three models show a slightly higher number of errors than the original corpus. In the Berkeley corpus, the n-gram model had the lowest number of errors, even lower than the original corpus, but the CBOW model had the highest number of errors and the GRU model had the highest number of errors.
dc.identifier.doi10.1109/icset53708.2021.9612541
dc.identifier.urihttps://dspace.kmitl.ac.th/handle/123456789/18691
dc.subjectNatural Language Processing Techniques
dc.subjectTopic Modeling
dc.subjectSpeech and dialogue systems
dc.titleWord-level Text Generation from Language Models
dc.typeArticle

Files

Collections