Papers
arxiv:2202.03829
TimeLMs: Diachronic Language Models from Twitter
Published on Feb 8, 2022
Authors:
Abstract
Despite its importance, the time variable has been largely neglected in the NLP and language model literature. In this paper, we present TimeLMs, a set of language models specialized on diachronic Twitter data. We show that a continual learning strategy contributes to enhancing Twitter-based language models' capacity to deal with future and out-of-distribution tweets, while making them competitive with standardized and more monolithic benchmarks. We also perform a number of qualitative analyses showing how they cope with trends and peaks in activity involving specific named entities or concept drift.
Models citing this paper 28
Browse 28 models citing this paperDatasets citing this paper 0
No dataset linking this paper
Cite arxiv.org/abs/2202.03829 in a dataset README.md to link it from this page.
Spaces citing this paper 153
Collections including this paper 0
No Collection including this paper
Add this paper to a
collection
to link it from this page.