Raising Bars, Not Parameters: LilMoo Compact Language Model for Hindi
Paper β’ 2603.03508 β’ Published β’ 3
A 0.6-billion-parameter Hindi language models trained entirely from scratch.
Note π§± Base model pretrained only with Hindi text.
Note π§± Base model pretrained with a Hindi+ English mixture.
Note π Pretraining dataset.
Note π Annotations to train classifiers/filters (Educational).
Note π Annotations to train classifiers/filters (Toxicity).
Note π― Quality Filter (Educational)
Note π― Quality Filter (Toxicity)
Note π Data used to train the LilMoo tokenizer.