Blogs
More blogs are coming soon, stay tuned!
Feb 7, 2024
Over the past years, foundation models have caused a paradigm shift in machine learning due to their unprecedented capabilities for zero-shot and few-shot generalization. However, despite the success in modalities such as natural language processing and computer vision, foundation models for time series forecasting has lagged behind. We introduce Lag-Llama: the first open-source foundation model for time series forecasting!
Model weights Demo GitHub Paper Tweet
Model weights Demo GitHub Paper Tweet
Dec 30, 2023
This is a progress update (stay tuned - next update mid-January) on our CL-FoMo (Continually-Learned Foundation Models) suite of open-source LLMs, containing four 410M models, and four 9.6B models, trained on Pile, SlimPajama (SP), Mix Pile+SP, and Continual (Pile, SP). On most benchmarks, our 9.6B continual model is competitive with the performance of the baseline joint model trained on mixed Pile+SP data (we call it IID Pile+SP to underscore that samples in the mix are identically distributed), and attains similar performance to RedPajama 7B and Pythia 12B on featured benchmarks.
Dec 24, 2023
The Robin team is proud to present Robin - a suite of Visual Language Models (VLMs). These models are competitive with the stateof the art models of similar scale.
Models: https://huggingface.co/agi-collective
Code: https://github.com/AGI-Collective/Robin/releases/tag/v1.0.0
Oct 31, 2023
Hindi is one of the most widely used languages, spoken by more than 600 Million people; however, no open-source high-quality Hindi LLM was built so far, perhaps due to the relative scarcity of available training datasets. We use continual pretraining with 600B Hindi/English tokens on top of 9.6B model from the CL-FoMo Suite pretrained on English language Pile dataset; as a result:
We built the best open-source Hindi-English LLM of its size, which outperforms all models of comparable sizes, such as Pythia 12B, LLaMa-2 and multilingual BLOOM models across various standard LLM Evaluation benchmark, including HellaSwag, ARC Reasoning Challenge and HellaSwag in the Hindi language.
Our model extends ability to a new language while further improving performance on English and Code domains. Continual pre training also improved the performance of the model on English domain, where it already starts to significantly outperform a larger 12B parameter Pythia model and closes the gap with the 7B LLaMa-2 across the TruthfulQA, ARC Easy, ARC Challenge and HellaSwag benchmarks for English languages. Similar trend - i.e., improved coding abilities - was observed when using the HumanEval benchmark in Python programming language.