Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts Paper • 2409.16040 • Published 6 days ago • 9
DataGemma Release Collection A series of pioneering open models that help ground LLMs in real-world data through Data Commons. • 2 items • Updated 18 days ago • 75
LLM2Encoder Collection Collection of initial models and models that use converted decoders to encoders as backbones • 11 items • Updated 20 days ago • 5
GLiNER bi-encoders Collection Bi-encoder and poly-encoder architectures of GLiNER • 5 items • Updated 20 days ago • 11
view article Article LAVE: Zero-shot VQA Evaluation on Docmatix with LLMs - Do We Still Need Fine-Tuning? Jul 25 • 18
NuNerZero - Zero Shot NER Collection The best compact Zero-Shot NER models with MIT license • 4 items • Updated Jul 3 • 17
An Emulator for Fine-Tuning Large Language Models using Small Language Models Paper • 2310.12962 • Published Oct 19, 2023 • 14
LongLoRA: Efficient Fine-tuning of Long-Context Large Language Models Paper • 2309.12307 • Published Sep 21, 2023 • 86
MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models Paper • 2309.12284 • Published Sep 21, 2023 • 18
Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference Using Sorted Fine-Tuning (SoFT) Paper • 2309.08968 • Published Sep 16, 2023 • 22
Clinical Text Summarization: Adapting Large Language Models Can Outperform Human Experts Paper • 2309.07430 • Published Sep 14, 2023 • 27
AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 16
In-context Autoencoder for Context Compression in a Large Language Model Paper • 2307.06945 • Published Jul 13, 2023 • 27