๐ Welcome the New and Improved GLiNER-Multitask! ๐
Since the release of our beta version, GLiNER-Multitask has received many positive responses. It's been embraced in many consulting, research, and production environments. Thank you everyone for your feedback, it helped us rethink the strengths and weaknesses of the first model and we are excited to present the next iteration of this multi-task information extraction model.
๐ก Whatโs New? Here are the key improvements in this latest version: ๐น Expanded Task Support: Now includes text classification and other new capabilities. ๐น Enhanced Relation Extraction: Significantly improved accuracy and robustness. ๐น Improved Prompt Understanding: Optimized for open-information extraction tasks. ๐น Better Named Entity Recognition (NER): More accurate and reliable results.
๐ง How We Made It Better: These advancements were made possible by: ๐น Leveraging a better and more diverse dataset. ๐น Using a larger backbone model for increased capacity. ๐น Implementing advanced model merging techniques. ๐น Employing self-learning strategies for continuous improvement. ๐น Better training strategies and hyperparameters tuning.
Auto-regressive LMs have ruled, but encoder-based architectures like GLiNER are proving to be just as powerful for information extraction while offering better efficiency and interpretability. ๐โจ
Past encoder backbones were limited by small pre-training datasets and old techniques, but with innovations like LLM2Vec, we've transformed decoders into high-performing encoders! ๐๐ก
Whatโs New? ๐นConverted Llama & Qwen decoders to advanced encoders ๐นImproved GLiNER architecture to be able to work with rotary positional encoding ๐นNew GLiNER (zero-shot NER) & GLiClass (zero-shot classification) models
๐ Meet the new GLiNER architecture ๐ GLiNER revolutionized zero-shot NER by demonstrating that lightweight encoders can achieve excellent results. We're excited to continue R&D with this spirit ๐ฅ. Our new bi-encoder and poly-encoder architectures were developed to address the main limitations of the original GLiNER architecture and bring the following new possibilities:
๐น An unlimited number of entities can be recognized at once. ๐นFaster inference when entity embeddings are preprocessed. ๐นBetter generalization to unseen entities.
While the bi-encoder architecture can lack inter-label understanding, we developed a poly-encoder architecture with post-fusion. It achieves the same or even better results on many benchmarking datasets compared to the original GLiNER, while still offering the listed advantages of bi-encoders. Now, itโs possible to run GLiNER with hundreds of entities much faster and more reliably.