Rohith K Bobby's picture

Rohith K Bobby

Rohith04
Ā·

AI & ML interests

LLM

Recent Activity

reacted to nicolay-r's post with ā¤ļø 9 days ago
šŸ“¢ Deligted to share the most recent milestone on quick deployment of Named Entity Recognition (NER) in Gen-AI powered systems. Releasing the bulk-ner 0.25.0 which represent a tiny framework that would save you time for deploing NER with any model. šŸ’Ž Why is this important? In the era of GenAI the handling out textual output might be challenging. Instead, recognizing named-entities via domain-oriented systems for your donwstream LLM would be preferable option. šŸ“¦: https://pypi.org/project/bulk-ner/0.25.0/ šŸŒŸ: https://github.com/nicolay-r/bulk-ner I noticed that the direct adaptaion of the LM for NER would result in spending signifcant amount of time on formatting your texts according to the NER-model needs. In particular: 1. Processing CONLL format with B-I-O tags from model outputs 2. Input trimming: long input content might not be completely fitted To cope with these problems, in version 0.25.0 I made a huge steps forward by providing: āœ… šŸ Python API support: see screenshot below for a quick deployment (see screenshot below šŸ“ø) āœ… šŸŖ¶ No-string: dependencies are now clear, so it is purely Python implementation for API calls. āœ… šŸ‘Œ Simplified output formatting: we use lists to represent texts with inner lists that refer to annotated objects (see screenshot below šŸ“ø) šŸ“’ We have a colab for a quick start here (or screenshot for bash / Python API šŸ“ø) https://colab.research.google.com/github/nicolay-r/ner-service/blob/main/NER_annotation_service.ipynb šŸ‘ The code for pipeline deployment is taken from the AREkit project: https://github.com/nicolay-r/AREkit
View all activity

Organizations

None yet

Rohith04's activity

reacted to nicolay-r's post with ā¤ļø 9 days ago
view post
Post
2084
šŸ“¢ Deligted to share the most recent milestone on quick deployment of Named Entity Recognition (NER) in Gen-AI powered systems.

Releasing the bulk-ner 0.25.0 which represent a tiny framework that would save you time for deploing NER with any model.

šŸ’Ž Why is this important? In the era of GenAI the handling out textual output might be challenging. Instead, recognizing named-entities via domain-oriented systems for your donwstream LLM would be preferable option.

šŸ“¦: https://pypi.org/project/bulk-ner/0.25.0/
šŸŒŸ: https://github.com/nicolay-r/bulk-ner

I noticed that the direct adaptaion of the LM for NER would result in spending signifcant amount of time on formatting your texts according to the NER-model needs.
In particular:
1. Processing CONLL format with B-I-O tags from model outputs
2. Input trimming: long input content might not be completely fitted

To cope with these problems, in version 0.25.0 I made a huge steps forward by providing:
āœ… šŸ Python API support: see screenshot below for a quick deployment (see screenshot below šŸ“ø)
āœ… šŸŖ¶ No-string: dependencies are now clear, so it is purely Python implementation for API calls.
āœ… šŸ‘Œ Simplified output formatting: we use lists to represent texts with inner lists that refer to annotated objects (see screenshot below šŸ“ø)

šŸ“’ We have a colab for a quick start here (or screenshot for bash / Python API šŸ“ø)
https://colab.research.google.com/github/nicolay-r/ner-service/blob/main/NER_annotation_service.ipynb

šŸ‘ The code for pipeline deployment is taken from the AREkit project:
https://github.com/nicolay-r/AREkit
reacted to Xenova's post with ā¤ļø about 2 months ago
view post
Post
5572
Have you tried out šŸ¤— Transformers.js v3? Here are the new features:
āš” WebGPU support (up to 100x faster than WASM)
šŸ”¢ New quantization formats (dtypes)
šŸ› 120 supported architectures in total
šŸ“‚ 25 new example projects and templates
šŸ¤– Over 1200 pre-converted models
šŸŒ Node.js (ESM + CJS), Deno, and Bun compatibility
šŸ” A new home on GitHub and NPM

Get started with npm i @huggingface/transformers.

Learn more in our blog post: https://huggingface.co/blog/transformersjs-v3
  • 3 replies
Ā·
reacted to nicolay-r's post with šŸ”„ 7 months ago
view post
Post
2183
The most recent LLaMA-3-70B Instruct showcases the beast performance in zero-shot-learning mode in Target-Sentiment-Analsys (TSA) šŸ”„šŸš€ In particular we experiment with sentence-level analysis, with sentences fetched from the WikiArticles that were formed into RuSentNE-2023 dataset.

The key takeaways out of LLaMA-3-70B performance on original (šŸ‡·šŸ‡ŗ) texts and translated into English are as follows:
1. Outperforms all ChatGPT-4 and all predecessors on non-english-texts (šŸ‡·šŸ‡ŗ)
2. Surpasses all ChatGPT-3.5 / nearly performs as good as ChatGPT-4 on english texts šŸ„³

Benchmark: https://github.com/nicolay-r/RuSentNE-LLM-Benchmark
Model: meta-llama/Meta-Llama-3-70B-Instruct
Dataset: https://github.com/dialogue-evaluation/RuSentNE-evaluation
Related paper: Large Language Models in Targeted Sentiment Analysis (2404.12342)
Collection: https://huggingface.co/collections/nicolay-r/sentiment-analysis-665ba391e0eba729021ea101