Update README.md
Browse files
README.md
CHANGED
@@ -3,11 +3,65 @@ license: apache-2.0
|
|
3 |
tags:
|
4 |
- trl
|
5 |
- sft
|
|
|
6 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
@inproceedings{cao2024catmemo,
|
8 |
title={CatMemo at the FinLLM Challenge Task: Fine-Tuning Large Language Models using Data Fusion in Financial Applications},
|
9 |
author={Cao, Yupeng and Yao, Zhiyuan and Chen, Zhi and Deng, Zhiyang},
|
10 |
booktitle={Joint Workshop of the 8th Financial Technology and Natural Language Processing (FinNLP) and the 1st Agent AI for Scenario Planning (AgentScen) in conjunction with IJCAI 2023},
|
11 |
pages={174},
|
12 |
year={2024}
|
13 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
tags:
|
4 |
- trl
|
5 |
- sft
|
6 |
+
library_name: transformers
|
7 |
---
|
8 |
+
|
9 |
+
# CatMemo: Fine-Tuning Large Language Models for Financial Applications
|
10 |
+
|
11 |
+
## Model Overview
|
12 |
+
This model, **CatMemo**, is fine-tuned using **Data Fusion** techniques for financial applications. It was developed as part of the FinLLM Challenge Task and focuses on enhancing the performance of large language models in finance-specific tasks such as question answering, document summarization, and sentiment analysis.
|
13 |
+
|
14 |
+
### Key Features
|
15 |
+
- Fine-tuned on financial datasets using **Supervised Fine-Tuning (SFT)** techniques.
|
16 |
+
- Optimized for **Transfer Reinforcement Learning (TRL)** workflows.
|
17 |
+
- Specialized for tasks requiring domain-specific context in financial applications.
|
18 |
+
|
19 |
+
## Usage
|
20 |
+
You can use this model with the [Hugging Face Transformers library](https://huggingface.co/docs/transformers/) to perform financial text analysis. Below is a quick example:
|
21 |
+
|
22 |
+
```python
|
23 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
24 |
+
|
25 |
+
# Load the model and tokenizer
|
26 |
+
model_name = "zeeshanali01/cryptotunned"
|
27 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
28 |
+
model = AutoModelForCausalLM.from_pretrained(model_name)
|
29 |
+
|
30 |
+
# Tokenize input
|
31 |
+
inputs = tokenizer("What are the key takeaways from the latest earnings report?", return_tensors="pt")
|
32 |
+
|
33 |
+
# Generate output
|
34 |
+
outputs = model.generate(**inputs)
|
35 |
+
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
36 |
+
```
|
37 |
+
|
38 |
+
## Training Details
|
39 |
+
This model was fine-tuned using **Data Fusion** methods on domain-specific financial datasets. The training pipeline includes:
|
40 |
+
- Preprocessing financial documents and datasets to enhance model understanding.
|
41 |
+
- Applying **Supervised Fine-Tuning (SFT)** to optimize the model for financial NLP tasks.
|
42 |
+
- Testing and evaluation on FinLLM benchmark tasks.
|
43 |
+
|
44 |
+
## Citation
|
45 |
+
If you use this model, please cite our work:
|
46 |
+
|
47 |
+
```
|
48 |
@inproceedings{cao2024catmemo,
|
49 |
title={CatMemo at the FinLLM Challenge Task: Fine-Tuning Large Language Models using Data Fusion in Financial Applications},
|
50 |
author={Cao, Yupeng and Yao, Zhiyuan and Chen, Zhi and Deng, Zhiyang},
|
51 |
booktitle={Joint Workshop of the 8th Financial Technology and Natural Language Processing (FinNLP) and the 1st Agent AI for Scenario Planning (AgentScen) in conjunction with IJCAI 2023},
|
52 |
pages={174},
|
53 |
year={2024}
|
54 |
+
}
|
55 |
+
```
|
56 |
+
|
57 |
+
## License
|
58 |
+
This model is licensed under the Apache 2.0 License. See the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) file for details.
|
59 |
+
|
60 |
+
## Acknowledgments
|
61 |
+
We thank the organizers of the FinLLM Challenge Task for providing the benchmark datasets and tasks used to develop this model.
|
62 |
+
|
63 |
+
---
|
64 |
+
### Model Card Metadata
|
65 |
+
- **License:** Apache 2.0
|
66 |
+
- **Tags:** TRL, SFT
|
67 |
+
- **Library Used:** Transformers
|