zeeshanali01 commited on
Commit
455b3a4
·
verified ·
1 Parent(s): 8e2e4d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -1
README.md CHANGED
@@ -3,11 +3,65 @@ license: apache-2.0
3
  tags:
4
  - trl
5
  - sft
 
6
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  @inproceedings{cao2024catmemo,
8
  title={CatMemo at the FinLLM Challenge Task: Fine-Tuning Large Language Models using Data Fusion in Financial Applications},
9
  author={Cao, Yupeng and Yao, Zhiyuan and Chen, Zhi and Deng, Zhiyang},
10
  booktitle={Joint Workshop of the 8th Financial Technology and Natural Language Processing (FinNLP) and the 1st Agent AI for Scenario Planning (AgentScen) in conjunction with IJCAI 2023},
11
  pages={174},
12
  year={2024}
13
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  tags:
4
  - trl
5
  - sft
6
+ library_name: transformers
7
  ---
8
+
9
+ # CatMemo: Fine-Tuning Large Language Models for Financial Applications
10
+
11
+ ## Model Overview
12
+ This model, **CatMemo**, is fine-tuned using **Data Fusion** techniques for financial applications. It was developed as part of the FinLLM Challenge Task and focuses on enhancing the performance of large language models in finance-specific tasks such as question answering, document summarization, and sentiment analysis.
13
+
14
+ ### Key Features
15
+ - Fine-tuned on financial datasets using **Supervised Fine-Tuning (SFT)** techniques.
16
+ - Optimized for **Transfer Reinforcement Learning (TRL)** workflows.
17
+ - Specialized for tasks requiring domain-specific context in financial applications.
18
+
19
+ ## Usage
20
+ You can use this model with the [Hugging Face Transformers library](https://huggingface.co/docs/transformers/) to perform financial text analysis. Below is a quick example:
21
+
22
+ ```python
23
+ from transformers import AutoTokenizer, AutoModelForCausalLM
24
+
25
+ # Load the model and tokenizer
26
+ model_name = "zeeshanali01/cryptotunned"
27
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
28
+ model = AutoModelForCausalLM.from_pretrained(model_name)
29
+
30
+ # Tokenize input
31
+ inputs = tokenizer("What are the key takeaways from the latest earnings report?", return_tensors="pt")
32
+
33
+ # Generate output
34
+ outputs = model.generate(**inputs)
35
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
36
+ ```
37
+
38
+ ## Training Details
39
+ This model was fine-tuned using **Data Fusion** methods on domain-specific financial datasets. The training pipeline includes:
40
+ - Preprocessing financial documents and datasets to enhance model understanding.
41
+ - Applying **Supervised Fine-Tuning (SFT)** to optimize the model for financial NLP tasks.
42
+ - Testing and evaluation on FinLLM benchmark tasks.
43
+
44
+ ## Citation
45
+ If you use this model, please cite our work:
46
+
47
+ ```
48
  @inproceedings{cao2024catmemo,
49
  title={CatMemo at the FinLLM Challenge Task: Fine-Tuning Large Language Models using Data Fusion in Financial Applications},
50
  author={Cao, Yupeng and Yao, Zhiyuan and Chen, Zhi and Deng, Zhiyang},
51
  booktitle={Joint Workshop of the 8th Financial Technology and Natural Language Processing (FinNLP) and the 1st Agent AI for Scenario Planning (AgentScen) in conjunction with IJCAI 2023},
52
  pages={174},
53
  year={2024}
54
+ }
55
+ ```
56
+
57
+ ## License
58
+ This model is licensed under the Apache 2.0 License. See the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) file for details.
59
+
60
+ ## Acknowledgments
61
+ We thank the organizers of the FinLLM Challenge Task for providing the benchmark datasets and tasks used to develop this model.
62
+
63
+ ---
64
+ ### Model Card Metadata
65
+ - **License:** Apache 2.0
66
+ - **Tags:** TRL, SFT
67
+ - **Library Used:** Transformers