cosimoiaia
commited on
Commit
•
3ca3bfe
1
Parent(s):
4442d7b
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,65 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-2.0
|
3 |
+
datasets:
|
4 |
+
- cosimoiaia/Loquace-102k
|
5 |
+
language:
|
6 |
+
- it
|
7 |
+
pipeline_tag: conversational
|
8 |
+
---
|
9 |
+
|
10 |
+
Model Card for Loquace-7B
|
11 |
+
|
12 |
+
## Model Details
|
13 |
+
|
14 |
+
- Model Name: Loquace-7B
|
15 |
+
- Model Version: 1.0
|
16 |
+
- Hugging Face Model Hub Link: [Link to the model on the Hugging Face Model Hub]
|
17 |
+
- License: CC-BY-NC (Creative Commons Attribution-NonCommercial)
|
18 |
+
|
19 |
+
## Model Description
|
20 |
+
|
21 |
+
Loquace-7B is a fine-tuned conversational model for the Italian language. It has been trained on a dataset of 102,000 question/answer examples in the Alpaca style. The model is based on the Falcon-7B architecture and was fine-tuned using the qload framework.
|
22 |
+
|
23 |
+
## Intended Use
|
24 |
+
|
25 |
+
Loquace-7B is designed to facilitate Italian language conversations. It can be used by developers, researchers, or anyone interested in building conversational systems, chatbots, or dialogue-based applications in Italian.
|
26 |
+
|
27 |
+
## Model Inputs
|
28 |
+
|
29 |
+
The model expects input in the form of text strings representing questions or prompts in Italian. The input should follow natural language conventions, and longer inputs may need to be truncated or split into multiple parts to fit the model's maximum sequence length.
|
30 |
+
|
31 |
+
## Model Outputs
|
32 |
+
|
33 |
+
The model generates responses as text strings in Italian, providing answers or replies based on the given input. The outputs can be post-processed or presented as-is, depending on the desired application.
|
34 |
+
|
35 |
+
## Training Data
|
36 |
+
|
37 |
+
Loquace-7B was trained on a conversational dataset comprising 102,000 question/answer pairs in Italian. The training data was formatted in the Alpaca style, which emphasizes conversational exchanges. The specific sources and characteristics of the training data are not disclosed.
|
38 |
+
|
39 |
+
## Evaluation Data
|
40 |
+
|
41 |
+
The model's performance was evaluated using a separate evaluation dataset, which consisted of human-labeled assessments and metrics tailored to the conversational nature of the model. The specific details of the evaluation data, such as size and sources, are not provided.
|
42 |
+
|
43 |
+
## Ethical Considerations
|
44 |
+
|
45 |
+
As with any language model, Loquace-7B may reflect biases present in the training data. Care should be taken when using the model to ensure fair and unbiased interactions. Additionally, as the model is released under the CC-BY-NC license, it should not be used for commercial purposes without proper authorization.
|
46 |
+
|
47 |
+
## Limitations
|
48 |
+
|
49 |
+
- Loquace-7B may not handle complex or nuanced queries well and may struggle with ambiguous or poorly formatted inputs.
|
50 |
+
- The model may generate responses that are factually incorrect or nonsensical. It should be used with caution, and outputs should be carefully verified.
|
51 |
+
- The training data primarily consists of conversational examples and may not generalize well to other types of tasks or domains.
|
52 |
+
|
53 |
+
## Dependencies
|
54 |
+
|
55 |
+
- PyTorch
|
56 |
+
- Transformers library by Hugging Face
|
57 |
+
|
58 |
+
## Contact Information
|
59 |
+
|
60 |
+
For any questions, issues, or inquiries related to Loquace-7B, please contact the developers at [contact email or link].
|
61 |
+
|
62 |
+
## Citation
|
63 |
+
|
64 |
+
[If the model is based on or inspired by a research paper, provide the citation here.]
|
65 |
+
|