ranamhamoud
commited on
Commit
•
400901b
1
Parent(s):
d1c461d
Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,9 @@ license: mit
|
|
5 |
language:
|
6 |
- en
|
7 |
pipeline_tag: text-generation
|
|
|
|
|
|
|
8 |
---
|
9 |
|
10 |
# Model Card for Educational Storytelling in Computer Science
|
@@ -15,13 +18,13 @@ This model, developed using Hugging Face’s transformer library, is designed fo
|
|
15 |
|
16 |
### Model Description
|
17 |
|
18 |
-
This model is an innovative tool for teaching fundamental computer science concepts via educational storytelling. It generates interactive
|
19 |
|
20 |
-
- **Developed by:**
|
21 |
- **Model type:** PEFT adapter model using LoRA from Meta's Llama2 7B
|
22 |
- **Language(s) (NLP):** English
|
23 |
-
- **License:**
|
24 |
-
- **Finetuned from model
|
25 |
|
26 |
### Model Sources
|
27 |
|
@@ -35,7 +38,7 @@ This model is an innovative tool for teaching fundamental computer science conce
|
|
35 |
|
36 |
The model is designed to be used directly via an interactive interface where users can ask for stories about specific computer science topics. It's suitable for educational purposes, particularly in learning environments or as a supplementary learning tool.
|
37 |
|
38 |
-
### Downstream Use
|
39 |
|
40 |
While primarily designed for educational storytelling, the model could potentially be adapted for other educational applications or interactive learning tools that require narrative generation.
|
41 |
|
@@ -61,12 +64,6 @@ Here's a general framework for initializing and running the model, detailed in t
|
|
61 |
|
62 |
The model was trained on a custom dataset generated specifically for this project, aimed at creating educational content related to computer science topics. The data generation scripts and datasets are available at the linked GitHub repository.
|
63 |
|
64 |
-
### Training Procedure
|
65 |
-
|
66 |
-
#### Preprocessing
|
67 |
-
|
68 |
-
Specific preprocessing details were not provided but would typically include data cleaning and formatting to fit the model's input requirements.
|
69 |
-
|
70 |
#### Training Hyperparameters
|
71 |
|
72 |
The model was trained on an NVIDIA A100 machine using quantization techniques to optimize performance. Training involved configurations like LoRA adaptation and fine-tuning of Meta's Llama2 7B model under specified training arguments.
|
@@ -75,18 +72,17 @@ The model was trained on an NVIDIA A100 machine using quantization techniques to
|
|
75 |
|
76 |
### Testing Data, Factors & Metrics
|
77 |
|
78 |
-
Further details on testing data and evaluation metrics
|
79 |
|
80 |
### Results
|
81 |
|
82 |
-
Results of the training and subsequent evaluations
|
83 |
|
84 |
## Environmental Impact
|
85 |
|
86 |
- **Hardware Type:** NVIDIA A100
|
87 |
-
- **Hours used:**
|
88 |
- **Cloud Provider:** RunPod
|
89 |
-
- **Compute Region:** Not specified (please provide if available)
|
90 |
- **Carbon Emitted:** Estimates not provided
|
91 |
|
92 |
[More Information Needed]
|
|
|
5 |
language:
|
6 |
- en
|
7 |
pipeline_tag: text-generation
|
8 |
+
tags:
|
9 |
+
- educational
|
10 |
+
- storytelling
|
11 |
---
|
12 |
|
13 |
# Model Card for Educational Storytelling in Computer Science
|
|
|
18 |
|
19 |
### Model Description
|
20 |
|
21 |
+
This model is an innovative tool for teaching fundamental computer science concepts via educational storytelling. It generates interactive stories tailored to specific CS topics requested by the user, such as algorithms, programming basics & more, incorporating assessments to enhance learning and engagement.
|
22 |
|
23 |
+
- **Developed by:** Ranam Hamoud & George Kanaan
|
24 |
- **Model type:** PEFT adapter model using LoRA from Meta's Llama2 7B
|
25 |
- **Language(s) (NLP):** English
|
26 |
+
- **License:** MIT License
|
27 |
+
- **Finetuned from model :** Finetuned from Meta's Llama-2 7B model
|
28 |
|
29 |
### Model Sources
|
30 |
|
|
|
38 |
|
39 |
The model is designed to be used directly via an interactive interface where users can ask for stories about specific computer science topics. It's suitable for educational purposes, particularly in learning environments or as a supplementary learning tool.
|
40 |
|
41 |
+
### Downstream Use
|
42 |
|
43 |
While primarily designed for educational storytelling, the model could potentially be adapted for other educational applications or interactive learning tools that require narrative generation.
|
44 |
|
|
|
64 |
|
65 |
The model was trained on a custom dataset generated specifically for this project, aimed at creating educational content related to computer science topics. The data generation scripts and datasets are available at the linked GitHub repository.
|
66 |
|
|
|
|
|
|
|
|
|
|
|
|
|
67 |
#### Training Hyperparameters
|
68 |
|
69 |
The model was trained on an NVIDIA A100 machine using quantization techniques to optimize performance. Training involved configurations like LoRA adaptation and fine-tuning of Meta's Llama2 7B model under specified training arguments.
|
|
|
72 |
|
73 |
### Testing Data, Factors & Metrics
|
74 |
|
75 |
+
Further details on testing data and evaluation metrics will be provided.
|
76 |
|
77 |
### Results
|
78 |
|
79 |
+
Results of the training and subsequent evaluations will be provided to understand the effectiveness of the model in educational storytelling.
|
80 |
|
81 |
## Environmental Impact
|
82 |
|
83 |
- **Hardware Type:** NVIDIA A100
|
84 |
+
- **Hours used:** 8 hours
|
85 |
- **Cloud Provider:** RunPod
|
|
|
86 |
- **Carbon Emitted:** Estimates not provided
|
87 |
|
88 |
[More Information Needed]
|