Shanks9 commited on
Commit
abaf2a0
·
verified ·
1 Parent(s): a2692bc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -34
README.md CHANGED
@@ -47,9 +47,6 @@ The model can be directly used to:
47
  - Forecast server health based on time-series metrics like temperature, power consumption, utilization and throughput.
48
  - Predict potential causes of failures using historical data.
49
 
50
-
51
- [More Information Needed]
52
-
53
  ### Downstream Use [optional]
54
 
55
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
@@ -57,16 +54,12 @@ The model is ideal for integration into platforms such as Splunk and Grafana to:
57
  - Monitor server health in real-time.
58
  - Support decision-making in preventive maintenance.
59
 
60
- [More Information Needed]
61
-
62
  ### Out-of-Scope Use
63
 
64
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
65
  - This model is not designed for general time-series forecasting outside server health monitoring.
66
  - It may not perform well on non-server-related data or domains significantly different from its training dataset.
67
 
68
- [More Information Needed]
69
-
70
  ## Bias, Risks, and Limitations
71
 
72
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
@@ -82,8 +75,6 @@ The model is ideal for integration into platforms such as Splunk and Grafana to:
82
  1. Limited to time-series metrics related to server health (e.g., temperature, power, throughput).
83
  2. Performance may degrade for very sparse or noisy datasets.
84
 
85
- [More Information Needed]
86
-
87
  ### Recommendations
88
 
89
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
@@ -93,8 +84,6 @@ The model is ideal for integration into platforms such as Splunk and Grafana to:
93
 
94
  ## How to Get Started with the Model
95
 
96
- [More Information Needed]
97
-
98
  The Mistral 7B - Time Series Predictor can process time-series queries such as server health metrics and predict failure probabilities and causes. The following Python script demonstrates how to load the model and generate responses.
99
  ### Code
100
  - from transformers import AutoModelForCausalLM, AutoTokenizer
@@ -109,7 +98,6 @@ The Mistral 7B - Time Series Predictor can process time-series queries such as s
109
  - response = tokenizer.decode(output[0])
110
  - print(response)
111
 
112
-
113
  **Example Prompt**
114
  - What is the failure probability and Cause for Server 'x' on Date : [mm/dd/yy]?
115
  - *Expected Ouptut*: The failure probability for ET-1 on 11th July is 0.72. The likely cause is overheating due to sustained high temperatures over the past week.
@@ -126,8 +114,6 @@ The Mistral 7B - Time Series Predictor can process time-series queries such as s
126
  **Source:** Synthetic and real-world server metrics from Esperanto servers.
127
  **Dataset:** Synthetic data generated with periodic patterns (e.g., cosine functions) combined with operational zones (green, yellow, red).
128
 
129
- [More Information Needed]
130
-
131
  ### Training Procedure
132
 
133
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
@@ -148,9 +134,6 @@ Prompts were carefully designed to incorporate domain knowledge, guiding the mod
148
 
149
  *These prompts ensured the model understood the critical relationships between input metrics and their operational implications.*
150
 
151
- [More Information Needed]
152
-
153
-
154
  #### Training Hyperparameters
155
 
156
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
@@ -161,9 +144,6 @@ Prompts were carefully designed to incorporate domain knowledge, guiding the mod
161
  - Training time: ~30 hours on NVIDIA A100 GPUs
162
  - Model size: ~7B parameters
163
 
164
-
165
- [More Information Needed]
166
-
167
  ## Evaluation
168
 
169
  <!-- This section describes the evaluation protocols and provides the results. -->
@@ -175,8 +155,6 @@ Prompts were carefully designed to incorporate domain knowledge, guiding the mod
175
  <!-- This should link to a Dataset Card if possible. -->
176
  *Validation set:* 10% of synthetic and real-world server performance data.
177
 
178
- [More Information Needed]
179
-
180
  #### Factors
181
 
182
  <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
@@ -187,29 +165,20 @@ Model evaluated for:
187
 
188
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6659207a17951b5bd11a91fa/UgK2hf8rK9gTw_1AAUuo7.png)
189
 
190
-
191
- [More Information Needed]
192
-
193
  #### Metrics
194
 
195
  <!-- These are the evaluation metrics being used, ideally with a description of why. -->
196
 
197
- [More Information Needed]
198
-
199
  ### Results
200
 
201
- [More Information Needed]
202
 
203
  #### Summary
204
 
205
 
206
-
207
  ## Model Examination [optional]
208
 
209
  <!-- Relevant interpretability work for the model goes here -->
210
 
211
- [More Information Needed]
212
-
213
  ## Environmental Impact
214
 
215
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
@@ -242,6 +211,4 @@ Sivakrishna Yaganti and Shankar Jayaratnam
242
 
243
  ## Model Card Contact
244
 
245
- shankar.jayaratnam@esperantotech.com
246
-
247
- [More Information Needed]
 
47
  - Forecast server health based on time-series metrics like temperature, power consumption, utilization and throughput.
48
  - Predict potential causes of failures using historical data.
49
 
 
 
 
50
  ### Downstream Use [optional]
51
 
52
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
 
54
  - Monitor server health in real-time.
55
  - Support decision-making in preventive maintenance.
56
 
 
 
57
  ### Out-of-Scope Use
58
 
59
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
  - This model is not designed for general time-series forecasting outside server health monitoring.
61
  - It may not perform well on non-server-related data or domains significantly different from its training dataset.
62
 
 
 
63
  ## Bias, Risks, and Limitations
64
 
65
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
 
75
  1. Limited to time-series metrics related to server health (e.g., temperature, power, throughput).
76
  2. Performance may degrade for very sparse or noisy datasets.
77
 
 
 
78
  ### Recommendations
79
 
80
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
 
84
 
85
  ## How to Get Started with the Model
86
 
 
 
87
  The Mistral 7B - Time Series Predictor can process time-series queries such as server health metrics and predict failure probabilities and causes. The following Python script demonstrates how to load the model and generate responses.
88
  ### Code
89
  - from transformers import AutoModelForCausalLM, AutoTokenizer
 
98
  - response = tokenizer.decode(output[0])
99
  - print(response)
100
 
 
101
  **Example Prompt**
102
  - What is the failure probability and Cause for Server 'x' on Date : [mm/dd/yy]?
103
  - *Expected Ouptut*: The failure probability for ET-1 on 11th July is 0.72. The likely cause is overheating due to sustained high temperatures over the past week.
 
114
  **Source:** Synthetic and real-world server metrics from Esperanto servers.
115
  **Dataset:** Synthetic data generated with periodic patterns (e.g., cosine functions) combined with operational zones (green, yellow, red).
116
 
 
 
117
  ### Training Procedure
118
 
119
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
 
134
 
135
  *These prompts ensured the model understood the critical relationships between input metrics and their operational implications.*
136
 
 
 
 
137
  #### Training Hyperparameters
138
 
139
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
144
  - Training time: ~30 hours on NVIDIA A100 GPUs
145
  - Model size: ~7B parameters
146
 
 
 
 
147
  ## Evaluation
148
 
149
  <!-- This section describes the evaluation protocols and provides the results. -->
 
155
  <!-- This should link to a Dataset Card if possible. -->
156
  *Validation set:* 10% of synthetic and real-world server performance data.
157
 
 
 
158
  #### Factors
159
 
160
  <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
 
165
 
166
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6659207a17951b5bd11a91fa/UgK2hf8rK9gTw_1AAUuo7.png)
167
 
 
 
 
168
  #### Metrics
169
 
170
  <!-- These are the evaluation metrics being used, ideally with a description of why. -->
171
 
 
 
172
  ### Results
173
 
 
174
 
175
  #### Summary
176
 
177
 
 
178
  ## Model Examination [optional]
179
 
180
  <!-- Relevant interpretability work for the model goes here -->
181
 
 
 
182
  ## Environmental Impact
183
 
184
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
 
211
 
212
  ## Model Card Contact
213
 
214
+ shankar.jayaratnam@esperantotech.com