File size: 1,262 Bytes
52181bf |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 |
---
# For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
# Doc / guide: https://huggingface.co/docs/hub/model-cards
{}
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This model is an instruction-tuned version of ⭐️ StarCoder. The instruction dataset involved is [Self-instruct-starcoder](https://huggingface.co/datasets/codeparrot/self-instruct-starcoder)
which was built by boostrapping on StarCoder's generations.
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
The model was fine-tuned with the following template
```
Question: <instruction>
Answer: <output>
```
For example, your prompt can look like
```python
instruction = "Write a function to compute the GCD between two integers a and b"
prompt = f"Question:{instruction}\n\nAnswer:"
input_ids = tokenizer(prompt, return_tensors="pt")["input_ids"]
completion = model.generate(input_ids)
print(tokenizer.decode(completion[len(input_ids):])[0])
```
|