What is the actual context size of mistralai/Mixtral-8x7B-Instruct-v0.1 model
3
#186 opened 12 months ago
by
Pradeep1995

How to All Utilize all GPU's when device="balanced_low_0" in GPU setting
2
#185 opened 12 months ago
by
kmukeshreddy

Update README.md
#184 opened 12 months ago
by
alamati
Is function calling (tools) supported?
1
#183 opened 12 months ago
by
TomerRobusta
Getting cut-off responses with Mixtral 8x7B-Instruct-v0.1 mostly in Date of Birth years
3
#182 opened 12 months ago
by
keskival
How can I run it on multiple GPUs?
11
#181 opened 12 months ago
by
barbery
Where is the mixtral-8x7b's tokenizer encoder? Is there a specific repository or node module?
1
#180 opened 12 months ago
by
RamanSB
What is the max token limit on this model?
2
#179 opened 12 months ago
by
RamanSB
Finetuning Mixtral 8x7B Instruct-v0.1 using Transformers
2
#178 opened 12 months ago
by
Ateeqq

Update chat template to resemble the prompt as stated in the model card.
7
#176 opened 12 months ago
by
nilsec
max_sequence_length
1
#175 opened 12 months ago
by
Ravnoor1
Awesome. I Got Very Good Responses, However...
#174 opened 12 months ago
by
deleted
How to run the full model ?
2
#171 opened about 1 year ago
by
dounykim
Is there a working/quantized/exl2 (etc) version that will fit on a single 24GB video card (4090)
2
#170 opened about 1 year ago
by
cleverest
403 error
1
#169 opened about 1 year ago
by
minhphan-qbe
Adding Evaluation Results
#168 opened about 1 year ago
by
leaderboard-pr-bot

Rename README.md to RegulusOne
#167 opened about 1 year ago
by
Theguy666
Help: CUDA Out of Memory. Hardware requirements.
2
#147 opened about 1 year ago
by
zebfreeman
Update README.md
#146 opened about 1 year ago
by
frank76rm
Experimental use
#144 opened about 1 year ago
by
yassineelkhadiri14

TemplateError: Conversation roles must alternate user/assistant/user/assistant/...
4
#143 opened about 1 year ago
by
quamer23
Is instruction format necessary
2
#142 opened about 1 year ago
by
supercharge19
[AUTOMATED] Model Memory Requirements
3
#141 opened about 1 year ago
by
model-sizer-bot
Update README.md
#140 opened about 1 year ago
by
woodyk
Cuda Out of memory issue when deploying mistralai/Mixtral-8x7B-Instruct-v0.1 on AWS "ml.g5.48xlarge"
1
#139 opened about 1 year ago
by
sonalisbapte
slow response
1
#138 opened about 1 year ago
by
bhavanam2809
Sparsity in mixtral
#137 opened about 1 year ago
by
dpk17
Request: DOI
#136 opened about 1 year ago
by
Sonny03
Running in Multi-gpu's
5
#134 opened about 1 year ago
by
kmukeshreddy

Update README.md
#133 opened about 1 year ago
by
gmverbas
How to format custom dataset to finetune Mixtral with TRL SFT script?
#132 opened about 1 year ago
by
icpro
How to use run the code on Colab Free Tier or Mac OS?
16
#131 opened about 1 year ago
by
dounykim
Different answer after each request
2
#130 opened about 1 year ago
by
amin2557
How to finetune the model?
2
#129 opened about 1 year ago
by
akasranjan
How much Resource is needed to run the Mixtral ?
1
#128 opened about 1 year ago
by
rkhapre
Update README.md
#126 opened about 1 year ago
by
mariakatosvich

The inference API Endpoint gives wrongly formatted answer based on the context given but works well in example Spaces. How we can fix this?
9
#125 opened about 1 year ago
by
rkhapre
Request: DOI
#124 opened about 1 year ago
by
jsr2
Update README.md
#123 opened about 1 year ago
by
Pawamami

what is max input token limit of this model?
1
#122 opened about 1 year ago
by
vaidehirao
Request: SDFSDFSD
1
#120 opened about 1 year ago
by
seedeera
Consistency check failed - model-00019-of-00019.safetensors
#118 opened about 1 year ago
by
br1-pist
Difference in EOS token between Mistral/Mixtral and LLAMA.
1
#117 opened about 1 year ago
by
xkszltl
Model Output is Changed
9
#116 opened about 1 year ago
by
AnzaniAI
The chat template doesn't support a system prompt
6
#114 opened about 1 year ago
by
sam-kap
How to get 'output_router_logits'
1
#113 opened about 1 year ago
by
cts13
Run inference on 2 GPUs
1
#112 opened about 1 year ago
by
bweinstein123