pmking27's picture
Update README.md
7b8179e verified
|
raw
history blame
3.04 kB
metadata
language:
  - en
license: apache-2.0
tags:
  - text-generation-inference
  - transformers
  - gemma
  - gguf
  - llama.cpp
base_model: pmking27/PrathameshLLM-7B

Uploaded model

  • Developed by: pmking27
  • License: apache-2.0
  • Finetuned from model : pmking27/PrathameshLLM-7B

Provided Quants Files

Name Quant method Bits Size
PrathameshLLM-7B.IQ3_M.gguf IQ3_M 3 1.31 GB
PrathameshLLM-7B.IQ3_S.gguf IQ3_S 3 1.29 GB
PrathameshLLM-7B.IQ3_XS.gguf IQ3_XS 3 1.24 GB
PrathameshLLM-7B.IQ4_NL.gguf IQ4_NL 4 1.56 GB
PrathameshLLM-7B.IQ4_XS.gguf IQ4_XS 4 1.5 GB
PrathameshLLM-7B.Q2_K.gguf Q2_K 2 1.16 GB
PrathameshLLM-7B.Q3_K_L.gguf Q3_K_L 3 1.47 GB
PrathameshLLM-7B.Q3_K_M.gguf Q3_K_M 3 1.38 GB
PrathameshLLM-7B.Q3_K_S.gguf Q3_K_S 3 1.29 GB
PrathameshLLM-7B.Q4_0.gguf Q4_0 4 1.55 GB
PrathameshLLM-7B.Q4_K_M.gguf Q4_K_M 4 1.63 GB
PrathameshLLM-7B.Q4_K_S.gguf Q4_K_S 4 1.56 GB
PrathameshLLM-7B.Q5_0.gguf Q5_0 5 1.8 GB
PrathameshLLM-7B.Q5_K_M.gguf Q5_K_M 5 1.84 GB
PrathameshLLM-7B.Q5_K_S.gguf Q5_K_S 5 1.8 GB
PrathameshLLM-7B.Q6_K.gguf Q6_K 6 2.06 GB
PrathameshLLM-7B.Q8_0.gguf Q8_0 8 2.67 GB