--- license: apache-2.0 library_name: transformers pipeline_tag: text-generation tags: - 8bit - sharded - open_llama inference: False --- # open_llama_13b-sharded-8bit This is [open_llama_13b](https://huggingface.co/openlm-research/open_llama_13b) sharded into 2 GB shards, and in 8-bit precision using `bitsandbytes==0.38.0`. Please refer to the original model card for details.