--- language: - en library_name: llama.cpp pipeline_tag: text-generation license: apache-2.0 quantization: q4_k_m --- # LLaMa_3.2_3B_Instruct_Text2SQL-Q4_K_M-GGUF.gguf This is a GGUF quantized version of the LLaMa 3.2 3B Text2SQL model. ## Model Details - **Architecture:** LLaMa 3.2 3B - **Task:** Text to SQL Generation - **Quantization:** Q4_K_M - **Context Length:** 65536 tokens (2^16) - **Format:** GGUF (Compatible with llama.cpp) ## Usage ```python from llama_cpp import Llama # Initialize model llm = Llama( model_path="downloaded_model.gguf", n_ctx=65536, # 64K context n_threads=8 # Adjust based on your CPU ) # Generate SQL response = llm( "Convert this to SQL: Find all users who signed up in January 2024", max_tokens=1024, temperature=0.7 ) print(response['choices'][0]['text']) ``` ## Model Source This is a quantized version of [XeAI/LLaMa_3.2_3B_Instruct_Text2SQL](https://huggingface.co/XeAI/LLaMa_3.2_3B_Instruct_Text2SQL)