-
Attention Is All You Need
Paper • 1706.03762 • Published • 44 -
LLaMA: Open and Efficient Foundation Language Models
Paper • 2302.13971 • Published • 13 -
Efficient Tool Use with Chain-of-Abstraction Reasoning
Paper • 2401.17464 • Published • 16 -
MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts
Paper • 2407.21770 • Published • 22
Justin PRO
jxtngx
AI & ML interests
None yet
Organizations
Collections
15
models
27
jxtngx/Nemotron-Mini-4B-Instruct-Q4_K_M-GGUF
Updated
•
6
jxtngx/Meta-Llama-3.2-3B-Instruct-Q4_K_M-GGUF
Text Generation
•
Updated
•
61
jxtngx/Llama-3.2-3B-Q4_K_M-GGUF
Text Generation
•
Updated
•
8
jxtngx/Meta-Llama-3.2-1B-Instruct-Q4_K_M-GGUF
Text Generation
•
Updated
•
23
jxtngx/Meta-Llama-3.2-1B-Q4_K_M-GGUF
Text Generation
•
Updated
•
336
•
2
jxtngx/Llama-3.1-Minitron-4B-Width-Base-Q4_K_M-GGUF
Updated
•
7
jxtngx/Meta-Llama-3.1-8B-Q4_K_M-GGUF
Text Generation
•
Updated
•
5
jxtngx/Meta-Llama-3.1-8B-Instruct-Q4_0-GGUF
Text Generation
•
Updated
•
23
jxtngx/Meta-Llama-3.1-8B-Instruct-Q4_K_M-GGUF
Text Generation
•
Updated
•
11
jxtngx/Hermes-2-Pro-Mistral-7B-Q4_0-GGUF
Updated
•
62
datasets
None public yet