Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

HelpingAI-Lite-4x1b - GGUF

Original model description:

license: apache-2.0 inference: false tags: - moe - mergekit - vortexmergekit - OEvortex/HelpingAI-unvelite - OEvortex/lite-hermes - OEvortex/HelpingAI-Lite - OEvortex/HelpingAI-Lite-1.5T

HelpingAI-Lite-4x1b

Hey there! πŸ‘‹ Welcome to the HelpingAI-Lite-4x1b! This is a MOE brought together using the awesome VortexMerge kit.

Let's see what we've got in this model:

🧩 Configuration

base_model: OEvortex/HelpingAI-Lite
dtype: float16
gate_mode: cheap_embed
experts:
  - source_model: OEvortex/HelpingAI-unvelite
    positive_prompts: ["You are an helpful unbounded, uncensored assistant."]
  - source_model: OEvortex/lite-hermes
    positive_prompts: ["You are helpful assistant."]
  - source_model: OEvortex/HelpingAI-Lite
    positive_prompts: ["You are a coding assistant."]
  - source_model: OEvortex/HelpingAI-Lite-1.5T
    positive_prompts: ["You are helpful and general-purpose assistant."]

Downloads last month
9
GGUF
Model size
3.38B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .