QwQ-LCoT-7B-Instruct Model File

The QwQ-LCoT-7B-Instruct is a fine-tuned language model designed for advanced reasoning and instruction-following tasks. It leverages the Qwen2.5-7B base model and has been fine-tuned on the amphora/QwQ-LongCoT-130K dataset, focusing on chain-of-thought (CoT) reasoning.

File Name Size Description Upload Status
.gitattributes 1.57 kB Tracks large files with Git LFS. Uploaded
README.md 273 Bytes Contains initial documentation, likely minimal. Updated
added_tokens.json 657 Bytes Maps additional tokens for the tokenizer. Uploaded
config.json 848 Bytes Model configuration (basic setup). Uploaded
generation_config.json 281 Bytes Settings for text generation tasks. Uploaded
merges.txt 1.82 MB Tokenizer merges for byte-pair encoding (BPE). Uploaded
model-00001-of-00004.safetensors 4.88 GB First part of model weights (split for LFS). Uploaded (LFS)
model-00002-of-00004.safetensors 4.93 GB Second part of model weights. Uploaded (LFS)
model-00003-of-00004.safetensors 4.33 GB Third part of model weights. Uploaded (LFS)
model-00004-of-00004.safetensors 1.09 GB Fourth part of model weights. Uploaded (LFS)
model.safetensors.index.json 29.5 kB Index file for managing model shards. Uploaded
special_tokens_map.json 644 Bytes Maps special tokens like <pad> or <eos>. Uploaded
tokenizer.json 11.4 MB Pre-trained tokenizer file in JSON format. Uploaded (LFS)
tokenizer_config.json 7.73 kB Configuration details for the tokenizer. Uploaded
vocab.json 2.78 MB Tokenizer vocabulary. Uploaded

Sample Long CoT:

Screenshot 2024-12-13 211732.png


Key Features:

  1. Model Size:

    • 7.62B parameters (FP16 precision).
  2. Model Sharding:

    • The model weights are split into 4 shards (safetensors) for efficient storage and download:
      • model-00001-of-00004.safetensors (4.88 GB)
      • model-00002-of-00004.safetensors (4.93 GB)
      • model-00003-of-00004.safetensors (4.33 GB)
      • model-00004-of-00004.safetensors (1.09 GB)
  3. Tokenizer:

    • Byte-pair encoding (BPE) based.
    • Files included:
      • vocab.json (2.78 MB)
      • merges.txt (1.82 MB)
      • tokenizer.json (11.4 MB)
    • Special tokens mapped in special_tokens_map.json (e.g., <pad>, <eos>).
  4. Configuration Files:

    • config.json: Defines model architecture and hyperparameters.
    • generation_config.json: Settings for inference and text generation tasks.

Training Dataset:


Use Cases:

  1. Instruction Following:
    Handle user instructions effectively, even for multi-step tasks.

  2. Reasoning Tasks:
    Perform logical reasoning and generate detailed step-by-step solutions.

  3. Text Generation:
    Generate coherent, context-aware responses.


Downloads last month
94
Safetensors
Model size
7.62B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for prithivMLmods/QwQ-LCoT-7B-Instruct

Base model

Qwen/Qwen2.5-7B
Finetuned
(141)
this model
Quantizations
6 models

Dataset used to train prithivMLmods/QwQ-LCoT-7B-Instruct

Spaces using prithivMLmods/QwQ-LCoT-7B-Instruct 2

Collection including prithivMLmods/QwQ-LCoT-7B-Instruct