Text Generation
Transformers
GGUF
English
Inference Endpoints
wikichat-v2 / README.md
leafspark's picture
Update README.md
0aefaec verified
|
raw
history blame
1.34 kB
metadata
license: apache-2.0
datasets:
  - HuggingFaceTB/cosmopedia
  - databricks/databricks-dolly-15k
  - Open-Orca/OpenOrca
language:
  - en
metrics:
  - accuracy
library_name: transformers
pipeline_tag: text-generation

WikiChat-v0.2

Training in progress model to have conversations.

The GGUFs uploaded are full FP32 precision.

Using OpenOrca GPT-4 data + cosmopedia for some extra data + dolly15k for instruct

Model Details:

  • 71.7M parameters (71,775,700)
  • 8 attention heads
  • 32 layers (34 layers on final model)
  • 384 embeddings size
  • 2048/8192/16384 context (please use 4x RoPE scaling, may train a 16k finetuned version later)
  • Batch size 16
  • llama.cpp (train-text-from-scratch)

Prompt Format (Alpaca):

Instruction: {system}
Input: {prompt}
Response: {response}

Please structure your prompts in an instruct format for maximum performance.

Training Details:

  • 1x RTX 3070 8GB (Infrencing speed: 80tok/s, full GPU offload)
  • 1x Ryzen 3 3700x
  • 96gb RAM
  • 10 iterations
  • Loss Target = 2.5 to 3.0
  • Approx 30 samples (>0.0001 epoches)
  • Training data = Refer to OpenOrca page

Notes:

The model isn't ready yet; this is to test tokenization of OpenOrca and a balance between training speed and model size

Example output:

User: What is the square root of 4?
Assistant: The square root of 4 is 2.