OGAI-8x7b / README.md
tommytracx's picture
Update README.md
3cff67b verified
|
raw
history blame
6.94 kB
---
license: apache-2.0
datasets:
- GainEnergy/SMoE-Training
- GainEnergy/reasoner
- GainEnergy/ogai-8x7B
- GainEnergy/oilandgas-engineering-dataset
- GainEnergy/ogdataset
- GainEnergy/upstrimacentral
- open-r1/OpenR1-Math-220k
- unsloth/LaTeX_OCR
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
tags:
- oil-gas
- drilling-engineering
- mixtral-8x7b
- lora
- fine-tuned
- energy-ai
- pragmatic-ai
- gguf
- text-generation-inference
- text-generation
model-index:
- name: OGAI-8x7B
results:
- task:
type: text-generation
name: Drilling Engineering AI
dataset:
name: GainEnergy Oil & Gas Corpus
type: custom
metrics:
- name: Drilling Calculations Accuracy
type: accuracy
value: 95.2
- name: Engineering Document Retrieval Precision
type: precision
value: 91.8
- name: Context Retention
type: contextual-coherence
value: High
variants:
- name: OGAI-8x7B-4bit
pipeline_tag: text-generation
repo_name: GainEnergy/ogai-8x7b-4bit
- name: OGAI-8x7B-8bit-32k
pipeline_tag: text-generation
repo_name: GainEnergy/ogai-8x7b-8bit-32k
- name: OGAI-8x7b-Q4_K_M-GGUF
pipeline_tag: text-generation
repo_name: GainEnergy/OGAI-8x7b-Q4_K_M-GGUF
library_name: transformers
language:
- en
widget:
- text: >-
User: What is the recommended mud weight for drilling a well with a
formation pressure of 8,000 psi?
AI:
example_title: Mud Weight Calculation
- text: >-
User: Explain the differences between rotary steerable systems and
conventional directional drilling tools.
AI:
example_title: Directional Drilling Technologies
- text: >-
User: How does drilling fluid viscosity impact hole cleaning efficiency in
horizontal wells?
AI:
example_title: Drilling Fluid Rheology
- text: >-
User: What are the key factors affecting wellbore stability during drilling
operations?
AI:
example_title: Wellbore Stability Analysis
- text: >-
User: Describe the process of well control in the event of a kick during
drilling operations.
AI:
example_title: Well Control Procedures
pipeline_tag: text-generation
---
# OGAI-8x7B: Oil & Gas AI Model for Drilling Process Optimization
![Hugging Face](https://img.shields.io/badge/HuggingFace-OGAI--8x7B-orange)
[![License](https://img.shields.io/github/license/huggingface/transformers.svg)](LICENSE)
## Model Description
**OGAI-8x7B** is a **LoRA fine-tuned Mixtral-8x7B model**, engineered for **oil and gas engineering applications**. This version has been optimized for **drilling process automation, technical document understanding, and engineering problem-solving**.
The model is a core component of **GainEnergy's Upstrima AI Platform**, which provides **pragmatic AI agents, advanced workflows, and retrieval-augmented generation (RAG)-enhanced document processing**.
## Technical Architecture
### Base Model Specifications
- **Architecture**: Mixtral-8x7B Sparse Mixture of Experts (SMoE)
- **Parameters**: 8 experts with 7B parameters each (46.7B total parameters, 12.9B active per token)
- **Context Length**: Extended to 32,768 tokens for handling technical documentation
- **Attention Mechanism**: Sliding Window Attention with specialized oil & gas technical vocabulary
### Fine-tuning Approach
- **Method**: Low-Rank Adaptation (LoRA) with rank 16
- **Training Dataset**: 2.8M specialized oil & gas engineering datapoints (800K drilling-specific)
- **Hardware**: Trained on 16x NVIDIA A100 80GB GPUs
- **Training Time**: 2,800 GPU hours
- **Special Features**: Enhanced numerical precision for engineering calculations with reduced hallucination rates
### Performance Optimizations
- **Quantization**: Custom 4-bit and 8-bit quantization schemes preserving numerical accuracy
- **Inference Speed**: Optimized KV cache management for real-time drilling advisory
- **Memory Footprint**: Reduced to 12GB with 4-bit quantization while maintaining 95%+ calculation accuracy
- **Speculative Decoding**: Implemented for 2.7x faster response generation with technical queries
## Deployment-Optimized Versions
For flexible deployment options, we provide **quantized versions** of this model:
- **[4-bit Quantized Version (NF4)](https://huggingface.co/GainEnergy/ogai-8x7b-4bit)** – Lower memory usage with efficient performance.
- **[8-bit Quantized Version (Int8)](https://huggingface.co/GainEnergy/ogai-8x7b-8bit-32k)** – Balanced model size and precision.
- **[GGUF Version for llama.cpp](https://huggingface.co/GainEnergy/OGAI-8x7b-Q4_K_M-GGUF)** – Optimized for CPU-based inference and edge computing.
### Using the GGUF Model with llama.cpp
To run OGAI-8x7B using llama.cpp, use the following command:
```bash
./server \
--gguf-file-name ogai-8x7b-q4_k_m.gguf \
--repo-slug GainEnergy/OGAI-8x7b-Q4_K_M-GGUF \
--np 8
```
## Deployment Options
This model supports one-click deployment to various platforms directly from the Hugging Face Hub:
### Hugging Face Inference Endpoints
Click the "Deploy" button and select "Inference Endpoints" to deploy on Hugging Face's managed infrastructure.
### Local Deployment with vLLM
```bash
python -m vllm.entrypoints.openai.api_server \
--model GainEnergy/ogai-8x7b \
--tensor-parallel-size 2
```
## OGAI Model Family & Expansion Roadmap
OGAI-8x7B is part of the OGAI Model Family, designed for various energy-sector applications.
| **Model Name** | **Base Model** | **Fine-Tuning** | **Domain Focus** |
|----------------|----------------|-----------------|------------------|
| **OGAI 3.1 Engineer** | Custom Framework | Full fine-tuning | AI-powered Oil & Gas Engineering Assistance |
| **OGAI-Quantum** | Hybrid AI-Quantum | Hybrid Fine-Tuning | Reservoir Simulation & Seismic Data Processing |
| **OGAI-R1** | TinyR1-32B | Fine-tuned | Engineering AI Reasoning & Logical Analysis |
| **OGMOE** | Mixtral-8x7B MoE | Fine-tuned | Mixture of Experts (MoE) for Drilling Optimization |
| **OGAI-8x22B** | Mixtral-8x22B | Full fine-tuning | High-Performance LLM for Engineering Reasoning |
| **OGAI-8x7B** | Mixtral-8x7B | LoRA fine-tuning | Drilling Optimization, Well Planning & Document RAG |
## How to Use
### Run Inference in Python
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "GainEnergy/ogai-8x7b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
prompt = "Calculate the required casing depth for a well in a high-pressure formation."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Citing OGAI-8x7B
```
@article{ogai8x7b2025,
title={OGAI-8x7B: An AI Model for Oil & Gas Drilling Engineering},
author={GainEnergy AI Team},
year={2025},
publisher={Hugging Face Models}
}
```