Zenith V1
Collection
All V1 models of Zenith series • 4 items • Updated
• 1
Standard GPU-optimized language model with code generation and emotional intelligence capabilities.
# Clone and setup
cd Zenith/V1/7B
pip install -r requirements.txt
# Full fine-tuning
python train.py \
--base_model Qwen/Qwen2.5-Coder-7B \
--train_data path/to/train.json \
--epochs 3 \
--batch_size 4 \
--learning_rate 2e-5
# LoRA fine-tuning (recommended for most users)
python train.py \
--base_model Qwen/Qwen2.5-Coder-7B \
--train_data path/to/train.json \
--use_lora \
--lora_r 16 \
--lora_alpha 32 \
--epochs 3 \
--batch_size 8
# Interactive mode
python inference.py --checkpoint ./outputs/checkpoint-final
# Single prompt
python inference.py \
--checkpoint ./outputs/checkpoint-final \
--prompt "Write a Python function to reverse a linked list" \
--max_new_tokens 512
# Build and run with Ollama
ollama create zenith-7b -f Modelfile
ollama run zenith-7b "Explain quantum computing in simple terms"
Zenith/V1/7B/
├── configs/ # Configuration files
│ ├── zenith_config.py # Model architecture config
│ ├── data_config.py # Data processing config
│ └── training_config.py # Training hyperparameters
├── data/ # Data processing modules
│ ├── openthoughts_processor.py
│ ├── quality_filter.py
│ ├── curriculum_sampler.py
│ ├── advanced_tokenizer.py
│ └── preprocessing.py
├── src/ # Source code
│ ├── models/
│ │ ├── zenith_model.py
│ │ ├── dense_layer.py
│ │ └── moe_layer.py
│ └── utils/
├── scripts/ # Utility scripts
├── tests/ # Test suite
├── train.py # Main training script
├── inference.py # Inference and generation
├── test_model.py # Model validation tests
├── finetune_qwen.py # Qwen fine-tuning guide
├── Modelfile # Ollama configuration
├── requirements.txt # Python dependencies
└── README.md # This file
The model uses a unified configuration system in configs/zenith_config.py:
from configs.zenith_config import get_7b_config
config = get_7b_config()
# Parameters:
# - hidden_size: 4096
# - num_layers: 32
# - num_heads: 32
# - num_experts: 0 (dense only, set >1 for MoE)
# - use_eq_adapter: True (emotional intelligence)
# - max_seq_len: 8192
The data pipeline supports the OpenThoughts-1.2M dataset:
from data.openthoughts_processor import OpenThoughtsProcessor, OpenThoughtsConfig
config = OpenThoughtsConfig(
dataset_name="open-thoughts/OpenThoughts3-1.2M",
streaming=True,
quality_filtering=True,
curriculum_learning=True,
augmentation=True
)
processor = OpenThoughtsProcessor(config)
dataset = processor.load_dataset()
Multi-dimensional quality assessment:
Progressive training stages:
Enable sparse activation for better performance:
python train.py --use_moe --num_experts 8
Emotional intelligence module:
python train.py --use_eq_adapter --eq_loss_weight 0.1
Efficient fine-tuning with low-rank adaptation:
# LoRA
python train.py --use_lora --lora_r 16 --lora_alpha 32
# QLoRA (4-bit quantization)
python train.py --use_qlora --use_lora --lora_r 8
Run the test suite:
python test_model.py
Tests include:
See requirements.txt for full dependencies. Key packages:
--mixed_precision bf16 for faster training (Ampere+ GPUs)--max_seq_lengthIf you use Zenith-7B in your research, please cite:
@misc{zenith-7b-2025,
title={Zenith-7B: A Hybrid MoE Model for Code and Emotional Intelligence},
year={2025},
publisher={Zenith Project}
}
[Specify your license here]
For issues and questions, please open an issue on the project repository.