NexusLLM-Math-1B-v1

Model Details

NexusLLM-Math-1B-v1 is a fine-tuned version of Llama 3.2 (1B parameters) optimized specifically for solving advanced high-school mathematics problems, with a focus on JEE Main and Advanced syllabus topics.

  • Developed by: ZentithLLM
  • Model Type: Causal Language Model (Fine-tuned with LoRA)
  • Language: English
  • Base Model: meta-llama/Llama-3.2-1B-Instruct
  • Precision: FP16

Intended Use

This model is designed to act as an educational assistant for 11th-grade mathematics. It is trained to provide step-by-step reasoning and explanations for complex topics, rather than just outputting the final answer.

Primary Topics Covered:

  • Binomial Theorem
  • Geometry (Circle Theorems, cyclic quadrilaterals, tangents, etc.)

Training Data

The model was trained on a custom dataset of structured mathematics Q&A pairs. The dataset maps specific mathematical prompts to detailed completions, heavily utilizing an explanation field to teach the model the underlying mathematical logic and derivation steps.

Training Procedure

The model was fine-tuned using the standard Hugging Face trl and peft libraries on a single NVIDIA T4 GPU, utilizing strictly native FP16 precision to ensure mathematical gradient stability.

  • Training Framework: Pure Hugging Face (No Unsloth/Quantization)
  • Method: LoRA (Low-Rank Adaptation)
  • Rank (r): 32
  • Alpha: 32
  • Optimizer: adamw_torch
  • Learning Rate: 2e-4
  • Max Sequence Length: 2048

How to Use

Because this model was trained on a specific dataset structure, you must wrap your prompts in the ### Instruction: and ### Response: format for it to output the correct mathematical explanations.

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "ZentithLLM/NexusLLM-Math-1B-v1"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.float16,
    device_map="auto" 
)

question = "What is the general term in the expansion of (x+y)^n?"
formatted_prompt = f"### Instruction:\\n{question}\\n\\n### Response:\\n"

inputs = tokenizer(formatted_prompt, return_tensors="pt").to(model.device)

outputs = model.generate(
    **inputs,
    max_new_tokens=250,
    temperature=0.3,
    do_sample=True,
    pad_token_id=tokenizer.eos_token_id
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
16
Safetensors
Model size
1B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ZentithLLM/NexusLLM-Math-1B-v1

Adapter
(557)
this model
Adapters
1 model