| | --- |
| | language: |
| | - en |
| | tags: |
| | - math |
| | - education |
| | - llama-3 |
| | - peft |
| | - lora |
| | base_model: meta-llama/Llama-3.2-1B-Instruct |
| | license: apache-2.0 |
| | --- |
| | |
| | # NexusLLM-Math-1B-v1 |
| |
|
| | ## Model Details |
| | NexusLLM-Math-1B-v1 is a fine-tuned version of Llama 3.2 (1B parameters) optimized specifically for solving advanced high-school mathematics problems, with a focus on JEE Main and Advanced syllabus topics. |
| |
|
| | - **Developed by:** ZentithLLM |
| | - **Model Type:** Causal Language Model (Fine-tuned with LoRA) |
| | - **Language:** English |
| | - **Base Model:** meta-llama/Llama-3.2-1B-Instruct |
| | - **Precision:** FP16 |
| |
|
| | ## Intended Use |
| | This model is designed to act as an educational assistant for 11th-grade mathematics. It is trained to provide step-by-step reasoning and explanations for complex topics, rather than just outputting the final answer. |
| |
|
| | **Primary Topics Covered:** |
| | - Binomial Theorem |
| | - Geometry (Circle Theorems, cyclic quadrilaterals, tangents, etc.) |
| |
|
| | ## Training Data |
| | The model was trained on a custom dataset of structured mathematics Q&A pairs. The dataset maps specific mathematical prompts to detailed completions, heavily utilizing an `explanation` field to teach the model the underlying mathematical logic and derivation steps. |
| |
|
| | ## Training Procedure |
| | The model was fine-tuned using the standard Hugging Face `trl` and `peft` libraries on a single NVIDIA T4 GPU, utilizing strictly native FP16 precision to ensure mathematical gradient stability. |
| |
|
| | - **Training Framework:** Pure Hugging Face (No Unsloth/Quantization) |
| | - **Method:** LoRA (Low-Rank Adaptation) |
| | - **Rank (r):** 32 |
| | - **Alpha:** 32 |
| | - **Optimizer:** adamw_torch |
| | - **Learning Rate:** 2e-4 |
| | - **Max Sequence Length:** 2048 |
| | |
| | ## How to Use |
| | Because this model was trained on a specific dataset structure, you **must** wrap your prompts in the `### Instruction:` and `### Response:` format for it to output the correct mathematical explanations. |
| | |
| | ```python |
| | import torch |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | |
| | model_id = "ZentithLLM/NexusLLM-Math-1B-v1" |
| |
|
| | tokenizer = AutoTokenizer.from_pretrained(model_id) |
| | model = AutoModelForCausalLM.from_pretrained( |
| | model_id, |
| | torch_dtype=torch.float16, |
| | device_map="auto" |
| | ) |
| | |
| | question = "What is the general term in the expansion of (x+y)^n?" |
| | formatted_prompt = f"### Instruction:\\n{question}\\n\\n### Response:\\n" |
| | |
| | inputs = tokenizer(formatted_prompt, return_tensors="pt").to(model.device) |
| | |
| | outputs = model.generate( |
| | **inputs, |
| | max_new_tokens=250, |
| | temperature=0.3, |
| | do_sample=True, |
| | pad_token_id=tokenizer.eos_token_id |
| | ) |
| | |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |