πŸ’Ž Finance LLM Full β€” Next-Gen Financial Intelligence Model

Finance LLM Logo

πŸ’Ό Finance LLM Full

A next-generation Financial Intelligence Model
Fine-Tuned, Merged & Optimized for Real-World Finance


Finance LLM Full is a high-performance, fully merged financial Large Language Model (LLM)
designed to deliver crystal-clear, accurate, and structured financial reasoning.

It is trained using LoRA fine-tuning on top of Phi-3 Mini 4K Instruct, and later
merged into a single standalone model for seamless deployment.

This model specializes in Finance, Accounting, Banking, Investment, Stock Markets, and Business Analysis β€”
making it ideal for FinTech products, AI advisors, investment copilots, and enterprise bots.


⚑ Why Finance LLM Full is Special

πŸ”Ή 1. Purpose-Built For Finance

Unlike general LLMs, this model deeply understands:

  • Balance Sheet Interpretation
  • Profit & Loss Breakdown
  • Cashflow Logic
  • EBITDA / EPS / ROE / DCF
  • Risk & Return Analysis
  • Banking, Loans, Limits, Credit Rules
  • Valuation Basics
  • Investment & Portfolio Concepts

πŸ”Ή 2. Merged Model β†’ One File, Zero Hassle

βœ” No LoRA needed
βœ” No adapter loading
βœ” Direct plug-and-play
βœ” Works on CPU / GPU / Colab / Docker

πŸ”Ή 3. Small Model β†’ Big Capability

Powered by Phi-3 Mini, optimized for:

  • Low latency
  • Low VRAM/RAM usage
  • Clean, structured answers
  • High domain accuracy

πŸ§ͺ Quick Start (Copy & Run)

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "devAnurag/finance_llm_full"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)

prompt = "Explain the difference between EBITDA and Net Profit."
inputs = tokenizer(prompt, return_tensors="pt")

outputs = model.generate(**inputs, max_new_tokens=150)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
43
Safetensors
Model size
4B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ 1 Ask for provider support