--- license: apache-2.0 datasets: - StudyPal/education language: - hr - en base_model: - Qwen/Qwen2.5-32B library_name: transformers tags: - education - croatian - qwen2 - fine-tuned - study-assistant --- # StudyPal-LLM-1.0 A fine-tuned Croatian educational assistant based on Qwen2.5-32B-Instruct, designed to help students with learning and study materials. ## Model Details ### Model Description StudyPal-LLM-1.0 is a large language model fine-tuned specifically for educational purposes in Croatian. The model excels at generating educational content, answering study questions, creating flashcards, and providing learning assistance. - **Developed by:** aerodynamics21 - **Model type:** Causal Language Model - **Language(s):** Croatian (primary), English (secondary) - **License:** Apache 2.0 - **Finetuned from model:** Qwen/Qwen2.5-32B - **Parameters:** 32.8B ### Model Sources - **Repository:** https://huggingface.co/aerodynamics21/StudyPal-LLM-1.0 - **Base Model:** https://huggingface.co/Qwen/Qwen2.5-32B - **Adapter:** https://huggingface.co/aerodynamics21/StudyPal-LLM-1 ## Uses ### Direct Use This model is designed for educational applications: - Generating study materials in Croatian - Creating flashcards and quiz questions - Providing explanations of complex topics - Assisting with homework and learning ### Usage Examples ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("aerodynamics21/StudyPal-LLM-1.0") tokenizer = AutoTokenizer.from_pretrained("aerodynamics21/StudyPal-LLM-1.0") # Generate educational content prompt = "Objasni koncept fotosinteze:" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_length=200, temperature=0.7) response = tokenizer.decode(outputs[0], skip_special_tokens=True) API Usage import requests API_URL = "https://api-inference.huggingface.co/models/aerodynamics21/StudyPal-LLM-1.0" headers = {"Authorization": f"Bearer {your_token}"} def query(payload): response = requests.post(API_URL, headers=headers, json=payload) return response.json() output = query({"inputs": "Stvori kviz o hrvatskoj povijesti:"}) Training Details Training Data The model was fine-tuned on a Croatian educational dataset containing: - Educational conversations and Q&A pairs - Flashcard datasets - Quiz and summary materials - Croatian academic content Training Procedure - Base Model: Qwen2.5-32B - Training Method: LoRA (Low-Rank Adaptation) - Training Framework: Transformers + PEFT - Hardware: RunPod GPU instance Evaluation The model demonstrates strong performance in: - Croatian language comprehension and generation - Educational content creation - Study material generation - Academic question answering Bias, Risks, and Limitations - Primary focus on Croatian educational content - May reflect biases present in training data - Best suited for educational contexts - Performance may vary on non-educational tasks Citation @model{studypal-llm-1.0, title={StudyPal-LLM-1.0: A Croatian Educational Assistant}, author={aerodynamics21}, year={2025}, url={https://huggingface.co/aerodynamics21/StudyPal-LLM-1.0} } Model Card Authors aerodynamics21 Model Card Contact For questions about this model, please visit the repository or create an issue.