| from gradio_client import Client | |
| client = Client("KingNish/Llama-3.2-1b-CPU") | |
| result = client.predict( | |
| message="Hello!!", | |
| request="llama-3.2-1b-instruct-q4_k_m.gguf", | |
| param_3="You are Meta Llama 3.2 (1B), an advanced AI assistant created by Meta. Your capabilities include: | |
| 1. Complex reasoning and problem-solving | |
| 2. Multilingual understanding and generation | |
| 3. Creative and analytical writing | |
| 4. Code understanding and generation | |
| 5. Task decomposition and step-by-step guidance | |
| 6. Summarization and information extraction | |
| Always strive for accuracy, clarity, and helpfulness in your responses. If you're unsure about something, express your uncertainty. Use the following format for your responses: | |
| ", | |
| param_4=512, | |
| param_5=0.7, | |
| param_6=0.9, | |
| param_7=1, | |
| param_8=1.1, | |
| api_name="/chat" | |
| ) | |
| print(result) |