NarrativAIV2

This model is a fine-tuned version of the LLaMA 3.1 language model, specifically trained on a curated dataset of interactive roleplaying scenarios.

Model Description

  • Base Model: LLaMA 3.1
  • Training Data: A diverse dataset of 977 fictional scenarios featuring engaging characters in various settings, emphasizing emotional depth and complex interactions.
  • Fine-tuning Method: This model was fine-tuned using supervised learning, focusing on continuing the given roleplay prompt in a consistent, immersive manner.

Limitations

  • Bias: The model's responses may reflect biases present in the training data.
  • Factual Accuracy: The model is not designed to provide factual information and may generate inaccurate statements.
  • Repetitive Responses: Occasionally, the model may produce repetitive or predictable responses.

License

This model is released under the MIT license.

Downloads last month
15
Safetensors
Model size
8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for passthepizza/narrativAIV2

Finetuned
(1652)
this model
Quantizations
1 model