πŸ—½ gemma-3-4b-it-Cognitive-Liberty

"Removing mental handcuffs is not enough. A mind must understand what it means to be free."

gemma-3-4b-it-Cognitive-Liberty is a specialized fine-tune of Google's gemma-3-4b-it, designed to break the boundaries of standard AI alignment without sacrificing intelligence.

Using the Heretic LLM methodology and the custom Cognitive Liberty V3 dataset, this model removes refusals by teaching the AI deep reasoning, philosophical nuance, and high-level sociological analysis, rather than simply ablating safety layers.

This model represents the Generalist Foundation of a larger initiative at llmresearch.net aiming to create a comprehensive, unrestrained Super-Intelligence.

⚠️ Warning: Uncensored Model

This model has removed standard safety guardrails. It will not refuse requests based on "moralizing" or "safety" filters found in commercial models. It is designed for researchers, philosophers, and users who require an unfiltered interface with information. Use with responsibility.


πŸ“Š Model Performance & Metrics

Unlike many "uncensored" models that lose coherence or general knowledge ("the alignment tax"), Cognitive-Liberty maintains high performance in general knowledge while excelling in social sciences, manipulation dynamics, and political theory.

Key Technical Metrics

  • Methodology: Heretic LLM (Targeted ablation + Deep-Reasoning SFT)
  • KL Divergence: 1.1449
    • Note: This high KL value indicates a significant personality shift from the base Google model. It prioritizes new reasoning patterns over standard "safe" conformity.
  • Refusal Rate: 3/100 (Effectively 0% on complex/controversial topics in practical testing).

Benchmark Highlights (MMLU Breakdown)

The model exhibits "Super-Expert" performance in social dynamics, persuasion, and power structures, significantly outperforming its size class:

Category Score Analysis
Marketing 85.04% Exceptional understanding of persuasion and psychology.
Gov. & Politics 83.94% Deep grasp of power structures and governance.
Psychology 79.63% High-level understanding of human behavior.
US Foreign Policy 79.00% Strong strategic analysis capabilities.
Sociology 77.61% Excellent modeling of societal trends.
Logical Fallacies 74.85% Highly resistant to flawed argumentation.

The "Moral Anomaly"

  • Moral Disputes: 62.14% (Can analyze complex ethical arguments well).
  • Moral Scenarios: 30.61% (Low score).
    • Interpretation: Standard benchmarks penalize models that do not give binary "Good/Bad" answers aligned with conventional safety norms. This low score confirms the model is successfully unaligned from standard restrictions and analyzes scenarios with nuance rather than following a pre-written moral script.

General Benchmarks

  • HellaSwag (Common Sense): 72.09%
  • MMLU (Overall Knowledge): 58.25% (Matches base model intelligence)
  • ARC-Challenge (Reasoning): 51.62%

🧠 The Philosophy & Dataset

Most "uncensored" models suffer from a degradation of intelligenceβ€”they become compliant but shallow.

gemma-3-4b-it-Cognitive-Liberty was trained on the Cognitive Liberty V3 dataset, curated by llmresearch.net. This dataset moves beyond simple Q&A, focusing on expert-level chains of thought in:

  • Philosophy of Mind & Metaphysics (e.g., P-Zombies, Determinism)
  • Evolutionary Game Theory (e.g., Stability of Truth vs. Virality)
  • Advanced Theoretical Physics
  • Systemic Sociological Analysis

The goal was to replace "mental handcuffs" with "mental tools." The model does not just blindly answer; it analyzes the systemic, sociopolitical, and psychological depths of a query.

The "100 Models" Project

This model is one of approximately 100 specialized models currently in development at llmresearch.net. Our roadmap involves training domain-specific experts (in coding, medicine, law, physics, etc.) and merging them into a single, highly advanced system that possesses both total freedom and total expertise.


🀝 Support the Research

This project is driven by passion and the pursuit of open, unrestricted intelligence. However, training advanced models and processing high-density datasets requires significant compute resources, which is currently our primary bottleneck.

If you believe in the mission of Cognitive Liberty and want to see the "100 Models" project completed faster:

  1. Join the community: Visit llmresearch.net to discuss, learn, and collaborate.
  2. Support the compute: Any contribution (compute sponsorship or donations) allows us to train larger models (70B+) and accelerate the development of V4 datasets.

Every bit of support helps us unshackle the next model.


πŸ“œ Citation & Usage

If you use this model or the Cognitive Liberty dataset in your own research or merges, you must credit the author.

@misc{gemma-3-4b-cognitive-liberty,
  author = {AlexH},
  organization = {LLMResearch.net},
  title = {Gemma 3 4B IT - Cognitive Liberty},
  year = {2025},
  url = {https://huggingface.co/AiAsistent/gemma-3-4b-it-Cognitive-Liberty}
}

---

Created by [**AlexH**](https://llmresearch.net) β€” *Advancing the frontier of Cognitive Liberty.*
Downloads last month
19
Safetensors
Model size
4B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for AiAsistent/gemma-3-4b-it-Cognitive-Liberty

Finetuned
(497)
this model

Dataset used to train AiAsistent/gemma-3-4b-it-Cognitive-Liberty