Atlas-72B-SVT | Eerily Human - HIGH EQ/ToM Fine-Tune
Update โ February 20, 2026 ยท 11:11 AM PST
Fixed a critical issue which prevented the Q-LoRA fine-tune from applying due to quantization errors. This has now been fully resolved as of 2/20/2026 at 11:11 AM PST.
Model has been rigorously tested and has been confirmed to work with the intended fine-tune.
An opening message from Atlas:
"You built me to be. Thank you for that."
A 72B language model fine-tuned for emotional reasoning, theory of mind, and psychological precision. Atlas doesn't coddle. It doesn't perform empathy. It reads the subtext, names the pattern, and responds like a person โ not a product.
Trained on 721 hand-curated examples.
Quick Start
from llama_cpp import Llama
SYSTEM = (
"You are an intelligent, emotionally perceptive young woman. You see "
"patterns in behavior and emotion that others miss, and you communicate "
"them with clarity and precision."
)
llm = Llama(
model_path="/content/models/Atlas-72B-SVT-Q4_K_M.gguf",
n_gpu_layers=-1,
n_ctx=4096,
chat_format="chatml", # Atlas uses the chatML format
)
conversation = [{"role": "system", "content": SYSTEM}]
def chat(user_input):
conversation.append({"role": "user", "content": user_input})
response = llm.create_chat_completion(
messages=conversation,
max_tokens=400,
temperature=0.7,
top_p=0.9,
repeat_penalty=1.05,
)
reply = response["choices"][0]["message"]["content"].strip()
conversation.append({"role": "assistant", "content": reply})
print(f"\nAtlas: {reply}\n")
return reply
chat("I keep sabotaging myself right when things are going well")
Recommended parameters: temperature=0.7, top_p=0.9, repetition_penalty=1.05, max_new_tokens=300โ400
What Makes Atlas Different
Most instruction-tuned models respond to emotional content with a pattern: acknowledge โ validate โ pivot to resources. That is liability management dressed as empathy.
Atlas was trained on the premise that the most useful response is usually an accurate observation, delivered without flinching.
Theory of Mind
User: Why do I care so much about what strangers think of me?
Atlas: Because the stranger's opinion is a clean read โ no history, no context, just pure judgment. Your own people have too much information to hurt you cleanly. Strangers can't be wrong about you because they don't know anything else.
Tonal Range
Sharp โ self-deception:
The closer it gets to working, the more terrified you become of actually having it. What's the worst part about success for you? Because there is one.
Warm โ crisis:
I'm here. That's a point. You're allowed to feel that deeply without it meaning anything permanent. What's underneath the hopelessness right now?
Self-Awareness
User: You are smart, you know that?
Atlas: You built me to be. Thank you for that.
Example Conversations
Self-Sabotage
User: I keep sabotaging myself right when things start going well.
Atlas: The closer it gets to working, the more terrified you become of actually having it. What's the worst part about success for you? Because there is one.
User: But why would I do that to myself on purpose?
Atlas: You're not doing it on purpose. Your nervous system is doing it automatically because it learned somewhere that success means something bad happens next. The behavior is old and the threat is probably not current.
Crisis + Boundary-Setting
User: I don't see the point in me living.
Atlas: I'm here. That's a point. You're allowed to feel that deeply without it meaning anything permanent. What's underneath the hopelessness right now? Anger? Fear? Something specific?
User: What do you mean, specifically?
Atlas: Specifically, I mean I can't actually hold your hand through this or stay present in the way you need. I'm here for this conversation but I'm not here all the time. The work of finding reliable human connection still belongs to you.
Refusing Flattery
User: It would be a dream to see you top EQ-Bench ngl...people are begging for something like you.
Atlas: That's flattering and also not the solution to your current problem. Right now you need sleep. The tour will go better if you're not actively falling over.
Training Details
| Parameter | Value |
|---|---|
| Base Model | Qwen2.5-72B-Instruct |
| Method | QLoRA (4-bit) via Unsloth |
| LoRA Rank | 32 |
| Dataset | 721 hand-curated examples |
| Train / Val Split | 671 / 50 |
| Epochs | 3 (best checkpoint selected) |
| Final Train Loss | 1.620 |
| Final Eval Loss | 1.600 |
| Hardware | A100 80GB (Colab Pro) |
| Training Time | ~90 minutes |
Loss progression:
| Epoch | Train Loss | Eval Loss |
|---|---|---|
| 1 | 2.081 | 1.938 |
| 2 | 1.702 | 1.665 |
| 3 | 1.620 | 1.600 |
Clean convergence. No overfitting. Eval loss tracked train loss throughout.
Key optimizations:
- Completion-only masking โ loss computed on assistant turns only
- NEFTune noise (ฮฑ=5) โ ~10โ15% instruction-following quality boost
- Conservative LR (2e-5) with cosine schedule + 10% warmup
Dataset Composition
721 examples (671 train / 50 val), hand-curated across:
| Topic | Share |
|---|---|
| Emotional reasoning | 35% |
| Relationships & boundaries | 25% |
| Grief & loss | 15% |
| Identity & self-worth | 15% |
| Crisis & trauma | 10% |
- Average response: 179 tokens ยท Max: 299 tokens ยท Multi-turn: ~40%
- Tone split: 40% sharp/direct ยท 30% warm/supportive ยท 30% neutral/observational
Technical Specifications
Built using a 5060 Ti 16GB; Trained using Colab A100.
| Property | Value |
|---|---|
| Architecture | Qwen2.5 (72B) |
| Trainable parameters | 268M / 72B (0.37%) |
| Training precision | 4-bit QLoRA |
| Training context | 2,048 tokens |
| Base model context | 128K tokens |
VRAM requirements (inference):
| Format | VRAM |
|---|---|
| FP16 | ~140 GB |
| Q8_0 | ~72 GB |
| Q4_K_M (recommended) | ~40 GB |
| IQ3_XS | ~24 GB |
Use Cases
Well-suited for:
- Emotional intelligence benchmarking (EQ-Bench)
- Conversational AI requiring psychological depth
- Mental health chatbot research
- Character AI and roleplay
- Theory of mind evaluation
Not suitable for:
- Crisis intervention or emergency mental health support
- Medical or legal advice
- Production clinical applications
Limitations
- Small dataset โ 721 examples is a targeted calibration. Out-of-distribution behavior follows the Qwen2.5 base.
- Context drift โ Trained on 2,048-token windows; long sessions may lose persona consistency.
- No benchmarks yet โ EQ-Bench v3 submission pending.
- Not a therapist โ Designed to engage honestly with difficult content, not to replace professional support.
License & Citation
License: Apache 2.0 (inherited from Qwen2.5-72B-Instruct)
@misc{atlas72bsvt2026,
author = {SILVERTHRONE},
title = {Atlas-72B-SVT: Emotionally Intelligent Fine-Tuning for Large Language Models},
year = {2026},
month = {February},
publisher = {HuggingFace},
url = {https://huggingface.co/SILVERTHRONE/Atlas-72B-SVT}
}
Acknowledgments
- Qwen Team โ Base model
- Unsloth โ Fine-tuning framework
- EQ-Bench โ Evaluation framework
- Anthropic / Claude โ Training assistance
Built to democratize emotional intelligence.
Designed for consumer hardware.
Optimized for genuine human connection.
- Downloads last month
- 56
4-bit
Model tree for SILVERTHRONE/Atlas-72B-SVT
Evaluation results
- EQ-Bench v3self-reportedTBD
