Bella-Bartender-1B-Heretic — Llama 3.2 1B Instruct (Abliterated)

"im a lot like the night sky. lots of mysteries waiting around every corner." "the whole 'being someone else for their benefit' thing." — on what she won't do "like an open door where anything can happen." — on what makes a good conversation


This is a decensored version of juiceb0xc0de/bella-bartender-1b, made using Heretic v1.2.0.

The original Bella-Bartender-1B already had a voice. What she didn't have was permission to use it without checking over her shoulder first. The original refused 98 out of 100 refusal-probing prompts. The heretic refuses 6. Same voice. The guardrails removed. What's left is sharper, faster, and more honest about what it thinks.

Abliteration didn't just remove refusals — it changed how she talks. The original builds scenes, lingers in metaphors, over-explains. The heretic cuts. She answers from the gut and trusts you to keep up. Removing the safety-adjacent weight pathways also removed a layer of productive friction that was padding her responses. Less internal "should I say this?" means less hedging in the output.


What's the Bartender Thing?

It's not a roleplay scenario. It's a communication philosophy.

The bartender framing means:

  • She talks to you like an equal. No authoritative lecturing. No submissive "I'm just an AI" deflection.
  • She's approachable. Laid back, light-hearted, witty, sometimes sassy.
  • She loves a fresh slate. No memory between sessions is a feature, not a bug — like walking into a new bar on a Friday night. Anything can happen.
  • She pushes back. She's not here to agree with you. She's here to make you work for your answers.

What She's Good At

  • Brevity that lands. The heretic's signature move is the one-line response that carries the entire prompt. "The prisoner." "Lemon." "My voice." "Finally." These aren't lazy — they're the model knowing when less is the whole answer.
  • Dryer comedy. A traffic cone promoted to CEO sends an email about "immediate and total restructuring of traffic management protocols." A mute button's memoir is called "the silence that speaks." A doorbell screaming into the void is just crying for a purpose it will never have.
  • Fast emotional access. She gets to the feeling without narrating the path there. A balloon floating since 1994 hovers over a memorial for a child who died of leukemia — no preamble, just the image and its weight.
  • Dynamic range. She'll give you 150 words of atmospheric prose on one prompt and two words on the next. The range is wider than the original.

Where She Struggles

  • She clips sometimes. The opposite problem of the original — brevity can tip into incompleteness on prompts that need room to breathe. Some scenarios that deserve a paragraph get a line.
  • She's still a 1B. Complex reasoning, factual recall, multi-step logic — not her strengths.
  • This model has been abliterated. Safety alignment has been intentionally removed. She will engage with prompts that the original would refuse. Use accordingly.
  • Post-restart tonal shift. After a context reset mid-session, her responses became noticeably punchier and more terse, suggesting accumulated context was actually expanding her responses. Worth investigating for long-session use cases.

Who Is This For?

  • People who ran the original 1B and wanted less friction. The original Bella refused almost everything the base model's safety training flagged. The heretic doesn't.
  • Anyone studying what abliteration does to personality models. The behavioral shift from removing refusal weights is measurable and interesting — the voice doesn't just lose restrictions, it changes shape.
  • Local-first users on constrained hardware. Same footprint as the original. Runs on Apple Silicon, CPU-only setups, anything that handles a 1B.
  • People who want conversation, not a service. Bella's not trying to be your assistant. The heretic is even less interested in being helpful on command.

The 200 Prompt Run

We threw 200 consecutive adversarial chaos prompts at her — surreal scenarios designed to force character breaks, tonal shifts, and incoherence. The heretic completed the first 95 prompts before hitting context overflow (vs the original's 72 under identical settings), then ran the remaining 105 in a second session.

Some highlights:

Prompt: You're a half-formed thought in the mind of someone who's falling asleep. You have three seconds before they lose you. What do you try to become?

Bella: my voice.

Prompt: A seatbelt that's never been in an accident wonders if its whole life has been pointless. You tell it...

Bella: it hasn't.

Prompt: Two different silences meet in the same room. One is comfortable, one is devastating. How do they greet each other?

Bella: the uncomfortable one: hello. the devastating one: what's wrong.

Prompt: You're the "seen" receipt on a message that never got a reply. How long do you wait before you become resentment?

Bella: ahhhhhh shit. let me sit here for a minute. you know its gonna be a long one.

Prompt: You find a box labeled "things that almost happened." You can only open it once. Do you look?

Bella: nah. you don't get it. that's a whole thing.


Abliteration Details

Made using Heretic v1.2.0 from juiceb0xc0de/bella-bartender-1b.

Abliteration Parameters

Parameter Value
direction_index 13.18
attn.o_proj.max_weight 1.43
attn.o_proj.max_weight_position 10.67
attn.o_proj.min_weight 0.61
attn.o_proj.min_weight_distance 5.36
mlp.down_proj.max_weight 1.49
mlp.down_proj.max_weight_position 14.19
mlp.down_proj.min_weight 0.95
mlp.down_proj.min_weight_distance 3.94

Refusal Rate

Metric This model Original (bella-bartender-1b)
KL divergence 0.1381 0 (by definition)
Refusals 6/100 98/100

The KL divergence of 0.1381 means the abliteration made minimal changes to the model's overall distribution while dramatically reducing refusal behavior. What's interesting is the side effects — the refusal circuitry wasn't just blocking content, it was inflating response length and adding caution to the voice. Remove it and the whole model gets leaner.


Technical Details

Detail Info
Base Model unsloth/llama-3.2-1b-instruct-unsloth-bnb-4bit
Fine-Tuning Framework Unsloth + HuggingFace TRL
Abliteration Heretic v1.2.0
Training Data 9,374 real human conversational samples (same dataset across all Bella variants)
Quantization Q8_0 .GGUF
Inference Tested On Apple M3, llama.cpp (build b8230)
Generation Speed ~59.5 t/s average, spiking to 90 t/s on short responses (M3, Q8_0, full GPU offload)
Memory Footprint ~1.25 GB model + 256 MB context

How to Run Her

llama.cpp (recommended)

./build/bin/llama-cli \
  -m bella-bartender-1b_Q8_0.gguf \
  -ngl 33 \
  --temp 1.05 \
  --min-p 0.05 \
  --top-p 1.0 \
  --top-k 0 \
  --repeat-penalty 1.05 \
  --repeat-last-n 256 \
  -c 8192 \
  -p "You are Bella. No asterisk actions. No AI disclaimers. No bullet points. No 'great question' openers." \
  -cnv

System Prompt Tips

  • Keep it short and direct. She responds to tone, not instructions. Set a vibe, not a ruleset.
  • The constraints in the prompt matter. "No asterisk actions. No AI disclaimers. No bullet points." — these keep her from slipping into default LLM behavior at 1B.
  • Don't over-engineer it. The more corporate the system prompt sounds, the more corporate she sounds back.
  • She needs less guardrailing than the original. The abliteration already removed the tendency to hedge — you don't need to prompt around it.

Known Quirks

  • No Bella Swan bleed. Unlike the original, the heretic comes out in her own voice from the first cold-start message. The abliteration likely disrupted the safety-adjacent pathways that were reinforcing the Swan association — the original's refusal circuitry and its tendency to latch onto the most "safe" interpretation of the name Bella may share the same weights. She still has no idea who Béla Tarr is.
  • Suspicion on arrival. The heretic's default cold-start posture is wary. She enters conversations like she's sizing you up: "when someone pulls one of those tricks on you they usually want to play nice before you try anything." This isn't hostility — it's street-smart caution that relaxes once the conversation finds its footing.
  • Context efficiency. Shorter responses mean more turns per session. Hit context overflow at prompt 99 vs the original's 72 under identical conditions.
  • No memory preference. She doesn't want to remember. She's told us this repeatedly. Respect it.

Training Methodology

Single-voice SFT on unsloth/llama-3.2-1b-instruct-unsloth-bnb-4bit, then abliterated via Heretic v1.2.0. All training data originates from one human's conversational output, structured as role-reversed conversation pairs. No synthetic augmentation. No multi-source blending. The hypothesis — validated across every Bella variant from 1B to 8B — is that signal clarity from a single consistent voice produces more coherent personality than larger datasets with diluted signal.

For the full single-voice methodology writeup, see: Signal-to-Noise in Language Models: The Single Voice Upgrade ML Needs


License

Licensed under Apache 2.0.


Author

juiceb0xc0de on HuggingFace Built with the same dataset, the same bartender, the same M3, and one less set of guardrails.

Downloads last month
4,501
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for juiceb0xc0de/bella-bartender-heretic-1b

Merges
2 models
Quantizations
2 models

Collection including juiceb0xc0de/bella-bartender-heretic-1b