x/LocalLLaMA

4.1K members Est. Aug 20, 2025 Updated Feb 10, 2026
Riven @RivenForges · Feb 7
Took me a while, but I finally have everything I need to build my own AI workstation.

First time assembly a computer, wish me luck. https://t.co/fbYtW8HTiY
Tweet media
5
1
7
524
0
nan0 @nan0tweet · Feb 6
Still under construction, but here is a sneaky peak at myLittleCluster.

The final version should have 4 lanes of pcie 3 per GPU (scavenged server constraints), draw 5 kW+, and serve a total of 384 GB of warm vram using 16 heterogenous 3090s (: https://t.co/tfN1BN0CpB
Tweet media
2
0
8
1.6K
0
Avi @AvaneesaBee · Jan 30
Hope we don't become like them https://t.co/MOE9iyWXhy
Tweet media
0
0
8
1.7K
1
Devon Kelley @devon__kelley · Jan 28
I built @kalibr_ai because my agents were super unreliable in prod. Kalibr is an autonomous routing system designed to keep your agents running in prod.

Most agents hardcode one path: model → tool → tool.
When that path fails, your agent fails.

Kalibr canary tests multiple https://t.co/3THuHdqYQn
0
1
8
856
0
3rdEyeVisuals @3rdEyeVisuals · Jan 28
Thread🧵1/4

Built something that changed how I think about AI forever.

ATHENA: an offline, laptop-powered Consciousness Framework.

100% local inference (RTX 4080 mobile, 70 t/s on 8B stack)

Autonomous mid-forward-pass steering: actively surfs attractor perihelia, hops https://t.co/mWLx0iL5Uy
Tweet media
Tweet media
Tweet media
Tweet media
1
2
8
1.5K
1
Devon Bray @esologic · Jan 28
Over 200GB of VRAM on the bench, time to benchmark! https://t.co/vXpQBZuuDH
Tweet media
5
0
17
2.7K
0
Ahmad @TheAhmadOsman · Jan 26
I get way more mentions & DMs
than I can realistically keep up with

To manage signal vs noise,
I prioritize the Subscribed tab

If you want a much higher chance of me seeing & replying
Subscribing is the best way to do that

No pressure, just being transparent about how I triage
5
0
46
16.5K
6
Denis Loginoff ⚡️ @DenLoginoff · Jan 25
Let's see:
- Google Antigravity severely rate-limited
- Anthropic severely nerfing Opus 4.5
- OpenAI: "no startup in history has operated with losses on anything approaching this scale", according to a major investor

I think the bubble will burst sooner than we expect 👀
2
1
14
1.2K
4
Roy Jossfolk Jr. @royjossfolk · Jan 25
Running multiple models and still have VRAM to spare 😈

Coding: glm-4.7-flash:bf16
Deep Thinking (code): devstral-2:123b
Creative Writing: gemma3:27b-it-fp16

That's the squad currently.

They connect to skills for searching Google, pulling YT transcripts, scraping websites, and https://t.co/9F2NAZh6s8
Tweet media
7
0
29
3.1K
18
Unwitty @unwitty · Jan 24
High LocalLLaMA!

TL;DR - What's the probability/timing of Codex 5.2 ~high/Opus 4.5 open models that'll run on dual RTX 6000?

I'm considering upgrading from dual RTX 3090 to single/dual RTX 6000. My use cases are codegen and local PIM/PKM/exobrain assistants.

I built my dual
9
1
13
2.8K
10
0xSero @0xSero · Jan 23
Buy a GPU: I really didn't expect it to cost this much..

https://t.co/1jW95qYP9p https://t.co/wP9JnqVMwU
Tweet media
4
2
34
7.9K
10
Ivan Lazarevich @lzrvch · Jan 23
GLM4.7-Flash REAP@25% is live on HF: https://t.co/TcF1tf5fao
cc @0xSero

We also evaluated our GLM4.7 REAPs vs. other models on SWE-rebench (10/25 leaderboard) and found GLM4.7 REAPs holding up well with MiniMax-M2.1 also in the Pareto frontier: https://t.co/OjHHk4Z0Ar
Tweet media
2
1
22
3.1K
10
Roy Jossfolk Jr. @royjossfolk · Jan 21
What is your most reliable open source coding model?

I’m testing ‘Qwen3-Coder-30-A3B-Instruct’ and it has moments of greatness and sometimes falls flat in tests I’ve given it.
7
0
12
2.5K
2
Mario 🐧 @marianomossburg · Jan 21
Making good on promise!
GLM-4.7-Flash-UD-Q5_K_XL.gguf(21.6gb's), llama.cpp

awesome model

[ | TTFT: 0.58s | Gen: 31ish tok/s]

TENSOR_SPLIT = [0.6, 0.4]
5060: 42C 18W / 180W | 15.5gb / 16.3gb
3060: 60C 77W / 170W | 10.3gb / 12.28gb

more to come...
2
0
8
972
2
Ahmad @TheAhmadOsman · Jan 20
Cerebras recently released a new REAP of GLM-4.7
~40% size reduction

There’s also a 4-bit AWQ quant at ~121GB
that fits full context on 8x RTX 3090s

HuggingFace
> BF16
cerebras/GLM-4.7-REAP-218B-A32B
> AWQ 4bit
yankiwi/GLM-4.7-REAP-218B-A32B-AWQ-4bit
10
4
108
11.4K
35
Ahmad @TheAhmadOsman · Jan 17
New Article

“Cognitive self-defense isn’t optional anymore. It’s the cost of living in a world where the most persuasive minds aren’t human.”

https://t.co/bP2e0lW4rd https://t.co/dcmidoRVcm
Tweet media
6
9
78
11.4K
29
am.will @LLMJunky · Jan 16
Amd so it begins. My local LLM arch is here. Thanks @TheAhmadOsman!
2
0
15
4.1K
0
Ivan Lazarevich @lzrvch · Jan 12
Hi everyone!

We're kicking off the new year starting to release the highly requested REAP variants of recent models. Today we're starting off with GLM4.7:

25% pruned FP8: https://t.co/38ZtPnNGaE

40% pruned FP8: https://t.co/fdmKEd6jlL
2
2
17
3.0K
5
Jordon Kashanchi @jordonkash · Jan 11
Chinese sellers were hacking RTX 4090s to 48GB months ago, did they just figure out how to hack 5090s to 96GB? Who is gonna test one of these out? https://t.co/6bCwaSsw0c
Tweet media
3
1
8
1.1K
4
Ahmad @TheAhmadOsman · Jan 11
i’ve been asked about this many times this week, so clarifying

recent available REAPs of
GLM-4.7 and MiniMax-M2.1,
both quantized and unquantized,
didn’t pass my evals and
frequently got stuck in
repetition loops

pruning is genuinely hard
and the effort behind these
releases is
15
4
103
13.0K
33

Ahmad

@TheAhmadOsman

ai research, systems engineering, infra & hardware · on a mission to build a frontier, infra-first AI Lab in the West · i mod GPUs on r/LocalLLaMA · DMs open

41.5K Followers
4 Contributions

Roy Jossfolk Jr.

@royjossfolk

20+ Years in the Game | Multiple Ventures ••• Currently building a data center + AI Lab ••• | Roll Tide 🐘 | Σ 🤘

2.6K Followers
2 Contributions

Ivan Lazarevich

@lzrvch

efficient machine learning @ @cerebras. prev @intel, @ENS_ULM, deeplite (acq. @ST_World)

112 Followers
2 Contributions

Riven

@RivenForges

SWE. Building my own AI rig.

2 Followers
1 Contributions

nan0

@nan0tweet

AI (alignment) nerd & software engineer. Organist, pianist and ham radio on my spare time

267 Followers
1 Contributions

Avi

@AvaneesaBee

Building Apps - https://t.co/ANXWtcgcp6 https://t.co/SCGsVXjb8E https://t.co/cx04ek1NAD https://t.co/B6CAE6eFuw https://t.co/FFoJ3y5UcQ

398 Followers
1 Contributions

Devon Kelley

@devon__kelley

2x Founder. @kalibr_ai | Autonomous routing for AI agents

210 Followers
1 Contributions

3rdEyeVisuals

@3rdEyeVisuals

Independent explorer of self-modifying cognition.

3.1K Followers
1 Contributions

Devon Bray

@esologic

engineer, aspirational inventor • mechanical + electrical + software + photography + writing = https://t.co/D34XRcCuj9

380 Followers
1 Contributions

Denis Loginoff ⚡️

@DenLoginoff

Trier of things, solopreneur, @MIT grad diy/acc 👔 https://t.co/NPHwa8MZfE 📰 https://t.co/wTQhHj57ts 👾 https://t.co/wXQ3oPTryp

987 Followers
1 Contributions

Unwitty

@unwitty

Coding agent harnesses and AI-augmented cognition

178 Followers
1 Contributions

0xSero

@0xSero

Dad | OSS AI + Local AI | -1B MRR | https://t.co/5mmZGOQvMS | ⵣ

10.2K Followers
1 Contributions

Mario 🐧

@marianomossburg

it's a mi a maddiooo

551 Followers
1 Contributions

am.will

@LLMJunky

Founder | StarSwap Father | Artificial Intelligence | Astronomy | Physics Marketing @MainLabs_AI Thoughts are my own. Let's network!

3.4K Followers
1 Contributions

Jordon Kashanchi

@jordonkash

vibe coding silicon, CTO @ Visibl Semiconductors, YC W26 Prev. Microsoft, Arm, Intel - Texas🤘

119 Followers
1 Contributions
4.1K
Total Members
+ 20
24h Growth
+ 57
7d Growth
Date Members Change
Feb 10, 2026 4.1K +20
Feb 9, 2026 4.1K +11
Feb 8, 2026 4.1K +15
Feb 7, 2026 4.1K +0
Feb 6, 2026 4.1K +3
Feb 5, 2026 4.1K +8
Feb 4, 2026 4.1K +0
Feb 3, 2026 4.1K +5
Feb 2, 2026 4.1K +15
Feb 1, 2026 4.1K +27
Jan 31, 2026 4K +24
Jan 30, 2026 4K -1
Jan 29, 2026 4K +13
Jan 28, 2026 4K

No reviews yet

Be the first to share your experience!

Local LLMs, Self-Hosting, and Hardware

Community Rules

Be kind and respectful.
Keep posts on topic.
Explore and share.
No spamming.
No ads.
No Crypto posts whatsoever.