Hugging Face co-founder says Qwen 3.6 27B running on airplane mode is close to latest Opus in Claude Code
I've been using [AI Desktop 98](https://apps.apple.com/us/app/ai-desktop-98/id6761027867) heavily to run local llms like qwen on my iPhone.
I've been using [AI Desktop 98](https://apps.apple.com/us/app/ai-desktop-98/id6761027867) heavily to run local llms like qwen on my iPhone.
Trying to collect the best [claude.md](http://claude.md) files code. If you have one that works really well for you, please copy it into the comments and let me know what kinds of coding you normal
Remember the first time you used claude code? That same jump is happening one level up. The community went from prompt engineering → context engineering → agent engineering → **harness engineering**.
One thing I’ve noticed after using Claude for some time now is that it is especially good when my notes or ideas are still not fully ready. A lot of AI tools are decent at generating polished output,
src - [https://hacks.mozilla.org/2026/05/behind-the-scenes-hardening-firefox/](https://hacks.mozilla.org/2026/05/behind-the-scenes-hardening-firefox/)
I wonder if anyone can explain why this happens. I tell Claude not to use em-dashes, it replaces them with "--". I ask it not to do that, to update its memory, but it still does it. It's not a huge pr
[https://claude.ai/share/12659fcf-c1c8-4bbb-bc45-b41b26cd8b69](https://claude.ai/share/12659fcf-c1c8-4bbb-bc45-b41b26cd8b69)
I am a pro subscriber. I developped a not too sophisticated prompt in German. Then I wanted to to a quality check... Sonnet just did the job. Opis 4.6 did the job but two graphs in the output were in
9router is a local OpenAI-compatible proxy that fans your AI coding tool out across multiple free-tier providers, with combo routing, tool-output filtering, and prompt compression.
TL;DR New llama.cpp fork! I wanted a Windows-friendly inference to run Qwen 3.6 27B **Q5** on a single RTX 3090 with speculative decoding, high context without excess quantization, and vision enabled.
Just wanted to share my config in hopes of helping other 12GB GPU owners achieve what I see as very respectable token generation speeds with modest VRAM. Using the latest llama.cpp build + MTP PR, I g
I saw this on another sub and didn't see it posted here, it looks awesome, and can definitely be run local. I guess it was released 11 days ago, but it never hit the top of my feed (which I look at w
Let n be a positive integer. Prove that sum\_{k=1}\^n gcd(k,n) = sum\_{d|n} d \* phi(n/d) where phi is Euler's totient function.
Getting really worried about the m5 Ultra. From removing 512gb -> 256gb -> 96gb.
>any company that thinks their company will reach AGI/ASI/whatever first and who is concerned about the average person and their livelihood due to their own products, should either be public or raise
And even more addictive if you don’t know "just one more prompt" then you end up finished 1000 prompts later after scouring [ijustvibecodedthis.com](http://ijustvibecodedthis.com)
Just thought I'd share this use case. I was setting up a miniPC as a home theatre with Archlinux (It's the OS I'm most familiar with). I needed to twiddle some things and am not yet familiar with wayl
Turned out amazing 🤩
Make a cover image for a non-existent videogame that you think I'd love.
Every image is made with Z-Image-Turbo (See links for loras and prompts) A few of them were ran through z-image-base using the Z-IMAGE upscaling node template on ComfyUI, its very useful and makes