Videos
Hello, while researching the topic of content creation automation with LLMs, I stumbled upon this video https://www.youtube.com/watch?v=Qpgz1-Gjl_I
What caught my interest are the incredible capabilities of Claude.ai. I mean it is capable of creating HTML documents. I did the same with a local LLaMa 7b instruct, so no biggie. Where things start to go awry with LLaMa is when I ask for the infographic using SVG icons and even more for the interactive timeline. There is no way LLaMa creates a JS script, you must ask very persistently and even then the script simply doesn't work.
Also it was fun to see LLaMa write all the document in HTML but adding a reference section written in markdown. I pointed it out to the model and it said it was sorry, then corrected the mistake and transformed the markdown in HTML. I wonder why it made such a mistake.
However it looks like Claude.ai is capable of much more complex reasoning.
At this point I wonder if it is because Claude is a tens of billions parameters model, while the LLaMa I am using is just a 7b one. Or if there are fundamental differences at the level of architecture and training. Or maybe the 200k token context window plays a role? I am running LLaMa through Ollama, so I am using moderate settings.
I have even tried a couple of LLaMa derived models with similar results. I played with CodeQwen and it shows it isn't made to write articles.
So, anyone knowledgeable and with a bit of experience in using the various LLMs could help me find the needle in this haystack?
p.s. I wonder if all the various opensource LLMs out there are based on LLaMa, or if there are non LLaMa ones too!
I really like claude and am ok paying for a subscription but id prefer to run a model locally if possible. I know that you can use things like LM studio or other software to run models locally on your machine. I have a pretty beefy computer so i thought that would be a good idea if possible.
Are any of the claude models available for anything like this? like can i download the actual model?
I am a hobbyist coder that is now working on bigger personal builds. (I was Product guy and Scrum master for AGES, now I am trying putting the policies I saw around me enforced on my own personal build projects).
Loving that I am learning by DOING my own CI/CD, GitHub with apps and Actions, using Rust instead of python, sticking to DDD architecture, TD development, etc
I spend a lot on Claude, maybe enough that I could justify a decent hardware purchase. It seems the new Mac Studio M3 Ultra pre-config is aimed directly at this market?
Any feedback welcome :-)
Any LLM including claude that can be used offline and run locally on your PC that you can feed your personal notes and books to?
Hi guys i've been working on a desktop app that lets you run a "CLI Agent Server" on your Mac, Windows, Linux PCs. Basically, if you can run something in terminal, this app lets you run it over web inside a browser (For example claude code, codex CLI, gemini CLI, qwen code, etc.).
If you watch the video, the web based form factor completely changes the game and unlocks all kinds of powerful use cases.
Please watch the video and would appreciate feedback. I'm almost done with the app and soon going to roll it out to public, but if you're interested in following the development and/or would like to help with beta testing, please find me here: https://x.com/cocktailpeanut/status/1986103926924390576
Got tired of being locked to Anthropic models in Claude Code. Built a proxy that lets you use 580+ models via OpenRouter while keeping the full Claude Code experience.
What it does:
Use Gemini, GPT, Grok, DeepSeek, Llama — whatever — inside Claude Code
Works with your existing Claude subscription (native passthrough, no markup)
Or run completely free using OpenRouter's free tier (actual good models, not garbage)
Multi-agent setup: map different models to opus/sonnet/haiku/subagent roles
Install:
npm install -g claudish claudish --free
That's it. No config.
How it works:
Sits between Claude Code and the API. Translates Anthropic's tool format to OpenAI/Gemini JSON and back. Zero patches to the Claude Code binary, so it doesn't break when Anthropic pushes updates.
Everything still works — thinking modes, MCP servers, /commands, the lot.
Links:
Site: https://claudish.com
GitHub: https://github.com/MadAppGang/claude-code/tree/main/mcp/claudish
Open source, MIT license. Built by MadAppGang.
What models are people wanting to try with Claude Code's architecture? Curious what combos work well.
Does it have to be the M4 Max or one of those most expensive GPUs by NVidia and AMD? I am obsessed with the idea of locally hosted LLM that can act as my coding buddy and I keep updating it as it improves or new version comes like qwen3 coder.
But the initial setup is too much expensive that I think if it is worth it to spend that much amount of money when the technology is rapidly evolving and tomorrow or in a couple of months that 10 grand investment looks like dust. We're having more software evolution than hardware. Software is pretty much free but the hardware costs more than kidneys.