llm code environment as claude code alternative.
- Last 7 days
-
-
-
elite-ai-assisted-coding.dev elite-ai-assisted-coding.dev
-
OpenHands: Capable but Requiring InterventionI connected my repository to OpenHands through the All Hands cloud platform. I pointed the agent at a specific issue, instructing it to follow the detailed requirements and create a pull request when complete. The conversational interface displayed the agent's reasoning as it worked through the problem, and the approach appeared logical.
Also used openhands for a test. says it needs intervention (not fully delegated iow)
-
When an agent doesn't deliver what you expected, the temptation is to engage in corrective dialogue — to guide the agent toward the right solution through feedback. While some agents support this interaction model, it's often more valuable to treat failures as specification bugs. Ask yourself: what information was missing that caused the agent to make incorrect decisions? What assumptions did I fail to make explicit?This approach builds your specification-writing skills rapidly. After a few iterations, you develop an intuition for what needs to be explicit, what edge cases to cover, and how to structure instructions for maximum clarity. The goal isn't perfection on the first try, but rather continuous improvement in your ability to delegate effectively.
don't iterate for corrections. Redo and iterate the instructions. This is a bit like prompt engineering the oracle, no? AI isn't the issue, it's your instructions. Up to a point, but in flux too.
-
One effective technique for creating comprehensive specifications is to use AI assistants that have full awareness of your codeba
ah, turtles all the way down. using AI to generate the task specs.
-
A complete task specification goes beyond describing what needs to be done. It should encompass the entire development lifecycle for that specific task. Think of it as creating a mini project plan that an intelligent but literal agent can follow from start to finish.
A discrete task description to be treated like a project in the GTD sense (anything above 2 steps is a project). At what point is this overkill, as in templating this project description may well lead to having the solutions once you've done this.
-
we tend to underspecify because we're exploring, experimenting, and can provide immediate course corrections. We might type a quick prompt, see what the AI produces, and refine from there. This exploratory approach works when you're actively engaged
indeed. as mentioned above too. n:: My sense that this is a learning mode akin to the haptic feedback of working on things by hand.
-
The fundamental rule for working with asynchronous agents contradicts much of modern agile thinking: create complete and precise task definitions upfront. This isn't about returning to waterfall methodologies, but rather recognizing that when you delegate to an AI agent, you need to provide all the context and guidance that you would naturally provide through conversation and iteration with a human developer.
What I mentioned above: to delegate you need to be able to fully describe and provide context for a discrete task.
-
The ecosystem of asynchronous coding agents is rapidly evolving, with each offering different integration points and capabilities:GitHub Copilot Agent: Accessible through GitHub by assigning issues to the Copilot user, with additional VS Code integrationCodex: OpenAI's hosted coding agent, available through their platform and accessible from ChatGPTOpenHands: Open-source agent available through the All Hands web app or self-hosted deploymentsJules: Google Labs product with GitHub integration capabilitiesDevin: The pioneering coding agent from Cognition that first demonstrated this paradigmCursor background agents: Embedded directly in the Cursor IDECI/CD integrations: Many command-line tools can function as asynchronous agents when integrated into GitHub Actions or continuous integration scripts
A list of async coding agents in #2025/08 github, openai, google mentioned. OpenHands is the one open source mentioned. mentions that command line tools can be used (if integrated w e.g. github actions to tie into the coding environment) - [ ] check out openhands agent by All Hands
-
isn't just about saving time — it's about restructuring how software gets built.
not just time saving, but a restructuring. So, any description of how the structure changes (before / after style) further down?
-
several of these tasks running in parallel, each agent working independently on different parts of your codebas
do multiple things in parallel. note: The assumption here that the context is coding
-
why asynchronous agents deserve more attention than they currently receive, provides practical guidelines for working with them effectively, and shares real-world experience using multiple agents to refactor a production codebase.
3 things in this article: - why async agents deserve more attention - practical guidelines for effective deployment - real world examples
-
asynchronous coding agents represent a fundamentally different — and potentially more powerful — approach to AI-augmented software development. These background agents accept complete work items, execute them independently, and return finished solutions while you focus on other tasks.
Async coding agents is a diff kind of vibe coding: you give it a defined more complex tasks and it will work in the background and come back with an outcome.
-
https://web.archive.org/web/20260125124811/https://elite-ai-assisted-coding.dev/p/working-with-asynchronous-coding-agents Eleanor Berger, August 2025.
on asynchronous coding agents
-
-
kentdebruin.com kentdebruin.com
-
One of the people [[💎 Claude + Obsidian Got a Level Up]] mentioned. Based in AMS, Kent de Bruin.
-
-
www.eleanorkonik.com www.eleanorkonik.com
-
[[Eleanor Konik p]] on how her work in Obsidian with Claude Code is changing
-
-
cursusclaudecode.nl cursusclaudecode.nl
-
By [[Frank Meeuwsen p]], ivm #2026/01/30 sessie in Utrecht
-
- Jan 2026
-
www.oneusefulthing.org www.oneusefulthing.org
-
Ethan Mollick prompts Claude AI to come up with something that people will pay for and could make $1k/month
(via [[Stephen Downes p]])
- [ ] return
-
-
cursor.com cursor.com
-
Cursor is an AI using code editor. It connects only to US based models (OpenAI, Anthropic, Google, xAI), and your pricing tier goes piecemeal to whatever model you're using.
Both an editor, and a CLI environment, and integrations with things like Slack and Github. This seems a building block for US-centered agentic AI silo forming for dev teams.
Tags
Annotators
URL
-
-
simonwillison.net simonwillison.net
-
I have yet to try a local model that handles Bash tool calls reliably enough for me to trust that model to operate a coding agent on my device.
this. Need to understand better conceptually the diff set-ups I have, and how I might switch between them.
-
My excitement for local LLMs was very much rekindled. The problem is that the big cloud models got better too—including those open weight models that, while freely available, were far too large (100B+) to run on my laptop.
Cloud models got much better stil than local models. Coding agents made a huge difference, with it Claude Code becomes very useful
-
The year of programming on my phone # I wrote significantly more code on my phone this year than I did on my computer.
vibe coding leads to a shift in using your phone to code. (not likely me, I hardly try to do anything productive on the limited interface my phone provides, but if you've already made the switch to speaking instructions I can see how this shift comes about)
-
The year of vibe coding # In a tweet in February Andrej Karpathy coined the term “vibe coding”, with an unfortunately long definition (I miss the 140 character days) that many people failed to read all the way to the end:
ah, didn't know. Vibe-coding is a term coined by Andrej Karpathy in #2025/02 in a tweet. That took on an own life!
-
There’s a new kind of coding I call “vibe coding”, where you fully give in to the vibes, embrace exponentials, and forget that the code even exists. It’s possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like “decrease the padding on the sidebar by half” because I’m too lazy to find it. I “Accept All” always, I don’t read the diffs anymore. When I get error messages I just copy paste them in with no comment, usually that fixes it. The code grows beyond my usual comprehension, I’d have to really read through it for a while. Sometimes the LLMs can’t fix a bug so I just work around it or ask for random changes until it goes away. It’s not too bad for throwaway weekend projects, but still quite amusing. I’m building a project or webapp, but it’s not really coding—I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works.
vibecoding original description by Andrej Karpathy
Quickly distorted to mean any code created w llm assistance. Note: [[Martijn Aslander p]] follows this dev quite closely (dictation, accept always, it mostly works)
-
The year I built 110 tools # I started my tools.simonwillison.net site last year as a single location for my growing collection of vibe-coded / AI-assisted HTML+JavaScript tools. I wrote several longer pieces about this throughout the year: Here’s how I use LLMs to help me write code Adding AI-generated descriptions to my tools collection Building a tool to copy-paste share terminal sessions using Claude Code for web Useful patterns for building HTML tools—my favourite post of the bunch. The new browse all by month page shows I built 110 of these in 2025!
Simon Willison vibe coded over 100 personal tools in 2025. This chimes with what Frank and Martijn were suggesting. Up above he also indicates that it is something that became possible at this scale only in 2025 too.
-
I love the asynchronous coding agent category. They’re a great answer to the security challenges of running arbitrary code execution on a personal laptop and it’s really fun being able to fire off multiple tasks at once—often from my phone—and get decent results a few minutes later.
async coding agents: prompt and forget
-
Vendor-independent options include GitHub Copilot CLI, Amp, OpenCode, OpenHands CLI, and Pi. IDEs such as Zed, VS Code and Cursor invested a lot of effort in coding agent integration as well.
non-vendor related coding agents. - [ ] which of these can I run locally? / integrate into VS Code
-
The major labs all put out their own CLI coding agents in 2025 Claude Code Codex CLI Gemini CLI Qwen Code Mistral Vibe
list of command line coding agents by major vendors
-
-
tools.simonwillison.net tools.simonwillison.net
-
personal tools built with vibecoding by Simon Willison Resulting tools are mostly HTML and javascript, some python.
- [ ] return
Tags
Annotators
URL
-
- Dec 2025
-
world.hey.com world.hey.com
-
You need three things. A Mac with Xcode, which is free to download. A $99 per year Apple Developer account. And an AI tool that can write code based on your descriptions.
Three elements for making his iphone apps Xcode (which I use) Apple Developer account (99USD / yr) AI support in coding (he uses Claude Code, vgl [[Mijn vibe coding set-up 20251220143401]]
-
-
www.humanlayer.dev www.humanlayer.dev
-
Writing a good CLAUDE.md
- CLAUDE.md is a special onboarding file to familiarize Claude (an AI code assistant) with your codebase.
- It should clearly outline the WHY (purpose of the project), WHAT (tech stack, project structure, key components), and HOW (development process, running tests, build commands) for Claude.
- The file helps Claude understand your monorepo or multi-application project and know where to look for things without flooding it with unnecessary details.
- Keep CLAUDE.md concise and focused; ideally, it should be under 300 lines, with many recommending less than 60 lines for clarity and relevance.
- Use progressive disclosure: point Claude to where to find further information rather than including all details upfront, avoiding overwhelming the model’s context window.
- Complement CLAUDE.md with tools like linters, code formatters, hooks, and slash commands to separate concerns like implementation and formatting.
- CLAUDE.md is a powerful leverage point for getting better coding assistance but must be carefully crafted—not auto-generated.
- The file should include core commands, environment setup, guidelines, and unexpected behaviors relevant to the repository.
- Encouraging Claude to selectively read or confirm files before use can help maintain focus during sessions.
Hacker News Discussion
- Users emphasized the benefit of explicit instruction patterns like "This is what I'm doing, this is what I expect," which improves monitoring and recovery from errors.
- Some commenters felt these markdown files had marginal gains and that model quality mattered more than the presence of CLAUDE.md.
- A few highlighted the importance of writing documentation primarily for humans rather than solely for LLMs.
- Discussion included anticipation of more stateful LLMs with better memory, which would impact how such onboarding files evolve.
- Recommendations included hierarchical or recursive context structures in CLAUDE.md for large projects, allowing a root file plus targeted sub-files.
- Comments supported having Claude address the user specifically to verify it is following instructions properly.
- Some users noted improvements in model adherence compared to past versions, making CLAUDE.md files more effective now.
- Practical tips were shared for managing large monorepos and integrating CLAUDE.md with version control status.
Tags
Annotators
URL
-