Vibecoding #2
Posted by ibobev 3 days ago
Comments
Comment by jeingham 3 days ago
Bottom line is that I am extremely grateful for AI has a teammate. As a solopreneur even more so. I'm building an application that I know would have taken at least $10 to 20K to build but all I'm paying is $60 a month Cursor Pro+ and my public facing server. And only $60 because I ran into a Cursor Claude limit.
Buckle up guys and gals, the midwit you always feared has the keys to the tank now...
Comment by gtowey 3 days ago
Comment by ako 3 days ago
Comment by altmanaltman 3 days ago
Its always worth learning something if you enjoy it, the same applies to code and languages. You can definitely create better apps knowing the details of a specific language than not knowing it and I think its still worth doing if you care about the ultimate quality of your work.
Comment by thunky 2 days ago
This argument is repeated often but what I think you're missing is that if you want to listen to music you put on the radio, you don't record an album.
Sure if I want to enjoy playing guitar I'll do that, but that's not what I'm paid to do and you're not paid write code. Nobody but me wants to hear me play guitar and nobody but you wants to look at your beautiful code.
Comment by ako 3 days ago
When working with an LLM i care more about prompting it about software architecture, software UX, and the domain we're working on, than the details of the language it uses.
Comment by altmanaltman 2 days ago
hard disagree on both points. You're talking about "impact" but surely you'll be a better coder if you can actually, you know, code? The other stuff is important sure but if you literally cannot read the code and just pleasure yourself with dreams of architecture and UX, what you're generating is 99% bad quality.
But prove me wrong, would love to see something you've made.
Comment by ako 2 days ago
Comment by altmanaltman 12 hours ago
Comment by ElijahLynn 3 days ago
I'm confident I can do anything with enough time. But I only have so much.
AI is going to enable so many more ideas to come to fruition and a better world because of it!
Comment by sparky4pro 3 days ago
If someone is in their 30’ or 40’ planning to work the next 5+ years on a project is no problem, even if it takes 10+ years in the end.
For the ones over 65 or older, it’s a different story…
Comment by jeingham 3 days ago
Comment by jeingham 3 days ago
Comment by dcre 3 days ago
If you want to script in Rust, xshell (https://docs.rs/xshell/latest/xshell/) is explicitly inspired by dax.
Comment by suralind 3 days ago
Comment by liampulles 3 days ago
Comment by dudewhocodes 3 days ago
Are they really late? Has everyone started using agents and paying $200 subscriptions?
Am I the one wrong here or these expressions of "falling behind" are creating weird FOMO in the industry?
EDIT: I see the usefulness of these tools, however I can't estimate how many people use them.
Comment by NothingAboutAny 3 days ago
If anything in my small circle the promise is waning a bit, in that even the best models on the planet are still kinda shitty for big project work. I work as a game dev and have found agents to only be mildly useful to do more of what I've already laid out, I only pay for the $100 annual plan with jetbrains and that's plenty. I haven't worked at a big business in a while, but my ex-coworkers are basically the same. a friend only uses chat now because the agents were "entirely useless" for what he was doing.
I'm sure someone is getting use out of them making the 10 billionth node.js express API, but not anyone I know.
Comment by bunderbunder 3 days ago
That would be fine if our value delivery rate were also higher. But it isn’t. It seems to actually be getting worse, because projects are more likely to get caught in development hell. I believe the main problem there is poorer collective understanding of generated code, combined with apparent ease of vibecoding a replacement, leads to teams being more likely to choose major rewrites over surgical fixes.
For my part, this “Duke Nukem Forever as a Service” factor feels the most intractable. Because it’s not a technology problem, it’s a human psychology problem.
Comment by fulladder 3 days ago
Don't get me wrong, overall I really like having AI in my workflow and have gotten many benefits. But even when I ask it to check its own work by writing test cases to prove that properties A, B and C hold, I just end up with thousands more lines of unit and integration tests that then take even more time to analyze -- like, what exactly is being tested here?, are the properties these tests purport to prove even the properties that I care about and asked the agent for in the first place, etc.
I have tried (with at least modest success) to use a second or third agent to review the work of the original coding agent(s), but my general finding has been that there is no substitute for actual human understanding from a legitimate domain expert.
Part of my work involves silicon design, which requires a lot of precision and complex timing issues, and I'll add that the best AI success I've had in those cases is a test-first approach (TDD), where I hand write a boatload of testbenches (that's what we call functional tests in chip design land), then coach my various agents to write the Verilog until my `make test` runs with no errors.
Comment by agumonkey 3 days ago
Comment by rootnod3 3 days ago
Especially considering that these 200$ subscriptions are just the start because those companies are still mostly operating at a loss.
It's either going to be higher fees or Ads pushed into the responses. Last I need is my code sprinkled with Ads.
Comment by RobinL 3 days ago
At the very least, it can quickly build throwaway productivity enhancing tools.
Some examples from building a small education game: - I needed to record sound clips for a game. I vibe coded a webapp in <15 mins that had a record button, keyboard shortcuts to progress though the list of clips i needed, and outputted all the audio for over 100 separate files in the folder structure and with the file names i needed, and wrote the ffmpeg script to post process the files
- I needed json files for the path of each letter. gemini 3 converted images to json and then codex built me an interactive editor to tidy up the bits gemini go wrong by hand
The quality of the code didn't matter because all i needed was the outputs.
The final games can be found: https://www.robinlinacre.com/letter_constellations https://www.robinlinacre.com/bee_letters/ code: https://github.com/robinL/
Comment by brokencode 3 days ago
How long did it take to learn how to use your first IDE effectively? Or git? Or basically any other tool that is the bedrock of software engineering.
AI fools people into thinking it should be really easy to get good results because the interface is so natural. And it can be for simple tasks. But for more complex tasks, you need to learn how to use it well.
Comment by kemotep 3 days ago
They “type” faster than me, but they do not type out correct PowerShell.
Fake modules, out of date module versions, fake options, fake expectations of object properties. Debugging what they output makes them a significant speed down compared to just, typing and looking up PowerShell commands manually and using the -help and get-help functions in my terminal.
But again, I haven’t forked over money for the versions that cost hundreds of dollars a month. It doesn’t seem worth it, even after 3 years. Unless the paid version is 10 times smarter with significantly less hallucinations the quality doesn’t seem worth the price.
Comment by azuanrb 3 days ago
Comment by Aurornis 3 days ago
No, the $20/month plans are great for minimal use
> Because every time, without fail, the free ChatGPT, Copilot, Gemini, Mistral, Deepseek whatever chatbots, do not write PowerShell faster than I do.
The exact model matters a lot. It's critical to use the best model available to avoid wasting time.
The free plans generally don't give you the best model available. If they do, they have limited thinking tokens.
ChatGPT won't give you the Codex (programming) model. You have to be in the $20/month plan or a paid trial. I recommend setting it to "High" thinking.
Anthropic won't give you Opus for free, and so on.
You really have to use one of the paid plans or a trial if you want to see the same thing that others are seeing.
Comment by brokencode 3 days ago
Tools like GitHub copilot can access the CLI. It can look up commands for you. Whatever you do in the terminal, it can do.
You can encode common instructions and info in AGENTS.md to say how and where to look up this info. You can describe what tools you expect it to use.
There are MCPs to help hook up other sources of context and info the model can use as well.
These are the things you need to learn to make effective use of the technology. It’s not as easy as going to ChatGPT and asking a question. It just isn’t.
Too many people never get past this low level of knowledge, then blame the tool.
Comment by kemotep 3 days ago
Comment by brokencode 3 days ago
All I’m saying is that the vast majority of people who say that AI dev tools don’t work and are a waste of time/money don’t know how and really haven’t even made a serious attempt at learning how to use them.
Comment by 3vidence 3 days ago
So when people just Yolo the ladder they don't get the results they expect.
I'm personally in the middle, chat interface + scripts seems to be the best for my productivity. Agentic stuff feels like a rabbit hole to me.
Comment by kemotep 3 days ago
Comment by brokencode 3 days ago
Comment by rootnod3 17 hours ago
Comment by andai 3 days ago
But if you use them for more than a few minutes, the tokens start adding up, and the subscriptions are heavily discounted relative to the tokens used.
There are also API-neutral tools like Charm Crush which can be used with any AI provider with API keys, and work reasonably well (for simple tasks at least. If you're doing something bigger you will probably want to use Claude Code).
Although each AI appears to be "tailored" to the company's own coding tools, so you'll probably get better results "holding it right".
That being said, the $3/month Z.ai sub also works great in Claude Code, in my experience. It's a bit slower and dumber than actual Claude, so I just went for the real thing in the end. 60 cents a day is not so bad! That's like, 1/3 of my canned ice coffee... the greater cost is the mental atrophy I am now undergoing ;)
Comment by johnfn 3 days ago
I haven't had an issue with a hallucination in many months. They are typically a solved problem if you can use some sort of linter / static analysis tool. You tell the agent to run your tool(s) and fix all the errors. I am not familiar with PowerShell at all, but a quick GPT tells me that there is PSScriptAnalyzer, which might be good for this.
That being said, it is possible that PowerShell is too far off the beaten path and LLMs aren't good at it. Try it again with something like TypeScript - you might change your mind.
Comment by 8note 3 days ago
whatever you learn now is going to be invalid and wasteful in 6 months
Comment by brokencode 3 days ago
And I reject that anything you learn today will be invalid. It’ll be a base of knowledge that will help you understand and adopt new tools.
Comment by drw85 3 days ago
Comment by kibwen 3 days ago
Comment by mixermachine 3 days ago
I had ChatGPT Codex GPT5.2 high reasoning running on my side project for multiple hours the last nights. It created a server deployment for QA and PROD + client builds. It waited for the builds to complete, got the logs from Github Actions and fixed problems. Only after 4 days of this (around 2-4 hours) active coding I reached the weekly limit for the ChatGPT Plus Plan (23€). Far better value so far.
To be fully honest, it fucked up one flyway script. I have to fix this now my self :D. Will write a note in the Agent.md to never alter existing scripts. But the work otherwise was quite solid and now my server is properly deployed. If I would switch between High reasoning for Planing and Middle reasoning for coding, I would get even more usage.
Comment by moron4hire 3 days ago
"... brought to you by Costco."
But seriously, I can't help but think that this proliferation of massive numbers of iterations on these models and productizations of the models is an indication that their owners have no idea what they are doing with any of it. They're making variations and throwing them against the wall to see what sticks.
Comment by Aurornis 3 days ago
Codex = The model trained specifically for programming tasks. You want this if you're writing code.
GPT5.2 = The current version. You don't have to think about this, you just use the latest.
High Reasoning = A setting you select for balancing between longer thinking time or quicker answers. It's usually set and forget.
Comment by CurleighBraces 3 days ago
I think for me it's a case of fear of being left behind rather than missing out.
I've been a developer for over 20 years, and the last six months has blown me away with how different everything feels.
This isn't like JQuery hitting the scene, PHP going OO or one of the many "this is a game changer" experiences if I've had in my career before.
This is something else entirely.
Comment by rootnod3 3 days ago
Comment by vidarh 3 days ago
There are absolutely maintainability challenges. You can't just tell these tools to build X and expect to get away with not reviewing the output and/or telling it to revise it.
But if you loosen the reigns and review finished output rather than sit there and metaphorically look over its shoulder for every edit, the time it takes me to get it to revise its work until the quality is what I'd expect of myself is still a tiny fraction of what it'd take me to do things manually.
The time estimate above includes my manual time spent on reviews and fixes. I expect that time savings to increase, as about half of the time I spend on this project now is time spent improving guardrails and adding agents etc. to refine the work automatically before I even glance at the output.
The biggest lesson for me is that when people are not getting good results, most of the time it seems to me it is when people keep watching every step their agent takes, instead of putting in place a decent agent loop (create a plan for X; for each item on the plan: run tests until it works, review your code and fix any identified issues, repeat until the tests and review pass without any issues) and letting the agent work until it stops before you waste time reviewing the result.
Only when the agent repeatedly fails to do an assigned task adequately do I "slow it down" and have it do things step by step to figure out where it gets stuck / goes wrong. At which point I tell it to revise the agents accordingly, and then have it try again.
It's not cost effective to have expensive humans babysit cheap LLMs, yet a lot of people seem to want to babysit the LLMs.
Comment by CurleighBraces 3 days ago
I basically have two modes
1. "Snipe mode"
I need to solve problem X, here I fire up my IDE, start codex up and begin prompting to find the bug fix. Most of the time I have enough domain context about the code that once it's found and fixed the issue it's trivial for to reconcile that it's good code and I am shipping it. I can be sniping several targets at anyone time.
Most of my day-to-day work is in snipe mode.
2. "Feature mode"
This is where I get agents to build features/apps, I've not used this mode in anger for anything other than toy/side projects and I would not be happy about the long term prospects of maintaining anything I've produced.
It's stupidly stupidly fun/addictive and yes satisfying! :)
I rebuilt a game that I used to play when I was 11 and still had a small community of people actively wanting to play it, entirely by vibe coding, it works, it's live and honestly I've had some of the most rewarding feedback from making that I've had in my career from complete strangers!
I've also built numerous tools for myself and my kids that I'd never of had time to build before, and I now can. Again the level of reward for building apps etc that my kids ( and their friends ) are using, is very different from anything I've been career wise.
Comment by jannyfer 3 days ago
Comment by CurleighBraces 3 days ago
It doesn't work on mobile, and unless you played it back in the day the feedback from my friends who I've introduced it too, is that it's got quite the learning curve.
You can see all the horrible vibe coding here ( it's slop, it's utter utter slop, but it's working slop )
https://github.com/battlecity-remastered/battlecity-remaster...
Comment by CurleighBraces 3 days ago
Comment by esafak 3 days ago
Comment by CurleighBraces 3 days ago
I think ultimately I've succumbed to the fact that writing code is no longer a primary aspect of my job.
Reading/reviewing and being accountable for code that something else is written very much is.
Comment by AstroBen 3 days ago
I'm also fairly confident having it write my code is not a productivity boost, at least for production work I'd like to maintain long term
Comment by jjice 3 days ago
No, most programmers I know outside of my own work (friends, family, and old college pals) don't use AI at all. They just don't care.
I personally use Cursor at work and enjoy it quite a bit, but I think the author is maybe at the tail end of _their circle's_ adoption, but not the industry's.
Comment by rdiddly 3 days ago
Comment by sodapopcan 3 days ago
Comment by PlatoIsADisease 3 days ago
The $20/mo I pay is quite affordable given the ROI.
I could see jumping between various free models.
Comment by wongarsu 3 days ago
Comment by rhines 3 days ago
Not because I think either way is better, just because personally I work well with AI in the latter capacity and have been considering subscribing to Claude, but don't know how limiting the usage limits are.
Comment by Aurornis 3 days ago
It also goes very fast if you don't actively manage your context by clearing it frequently for new tasks and keeping key information in a document to reference each session. Claude will eat through context way too fast if you just let it go.
For true vibecoding-style dev where you just prompt the LLM over and over until things are done, I agree that $100 or $200 plans would be necessary though.
Comment by sodapopcan 3 days ago
EDIT: I also wasn't going to say it but it's not about the money for me, I just don't want to support any of these companies. I'm happy waste their resources for my benefit but I don't lean on it too often.
Comment by PlatoIsADisease 3 days ago
Its not even SOTA open source anymore, let alone competitive with GPT/Gemini/Grok.
Comment by sodapopcan 3 days ago
Comment by PlatoIsADisease 3 days ago
I couldnt use GPT3 for coding and deepseek is at GPT3 + COT levels.
Comment by sodapopcan 3 days ago
I'm not going to be sending money every month to billion dollars companies who capitulate to a goon threatening to annex my country. I accept whatever consequences that has on my programming career.
Comment by PlatoIsADisease 3 days ago
Comment by sodapopcan 3 days ago
Comment by zapnuk 2 days ago
A 10$ GitHub Copilot or 20$ ChatGPT/Claude subscription get you a long way.
And if the employer isn't willing to spend this little money to improve their workers productivity they're pretty dumb.
There are valid concerns like privacy and oss licences. But lack of value or gain in productivity isn't one of them.
Comment by quijoteuniv 3 days ago
Comment by sodapopcan 3 days ago
Comment by Insanity 3 days ago
(“Scrape kindle highlights from the kindle webpage, store it in a database, and serve it daily through an email digest”).
No success so far in getting it to do so without a lot of handholding and manually updating the web scraping logic.
It’s become something of a litmus test for me.
So, maybe there is some FOMO but in my experience it’s a lot of snake oil. Also at work, I manage a team of engineers and like 2 out of 12 clearly submit AI generated code. Others stopped using it, or just do a lot more wrangling of the output.
Comment by keeda 3 days ago
If you rephrase the question as "Are most engineers already using AI?" -- because it transcends the specific modality (agents vs chat vs autocomplete) and $200 subscriptions (because so many tools are available for free) -- signs point to "yes."
Adoption seems to be all the way upto 85% - 90% in 2025, but there is a lot of variance in the frequency of use:
https://dora.dev/research/2025/
https://survey.stackoverflow.co/2025/
https://newsletter.pragmaticengineer.com/p/the-pragmatic-eng...
If there is FOMO, I'm not sure it's "weird."
Comment by Aurornis 3 days ago
The $20/month subscriptions go a long way if you're using the LLM as an assistant. Having a developer in the loop to direct, review, and write some of the code is much more token efficient than trying to brute force it by having the LLM try things and rewrite until it looks like what you want.
If you jump to the other end of the spectrum and want to be in the loop as little as possible, the $100/$200 subscriptions start to become necessary.
My primary LLM use case is as a hyper-advanced search. I send the agent off to find specific parts of a big codebase I'm looking for and summarize how it's connected. I can hit the $20/month windowed limits from time to time on big codebases, but usually it's sufficient.
Comment by giancarlostoro 3 days ago
Comment by anonymous908213 3 days ago
It is the very definition of FOMO if there is an entire cult of people telling you that for a year, and yet after a year of hearing about how "everything has changed", there is still not a single example of amazing vibe-coded software capable of replacing any of the real-world software people use on a daily basis. Meanwhile Microsoft is shipping more critical bugs and performance regressions in updates than ever while boasting about 40% of their code being LLM-generated. It is especially strange to cite "Windows as a great example" when 2025 was perhaps one of the worst years I can remember for Windows updates despite, or perhaps because of, LLM adoption.
Comment by giancarlostoro 3 days ago
Comment by drw85 3 days ago
Azure, Office, Visual Studio, VS Code, Windows are all shipping faster than ever, but so much stuff is unfinished, buggy, incompatible to existing things, etc.
Comment by CodeMage 3 days ago
Enshittification is not primarily caused by "we can fix it later", because "we can fix it later" implies that there's something to fix. The changes we've seen in Windows and Google Search and many other products and services are there because that's what makes profit for Microsoft and Google and such, regardless of whether it's good for their users or not.
You won't fix that with AI. Hell, you couldn't even fix Windows with AI. Just because the company is making greedy, user-hostile decisions, it doesn't mean that their software is simple to develop. If you think Windows will somehow get better because of AI, then you're oversimplifying to an astonishing degree.
Comment by giancarlostoro 3 days ago
Comment by CodeMage 2 days ago
My point is that they will continue to do so no matter how easy it is to fix bugs. It's a people problem, not a tech problem.
Comment by andai 3 days ago
In those days already the attitude with regard to professional work was that if you aren't constantly advancing in your industry, you are falling behind.
Comment by kibwen 3 days ago
Comment by nurettin 3 days ago
Comment by ryanSrich 3 days ago
Comment by erdaniels 3 days ago
I just got native LSP working this past weekend and in sublime it's as much as: { "clients": { "remote-gopls": { "command": [ "tool", "lsp", "gopls" ], "enabled": false, "selector": "source.go", }, } }
From what you built so far, do you think there's any appetite in people paying for this type of tool which lets you spin up infra on demand and gives you all the capabilities built so far? I'm skeptical and I may just release it all as OSS when it gets closer to being v1.0.
Comment by lightandlight 3 days ago
(I'm not the author) The easiest way to charge for this kind of software is to make it SaaS, and I think that's pretty gross, especially for a CLI tool.
> I'm skeptical and I may just release it all as OSS
It doesn't have to be one or the other: you could sell the software under libre[1] terms, for example.
Comment by erdaniels 3 days ago
Comment by jasonjmcghee 3 days ago
It's very easy to get hit with a massive bill due to just leaving instances around.
Comment by hendiatris 3 days ago
sudo shutdown +15 (or other amount of minutes)
when I need a compute instance and don’t want to forget to turn it off. It’s a simple trick that will save you in some cases.
Comment by apawloski 3 days ago
/5 * * * [ -f /var/run/keepalive ] && [ $(( $(date +\%s) - $(stat -c \%Y /var/run/keepalive) )) -gt 7200 ] && shutdown -h now
Comment by tln 3 days ago
Comment by harry34 2 days ago
What's interesting is that the bottleneck is shifting. For experienced developers, the constraint was never typing speed or recalling syntax - it was understanding the problem domain, making architectural decisions, and maintaining systems over time. AI tools amplify this: they make the gap between 'can generate code' and 'can build maintainable systems' even wider.
The real question isn't whether to use AI tools, but how they change what's worth learning deeply. If AI can scaffold boilerplate, then understanding why certain patterns exist becomes more valuable, not less. The ability to evaluate AI-generated code, spot subtle bugs, or recognize when it's taking you down a bad architectural path - these skills require deep knowledge.
For solopreneurs and builders, the calculation is different. Getting something working that creates value for users is often better than perfect code that ships too late. The key is being honest about the tradeoffs: move fast with AI, but budget time to understand what you've built before scaling it.
Comment by jeingham 3 days ago
Comment by nromiun 3 days ago
Wait, this is how people vibe code? I thought it was just giving instruction line by line and refining your program. People are really creating a dense, huge spec for their project first?
I have not seen any benefit of AI in programming yet, so maybe I should try it with specs and like a auto-complete as well.
Comment by linsomniac 3 days ago
Lots of people are using PRD files for this. https://www.atlassian.com/agile/product-management/requireme...
I've been using checklists and asking it to check off items as it works.
Another nice feature of using these specs is that you can give the AI tools multiple kicks at the can and see which one you like the most, or have multiple tools work on competing implementations, or have better tools rebuild them a few months down the line.
So I might have a spec that starts off:
#### Project Setup
- [ ] Create new module structure (`client.py`, `config.py`, `output.py`, `errors.py`)
- [ ] Move `ApiClient` class to `client.py`
- [ ] Add PyYAML dependency to `pyproject.toml`
- [ ] Update package metadata in `pyproject.toml`
And then I just iterate with a prompt like: Please continue implementing the software described in the file "dashcli.md". Please implement the software one phase at a time, using the checkboxes to keep track of what has already been implemented. Checkboxes ("[ ]") that are checked ("[X]") are done. When complete, please do a git commit of the changes. Then run the skill "codex-review" to review the changes and address any findings or questions/concerns it raises. When complete, please commit that review changeset. Please make sure to implement tests, and use tests of both the backend and frontend to ensure correctness after making code changes.Comment by weakfish 3 days ago
Comment by candl 3 days ago
Comment by tmtvl 3 days ago
Comment by azuanrb 3 days ago
Comment by NicoJuicy 3 days ago
Ollama with qwen3 and starcoder2 are ok.
I'd recomment to experiment with the following models atm. (eg. with "open-webui"): - gpt-oss:20b ( fast) - nemotron-3-nano:30b ( good general purpose)
It doesn't compare to the large LLM's atm. though.
Comment by css_apologist 3 days ago
Comment by indigodaddy 3 days ago
Comment by yakshaving_jgt 3 days ago
Comment by cadamsdotcom 3 days ago
Author needed a thing, it didn’t exist, so they made that thing.
That’s incredible empowerment.
Comment by foobarqux 3 days ago
Comment by bilekas 3 days ago
> I personally don’t know how to solve it without wasting a day. So, I spent a day vibecoding my own square wheel.
This worries me, in the case of OP it seems was dillegent and reviewed everything thoroughly but I can bet that that's not the majority.. And pushing to prod an integral piece without fully knowing how it works just terrifies me.
Comment by wiredfool 3 days ago
Comment by darkwater 3 days ago
Using those as hint I bet CC would have one-shotted it pretty easily
Comment by fud101 3 days ago
Comment by jacobtomlinson 3 days ago
Comment by lukax 3 days ago
Comment by indigodaddy 3 days ago
Comment by indigodaddy 3 days ago