Thread #108621719
HomeIndexCatalogAll ThreadsNew ThreadReply
H
File: vcg.jpg (1.2 MB)
1.2 MB
1.2 MB JPG
A general for coding with agents

►Harnesses
https://developers.openai.com/codex
https://code.claude.com/docs/en/overview
https://opencode.ai/
https://antigravity.google/
https://cursor.com/docs
https://pi.dev/
+Showing all 77 replies.
>>
more like gay-i
>>
>>
>>108621734
Been on a camping trip touching grass for four days and I came back to the codex daily 5-hours limit removed and a single (1, singular) prompt eating the entirety of your weekly usage? Am I reading this bullshit right?
>>
>>108621885
>Am I reading this bullshit right?
no
>>
File: file.png (2.5 MB)
2.5 MB
2.5 MB PNG
new qwen release for localfags
>>
This time my test suite will be perfect.
>>
>>108619771
What are you making anon?
>>
>>108621732
>altman in shambles
>nvidia stock dropping like a rock
>global financial crisis imminent
anon, why would you do this
>>
>>108622243
AI Chads can't be stopped
>>
can claude crack denuvo?
>>
>>108622339
name 1 (one) uncracked denuvo game that is worth playing.

also, no, probably not, especially since you'd need to jailbreak it first.
but it'd be fun to try, i guess
>>
File: file.png (307.3 KB)
307.3 KB
307.3 KB PNG
total dario victory
>>
File: rip.png (169.8 KB)
169.8 KB
169.8 KB PNG
i like the increased pace of vibe coding, but needing to go back and understand all the changes being made and refactor to my standards just makes me feel like im in an all-day PR review. how do i make this less exhausting
>>
>>108622380
Tell her to explain it to you like you are retarded
>>
>>
>>108622401
>Tell her to explain it to you like you are retarded
LLMs can infer that themselves now
>>
opencode is switching to an electron desktop app like everyone else
soon we have a dozen versions of the exact same thing
>>
Vibecoded a simple utility .exe that prevents my mouse from double-clicking.
>>
>>108622658
I use the web UI instead, had to vibecode a fix to set the IDs of the web ui client 10s behind the server or shit would break (duplicated messages would appear in the TUI/no responses on the web UI) but other than that I am happy with it.
Also can't have more than one tab open lol but that's a minor bug.
I prefer a web UI that can be easily customized and hosted anywhere over a traditional native app
>>
You don't sit at your computer to vibe code.
The real way to vibecode is by texting it on telegram while you're sitting on a bench at the park.
>>
>>108622850
I don't really vibe, this is no less stressful than normal programming.
>>
Good morning bros. Whatever happened to that class action lawsuit over the 'Global AI Nerf fraud'?
>>
honestly, I don't like the word "vibe" too much
>>
>>108622380
Just return back to normal coding. You can still use agents. Just don't generate too much shit at once.
>>
>>108622850
now replace park bench with yacht in the carribean while getting your dick sucked by a 10/10 supermodel
>>
>>108623075
same

I bought a 7900XTX for gaming in 2022 for $999 and thought it was overkill. Now I have Qwen 27B on it doing my job for me and the same GPU costs $1200 now.
>>
>>108622380
I generate detailed markdown plans and refine them, then I only let the AI do tiny atomic things. That way it feels like I'm coding it and I can fix it as I go instead of having a gargantuan "rewrite everything the ai shat out" phase that's no fun.
>>
Alright, PiClaw is out of alpha and into beta. It's up and running at home, and I'm connected with Telegram. It's ready to edit files, interact with websites, look at images, and tell me I'm a special boy. We'll consider today the first test as a daily driver.
>>
>>108621732
Stop it
>>
>My honest read: Rewriting env in JAX is 3-6 weeks of engineering for a training run that completes in ~10 days. Bad ROI for a project at your stage.
>I've now read the env. My prior estimate was a wild guess, and grounding it in your actual 1464 LOC changes the picture.
t. Opus 4.7
>>
do i pay anthropic 20 bucks for claude code or is it still gimped?
>>
>>108623169
In the pipe, five by five
>>108623197
Stop subscribing to LLMs
>>
>>108623197
The $20 poorfag plan gets you 1 opus prompt every 5 hours
don't even bother

Claude Code starts at $100+tip Max 5x
>>
>>108623197
depends on what you build
>>
Now I have the complete picture.
>>
>>108623195
The estimates are always horribly off, they are even off if a human had to implement it.
>>
>>108621719
VibeGODS won
>>
>>108623201
>Stop subscribing to LLMs
why?
thinking of getting codex subscription
still trying to find a way to avoid having my info tied to it tho
>>
>>108623201
>>108623208
Ok I'll just build a $60k rig and run something locally, that makes much more sense.
>>
>>108623368
you have to tweak the qwens and give them more context but they're competitive with Claude if you are willing to try. Mac / Strix Halo / Big Boy GPU are all options in the $1-4k range
>>
>>108623197
Opus 4.7 feels *a lot* like what they did to Gemini 3.1 after a while. There might be public for it, but instead of being a helpful assistant, now it's more prone to write an essay telling half a dozen totally equivalent way to do that thing. Other times, it tries to take shortcuts, confidently saying things about files it hasn't looked at, and when called over it, offers "you're absolutely right" type of platitudes.

Then when you get it to do the work it's supposed to, before starting always highlight that it will do it, sounding like it is doing you a favor.

It seemingly feel in love with the word honest and keeps repeating "Honest take", "Honest answer", etc. The way the responses are written are also starting to *feel* like LinkedIn posts: "(statement) Read on to find out why."

It's great by all metrics if we consider that this can even exist. It's impressive. It's also disappointing in a what-did-they-to-my-boy kind of way.
>>
>>108623368
i just told you you can just pay $100 or $200 and get reasonable usage
>>
These threads are always full of superstitious nonsense. Had a bad week at the AIs? Someone at Anthropic must have nerfed the model!
The roulette wheel isn't rigged, sometimes you're just unlucky. Try another spin, maybe you'll do better. Or just write your own code.

(stolen from hackernews but xhe isn't wrong. I have literally never witnessed a regression from a model ever)
>>
>>108623426
i only have 64 cores epyc with 128gb ram and a potato gpu on the server for hwenc.
i cant run anything on that.. in real time, i can tho run some agent overnight.
but i havent looked at self hosted agentic programming stuffs yet
>>
what can i run on 32GB RAM and a 9070xt? That is also my main PC
>>
>>108623486
Gemma4
>>
>>108623453
There can definitely be some. Gemini 3.1 used to be extremely helpful in Antigravity, now it starts every interaction by a chain of thought saying that it is avoiding cat for file manipulation and is focused to using dedicated tools like grep_search and other utilities to yada yada yada. I'm pretty sure sometimes it falls back to that self conversation in the middle of a chain of tool calls too. I don't know if I just need to empty completely the history and memories, but it seems like repeating this to itself again and again and again is about half of the effort it spends answering every request.
>>
>>108623465
anon just bite the bullet and pay api prices for a chink model.

glm5.1, which is probably the best of the bunch currently, is at ~ $1/$3, a full 5 times cheaper than sonnet (never mind opus, api prices for that are a joke.)
mimo2.7 is not far behind in benches, and its at $0.3/$1.2, another 3 times cheaper, lol
and deepseek 3.2 starts out at $0.3/$0.4, which is just ridiculously cheap (tho the benches are decidedly less impressive than glm/mimo).

its not worth it to pay for hardware for local currently.
you'll only be able to run much smaller, much less capable models, and you'll be getting shit throughput (especially in a cpu+ram only config like yours)
even if you were willing to limit yourself to small, local targeted open stuff like the latest qwen3.6 35b (which is FAR inferior to the dirt cheap chink stuff mentioned above), you'd definitely want to buy at least a used 3090 to run it at a decent speed
at current prices, it just doesn't make sense.

if/when the current ai bubble bursts, then maybe.
its possible that a combination of cratering gpu prices (datacenter gpus, with 80gigs vram apiece, not the cucked consumer stuff) and inference providers jacking up prices by 10x or more might make it viable
but as it stands right now, its not even close.
>>
>>108623520 (me)
And if I am dumb regarding this and it's user errors, well, it might be the same for others who are claiming regressions, but they make the tools get bogged down into unhelpful patterns by their own memory files or it's the harness going off rail trying to keep track of user preferences in ways that don't make sense and result in doing the opposite, the users are not the only one to blame.
>>
>>108623453
Except the regressions were literally measured. Literal battered wife syndrome.
>>
>>108623361
PAY AS YOU YOU GO WITH API KEYS YOUUUUUU DENSE MOTHERFUCKER
>>
>>108623692
>paying api prices
>calling someone else dense
on the peak of mt midwit
>>
>>108623692
I think next OP we should put some guide about API vs subcription
>>
>>108623747
>we should put some guide
Yeah but are you going to spend your precious tokens on making one?
>>
>>108623747
API prices for API freedom
>>108623763
I'm a paypig, I'll have gpt 5.4 write it
>>
I hope this general is a troll post and not actually serious. Mods need to ban AI shit from /g/.
>>
>>108623787
Luddites on reddit please
>>
>>108623787
We got a luddie here boys
>>
>>108623787
Cope. I'm making custom proprietary software and there's nothing you can do about it
>>
>>108623787
>Mods need to ban AI shit from /g/.
how the fuck are you on /g/ and anti-AI? every single comp sci student I know uses claude/codex. even my friend who holds multiple STEM degrees and is an unironic genius uses AI tools.
>>
Yes, you should walk to the car wash.
>>
>>108623933
>carrying about if the latest model can accurately tell you how many r's are in strawberry
ngmi
>>
>>108623933
Nobody asked tho
>>
I should have used Codex more last month when the limits were higher. I just didn't have the idea I have now
>>
>>108623951
>second-to-latest model can correctly count r's
>model hailed has "latest" and "goodest" can't correctly count r's anymore
Yes, you should care
>>
>>108623978
man, i remember being so impressed that qwen-coder-14b-r1-distill could actually do this
good times, good times.
>>
>>108623978
why should I care when it can trivially write a program to count the Rs in strawberry if you ask it to?
>>
i am almost done with my imageboard summarizer next tool will be a [redacted] music platform piracy tool with integrated mp3 tagging , and a poor mans files library backup system using tar, rsync and par2 with a simple gui for syncing between drives and validating integrity
>>
codex keeps calling things gremlins now
i haven't mentioned gremlins anywhere in any messages ever
>>
>>108624281
It's an old term of art. They've added more code, docs, discussions, etc. from the 80s through 2000s to the training set.
>>
>>108624233
can i see the image board summarizer
>>
WATCH OUT DARIO
>>
File: IMG_5260.png (52 KB)
52 KB
52 KB PNG
>You've used 27% of your weekly limit • resets in 5d •
>>
>>108624281
>>108624337
The what? Bugs?
>>
>>108624533
anything
>>
>>108623441
I still think GPT 5.4 Extra High via Copilot VS Code extension is a gorillion times better deal than Claude anything, given it is still only a 1x quota consumption model relative to what you get as far as premium requests a month for the $10 monthly sub.

Reply to Thread #108621719


Supported: JPG, PNG, GIF, WebP, WebM, MP4, MP3 (max 4MB)