Ask HN: How much better are AI IDEs vs. copy pasting into chat apps?

I just wanted to hear peoples experiences with AI IDEs.

For context, I’m a heavy user of Gemini / ChatGPT for coding and Copilot. But I haven’t used Cursor / Windsurf / etc..

Copy pasting into chat apps is a first world problem: it will do the work for you, but you have to give it all the context in the prompt, which for a larger project, gets tedious.

The issue with Copilot is that it’s not as smart as the “thinking” chat apps.

This makes it clear why there’s such a need for AI IDEs. I don’t want to construct my context to a chat app. The context is already in my codebase, so the AI should pick up on it. But I also hear that it gets expensive because of the pay-per-use pricing, as opposed to effectively unlimited prompts for a thinking chat app if you pay the monthly subscription.

So I just wanted to get the lay of the land. How good are these IDEs on constructing your context to the LLMs? How much more expensive is it, and is it worth it for you? — Read More

#devops

Working with LLMs: A Few Lessons

An interesting part of working with LLMs is that you get to see a lot of people trying to work with them, inside companies both small and large, and fall prey to entirely new sets of problems. Turns out using them well isn’t just a matter of knowhow or even interest, but requires unlearning some tough lessons. So I figured I’d jot down a few observations. Here we go, starting with the hardest one, which is:

Perfect verifiability doesn’t exist

LLMs inherently are probabilistic. No matter how much you might want it, there is no perfect verifiability of what it produces. Instead what’s needed is to find ways to deal with the fact that occasionally it will get things wrong. — Read More

#devops

Why Developers Should Care About Generative AI (Even They Aren’t AI Expert)

Software development is about to undergo a generative change. What this means is that AI (Artificial Intelligence) has the potential to make developers more productive, as three systems on the market already provide this: GitHub Copilot, Anthropic’s Claude and OpenAI’s ChatGPT.

Hence, every developer, no matter if he or she specializes in AI or not, needs to understand and realize that as this technology is advancing so rapidly, any of us needs to know what it is, why it is relevant, and how to use it. — Read More

#devops

What «Shifting Left» Means and Why it Matters for Data Stacks

Moving Data Quality and Business Logic Upstream for More Efficient Data Systems

Shifting left is an interesting concept that’s gaining momentum in modern data engineering. SDF has been among those sharing this approach, even making “shifting left” one of their main slogans. As Elias DeFaria, SDF’s co-founder, describes it, shifting left means “improving data quality by moving closer toward the data source”.

However, the benefits extend beyond just data quality improvements. With dbt Labs’ recent acquisition of SDF, many are wondering: what does this mean for the shifting left movement, and more importantly, what exactly is shifting left in the data context?

In this article, we’ll explore the core principles behind shifting left, examine how code-first approaches have made moving logic upstream more efficient, and answer the questions: Why should data teams shift left? What elements need to be shifted? And how can your organization implement this approach to build more maintainable, efficient data systems? — Read More

#devops

Model Context Protocol (MCP)

MCP is an open protocol that standardizes how applications provide context to LLMs. Think of MCP like a USB-C port for AI applications. Just as USB-C provides a standardized way to connect your devices to various peripherals and accessories, MCP provides a standardized way to connect AI models to different data sources and tools. — Read More

#devops

How to Build an Agent

It’s not that hard to build a fully functioning, code-editing agent.

It seems like it would be. When you look at an agent editing files, running commands, wriggling itself out of errors, retrying different strategies – it seems like there has to be a secret behind it.

There isn’t. It’s an LLM, a loop, and enough tokens. It’s what we’ve been saying on the podcast from the start. The rest, the stuff that makes Amp so addictive and impressive? Elbow grease.

But building a small and yet highly impressive agent doesn’t even require that. You can do it in less than 400 lines of code, most of which is boilerplate.

I’m going to show you how, right now. We’re going to write some code together and go from zero lines of code to “oh wow, this is… a game changer.” — Read More

#devops

DeepCoder: A Fully Open-Source 14B Coder at O3-mini Level

Through a joint collaboration between the Agentica team and Together AI, we release DeepCoder-14B-Preview, a code reasoning model finetuned from Deepseek-R1-Distilled-Qwen-14B via distributed RL. It achieves an impressive 60.6% Pass@1 accuracy on LiveCodeBench (+8% improvement), matching the performance of o3-mini-2025-01-031 (Low) and o1-2024-12-17 with just 14B parameters. We’ve open-sourced our dataset, code, training logs, and systems optimizations for everyone to progress on scaling and accelerating intelligence with RL. — Read More

#devops

The day I taught AI to think like a Senior Developer

Is it just me, or are the code generation AIs we’re all using fundamentally broken?

For months, I’ve watched developers praise AI coding tools while silently cleaning up their messes, afraid to admit how much babysitting they actually need.

I realized that AI IDEs don’t actually understand codebases — they’re just sophisticated autocomplete tools with good marketing. The emperor has no clothes, and I’m tired of pretending otherwise.

After two years of frustration watching my AI assistants constantly “forget” where files were located, create duplicates, and use completely incorrect patterns, I finally built what the big AI companies couldn’t (or wouldn’t.)

I decided to find out: What if I could make AI actually understand how my codebase works? — Read More

#devops

Vibe Coding: Pairing vs. Delegation

In The Vibe Coding Handbook: How To Engineer Production-Grade Software With GenAI, Chat, Agents, and Beyond, Steve Yegge and I describe a spectrum of coding modalities with GenAI. On one extreme is “pairing,” where you are working with the AI to achieve a goal. It really is like pair programming with another person, if that person was like a “summer intern who believes in conspiracy theories” (as coined by Simon Willison) and the world’s best software architect.

On the other extreme is “delegating” (which I think many will associate with “agentic coding”), where you ask the AI to do something, and it does so without any human interaction.

… These dimensions dictate the frequency of reporting and feedback you need.  — Read More

#devops

Code is the new no-code

Most people can’t code. So if you’re running a business, for years you’ve had only two options when you wanted to improve your productivity with the tools and systems you used.

1. Buy better software
2. Pay someone to build better software

For years, we’ve been promised a future where anyone could build software without learning to code, giving us a third option. A promised third option was that you could just drag-and-drop some blocks, connect a few nodes, and voilà — you’ve built a fully functional app without writing a single line of code! — Read More

    #devops