Goog is even heavily subsidising this. Anthropic is likely doing it with their top tiers as well. Even the small ones @20$ most likely did in the beginning.
In specific, I'm really proud of "spec driven development", which is based on the internal processes that software development teams at Amazon use to build very large technical projects. Kiro can take your basic "vibe coding" prompt, and expand it into deep technical requirements, a design document (with diagrams), and a task list to break down large projects into smaller, more realistic chunks of work.
I've had a ton of fun not just working on Kiro, but also coding with Kiro. I've also published a sample project I built while working on Kiro. It's a fairly extensive codebase for an infinite crafting game, almost 95% AI coded, thanks to the power of Kiro: https://github.com/kirodotdev/spirit-of-kiro
Realistically, I don't think that Harper's statement of "I get to play cookie clicker" is achievable, at least not for nontrivial tasks. Current LLM's still need a skilled human SDE in the loop. But Kiro does help that loop run a lot smoother and on much larger tasks than a traditional AI agent can tackle.
This recent OpenAI presentation might resonate too then:
Prompt Engineering is dead (everything is a spec)
In an era where AI transforms software development, the most valuable skill isn't writing code - it's communicating intent with precision. This talk reveals how specifications, not prompts or code, are becoming the fundamental unit of programming, and why spec-writing is the new superpower.
Drawing from production experience, we demonstrate how rigorous, versioned specifications serve as the source of truth that compiles to documentation, evaluations, model behaviors, and maybe even code.
Just as the US Constitution acts as a versioned spec with judicial review as its grader, AI systems need executable specifications that align both human teams and machine intelligence. We'll look at OpenAI's Model Spec as a real-world example.
Have you considered a fourth file for Implemented such that Spec = Implemented + Design?
It would serve both as a check that nothing is missing from Design, and can also be an index for where to find things in the code, what architecture / patterns exist that should be reused where possible.
And what about coding standards / style guide? Where does that go?
>overage charges for agentic interactions will be $0.04 per interaction, and if enabled, will begin consuming overages once your included amounts are used (1,000 interactions for Pro tier, 3,000 for Pro+ tier). Limits are applied at the user level. For example, if you are a Pro tier customer who uses 1,200 requests, your bill would show an overage charge of $8 (200 × $0.04). Overages for agentic interactions must be enabled prior to use.
What is defined as an interaction?
EDIT: RTFM
>Whenever you ask Kiro something, it consumes an agentic interaction. This includes chat, a single spec execution, and/or every time an agent hook executes. However, the work Kiro does to complete your request—such as calling other tools, or taking multiple attempts—does not count towards your interactions.
And yes, Kiro is agentic, so it can (and often does) execute a long running multi-turn workflow in response to your interactions, however, the billing model is based on your manual interaction that kicks off the workflow (via chat, spec, or hook), even if that agent workflow takes many turns for Kiro to complete
How does Kirk deal with changes to the requirements? Are all the specs updated?
I find the best way to use specs is to progressively commit them into the repo as an append only "history" showing the gradual change of the project over time. You can use Kiro to modify an existing spec and update it to match the new intended state of the project, but this somehow feels a bit less valuable compared to having a historical record of all the design choices that led from where you started to where you now are.
I think in the long run Kiro will be able to serve both types of use: keeping a single authoritative library of specs for each feature, and keeping a historical record of mutations over time.
For example, you can use Kiro without having any AWS account at all. Kiro has social login through Google and GitHub. Basically, Kiro is backed by AWS, but is it's own standalone product, and we hope to see it grow and appeal to a broader audience than just AWS customers.
Redirect back to localhost:3128 is normal, that's where Kiro is watching for a callback, but the missing state is not normal. Something may have stripped the info out of the callback before it occurred, which is why I suspect an extension in your browser.
Will keep an eye on this though!
I think Auth can be a bit of mess, but yes Its still absolutely great that I can just login with github and it just works, I am trying out Kiro right as we speak!
Then in Kiro I see "There was an error signing you in. Please try again.".
FWIW, I've tried GitHub & Google, in different browsers, on different networks.
There is an AWS IAM Identity Center option for login as well: https://kiro.dev/docs/reference/auth-methods/#aws-iam-identi...
We really need to add some more step by step docs for setting this up, but it's very similar to the Amazon Q Developer integration with AWS IAM Identity Center if you are familiar with that: https://docs.aws.amazon.com/amazonq/latest/qdeveloper-ug/sec...
Companies would benefit a lot by creating better onboarding flows that migrate users from other applications. It should either bring in the rules 1:1 or have an llm agent transform them into a format that works better for the agent.
As for a hypothetical new "context setup" protocol like you posit, I suspect it'd benefit from the "cognitive tools" ideas in this awesome paper / project: <https://github.com/davidkimai/Context-Engineering>
^ inspiring stuff
Also these steering rules are just markdown files, so you can just drop your other rules files from other tools into the `.kiro/steering` directory, and they work as is.
“Kirk is actually quite good at this: you just have to do X”
“…”
Kiro will do it for you automatically.
In all seriousness, I'm sure this will become more standardized over time, in the same way that MCP has standardized tool use.
I've long been interested in something that can gather lightweight rules files from all your subdirectories as well, like a grandparent rule file that inherits and absorbs the rules of children modules that you have imported. Something kind of like this: https://github.com/ash-project/usage_rules
I think over time there will be more and more sources and entities that desire to preemptively provide some lightweight instructive steering content to guide their own use. But in the meantime we just have to deal with the standard proliferation until someone creates something amazing enough to suck everyone else in.
Then I add it as a git submodule to my projects and tell whatever agents to look at @llm-shared/ and update its own rule file(s) accordingly
My CLAUDE.md and GEMINI.md both just say "See AGENTS.md".
I've published a sample project that is medium sized, about 20k lines encompassing a game client, game server, and background service: https://github.com/kirodotdev/spirit-of-kiro This has been all developed by Kiro. The way Kiro is able to work in these larger projects is thanks to steering files like these:
- Structure, helps Kiro navigate the large project: https://github.com/kirodotdev/spirit-of-kiro/blob/main/.kiro...
- Tech, helps Kiro stay consistent with the tech it uses in a large project: https://github.com/kirodotdev/spirit-of-kiro/blob/main/.kiro...
And yes, the specs do help a lot. They help Kiro spend more time gathering context before getting to work, which helps the new features integrate into the existing codebase better, with less duplication, and more accuracy.
That said, thanks for being willing to demo what kinds of things it can do!
An agent running in the IDE can make use of all this context to provide better results. So, for example, you will see Kiro automatically notice and attempt to resolve problems from the "Problems" tab in the IDE. Kiro will look at what files you have open and attempt to use that info to jump to the right context faster.
The way I describe it is that the ceiling for an IDE agent is a lot higher than a CLI agent, just because the IDE agent has more context info to work with. CLI agents are great too, but I think the IDE can go a lot further because it has more tools available, and more info about what you are doing, where you are working, etc
In my opinion, CLIs have a higher ceiling, and then they are easy to integrate into CI/CD, run them in parallel, etc.
The advantage of something more purpose built for gathering context from the IDE is that you can skip a lot of roundtrips. Knowing the user's intent upfront, the IDE can gather all the necessary context data preemptively, filter it down to a token efficient representation of just the relevant stuff, add it in the context preemptively along with the user's prompt, and there is a single trip to the LLM before the LLM gets to work.
But yeah I agree with your point about CLI capabilities for running in parallel, integrating in other places. There is totally room for both, I just think that when it comes to authoring code in the flow, the IDE approach feels a bit smoother to me.
What people do to avoid what you discussed, is multi-agents. The main agent can build up context, plan, than delegate execution to other agents, etc.
In my opinion, the benefit of the IDE is really just in the possibility of an improved UI/UX over a TUI.
- cmd-t fuzzy finding files of cmd-p finding symbols to open the various files that are relevant
- selecting a few lines in each file using fast IDE shortcuts to move and add
- drag and drop an image or other json files into prompt
- not leave the editor im already working on
Not to mention:
- viewing the agents edits as a diff in the editor and all the benefits of easily switching between tabs and one click rejecting parts etc
- seeing the sidebar of the agents thoughts and progress async alongside the code as I keep looking at things
- pausing the agent and reversing back steps visually with the sidebar
- not having to reconfig or setup my entire dev environment for some CLI - for example the biome v2 lsp just works since it’s already working in code which has the best support for these things
And really the list of reasons an editor is far better just never ends. Claude is ok, but I’m way way faster with Cursor when I do need AI.
IDEs don't feel slow, they ARE slow
because written in HTML and Javascript
go and try Delphi from 2005, it's blazing fast (and more functional...)
It actually has a pretty decent free tier, and maybe the subscription is better value than Claude Code, but hard to tell.
Also,I don't mean to be rude to cursor but the fact that they are literally just a vscode wrapper still, to this day makes me really crazy thinking that the value of an AI editor could be so high..
I think it was the lack of competition really, Cursor (IMO) always felt like the biggest player, I think there was continue.dev before that, but that's all I know before Cursor.
After Cursor became a hit, there are lot more things now like (Void editor?) etc.
Also, if you Find Vscode editor slow, try zed. But like my brother said to me when I was shilling zed, Vscode is just for manipulating texts and using LSP. He personally didn't feel like it was there any meaningful slowness to Vscode even though he had tried zed. Zed has Ai stuff too iirc
Now Sure, they could've created CLI, but there are a lot of really decent CLI like SST/opencode and even gemini cli. Though I have heard good things about claude code too.
Honestly, I just think that any efforts in anything is cool. I just like it when there are a lot of options and so things stay a little competitive I guess.
ActiveX and Java Web Start, etc all tried to do this, and all of them ended up deprecated and out of favor for native web solutions.
Java IDEs did a lot of this for many years (Eclipse, IntelliJ, NetBeans, JDeveloper, etc) and they worked reasonably well on the desktop, but had no path to offering a web hosted solution (like gitpod or codespaces)
There are not a lot of options here, compiling down a native solution to wasm and running it in the browser would work, I'm not sure if the performance would be substantially better or more consistent across all OS'es and web unfortunately.
So we are where we are :)
1) It's normal for Kiro (and almost every AI editor) to use a lot more CPU when you first start it up, because it is indexing your codebase in the background, for faster and more accurate results when you prompt. That indexing should complete at some point
2) On initial setup of Kiro it will import and install your plugins from VS Code. If you have a large number of plugins this continues in the background, and can be quite CPU heavy as it extracts and runs the installs for each plugin. This is a one time performance hit though.
3) If your computer is truly idle, most modern CPU's get throttled back to save power. When the CPU is throttled, even a tiny amount of CPU utilization can show up as a large percentage of the CPU, but that's just because the CPU has been throttled back to a very slow clock speed.
In my setup (minimal plugins, medium sized codebase, computer set to never idle the processor clock) I rarely see Kiro helper go above .4% CPU utilization, so if you are seeing high CPU it is likely for one of the above reasons.
Is there any way to control this? I have my files.watcherExclude setting, does it respect that?
However with a large project, it seems that it indexed, then dropped CPU, then I started opening up files and working with them, then the CPU spiked again.
But in the meantime I'm also the author of the "Learn by Playing" guide in the Kiro docs. It goes step by step through using Kiro on this codebase, in the `challenge` branch. You can see how Kiro performs on a series of tasks starting with light things like basic vibe coding to update an HTML page, then slightly deeper things like fixing some bugs that I deliberately left in the code, then even deeper to a full fledged project to add email verification and password reset across client, server, and infrastructure as code. There is also an intro to using hooks, MCP, and steering files to completely customize the behavior of Kiro.
Guide link here: https://kiro.dev/docs/guides/learn-by-playing/
This, along with the "CHALLENGE.md" and "ROADMAP.md" document, is an incredibly cool way to show off your project and to give people a playground to use to try it out. The game idea itself is pretty interesting too.
It would be awesome if I ... didn't have to deal with AWS to use it. I guess maybe that might be a good use case for agentic coding: "Hey, Kiro - can you make this thing just use a local database and my Anthropic API key?"
Complaining aside though, I think that's just such a cool framework for a demo. Nice idea.
Neither VSCode nor Cursor do this, so even if it's an extension triggering it somehow, the behaviour in Kiro is different to those other two.
I wrote more about Spec Driven AI development here: https://lukebechtel.com/blog/vibe-speccing
I'm also thinking of creating a fork of the project that is designed to run entirely locally using your GPU. I believe with current quantized models, and a decent GPU, you can have an adequate enough fully local experience with this game, even the dynamic image generation part.
If you were referring to the prompts inside of the game, you might find those fun and interesting. This one in particular is the heart of the game: https://github.com/kirodotdev/spirit-of-kiro/blob/main/serve...
I integrated[1] the recently released Apple Container (instead of shell) to run codes generated by Kiro. It works great!
1. CodeRunner: https://github.com/BandarLabs/coderunner
To be clear, we have no intent to hide that Kiro is from Amazon / AWS, that's why you'll see Matt Garman, for example, posting about Kiro: https://www.linkedin.com/feed/update/urn:li:activity:7350558...
However, the long term goal is for Kiro to have it's own unique identity outside of AWS, backed by Amazon / AWS, but more friendly to folks who aren't all in on AWS. I'll admit that AWS hasn't been known in recent years for having the best new user or best developer experience. Kiro is making a fresh start from an outsider perspective of what's possible, not just what's the AWS tradition. So, for example, you can use Kiro without ever having an AWS account. That makes it somewhat unique, and we aim to keep it that way for now.
Unless it's literally a Cursor clone, I'd request to change it to describe the product category.
Cursor by no means defines the whole category. Not even close.
In a space that moves as quickly as "AI" does, it is inevitable that a better and cheaper solution will pop up at some point. We kinda already see it with Cursor and Windsurf. I guess Claude Code is all the rage now and I personally think CLI/TUI is the way to go for anyone that has a similar view.
That said, I'm sure there's a very big user base (probably bigger than terminal group) that will enjoy using this and other GUI apps.
You're basically advocating for GNU Emacs: https://github.com/karthink/gptel
For people wanting to get up and running with vanilla Emacs (instead of a distribution) so that they can try out gptel sometime this week, I recommend emacs-bedrock: https://codeberg.org/ashton314/emacs-bedrock
And for a gptel backend Gemini is the fastest route (excluding something local) from generating an API key to using a LLM in Emacs (for free).
Bonus points because Emacs is useful for things other than coding you can use gptel on your notes or any buffer really to ask/talk about stuff.
Is Claude Code good for the "ask" flow? No, right?
The old flow before agent mode got added. Select some code, ask questions about it or give an instruction on editing it and then choose to accept the change.
As I understand (I could be wrong), with agent mode, it edits the file for you, no way for you to accept before it does, so you have to manually check the diff, roll back parts you don't want, etc.
Am I right?
There was a paper recently where they had an LLM evolve tool harnesses and got ~20% more than w/ aider on the benchmark they used, so it's pretty clear that the models + tools (+better harness) are better than just aider.
> evolve tool harnesses
Claude code & Gemini cli etc. don't do this either
As for 1), I agree but you force the model to work within aider's constraints. Claude4 for example excels at the agentic flow and it's better at that than providing the diffs that aider expects.
As for the last sentence, I disagree. They are evolving the stack, and more importantly they are evolving both at the same time, stack + LLM. That's the main reason they all subsidise use atm, they are gathering data to improve both. If I were to place a bet right now, I'd say that provider_tool + provider_LLM > 3rd party tool + same model in the short, medium and long term.
In reality these tools would be best if they took a more socratic method, a more interactive pair programming approach. So instead of giving you a blanket diff to accept or refuse or "No, and here's some changes" -- it should be more dialog oriented.
Of all of them so far though, I think Claude Code is closest to this. IF you prompt it right you can have a much more interactive workflow, and I find that most productive.
You’re sort of technically correct but I wouldn’t really describe it this way exactly. You have to proactively accept or reject all changes to your files in some way.
It is almost impossible to accidentally commit code you don’t want.
It’s not really an edit in the same sense as an unstated change. It doesn’t even really do that until you accept the result.
It’s basically saving you a UI step compared to ask mode with basically no downside.
> Am I right?
With cursor you get reasonably flexible control at many levels. You can have it only suggest changes that you have to apply manually or you can have it make automatic changes with various ways to review, change, reject or accept. I usually have the changes made automatically but don’t accept the changes automatically. Cursor has a UI that lets you review each edit individually, for the whole file or all files. Depending on the situation I will use whichever level is appropriate. The UI also allows you to revert changes or you can ask the AI to undo or rework a change that you just approved so there’s plenty of ways to do large changes without giving up control. There’s also a stop button you can use to interrupt mid-stream if the work it’s doing isn’t what you want. It isn’t flawless but I haven’t found myself in a corner where I couldn’t get back to a happy path.
I think it is entirely possible to build a fantastic CLI tool for coding, and the CLI tools for coding already work well enough, but there is just more context info available inside of an IDE, therefore the ceiling is higher when working with an agent that runs inside of the IDE. Context is king for LLM results, and IDE's just have more context.
Over time I'm sure we'll see tools like Claude Code support everything that an IDE can do, but for now if you want to reach the same ceiling you still have to glue together a very custom setup with MCP tool use, and that has the downside of introducing additional tool use latency, compared to an IDE that is able to source context directly from the IDE's internal API, and provide that to the LLM nearly instantly.
It’s also not free or unlimited (though throttled) like Cursor and Claude Code using max plan.
In the short term though, I think CLI-based tools like Claude Code are taking off because hardcore developers see them as the last "vestige" they have in separating themselves from the "noobs." They know there's still a good portion of the public who don't know how to use the terminal, install packages, or even know what Linux is.
If these re-skinned vscode IDEs have any good ideas I'm sure Microsoft will steal them anyway.
I guess you lose tab-completion suggestions, but I am not a fan of those compared to 'normal' tab-complete (if backed by an lang server). If I want AI, I'll write a short comment and invoke the tool explicitly.
EDIT: Of course, it really depends an your usecase. I maintain/upgrade C code libs and utils; I really cannot speak to what works best for your env! Webdev is truly a different world.
EDIT2: Can't leave this alone for some reason, the backend thing is a big deal. Switching between Claude/Gemini/Deekseek and even rando models like Qwen or Kimi is awesome, they can fill in each other's holes or unblock a model which is 'stuck'.
It’s not just the IDE but the ML model you are selling yourself to. I see my colleagues atrophy before me. I see their tools melt in their hands. I am rapidly becoming the only person functionally capable of reason on my own. It’s very very weird.
When the model money dries up what’s going to happen?
I too am old enough to have seen a lot of unnecessary tech change cycles, and one thing I've noticed about this industry is no matter how foolish a trend was, we almost never unwind it.
As for trends, I've been around long enough to have seen this cycle a couple of times...
My point is there aren't many of you, are there?
All things considered, keeping up with the industry trends is generally a more reliable career path.
I’ve survived every single layoff season since 1995.
All the people I know in the US with those skills make huge amounts of money- at least, the ones who haven't already retired rich.
Why are they shipping them with different key bindings? Seems like the opposite of what you do to encourage product adoption.
AI seems to be a way to engage happy users to try new things. Kiro joins a growing list of projects:
- Kiro (AWS)
- VSCode + Copilot (Microsoft)
- Windsurf (OpenAI tried to get it)
- Cursor
- Trae (Alibaba)
- Zed
- etc.
I put Zed in a separate category in the past. Now with assistants / agents, it's playing on the same space.
The market is a bit saturated and tools like Claude Code gave some flexibility and an alternative for users. I tried Cursor in the past, and now I'm back to Helix / VSCode + Claude Code.
But at the same time, it's my biggest worry that they will continue on the AI and pollute the project with too much junk. I gotta trust the smart minds behind it will find a way to balance this trend.
Lately, I started putting it together due to all the AI excitement. People try it because of the AI capabilities to find an IDE that works for them.
I hope Zed continues providing new amazing features in all the areas.
And Google killed it.
Basically, the AI fast forwards through the easy stuff and I just spend all day jumping directly from hard problem to hard problem.
Overall I do believe this has accelerated our development and I'm interested to see where it goes. I don't think it's a direct comparison to claude code or cursor - its a different approach with some overlap.
As a customer I have no incentive to try it.
I think that reputation is 100% Amazon’s fault. When all you do is ship half-baked rushed products your customers will assume your next big thing sucks because that’s the reputation you built for yourself.
The Q Developer CLI, Q Developer IDE plugins, and now Kiro are pretty much just wrappers around Claude Sonnet 3.7/4, and work just as well as them.
How is one fork different from cursor or kiro or something else?
Arent these like what i assume skinning chromium or something more ?
Just this morning, Cursor was giving me a ton of incorrect tab completions. When I use prompts, it tends to break more than it fixes. It's still a lot faster to write by hand. Lots of libraries that take *arguments in Python also cannot be groked by AI.
I doubt these tools will ever convince every last person on every single use case, so the existence of those people isn't exactly an indictment.
I have found it extremely useful for spinning up personal projects though.
My wife bought us Claude subscriptions and she's been straight-up vibe coding an educational game for our son with impressive results (she is a UX designer so a lot more attuned to vibes than gen-pop). I'm picking up some computational physics research threads I dropped in grad school and Claude Code has been incredible at everything besides physics and HPC. Define and parse an input file format, integrate I/O libraries, turn my slapdash notes into LaTeX with nice TiKz diagrams, etc.
Hoping I can transfer over some insights to make it more helpful at work.
It's also interesting that the pricing is in terms of "interactions" rather than tokens. I don't believe I've seen that before.
In my experience using Kiro you are still going to be hands on with the code. Personally I choose to turn AI powered autocomplete off because when I do touch the code manually it's usually for the purposes of working on something tricky that AI would likely not get right in autocomplete either.
However, the Kiro autocomplete is average in capability in my experience, and you can absolutely use it to write code by hand as well.
When you make a tool that is "model agnostic" you also make a tool that is unable to play to the individual strengths of each model, or you set yourself up for a massive, multiplicative effort of trying to tune your tool to every single popular model out there, even though some of the models are drastically less capable than others.
Going to take a while before I trust any AWS AI related tooling won't just be abandoned / mis-managed after my prior experience.
Edit: I know there’s manual VSIX route.
What else does Kiro do differently?
Edit: The hooks feature looks nifty. How is the memory management handled? Any codebase indexing etc? Support to add external MCP servers like context7 etc?
I always keep the readme and some basic architecture docs (using markdown/mermaid) updated as I go, and I often just work on those rather than on code with Claude, because I find the value it offers is less in code generation and more in helping me document the rubber ducking process into useful schematics and architecture.
What can Kiro offer that's meaningfully better than what I'm already doing? I can take my system anywhere Claude Code and my repos can go, using whatever editor I like. Does Kiro have some special sauce for making this approach work better? Maybe some DSL it uses for more succinct and actionable diagrams and plans?
As much as I like the idea, I find it so hard to abandon a process I've been working on for months, using tools I'm already productive with.
Also, will pricing essentially be bedrock pricing, or will there be a value-add margin tacked on?
I'd like to think so, but you'd have to compare the results to what you are currently doing to see how you feel about it. I personally love the format that it uses to define requirements, and the details of the software design docs that it writes (including mermaid diagrams)
> will pricing essentially be bedrock pricing, or will there be a value-add margin tacked on?
The pricing is a flat rate, with a cap on number of interactions per month. Each human driven "push" for Kiro to do something is an interaction toward your limit, but Kiro may work autonomously for many turns based on an interaction, and will produce significant amounts of code from a single interaction.
More details here: https://kiro.dev/pricing/
At $39/month, is 3000 interactions a high limit? I use Claude Code on the $30 plan (I think), and routinely hit limits. I'm not ready to jump to the next tier, though. I think it's $200/month, and the NGO I work for isn't prepared to throw that kind of cash at developers (I'm second-rate here; the science comes first)
For reference 3000 interactions * assumed 3 mins of AI work per interaction / 60 mins per hour / 8 working hours per day equals 18.75 working days of nonstop back to back AI coding. Typical month has 20-23 working days. But realistically you likely won't be using Kiro nonstop all day back to back, so 3000 interactions per month should more than cover your work month.
Lets see EU US data privacy shield gone. Aws a bunch of open source tools gobbled together with proprietary source. Trust in made in US cloud platform tools is gone!
It has a pretty decent free tier, and maybe the subscription is better value than Claude Code, but hard to tell.
It supports MCP as well.
Amazon Q also has a VC Code and IntelliJ Idea plugin too, Kiro goes beyond what you can do as a plugin in VS Code though, similar to why Cursor had to fork VS Code.
Not as polished as Claude Code but also a bit price difference
> Starting August 1, 2025, we’re introducing a new pricing plan for Amazon Q Developer designed to make things simpler and more valuable for developers.
> Pro Tier: Expanded limits $19/mo. per user
> 1,000 agentic requests per month included (starting 8/1/2025)
- https://aws.amazon.com/q/developer/pricing/
Previously agentic use was apparently "free", but with a set deadline in June, so it seems like this was just for a testing phase?
I read it (I think) in one of the comment that There is a model picker that currently allows you to switch between Claude Sonnet 4.0 and Claude Sonnet 3.7
So is this just using Claude?
I really thought that the advantages of using Kiro might really be that of the leverage that Amazon Gpu's infrastructure could provide, maybe even some discounts to lure people to Kiro.
I am pretty sure that a lot of people will ask you the same question, But I would really appreciate it if you could answer me this question in preferably simple terms: "Why Kiro? Why not all the other stuff that has come before it and the stuff that will come after it"
Also I am really having some dejavu but while writing this comment, has the title of this post changed, I swear I saw something written in the header with Amazon and now I don't see it. honestly, I am really being so off-topic but after seeing this name change of the post, I really wish if that there was some website that could track all the name changes of posts that happen in HN, because I was completely baffled by this name change or I am being totally paranoid.
Claude 4 can do it all already.
- Created by an AWS team but aws logo is barely visible at the bottom.
- Actually cute logo and branding.
- Focuses on the lead devs front and center (which HN loves). Makes it seem less like a corporation and more like 2 devs working on their project / or an actual startup.
- The comment tone of "hey ive been working on this for a year" also makes it seem as if there weren't 10 6-pagers written to make it happen (maybe there weren't?).
- flashy landing page
Props to the team. Wish there were more projects like this to branch out of AWS. E.g. Lightsail should've been launched like this.
In some sense, we are starting with a very high-level and gradually refining the idea to a lower and lower levels of detail. It is structured hierarchical thinking. Right now we are at 3 levels: requirement -> spec -> code. Exposing each of these layers as structured text documents (mostly Markdown right now it seems) is powerful since each level can be independently reviewed. You can review the spec before the code is written, then review the code before it gets checked in.
My intuition is that this pattern will be highly effective for coding. And if we prove that out at scale, we should start asking: how does this pattern translate to other activities? How will this affect law, medicine, insurance, etc. Software is the tip of the iceberg and if this works then there are many possible avenues to expand this approach, and many potential startups to serve a growing market.
The key will be managing all of the documents, the levels of abstraction and the review processes. This is a totally tractable problem.
It is a huge hassle to match my existing settings, which I've spent countless hours tweaking over the years, with a new editor that can't import them. :(
Or, you know, stop chasing the latest trends, and use whatever you're most comfortable with.
It is my new daily driver.
> For users who access Kiro with Pro or Pro+ tiers once they are available, your content is not used to train any underlying foundation models (FMs). AWS might collect and use client-side telemetry and usage metrics for service improvement purposes. You can opt out of this data collection by adjusting your settings in the IDE. For the Kiro Free tier and during preview, your content, including code snippets, conversations, and file contents open in the IDE, unless explicitly opted out, may be used to enhance and improve the quality of FMs. Your content will not be used if you use the opt-out mechanism described in the documentation. If you have an Amazon Q Developer Pro subscription and access Kiro through your AWS account with the Amazon Q Developer Pro subscription, then Kiro will not use your content for service improvement. For more information, see Service Improvement.
1. Open Settings in Kiro.
2. Switch to the User sub-tab.
3. Choose Application, and from the drop-down choose Telemetry and Content.
4. In the Telemetry and Content drop-down field, select Disabled to disable all product telemetry and user data collection.
source: https://kiro.dev/docs/reference/privacy-and-security/#opt-ou...
Clearly, companies view the context fed to these tools as valuable. And it certainly has value in the abstract, as information about how they're being used or could be improved.
But is it really useful as training data? Sure, some new codebases might be fed in... but after that, the way context works and the way people are "vibe coding", 95% of the novelty being input is just the output of previous LLMs.
While the utility of synthetic data proves that context collapse is not inevitable, it does seem to be a real concern... and I can say definitively based on my own experience that the _median_ quality of LLM-generated code is much worse than the _median_ quality of human-generated code. Especially since this would include all the code that was rejected during the development process.
Without substantial post-processing to filter out the bad input code, I question how valuable the context from coding agents is for training data. Again, it's probably quite useful for other things.
The docs state at https://kiro.dev/docs/reference/privacy-and-security/#servic... that "Kiro is an AWS application that works as a standalone agentic IDE."
But nowhere on the landing page or other pages it states that this is an Amazon product.
What is going on?
Edit: I see that @nathanpeck is the "author" and he works for Amazon, why are they trying to hide that fact?
From the about page.
> Kiro is built and operated by a small, opinionated team within AWS.
Disclaimer: I work at AWS, different org though.
Looking at the brief for this, it likely involves painstaking work to review and refine the specs produced. Not that there is anything wrong with that; as I said before in a comment on another story, coding assistants may reduce quite a bit of drudgery, but to get the best out of them you still need to do lots of work.
The more I use agentic coding tools, the more I come to the realization that speccing is where you add value as an experienced skilled engineer. And I think this bodes well for software engineering, as a bifurcation emerges between the vibe coders (who will probably merge with the mythical end-user programmers) and serious engineers whose are skilled at using LLMs via high quality specs to create software of higher quality and maintainability.
So the vibe coders would probably not take to this tool that much, but that's fine.
2. At every task it tried to compile the code but failed for dependency errors
3. It still marked the task being complete and passed the onus of failures on the downstream tasks
4. Kept moving with the tasks where the original error were still not fixed but the tasks were being marked as done
5. After some point of time I got tired to a degree that I stopped reading the exact commands being executed, the fatigue of doing something that you are not involved in is for real 6. I made a naive assumption that I can sandbox it by giving permissions to the project folder only. It executed some CLI commands for java that looked simple enough in the beginning.
7. Turns out my environment variables got messed up and other simple things related to git, gradle stopped working
Ended my experiment, reverted the code changes, fixed my environment
Key takeaways:
1. Its giving a sense of work being executed, the quality and concreteness of work is hard to measure unless you have already done that in past. Its creating classes, tests which are not needed instead of focussing on the actual use case.
2. Sandboxes are MUST, there is a real risk of corruption, environment commands are not just simple file changes which could be easily reverted.
OSS option will be there soon and will outsmart Kiro.
Why not mention it on website and be explicit that this is from Amazon.
As someone who is spending enormously on agents, ML, and AI, I don't see anything good coming out from this.
They priced it at $19 because Cursor is $20? Biggest LOL.
Amazon is biggest investor in Anthropic who is bleeding money to get Claude Code usage up.
Why not stick with Anthripic and make these backhanded bets?
What if I want to set preferences for the underlying LLM for different usage scenarios? For example, for a quick and snappy understanding of a single file id want to use a fast model that doesn't cost me an arm and a leg. Recent research on preference-aligned LLM routing here: https://arxiv.org/abs/2506.16655
- Uses ripgrep under the hood
- VSCode fork (thus suffers from the https://ghuntley.com/fracture problem)
- There are 14 different ways defined to edit a file due to its multi-modal design. Tuning this is going to be a constant source of headaches for the team.
- Kiro utilises a https://ghuntley.com/specs based workflow.
They can try to market grab with low %, but will find themselves in the boat as Cursor and eventually be forced to raise their prices. Except their market grab will be significantly less effective because they're not a stand-out product. Cursor was.
QuinnyPig•8h ago
NathanKP•6h ago
mh-•5h ago
(I initially started writing this as a joke upon recognizing your name, but now I think I'm serious..)
QuinnyPig•4h ago
mh-•4h ago