frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Ask HN: The government of my country blocked VPN access. What should I use?

700•rickybule•8h ago•401 comments

Python: The Documentary

https://lwn.net/Articles/1035537/
28•chmaynard•1h ago•2 comments

Fuck up my site – Turn any website into beautiful chaos

https://www.fuckupmysite.com/?url=https%3A%2F%2Fnews.ycombinator.com&torchCursor=true&comicSans=t...
130•coloneltcb•3h ago•40 comments

Some thoughts on LLMs and software development

https://martinfowler.com/articles/202508-ai-thoughts.html
174•floverfelt•6h ago•159 comments

My startup banking story (2023)

https://mitchellh.com/writing/my-startup-banking-story
155•dvrp•5h ago•66 comments

Uncertain<T>

https://nshipster.com/uncertainty/
237•samtheprogram•7h ago•50 comments

Death by PowerPoint: the slide that killed seven people

https://mcdreeamiemusings.com/blog/2019/4/13/gsux1h6bnt8lqjd7w2t2mtvfg81uhx
44•scapecast•3h ago•10 comments

Expert LSP the official language server implementation for Elixir

https://github.com/elixir-lang/expert
47•pimienta•3h ago•8 comments

RSS Is Awesome

https://evanverma.com/rss-is-awesome
58•edverma2•1h ago•12 comments

Building your own CLI coding agent with Pydantic-AI

https://martinfowler.com/articles/build-own-coding-agent.html
102•vinhnx•6h ago•21 comments

TuneD is a system tuning service for Linux

https://tuned-project.org/
27•tanelpoder•3d ago•8 comments

Are OpenAI and Anthropic losing money on inference?

https://martinalderson.com/posts/are-openai-and-anthropic-really-losing-money-on-inference/
431•martinald•14h ago•414 comments

AI adoption linked to 13% decline in jobs for young U.S. workers: study

https://www.cnbc.com/2025/08/28/generative-ai-reshapes-us-job-market-stanford-study-shows-entry-l...
170•pseudolus•10h ago•260 comments

Launch HN: Dedalus Labs (YC S25) – Vercel for Agents

43•windsor•8h ago•11 comments

Rupert's Property

https://johncarlosbaez.wordpress.com/2025/08/28/a-polyhedron-without-ruperts-property/
19•robinhouston•2h ago•1 comments

A forgotten medieval fruit with a vulgar name (2021)

https://www.bbc.com/future/article/20210325-the-strange-medieval-fruit-the-world-forgot
65•ohjeez•1d ago•27 comments

Dependent types I › Universes, or types of types

https://www.jonmsterling.com/01ET/index.xml
7•matt_d•1d ago•0 comments

Bad Craziness

https://www.math.columbia.edu/~woit/wordpress/?p=15191
13•jjgreen•1h ago•2 comments

You no longer need JavaScript: an overview of what makes modern CSS so awesome

https://lyra.horse/blog/2025/08/you-dont-need-js/
82•todsacerdoti•4h ago•30 comments

Thrashing

https://exple.tive.org/blarg/2025/08/26/thrashing/
12•pch00•1d ago•1 comments

Speed-coding for the 6502 – a simple example

https://www.colino.net/wordpress/en/archives/2025/08/28/speed-coding-for-the-6502-a-simple-example/
18•mmphosis•3h ago•7 comments

Will AI Replace Human Thinking? The Case for Writing and Coding Manually

https://www.ssp.sh/brain/will-ai-replace-humans/
110•articsputnik•10h ago•90 comments

VLT observations of interstellar comet 3I/ATLAS II

https://arxiv.org/abs/2508.18382
44•bikenaga•6h ago•32 comments

Optimising for maintainability – Gleam in production at Strand

https://gleam.run/case-studies/strand/
87•Bogdanp•9h ago•21 comments

Show HN: SwiftAI – open-source library to easily build LLM features on iOS/macOS

https://github.com/mi12labs/SwiftAI
52•mi12-root•11h ago•11 comments

Web Bot Auth

https://developers.cloudflare.com/bots/reference/bot-verification/web-bot-auth/
38•ananddtyagi•6h ago•37 comments

In Search of AI Psychosis

https://www.astralcodexten.com/p/in-search-of-ai-psychosis
85•venkii•2d ago•48 comments

RFC 8594: The Sunset HTTP Header Field (2019)

https://datatracker.ietf.org/doc/html/rfc8594
24•aiven•5h ago•9 comments

That boolean should probably be something else

https://ntietz.com/blog/that-boolean-should-probably-be-something-else/
84•vidyesh•12h ago•94 comments

Microbial metabolite repairs liver injury by restoring hepatic lipid metabolism

https://journals.asm.org/doi/10.1128/mbio.01718-25
108•PaulHoule•13h ago•28 comments
Open in hackernews

Launch HN: Dedalus Labs (YC S25) – Vercel for Agents

43•windsor•8h ago
Hey HN! We are Windsor and Cathy of Dedalus Labs (https://www.dedaluslabs.ai/), a cloud platform for developers to build agentic AI applications. Our SDK allows you to connect any LLM to any MCP tools – local or hosted by us. No Dockerfiles or YAML configs required.

Here’s a demo: https://youtu.be/s2khf1Monho?si=yiWnZh5OP4HQcAwL&t=11

Last October, I was trying to build a stateful code execution sandbox in the cloud that LLMs could tool-call into. This was before MCP was released, and let’s just say it was super annoying to build… I was thinking to myself the entire time “Why can’t I just pass in `tools=code_execution` to the model and just have it…work?

Even with MCP, you’re stuck running local servers and handwiring API auth and formatting across OpenAI, Anthropic, Google, etc. before you can ship anything. Every change means redeploys, networking configs, and hours lost wrangling AWS. Hours of reading docs and wrestling with cloud setup is not what you want when building your product!

Dedalus simplifies this to just one API endpoint, so what used to take 2 weeks of setup can take 5 minutes. We allow you to upload streamable HTTP MCP servers to our platform. Once deployed, we offer OpenAI-compatible SDKs that you can drop into your codebase to use MCP-powered LLMs. The idea is to let anyone, anywhere, equip their LLMs with powerful tools for function calling.

The code you write looks something like this:

  python
  client = Dedalus()
  runner = DedalusRunner(client)
  
  result = runner.run(
    input=prompt,
    tools=[tool_1, tool_2],
    mcp_servers=["author/server-1”, “author/server-2”],
    model=["openai/gpt-4.1”, “anthropic/claude-sonnet-4-20250514”],  # Defaults to first model in list
    stream=True,
  )
  stream_sync(result)  # Streams result, supports tool calling too
Our docs start at https://docs.dedaluslabs.ai. Here’s a simple Hello World example: https://docs.dedaluslabs.ai/examples/01-hello-world. For basic tool execution, see https://docs.dedaluslabs.ai/examples/02-basic-tools. There are lots more examples on the site, including more complex ones like using the Open Meteo MCP to do weather forecasts: https://docs.dedaluslabs.ai/examples/use-case/weather-foreca....

There are still a bunch of issues in the MCP landscape, no doubt. One big one is authentication (we joke that the “S” in MCP stands for “security”). MCP servers right now are expected to act as both the authentication server and the resource server. That is too much to ask of server writers. People just want to expose a resource endpoint and be done.

Still, we are bullish on MCP. Current shortcomings are not irrecoverable, and we expect future amendments to resolve them. We think that useful AI agents are bound to be habitual tool callers, and MCP is a pretty decent way to equip models with tools.

We aren’t quite yet at the stateful code execution sandbox that I wanted last October, but we’re getting there! Shipping secure and stateful MCP servers is high on our priority list, and we’ll be launching our auth solution next month. We’re also working on an MCP marketplace, so people can monetize their tools, while we handle billing and rev-share.

We’re big on open sourcing things and have these SDKs so far (MIT licensed):

https://github.com/dedalus-labs/dedalus-sdk-python

https://github.com/dedalus-labs/dedalus-sdk-typescript

https://github.com/dedalus-labs/dedalus-sdk-go

https://github.com/dedalus-labs/dedalus-openapi

We would love feedback on what you guys think are the biggest barriers that keep you from integrating MCP servers or using tool calling LLMs into your current workflow.

Thanks HN!

Comments

muratsu•7h ago
Congrats on the launch. I see that you're charging 5% on Balance Reloads. This pricing model seems to be getting popular across multi-LLM applications. Was curious to know how did you go about implementing it? or are you just passing on the 5% of openrouter
windsor•7h ago
Good eye. A ~5% surcharge on prepaid credits is the standard model right now for most multi-LLM services. We actually do not use OpenRouter internally, so this number is flexible. One thing I'll note is that we try to be as upfront and transparent about our platform fee as possible so that no one is surprised.
muratsu•6h ago
Oh interesting. I've previously looked into implementing it myself but seemed like it would require a lot of effort. I would love to connect and learn more about your implementation. What's the best way to reach out to you? My email is available on my profile.
mifydev•4h ago
Congrats on the launch! I’m curious, do you have to store the tool inputs and outputs on the server side while either of the sides are waiting for response? I’m building a specialized coding agent for integrations and I had to avoid stateful api, because I don’t want to store user code.
windsor•4h ago
We tried to keep things simple, so our runner utility class is currently stateless.

However, it's highly extensible, and we can support stateful logic if we wanted to. For instance, we have a cool concept called "Policy" in our SDKs, which is basically a user-defined callback function that is run after each runner step (see our docs for more info). You can build some pretty advanced use cases with this, e.g. executing conditional database calls on a per-step basis.

The code for the runner is open source, in case you wanted to check out how we did it: https://github.com/dedalus-labs/dedalus-sdk-python/blob/main...

just_human•4h ago
Congrats on the launch and looks interesting! I love how easy it is to combine local code "tools" with remote mcp servers. The marketplace looks promising, but would be helpful to have some curation as many of the servers don't have descriptions and link to private github repo's. Neat vision and look forward to trying this.
windsor•3h ago
Thanks! Really glad you noticed this feature. Mixing client-side and server-side tool calls was something we spent a lot of time thinking through.

The current SOTA, e.g. OpenAI’s Responses API or Anthropic’s Computer Use API, basically mandates that server-side tool results return directly, while client-side tool results have to be manually parsed and executed by the user (for obvious security reasons). As a result, it was extremely unclear how a user would be able to chain together tool calls that mixed local and remote tools.

We wanted to close this DX gap, which surprisingly had no real incumbent solution. Users should be able to just define tools and get back clean responses. For power users, we still support manual JSON parsing for full low-level control, but our belief is simple: developers should spend their time building, not doing plumbing work like post-processing tool results.

mnafees•4h ago
Congratulations on the launch!

I’ve been writing Go for the past 4 years, and I’d strongly suggest avoiding Stainless for auto-generating Go SDKs. Some of the issues I’ve run into: - Awkward package naming (e.g., githubcomdedaluslabsdedalussdkgo) - Methods with unnecessary complexity, including excessive use of reflection for JSON handling - Other codegen quirks that make the SDK harder to use and maintain

From experience, I’d recommend either using another code generator or hand-writing the Go SDK for a cleaner and more idiomatic developer experience.

windsor•3h ago
Thanks! What's your favorite SDK generator for Go? What are the most important things for you when interacting with Go SDKs?
kitan232•3h ago
Congratulations on the launch. Been trying this out for the past hour, and really like how easy it is to host your own MCP servers. Would love to see more public MCPs on the marketplace. Also, I was wondering when you plan to support auth?
windsor•3h ago
Thanks for trying us out! Let me know if you run into any issues with deployment or the like. If you want a clean MCP template to get up and running, check out this example: https://github.com/dedalus-labs/brave-search-mcp

Auth is a big one for us, and we're working really hard to provide a robust auth experience that is easy to use for both LLM agents and human users. One of our goals is to help mold the shape that the community takes with MCP. We’ll be launching our auth solution around end of September. I’m personally really excited to tackle this problem.