frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

P2P crypto exchange development company

1•sonniya•5m ago•0 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
1•jesperordrup•10m ago•0 comments

Write for Your Readers Even If They Are Agents

https://commonsware.com/blog/2026/02/06/write-for-your-readers-even-if-they-are-agents.html
1•ingve•11m ago•0 comments

Knowledge-Creating LLMs

https://tecunningham.github.io/posts/2026-01-29-knowledge-creating-llms.html
1•salkahfi•11m ago•0 comments

Maple Mono: Smooth your coding flow

https://font.subf.dev/en/
1•signa11•18m ago•0 comments

Sid Meier's System for Real-Time Music Composition and Synthesis

https://patents.google.com/patent/US5496962A/en
1•GaryBluto•26m ago•1 comments

Show HN: Slop News – HN front page now, but it's all slop

https://dosaygo-studio.github.io/hn-front-page-2035/slop-news
4•keepamovin•27m ago•2 comments

Show HN: Empusa – Visual debugger to catch and resume AI agent retry loops

https://github.com/justin55afdfdsf5ds45f4ds5f45ds4/EmpusaAI
1•justinlord•29m ago•0 comments

Show HN: Bitcoin wallet on NXP SE050 secure element, Tor-only open source

https://github.com/0xdeadbeefnetwork/sigil-web
2•sickthecat•31m ago•1 comments

White House Explores Opening Antitrust Probe on Homebuilders

https://www.bloomberg.com/news/articles/2026-02-06/white-house-explores-opening-antitrust-probe-i...
1•petethomas•32m ago•0 comments

Show HN: MindDraft – AI task app with smart actions and auto expense tracking

https://minddraft.ai
2•imthepk•37m ago•0 comments

How do you estimate AI app development costs accurately?

1•insights123•38m ago•0 comments

Going Through Snowden Documents, Part 5

https://libroot.org/posts/going-through-snowden-documents-part-5/
1•goto1•38m ago•0 comments

Show HN: MCP Server for TradeStation

https://github.com/theelderwand/tradestation-mcp
1•theelderwand•41m ago•0 comments

Canada unveils auto industry plan in latest pivot away from US

https://www.bbc.com/news/articles/cvgd2j80klmo
3•breve•42m ago•1 comments

The essential Reinhold Niebuhr: selected essays and addresses

https://archive.org/details/essentialreinhol0000nieb
1•baxtr•44m ago•0 comments

Rentahuman.ai Turns Humans into On-Demand Labor for AI Agents

https://www.forbes.com/sites/ronschmelzer/2026/02/05/when-ai-agents-start-hiring-humans-rentahuma...
1•tempodox•46m ago•0 comments

StovexGlobal – Compliance Gaps to Note

1•ReviewShield•49m ago•1 comments

Show HN: Afelyon – Turns Jira tickets into production-ready PRs (multi-repo)

https://afelyon.com/
1•AbduNebu•50m ago•0 comments

Trump says America should move on from Epstein – it may not be that easy

https://www.bbc.com/news/articles/cy4gj71z0m0o
6•tempodox•51m ago•3 comments

Tiny Clippy – A native Office Assistant built in Rust and egui

https://github.com/salva-imm/tiny-clippy
1•salvadorda656•55m ago•0 comments

LegalArgumentException: From Courtrooms to Clojure – Sen [video]

https://www.youtube.com/watch?v=cmMQbsOTX-o
1•adityaathalye•58m ago•0 comments

US moves to deport 5-year-old detained in Minnesota

https://www.reuters.com/legal/government/us-moves-deport-5-year-old-detained-minnesota-2026-02-06/
8•petethomas•1h ago•3 comments

If you lose your passport in Austria, head for McDonald's Golden Arches

https://www.cbsnews.com/news/us-embassy-mcdonalds-restaurants-austria-hotline-americans-consular-...
1•thunderbong•1h ago•0 comments

Show HN: Mermaid Formatter – CLI and library to auto-format Mermaid diagrams

https://github.com/chenyanchen/mermaid-formatter
1•astm•1h ago•0 comments

RFCs vs. READMEs: The Evolution of Protocols

https://h3manth.com/scribe/rfcs-vs-readmes/
3•init0•1h ago•1 comments

Kanchipuram Saris and Thinking Machines

https://altermag.com/articles/kanchipuram-saris-and-thinking-machines
1•trojanalert•1h ago•0 comments

Chinese chemical supplier causes global baby formula recall

https://www.reuters.com/business/healthcare-pharmaceuticals/nestle-widens-french-infant-formula-r...
2•fkdk•1h ago•0 comments

I've used AI to write 100% of my code for a year as an engineer

https://old.reddit.com/r/ClaudeCode/comments/1qxvobt/ive_used_ai_to_write_100_of_my_code_for_1_ye...
3•ukuina•1h ago•1 comments

Looking for 4 Autistic Co-Founders for AI Startup (Equity-Based)

1•au-ai-aisl•1h ago•1 comments
Open in hackernews

Launch HN: Plexe (YC X25) – Build production-grade ML models from prompts

https://www.plexe.ai/
85•vaibhavdubey97•3mo ago
Hey HN! We're Vaibhav and Marcello, founders of Plexe (https://www.plexe.ai). We create production-ready ML models from natural language descriptions. Tell Plexe what ML problem you want to solve, point it at your data, and it handles the entire pipeline from feature engineering to deployment.

Here’s a walkthrough: https://www.youtube.com/watch?v=TbOfx6UPuX4.

ML teams waste too much time on generic heavy lifting. Every project follows the same pattern: 20% understanding objectives, 60% wrangling data and engineering features, 20% experimenting with models. Most of this is formulaic but burns months of engineering time. Throwing LLMs at it isn't the answer as that just trades engineering time for compute costs and worse accuracy. Plexe automates this repetitive 80%, so your team can work faster on what actually has value.

You describe your problem in plain English ("fraud detection model for transactions" or "product embedding model for search"), connect your data (Postgres, Snowflake, S3, direct upload, etc), and then Plexe: - Analyzes data and engineers features automatically - Runs experiments across multiple architectures (logistic regression to neural nets) - Generates comprehensive evaluation reports with error analysis, robustness testing, and prioritized recommendations to provide actionable guidance - Deploys the best model with monitoring and automatic retraining

We did a Show HN for our open-source library five months ago (https://news.ycombinator.com/item?id=43906346). Since then, we've launched our commercial platform with interactive refinement, production-grade model evaluations, retraining pipeline, data connectors, analytics dashboards, and deployment for online and batch inference.

We use a multi-agent architecture where specialized agents handle different pipeline stages. Each agent focuses on its domain: data analysis, feature engineering, model selection, deployment, and so on. The platform tracks all experiments and generates exportable Python code.

Our open-source core (https://github.com/plexe-ai/plexe, Apache 2.0) remains free for local development. For the paid product, our pricing is usage-based, with a minimum top up of $10. Enterprises can self-host the entire platform. You can sign up on https://console.plexe.ai. Use promo code `LAUNCHDAY20` to get $20 to try out the platform.

We’d love to hear your thoughts on the problem and feedback on the platform!

Comments

johnsillings•3mo ago
very cool – I like how opinionated the product approach is vs. a bunch of disconnected tools for specialists to use (which seems more common for this space).
marcellodb•3mo ago
Thanks, we're pretty opinionated on "this should make sense to non-ML practitioners" being a defining aspect of the product vision. Behind the scenes, we've had quite a few conversations specifically about how to avoid features feeling "disconnected", which is always challenging at an early stage when you're getting pulled in several directions by users with different use cases. Happy to hear it came across that way to you.
oxml•3mo ago
Great product!
vaibhavdubey97•3mo ago
Thank you! :)
tnt128•3mo ago
In the demo, you didn’t show the process of cleaning and labeling data, does your product do that somehow, or do you still expect the user to provide that after connecting the data source.
vaibhavdubey97•3mo ago
We have a data enricher feature (still in a beta mode) which uses LLMs to generate labels for your data. For cleaning and feature engineering, we use agents that automatically handle it for you once you've connected your data and defined your ML problem.

P.S. Thanks for the feedback on the video! We'll update it to show the cleaning and labelling process :)

marcellodb•3mo ago
Great question, this is super important. The agents in the platform have the ability to do some degree of cleaning on your data when building a model (for example, imputing missing values). However, major improvements to data quality are generally not possible without an understanding of the data domain (i.e. business context), so you'll get better results if you "help" the platform by providing data in a reasonably clean state, answering the agent's follow-up questions in the chat, etc. By doing so you can give the agent better context and help it understand your data better, in which case it will also be more capable of dealing with things like missing values, misnamed columns etc.

This also highlights the important role of the user as a (potentially non-technical) domain expert. Hope that makes sense!

brightstar18•3mo ago
Product seems cool. But can you help me understand if what you are doing is different from the following: > you put a prompt > Plexe glorifies that prompt into a bigger prompt with more specific instructions (augmented by schema definitions, intent and whatnot) > plug it into the provided model/LLM > .predict() gives me the output (which was heavily guardrailed by the glorified prompt in the step 2)
marcellodb•3mo ago
Great question, and yes, it's quite different: Plexe generates code for a pipeline that processes your dataset (analysis, feature engineering, etc) and trains a custom ML model for your use case. When you call `.predict()`, it is that trained custom model that provides the response, not an LLM. The model is also hosted for you, and Plexe takes care of MLOps things like letting you retrain the model on new data, evaluating the model performance for you, etc. Using custom specialised models is generally more effective, faster and cheaper compared to running your predictions through an LLM when you have a lot of data specific to your business.
ryanmerket•3mo ago
Really diggin this. Can't wait to try it out.
vaibhavdubey97•3mo ago
Thanks a lot! Excited for you to try it out and get your feedback :)
sinanuozdemir•3mo ago
Sounds interesting! I'm trying to train a model but it's still "processing" after a bit but fine-tuning takes a while I get it. I'm having trouble understanding how it's inferring schema. I used a sample dataset and yet the sample inference curl uses a blank json?

curl -X POST "XXX/infer" \ -H "Content-Type: application/json" \ -H "x-api-key: YOUR_API_KEY" \ -d '{}'

How do I know what the inputs/outputs are for one of my models? I see I could have set the response variable manually before training but I was hoping the auto-infer would work.

Separately it'd be ideal if when I ask for models that you seem to not be able to train (I asked for an embedding model as a test) the platform would tell me it couldn't do that instead of making me choose a dataset that isn't anything to do with what I asked for.

All in all, super cool space, I can't wait to see more!

I'm a former YC founder turned investor living in Dogpatch. I'd love to chat more if you're down!

marcellodb•3mo ago
Thanks for the great feedback! To your points:

1. Depending on your dataset the training could take from 45 mins to a few hours. We do need add an ETA on the build in the UI.

2. The input schema is inferred towards the end of the model building process, not right at the start. This is because the final schema depends on the decisions made regarding input features, model architecture etc during the building process. You should see the sample curl update soon, with actual input fields.

3. Great point about upfront rejecting builds for types of models we don't yet support. We'll be sure to add this soon!

We're in London at the moment, but we'd love to connect with you and/or meet in person next time we're in SF - drop us a note on LinkedIn or something :)

vaibhavdubey97•3mo ago
Thanks for the great feedback! We've added a `baseline_deployed` status where the agents create an initial baseline and deploy it so you have something to play around with quickly. This is why you're seeing a blank json there. Once your final model is deployed, it creates an input and output schema from the features used for the model build :)
lcnlvrz•3mo ago
How does it perform when build computer vision models?
marcellodb•3mo ago
Unfortunately we don't officially support image, video or audio yet - only tabular data for now. We do plan to add that capability at some point in the coming weeks depending on popular demand. Do you have any particular use case in mind?

Caveat: as a more technical user, you can currently "hack" around this limitation by storing your images as byte arrays in a parquet file, in which case the platform can ingest your data and train a CV model for you. We haven't tested the performance extensively though, so your mileage may vary here.

alansaber•3mo ago
Problem space is very interesting. Sounds like most of the work will be the data handling which is an evergreen problem
vaibhavdubey97•3mo ago
Absolutely! Getting AI agents to determine the right set of features from raw data has been a very interesting problem
pplonski86•3mo ago
Amazing! Great work. Congratulations on launch.

Few questions: 1. Can it work with tabular data, images, text and audio? 2. Data preprocessing code is deployed with the model? 3. Have you tested use cases when ML model was not needed? For example, you can simply go with average. I'm curious if agent can propose not to use ML in such case. 4. Do you have agent for model interpretation? 5. Are you using generic LLM or have your own LLM tuned on ML tasks?

marcellodb•3mo ago
Thanks! Great set of questions:

1. Tabular data only, for now. Text/images also work if they're in a table, but unfortunately not unstructured text or folders of loose image files. Full support for images, video, audio etc coming sometime in the near future.

2. Input pre-processing is deployed in the model endpoint to ensure feature engineering is applied consistently across training and inference. Once a model is built, you can see the inference code in the UI and you'll notice the pre-processing code mirrors the feature engineering code. If you meant something like deploying scheduled batch jobs for feature processing, we don't support that yet, but it's in our plans!

3. The agent isn't explicitly instructed to "push back" on using ML, but it is instructed to develop a predictor that is as simple and lightweight as possible, including simple baseline heuristics (average, most popular class, etc). Whatever performs best on the test set is selected as the final predictor, and this could just be the baseline heuristic, if none of the models outperform it. I like the idea of explicitly pushing back on developing a model if the use case clearly doesn't call for it!

4. Yes, we have a model evaluator agent that runs an extensive battery of tests on the final model to understand things like robustness to missing data, feature importance, biases, etc. You can find all the info in the "Evaluations" tab of a built model. I'm guessing this is close to what you meant by "model interpretation"?

5. A mix of generic and fine-tuned, and we're actively experimenting with the best models to power each of the agents in the workflow. Unsurprisingly, our experience has been that Anthropic's models (Sonnet 4.5 and Haiku 4.5) are best at the "coding-heavy" tasks like writing a model's training code, while OpenAI's models seem to work better at more "analytical" tasks like reviewing results for logical correctness and writing concise data analysis scripts. Fine-tuning for our specific tasks is, however, an important part of our implementation strategy.

Hope this covers all your questions!

vaibhavdubey97•3mo ago
Thanks a lot! On a side note: big fan of mljar here. When we were initially playing around with using agents for automating ML tasks, we had used problems from the openml's automl benchmark which you had posted about on Reddit for our initial tests
throwacct•3mo ago
Interesting. Are you guys only charging for the model building using tokens or overall usage (model building, inferences, etc)?
marcellodb•3mo ago
It's a combination of tokens consumed, dataset + model storage cost, and inference + training compute cost.
canada_dry•3mo ago
The tool gave me advice and code to do what I asked... but when I used the "export analysis" it did NOT include the code. It was simply an overview.

It would be more useful for the export to have an option (or by default) to include everything from the session.

vaibhavdubey97•3mo ago
Thanks for your feedback! The "export analysis" functionality is built to enable you to get detailed data insights and get a report generated from the insights. Would you prefer to see the entire chat in your export or would it be helpful for it to simply include code snippets in the same format you received right now?
canada_dry•3mo ago
A way to select all (or by Q/A paragraph) for export (e.g. to pdf or how about a .ipynb) would probably be the most useful. Perhaps as just a plain "export" option in addition to an analysis/insights summary.

p.s. kudos on the promo code that enable folks to kick the tires with as little friction as possible.

vaibhavdubey97•3mo ago
Makes sense! We'll be sure to make this available very soon :) Thank you!
liqilin1567•3mo ago
> Each agent focuses on its domain: data analysis, feature engineering, model selection, deployment, and so on

Sounds very practical in real-world use cases. I trained a ML model couple months ago, I think it's a good case to test this product.

vaibhavdubey97•3mo ago
Thanks a lot! We’ve tried to mimic the workflow of an ML engineer and built agents that can own specific functions of the workflow. Good to hear that the idea resonates with you!
abhishekbasu•3mo ago
Great product and congratulations on the launch. Who is the target user vs customer? On the surface, and I may be wrong here, this feels like a LLM layered on top of a typical AutoML structure eg: TPOT, Caret. Is that the correct mental model for a tool like this? And if so, do you see a similar problem that these tools faced in broader adoption at companies?
marcellodb•3mo ago
I think "agents layered on top of AutoML" is a reasonable simple mental model for Plexe's model building capabilities, but it also masks some important qualitative differences between Plexe and traditional AutoML tools:

1. AutoML tools work on clean data. Data preparation requires an understanding of business context, the ability to reason on the data in that context, and then produce code for the required data transformations. Given that this process could not be automated with "templated" pipelines, teams using AutoML still have to do the hardest - and arguably most important - part of the data science job themselves.

2. AutoML tools use "templated" models for regression, classification, etc, which may not result in as good a "task-data-model fit" as the sort of purpose-written ML code a data scientist or ML engineer might produce.

3. AutoML tools still require a working understanding of data science technicalities. They automate the running of ML training experiments, but not the task of deciding what to do in the first place, or the task of understanding whether what was done actually fits the task.

With this in mind, we've seen that most ML teams don't find traditional AutoML tools useful (they only automate the "easy" part), while software teams don't find them accessible (data science knowledge is still required).

Plexe addresses both of these issues: the agents' reasoning capabilities enable it to work with messy data (as long as you provide business context), and to ENTIRELY abstract the deeper technicalities of building custom models fitting the task and the data. We believe this makes Plexe both useful to ML teams and accessible to non-ML teams.

Does this line up with your experience of AutoML tools?