frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Medinilla – an OCPP compliant .NET back end (partially done)

https://github.com/eliodecolli/Medinilla
1•rhcm•1m ago•0 comments

How Does AI Distribute the Pie? Large Language Models and the Ultimatum Game

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6157066
1•dkga•1m ago•1 comments

Resistance Infrastructure

https://www.profgalloway.com/resistance-infrastructure/
2•samizdis•5m ago•0 comments

Fire-juggling unicyclist caught performing on crossing

https://news.sky.com/story/fire-juggling-unicyclist-caught-performing-on-crossing-13504459
1•austinallegro•6m ago•0 comments

Restoring a lost 1981 Unix roguelike (protoHack) and preserving Hack 1.0.3

https://github.com/Critlist/protoHack
1•Critlist•7m ago•0 comments

GPS and Time Dilation – Special and General Relativity

https://philosophersview.com/gps-and-time-dilation/
1•mistyvales•11m ago•0 comments

Show HN: Witnessd – Prove human authorship via hardware-bound jitter seals

https://github.com/writerslogic/witnessd
1•davidcondrey•11m ago•1 comments

Show HN: I built a clawdbot that texts like your crush

https://14.israelfirew.co
2•IsruAlpha•13m ago•1 comments

Scientists reverse Alzheimer's in mice and restore memory (2025)

https://www.sciencedaily.com/releases/2025/12/251224032354.htm
1•walterbell•16m ago•0 comments

Compiling Prolog to Forth [pdf]

https://vfxforth.com/flag/jfar/vol4/no4/article4.pdf
1•todsacerdoti•17m ago•0 comments

Show HN: Cymatica – an experimental, meditative audiovisual app

https://apps.apple.com/us/app/cymatica-sounds-visualizer/id6748863721
1•_august•19m ago•0 comments

GitBlack: Tracing America's Foundation

https://gitblack.vercel.app/
2•martialg•19m ago•0 comments

Horizon-LM: A RAM-Centric Architecture for LLM Training

https://arxiv.org/abs/2602.04816
1•chrsw•19m ago•0 comments

We just ordered shawarma and fries from Cursor [video]

https://www.youtube.com/shorts/WALQOiugbWc
1•jeffreyjin•20m ago•1 comments

Correctio

https://rhetoric.byu.edu/Figures/C/correctio.htm
1•grantpitt•20m ago•0 comments

Trying to make an Automated Ecologist: A first pass through the Biotime dataset

https://chillphysicsenjoyer.substack.com/p/trying-to-make-an-automated-ecologist
1•crescit_eundo•24m ago•0 comments

Watch Ukraine's Minigun-Firing, Drone-Hunting Turboprop in Action

https://www.twz.com/air/watch-ukraines-minigun-firing-drone-hunting-turboprop-in-action
1•breve•25m ago•0 comments

Free Trial: AI Interviewer

https://ai-interviewer.nuvoice.ai/
1•sijain2•25m ago•0 comments

FDA intends to take action against non-FDA-approved GLP-1 drugs

https://www.fda.gov/news-events/press-announcements/fda-intends-take-action-against-non-fda-appro...
21•randycupertino•27m ago•10 comments

Supernote e-ink devices for writing like paper

https://supernote.eu/choose-your-product/
3•janandonly•29m ago•0 comments

We are QA Engineers now

https://serce.me/posts/2026-02-05-we-are-qa-engineers-now
1•SerCe•29m ago•0 comments

Show HN: Measuring how AI agent teams improve issue resolution on SWE-Verified

https://arxiv.org/abs/2602.01465
2•NBenkovich•29m ago•0 comments

Adversarial Reasoning: Multiagent World Models for Closing the Simulation Gap

https://www.latent.space/p/adversarial-reasoning
1•swyx•30m ago•0 comments

Show HN: Poddley.com – Follow people, not podcasts

https://poddley.com/guests/ana-kasparian/episodes
1•onesandofgrain•38m ago•0 comments

Layoffs Surge 118% in January – The Highest Since 2009

https://www.cnbc.com/2026/02/05/layoff-and-hiring-announcements-hit-their-worst-january-levels-si...
13•karakoram•38m ago•0 comments

Papyrus 114: Homer's Iliad

https://p114.homemade.systems/
1•mwenge•38m ago•1 comments

DicePit – Real-time multiplayer Knucklebones in the browser

https://dicepit.pages.dev/
1•r1z4•38m ago•1 comments

Turn-Based Structural Triggers: Prompt-Free Backdoors in Multi-Turn LLMs

https://arxiv.org/abs/2601.14340
2•PaulHoule•40m ago•0 comments

Show HN: AI Agent Tool That Keeps You in the Loop

https://github.com/dshearer/misatay
2•dshearer•41m ago•0 comments

Why Every R Package Wrapping External Tools Needs a Sitrep() Function

https://drmowinckels.io/blog/2026/sitrep-functions/
1•todsacerdoti•41m ago•0 comments
Open in hackernews

Launch HN: Promi (YC S24) – Personalize e-commerce discounts and retail offers

25•pmoot•6mo ago
Hey HN! I’m Peter from Promi (https://www.promi.ai/). We’re building a platform for ecommerce merchants to send realtime personalized discounts, optimized with AI (obviously)

Demo: https://youtu.be/BCYNCqb4fUc, Sales Video: https://www.youtube.com/watch?v=WiO1S7RBn-o

All the big tech companies send personalized discounts - Uber, DoorDash, Google, etc. In fact, I was the product lead overseeing discounts at Uber, so if you’ve gotten a promotion on Uber Rides or Eats, that was our tech. These personalization models often generate >30% more revenue vs. non-personalized discounts (cost-neutral that is), so this is a hugely impactful product.

It’s no surprise then that other merchants want to follow suit. Merchants don’t want to waste discounts on customers who would have purchased anyway. Frankly it’s not a new idea to offer a software solution to personalize discounts - plenty of other startups have entered this space with a similar product.

The biggest problem with personalizing discounts for mid-size and smaller companies has been that traditionally you rely on ‘explore’ data - data from randomly sending out discounts to a portion of the user base. But this has a lot of problems: merchants need to be large, collecting this data is expensive, training data really should be fresh (so explores should constantly be running), and if you want to try a different discount structure (e.g. BOGO instead of 20% off) you’ll need to run a new explore with the new structure.

So what does Promi do differently? We train on regular traffic and simplify the problem by just focusing on conversion rate. If we can accurately predict who is unlikely to convert and which products are unlikely to be bought, we can issue discounts without the fear of burning money on an order that would have happened anyway. One of my major takeaways from my time at Uber was that our model was mostly targeting users who had a low likelihood of converting in a given week. Quantifying how much more likely they were to convert when given a discount via explores was helpful, but not as impactful as understanding starting conversion rate.

Side note - It’s been a bit interesting launching an AI company during this hype cycle that isn’t actually using the latest and greatest LLMs. We believe more traditional machine learning still has a lot of value to add. I don’t want to say we won’t use LLMs down the road (there may be some interesting applications for developing additional features), but starting this way has worked out well for us.

There have been plenty of other challenges (as with any startup). We’ve had to figure out how to automate integrations when so many websites have custom code. We’ve had to make the model work without rich user data since the majority of website visitors aren’t logged in. A quick note in this one - we can use first party cookies to more or less track the view and transaction history, but we’ve found that one big predictor of conversion is traffic source: whether a visitor is coming from ads, email, direct traffic, google search, etc. That traffic source isn’t something as valuable at Uber (since everyone uses the app), so it’s been a bit of a tradeoff in the types of features that are most impactful.

Our model seems to be working well! We have case studies on our website showing the typical revenue and profit lift we see. We currently have tiered pricing with different quotas for the amount of revenue managed by Promi discounts.

I’d love to get thoughts from the machine learning experts in this community, though full disclosure I’m the non-technical founder. Let us know what you think!

Comments

lazyninja987•6mo ago
Does a merchant has to give your tool access to their user data to generate personalized discounts? Apart from user activity data, what data do you need for maximum effectiveness?
pmoot•6mo ago
Yes. We're going through Shopify, so merchants agree to terms when they install the app.

There's user activity data, but also contextual data and shop data that we use. 'Contextual' data refers to things like device type, traffic source, time of day, day of week (there have been some interesting trends with corporate vs. non-corporate customers in this one).

Shop data includes things like product profit margin and product conversion rate. Obviously we can go deeper with our discounts on products that are very profitable, and it's typically more efficient to give a discount on products with lower conversion. Merchants also like boosting items that haven't been selling well.

klaussilveira•6mo ago
> If we can accurately predict who is unlikely to convert

Do you use historical purchase data to make that assumption? Or someone that frequently abandons carts?

pmoot•6mo ago
We use historical purchase data, as well as view history, traffic source, device type, etc.

Traffic source a lot of times is the most impactful. People coming from ads are often more in a browsing mindset, vs. people typing in the url directly have a higher purchase intent.

We don't have abandoned cart rate as a feature in our model, but actually might be something worth looking into adding.

malshe•6mo ago
If I understand it correctly, you estimate the probability of purchase given the user characteristics, behaviors, etc. If this probability is below a cutoff, you offer a discount. Did I get it right?

Is the cutoff itself a function of other variables in the data?

pmoot•6mo ago
Yes, that's mostly right. We also vary the discount value, so it's less a binary discount/no discount and more a range. There is often a cutoff though. Merchants can input a hard cutoff e.g. if they want to ensure everyone gets a discount (great if they also have marketing assets for a sale), or if they want to avoid making their sites feel too 'sales-y'. Otherwise the cutoff is defined by conversion prediction, inventory levels, and a few other inputs.

There's actually a lot more we could do to make this cutoff more intelligent though - e.g. at Uber the cutoff was set to exhaust a certain promotional budget. Or we could target a specific ROI if we eventually have good enough predictions.

malshe•6mo ago
Thanks for the reply. Do you use Bayesian models for this? Btw, Pete Fader[1] has done so much work in customer valuation where estimating the probability of purchase is a crucial aspect. Maybe you already use them.

[1] https://marketing.wharton.upenn.edu/profile/faderp/#overview

pmoot•6mo ago
We're using a neural network, not a bayesian model. And we haven't used Pete Fader's work, but thank you for the resource.
kristianc•6mo ago
> We’ve had to make the model work without rich user data since the majority of website visitors aren’t logged in.

Be aware that this sentence largely disqualifies you from doing business in Europe.

pmoot•6mo ago
We faced a similar issue with GDPR at Uber. We will definitely need to be careful, but many merchants already have customer opt ins (e.g. the cookie consent pop up) for data processing that we ideally should be able to piggy-back off of.