frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Ask HN: How to overcome the limit of roles in LLM's

1•weli•1h ago
Our use case is not uncommon, we are developing tools so that people can install LLM's on their e-commerces.

But there are some interesting challenges that I feel can't be solved unless inference providers allow us to include the concept additional entities in a conversation.

As far as I know the three most basic ones shared alongside all providers are:

- System

- Assistant

- User

That's fine and it allows for simple conversational-based approaches (ChatGPT, Claude, Gemini, etc). However in our use case we allow our customers (not the final user who is talking with the AI) to configure the AI in different ways (personality, RAG, etc), which poses a problem.

If we inject those customer settings in the System prompt then that's a risk because there might be conflicting prompts with our internal rules. So the easiest option is to "clean" the customer prompts before injecting them, but that feels hacky and just adds one more level of indirection. Cleaning the prompt and injecting it with common patterns like XML tags seems to help a bit but still feels extremely risky for some reason.

Injecting it in the assistant or user also seems flaky and prone to prompt injection.

Creating a fake tool call and result like "getPersonalityConfiguration" seems to work the best, from our testing it is treated as something between the System and Assistant roles. And our top system prompt rules are still respected while allowing the customer some freedom to configure the AI.

The problem comes when you need to add more parties to what essentially is a 2 entity conversation. Sometimes we want external agents to chime in a conversation (via subagents or other methods) and there is no good way to do that AFAIK. It gets the occasional confusion and starts mixing up who is who.

One of our typical scenarios that we need to model:

System: Your rules are: You will never use foul language...

Store owner: You are John the customer agent for store Foo...

User: Do you have snowboards in stock?

Assistant->User: Let me check with the team. I'll get back to you soon.

System->Team: User is asking if we have snowboards in stock. Do we?

Team: We do have snowboards in stock.

Team->User: Yes we do have snowboards in stock!

User: Perfect, if I buy them will the courier send it to my country? [country name].

Assistant->User: Let me check, I need to see if our courier can ship a snowboard to your country.

Assistant->Third party logistics: I have a user from [country] interested in buying a snowboard. The dimensions are X by Y and the weight is Z. We would send it from our logistics center located at [address].

Third party logistics -> Assistant: Yes we can do it, it will be 29.99 for the shipping.

Assistant->User: Yes they can ship it to [country] but it does incur in 29.99 extra charge...

I obviated tool calls and responses, but that's basically the gist of it. Spawning sub-agents that have the context of the main conversation works but at some point it is limiting (we need to copy all personality traits and relevant information via summarization or injecting the conversation in a manner that the sub-agent won't get confused). It feels like an anti-pattern and trying to fight the intended use case of LLM's, which seems to be focused in conversation between two entities with the occasional external information going in through System or tool calling.

It would be amazing if we could add custom roles to model messages, still with special cases like agent or assistant.

Has anyone worked with similar problems? How did you solve it? Is this solved in the model lab or at the inference provider level (post-training)?

Show HN: Turn Steam reviews into personas and insights, without agent chatting

https://steam-review.dexmage.com/
1•dexteryy•48s ago•0 comments

Show HN: A-MEM – Memory for Claude Code that links and evolves on its own

https://github.com/DiaaAj/a-mem-mcp
1•AttentionBlock•53s ago•0 comments

Yasu – AI agents that fix cloud waste, not just report it

https://yasu.cloud/
1•johninthout•2m ago•1 comments

Where does data help in real estate – and where does it fail?

1•D___R___•3m ago•0 comments

Important Paper

https://zenodo.org/records/18234712
1•KaoruAK•5m ago•0 comments

Cursor may be switching from Solid to React

https://cursor.com/blog/scaling-agents#running-for-weeks
1•benatkin•5m ago•1 comments

AI as life coach: experts say what works, what doesn't and what to look out for

https://www.theguardian.com/wellness/2026/jan/15/ai-life-coach
3•billybuckwheat•6m ago•0 comments

Open Responses

https://www.openresponses.org/
1•davidbarker•7m ago•0 comments

Is Greenland USA?

https://isgreenlandusa.com/
4•Javimoya•8m ago•0 comments

Nothing new under the sun: everything is a file

https://turso.tech/blog/nothing-new-under-the-sun
1•weinzierl•9m ago•0 comments

Every journey feels better when you're part of it

https://otieu.com/4/8908706
1•abdpatil•9m ago•0 comments

Paul Graham Claude Code Skill

https://www.aibuilder.sh/skills/acossta/paul-graham
1•acossta•9m ago•0 comments

Ask HN: Anyone have a good solution for modern Mac to legacy SCSI converters?

5•stmw•10m ago•1 comments

You can't tell people anything

https://habitatchronicles.com/2004/04/you-cant-tell-people-anything/
2•evo_9•11m ago•0 comments

Analysis of ServiceNow's AI Vulnerability (85% of Fortune 500 Affected)

https://opena2a.org/blogs/servicenow-ai-vulnerability
1•abdelfane•11m ago•1 comments

Show HN: Okiro – spin up ephemeral codebases for parallel AI coding

https://github.com/ygwyg/okiro
3•burcs•12m ago•0 comments

Scrollbars in Scrollbars

https://matoseb.com/scrollbars-scrollbars/
1•birdculture•13m ago•0 comments

Solving TaxCalcBench: LLMs solve taxes by navigating IRS taxcode like a codebase [pdf]

https://prime-meridian-papers.s3.us-west-2.amazonaws.com/solving_taxes_like_code.pdf
1•rishabhjain1198•14m ago•2 comments

Within months, UK smartphones could be subjected to client scanning surveillance

https://www.telegraph.co.uk/news/2026/01/13/starmer-hell-bent-destroying-your-right-to-private-life/
2•walterbell•15m ago•0 comments

Finding Solar Farms with a 42k-Parameter Model

https://toao.com/blog/earth-observation-budget-solar-farms-tiny-model
1•avsm•15m ago•0 comments

AntiRender: See what architect renders will look like

https://antirender.com/
1•tontonius•18m ago•0 comments

Gas Town

https://github.com/steveyegge/gastown
1•tosh•18m ago•1 comments

Denmark's struggle to break up with Silicon Valley

https://www.politico.eu/article/denmark-declared-war-against-big-tech-digital-sovereignty/
12•giuliomagnifico•19m ago•2 comments

AI #151: While Claude Coworks

https://thezvi.substack.com/p/ai-151-while-claude-coworks
1•paulpauper•20m ago•0 comments

Sustainable Production of a Promising Protein Source from Carrot Side Streams

https://pubs.acs.org/doi/10.1021/acs.jafc.5c11223
1•PaulHoule•20m ago•0 comments

Proving (literally) that ChatGPT isn't conscious

https://www.theintrinsicperspective.com/p/proving-literally-that-chatgpt-isnt
3•paulpauper•21m ago•0 comments

The competitive claw machine boom

https://thehustle.co/originals/the-competitive-claw-machine-boom
1•paulpauper•21m ago•0 comments

Mathematical Recreations and Essays by W. W. Rouse Ball [pdf]

https://www.gutenberg.org/files/26839/26839-pdf.pdf
2•ibobev•22m ago•0 comments

CVEs Affecting the Svelte Ecosystem

https://svelte.dev/blog/cves-affecting-the-svelte-ecosystem
18•tobr•24m ago•0 comments

Show HN: Cache Explorer – The Compiler Explorer for CPU Cache Behavior

https://github.com/AveryClapp/Cache-Explorer
2•AveryClapp•26m ago•0 comments