frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Monetising an API by simply emailing public keys

https://img.arible.co/
1•sim04ful•42s ago•0 comments

Sweet Alert++ – A Modern, Accessible Modal Library (SweetAlert2 Alternative)

https://raiank.github.io/sweetalertplusplus.github.io/
1•sueraccount•8m ago•0 comments

Sterilization and contraception increase lifespan across vertebrates

https://www.nature.com/articles/s41586-025-09836-9
1•thunderbong•8m ago•0 comments

How to Make a Crypto Coin

https://coinheadlines.com/news/features/how-to-make-a-crypto-coin/article-22639/
1•RitaDfouni•10m ago•1 comments

ChatGPT and search ads used for malware distribution

https://eclecticlight.co/2025/12/11/how-online-search-and-ai-can-install-malware/
2•louis-paul•11m ago•0 comments

Feedback on Integrated Security Platform

https://substack.com/home/post/p-181243655
2•alex-dozer•12m ago•1 comments

The Waffle Singularity

https://www.tetraslam.world/blog/the_waffle_singularity
2•Tetraslam•13m ago•0 comments

Show HN: Simo.io – Security-first wired open-source Smart Home System for pros

https://simo.io
2•pysupremacy•21m ago•0 comments

I Stopped Coding

https://www.youtube.com/watch?v=KBL_RkTx5eI
2•avivby•24m ago•0 comments

The Enterprise AI Revolution: From Chatbots to Autonomous Agentic Architectures

https://medium.com/@mohan.khilariwal/the-enterprise-ai-revolution-moving-beyond-chatbots-to-auton...
2•avivby•24m ago•0 comments

Good Leadership Hinges on "Organizational Intelligence" (2020)

https://hbr.org/2020/06/good-leadership-hinges-on-organizational-intelligence
3•avivby•25m ago•0 comments

Show HN: Turn Git commits into Linear-style release note

https://www.updated.dev/
2•hyun_kim•31m ago•1 comments

$5 whale listening hydrophone making workshop

https://exclav.es/2025/08/03/dinacon-2025-passive-acoustic-listening/
2•gsf_emergency_6•32m ago•0 comments

Parakeets make their home in German trees

https://angiesweb.com/rose-ringed-parakeets-in-germany/
4•doruk101•39m ago•1 comments

Self-hosted Gits battered in 0-day attacks with no fix imminent

https://www.theregister.com/2025/12/10/gogs_0day_under_active_exploitation/
4•Brajeshwar•39m ago•0 comments

Qwen's API platform for image/video generation

https://www.mulerouter.ai
2•dr_dshiv•40m ago•1 comments

Native Parallel Reasoner: Self-Evolving to Learn Parallel Reasoning

2•jacklanda•40m ago•0 comments

Show HN: BJH OS – A Web-Based OS That Works Without Back End or Frameworks

https://github.com/Haris16-code/BJH-OS
1•Haris18•41m ago•1 comments

Skydiver dangles from plane 15,000ft in dramatic new footage of parachute snag

https://www.theaustralian.com.au/business/aviation/skydiver-dangles-from-plane-at-15000ft-in-dram...
1•asdefghyk•45m ago•1 comments

Rails MCP Server: Context-Efficient Tool Architecture

https://mariochavez.io/desarrollo/2025/12/10/rails-mcp-server-context-efficient-refactoring/
1•amalinovic•45m ago•0 comments

Startupideasdb,com is where I got my dream AI Tech Startup Idea. You can Google

5•peterbricks•46m ago•0 comments

ODAM Memory for Cursor – Long-Term Project Memory for Your AI Coding Assistant

https://github.com/aipsyhelp/Cursor_ODAM
1•AndrewMPT•48m ago•1 comments

Show HN: 8B Parallel Coordinated Reasoning Model

https://github.com/stepfun-ai/PaCoRe
1•hzwer•49m ago•0 comments

Console.text() – SMS alerts when code executes

2•Noel04•59m ago•1 comments

List of Mcface Spoofs

https://en.wikipedia.org/wiki/List_of_Mcface_spoofs
3•minajevs•1h ago•0 comments

CVE-2025-66491: Traefik's "Verify=on" Turned TLS Off

https://aisle.com/blog/cve-2025-66491-traefiks-verifyon-turned-tls-off
1•abhisek•1h ago•0 comments

Password Complexity is a Lie – Here’s What Actually Keeps You Safe

https://www.youtube.com/watch?v=6R1DfG8HiCs
1•sipofwater•1h ago•0 comments

Live by the Law or Die on the Cross

https://www.tabletmag.com/sections/news/articles/live-law-die-cross-israel
1•KnuthIsGod•1h ago•2 comments

How to Use N8n to Sync Google Calendar Automatically

https://www.indiehackers.com/post/how-to-use-n8n-to-sync-google-calendar-automatically-a6BYizPK8i...
2•plakhlani2•1h ago•0 comments

•AI Surveys• New Startup - Surveyi

https://surveyi.app/
2•dickiedogbrands•1h ago•1 comments
Open in hackernews

AI chatbots can sway voters with remarkable ease

https://www.nature.com/articles/d41586-025-03975-9#ref-CR1
31•marojejian•12h ago

Comments

marojejian•12h ago
While I'm as paranoid about LLMs as the next HN'er, there are some silver linings to this research:

1) the LLMs mostly used factual information to influence people (vs. say emotional or social influence) 2) the fact were mostly accurate

I'm not saying we shouldn't worry. But I expected the results to be worse.

Overall, the interesting finding here is that that political opinions can be changed by new information at all. I'm curious how this effect would compare to comparably informed human discussions. I would not be surprised if the LLMs were more effect for at least two reasons:

1) Cost-efficiency, in terms of the knowledge required, and effort/skill to provide personalized arguments. 2) Reduction in the emotional barrier to changing your mind: people don't want to "lose" by being wrong about politics to someone else. But perhaps the machine doesn't trigger this social/tribal response.

Cited papers:

https://www.nature.com/articles/s41586-025-09771-9

https://www.science.org/doi/10.1126/science.aea3884

techblueberry•12h ago
I’ll add a third reason, which is I think in general, people are very bad at understanding how to make an argument to someone with a different value system. I’m liberal, I have family members who are conservative, and I’ll read conservative books and I’m genuinely a person who is curious to new ideas, but most people I know(and I’m sure this works vice versa) are only good at expressing political opinions in the language of people who share their values. Republicans and Democrats don’t just talk about different things, they talk about them in very different ways.

I find this online as well, like I hate being “out of my echo chamber” because those arguments are just uniformly pointless. (This is in all directions by the way, people to the right or left of me).

Though I also interestingly find trying to talk to LLMs about competing values challenging too, if I try to get the LLM to explain a conservative position, then I make counter-arguments to that position, it will almost never tell me my counter argument is wrong, just “you’ve hit the nail on the head! Boy are you smart!”

dlivingston•10h ago
I had a friend in grad school who influenced my political beliefs more than anyone I'd met.

He never engaged in political conversation with "here's what I believe, and here's why you should too." His approach was more Socratic; to listen to me talk, and then offer an additional viewpoint or context.

I never got the impression from him that he was trying to convince me of something, or that he thought I was wrong about X/Y/Z, but rather, that we were on an intellectual journey together to identify what the problems actually were and what nuanced solutions might look like.

I still have no idea to this day what his ACTUAL political party is (or if he even has one). I genuinely could not tell you if he was left, right, or center.

apercu•10h ago
As far as I can tell most conservative argument points seem to be about the price of gas. If there is a democrat in the whitehouse, the price of gas is astronomical. If there is a republican in office, gas is far cheaper somehow than what I always end up paying.

Gasoline is like the least important cost metric in my life.

techblueberry•9h ago
I've been going down a bit of a rabbit hole on "what conservatives believe" and weirdly, and this is from both Roger Scruton, and the book "The conservative mind". is it's a bit like porn, you can't define it, but you know it when you see it. I mean this is sort of a tangible points conservatives make about believing in "common sense" that there's basically a higher truth that we all know exists that should guide us.

Roger Scruton in I think this video: https://www.youtube.com/watch?v=1eD9RDTl6tM. Says that basically conservatism in the 80's in the UK was whatever Margaret Thatcher believed. This really I think helped me understand why the conservative transition from Reagan/Bush to Trump went more smoothly than I thought it would among trad conservatives.

paulryanrogers•7h ago
Growing up indoctrinated into conversative evangelism, I saw that the Midwestern flavor valued freedom of individuals from government. It was a shallow flavor of self sufficiency, which discounted all social support except family and churches. Abortion was a wedge issue preached from every platform.

Tribalism was a key substrate. This often manifested as a near blind loyalty to the party and chosen thought leaders like Bill Graham, Rush Limbaugh, Bill O'Reilly, and now Tucker Carlson. They told us how to interpret events and we repeated the talking points. They gave us the (often contradictory) rules and principles we were to use to view everything in life.

zem•9h ago
the scenario that worries me is "fox news but personalised", e.g. fox can run a dozen pieces on "immigrants are taking your jobs" but an LLM hooked into your google profile could generate an article on how "plumbers in nashville are being displaced by low-paid mexicans" that is specifically designed to make you personally fear for your job if the nazi du jour isn't elected.
ekjhgkejhgk•9h ago
> the LLMs mostly used factual information to influence people

No, you see. This is how I used to think when I was a teenager.

Democracy isn't about being factually correct. It's about putting in place rules to make accumulation of power to the point that it can bend the rules themselves, very difficult.

It's not a silver lining that LLMs are persuasive by being mostly accurate, if they're used to increase the power of their owner further.

TomasBM•10h ago
I looked at the original study [1], and it seems to be a very well-supported piece of research. All the necessary pieces are there, as you would expect from a Nature publication. And overall, I am convinced there's an effect.

However, I'm still skeptical of the effects or size of the change. First, a point that applies to the Massachusetts ballot on psychedelics in particular, putting views into percentages, and getting accurate results from political polls are notoriously difficult tasks [2]. Therefore, the size of any effect is faced with whatever confounding variables make those tasks difficult.

Second, there could be some level Hawthorne effect [3] at play here, such that participants may report being (more) convinced because that's what (they think) is expected of them. I'm not familiar with the recruiting platforms they used, but if they're specialized in paid or otherwise professional surveys, I wonder if participants feel an obligation to perform well.

Third, and somewhat related to the above, participants could state they'd vote Y after initially reporting X preference, because they know it's a low-cost no-commitment claim. In other words, they can claim they'd now vote for Y without fear of judgement because it's a lab environment and an anonymous activity, but they can always go back to their original position once the actual vote happens. To show the size of the effect with respect to other things, researchers will have to make the stakes higher, or follow-up with participants after the vote and find out if/why they changed their mind (again).

Fourth, if one 6-minute-average conversation with a chatbot could convince an average voter, I wonder how much did they know about the issue/candidate being voted on. More cynically for the study, there may be much more at play with actual vote preference than a single dialectic presentation of facts. For example: salient events that happen in the period up to the election; emotional connection with the issue/candidate; personal experiences.

Still, this does not make the study flawed for not covering everything. We can learn a lot from this work, and kudos to the authors for publishing it.

[1] https://www.nature.com/articles/s41586-025-09771-9

[2] For example: https://www.brookings.edu/articles/polling-public-opinion-th...

[3] https://en.wikipedia.org/wiki/Hawthorne_effect

jacknews•6h ago
Reminds me of the Hypnodrones in the Universal Papaerclips clicker game.
ChrisArchitect•5h ago
Related:

Chatbots can sway political opinions but are 'substantially' inaccurate: study

https://news.ycombinator.com/item?id=46154066