frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Are AI agents ready for the workplace? A new benchmark raises doubts

https://techcrunch.com/2026/01/22/are-ai-agents-ready-for-the-workplace-a-new-benchmark-raises-do...
1•PaulHoule•4m ago•0 comments

Show HN: AI Watermark and Stego Scanner

https://ulrischa.github.io/AIWatermarkDetector/
1•ulrischa•5m ago•0 comments

Clarity vs. complexity: the invisible work of subtraction

https://www.alexscamp.com/p/clarity-vs-complexity-the-invisible
1•dovhyi•6m ago•0 comments

Solid-State Freezer Needs No Refrigerants

https://spectrum.ieee.org/subzero-elastocaloric-cooling
1•Brajeshwar•6m ago•0 comments

Ask HN: Will LLMs/AI Decrease Human Intelligence and Make Expertise a Commodity?

1•mc-0•7m ago•1 comments

From Zero to Hero: A Brief Introduction to Spring Boot

https://jcob-sikorski.github.io/me/writing/from-zero-to-hello-world-spring-boot
1•jcob_sikorski•8m ago•0 comments

NSA detected phone call between foreign intelligence and person close to Trump

https://www.theguardian.com/us-news/2026/feb/07/nsa-foreign-intelligence-trump-whistleblower
5•c420•8m ago•0 comments

How to Fake a Robotics Result

https://itcanthink.substack.com/p/how-to-fake-a-robotics-result
1•ai_critic•9m ago•0 comments

It's time for the world to boycott the US

https://www.aljazeera.com/opinions/2026/2/5/its-time-for-the-world-to-boycott-the-us
1•HotGarbage•9m ago•0 comments

Show HN: Semantic Search for terminal commands in the Browser (No Back end)

https://jslambda.github.io/tldr-vsearch/
1•jslambda•9m ago•1 comments

The AI CEO Experiment

https://yukicapital.com/blog/the-ai-ceo-experiment/
2•romainsimon•11m ago•0 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
3•surprisetalk•14m ago•0 comments

MS-DOS game copy protection and cracks

https://www.dosdays.co.uk/topics/game_cracks.php
3•TheCraiggers•15m ago•0 comments

Updates on GNU/Hurd progress [video]

https://fosdem.org/2026/schedule/event/7FZXHF-updates_on_gnuhurd_progress_rump_drivers_64bit_smp_...
2•birdculture•16m ago•0 comments

Epstein took a photo of his 2015 dinner with Zuckerberg and Musk

https://xcancel.com/search?f=tweets&q=davenewworld_2%2Fstatus%2F2020128223850316274
7•doener•16m ago•2 comments

MyFlames: Visualize MySQL query execution plans as interactive FlameGraphs

https://github.com/vgrippa/myflames
1•tanelpoder•18m ago•0 comments

Show HN: LLM of Babel

https://clairefro.github.io/llm-of-babel/
1•marjipan200•18m ago•0 comments

A modern iperf3 alternative with a live TUI, multi-client server, QUIC support

https://github.com/lance0/xfr
3•tanelpoder•19m ago•0 comments

Famfamfam Silk icons – also with CSS spritesheet

https://github.com/legacy-icons/famfamfam-silk
1•thunderbong•20m ago•0 comments

Apple is the only Big Tech company whose capex declined last quarter

https://sherwood.news/tech/apple-is-the-only-big-tech-company-whose-capex-declined-last-quarter/
2•elsewhen•23m ago•0 comments

Reverse-Engineering Raiders of the Lost Ark for the Atari 2600

https://github.com/joshuanwalker/Raiders2600
2•todsacerdoti•24m ago•0 comments

Show HN: Deterministic NDJSON audit logs – v1.2 update (structural gaps)

https://github.com/yupme-bot/kernel-ndjson-proofs
1•Slaine•28m ago•0 comments

The Greater Copenhagen Region could be your friend's next career move

https://www.greatercphregion.com/friend-recruiter-program
2•mooreds•28m ago•0 comments

Do Not Confirm – Fiction by OpenClaw

https://thedailymolt.substack.com/p/do-not-confirm
1•jamesjyu•29m ago•0 comments

The Analytical Profile of Peas

https://www.fossanalytics.com/en/news-articles/more-industries/the-analytical-profile-of-peas
1•mooreds•29m ago•0 comments

Hallucinations in GPT5 – Can models say "I don't know" (June 2025)

https://jobswithgpt.com/blog/llm-eval-hallucinations-t20-cricket/
1•sp1982•29m ago•0 comments

What AI is good for, according to developers

https://github.blog/ai-and-ml/generative-ai/what-ai-is-actually-good-for-according-to-developers/
1•mooreds•29m ago•0 comments

OpenAI might pivot to the "most addictive digital friend" or face extinction

https://twitter.com/lebed2045/status/2020184853271167186
1•lebed2045•30m ago•2 comments

Show HN: Know how your SaaS is doing in 30 seconds

https://anypanel.io
1•dasfelix•31m ago•0 comments

ClawdBot Ordered Me Lunch

https://nickalexander.org/drafts/auto-sandwich.html
3•nick007•32m ago•0 comments
Open in hackernews

Meta Is Going to Let Job Candidates Use AI During Coding Tests

https://www.wired.com/story/meta-ai-job-interview-coding/
55•geox•6mo ago

Comments

joshstrange•6mo ago
I'm currently running the interview process for the company I work for and I allow AI on the coding tests for the same reason I allow googling/etc:

I want to simulate as close to the real environment you will be coding in if you come and work here.

My "rules" on AI/LLMs at work is that you can use whatever tools you want (much like SO before) but any code you commit, you are responsible for (also unchanged from before LLMs). I never want to hear "the LLM wrote that" when asked about how something works.

If you can explain, modify, build on top of the code that an LLM spits out then fine by me. We don't require LLM usage (no token quotas or silly things like that) nor do we disallow LLMs. At the end of the day they are all tools.

I've always found leet coding, whiteboarding, memorizing algorithms, etc to just be a silly form of hazing and a bad indicator of how well the employee will perform if you hire them. In the same way I still think my college was stupid for making us write, on paper, a C program with all the correct headers/syntax/etc for an exam or how I got a 90/100 on my final SQL exam because I didn't end my, otherwise perfect, queries with a semicolon.

freedomben•6mo ago
I run our interviewing process as well and fully agree with you. An interview process that most closely mirrors "real work" gets much better results. Some of the best people we've had were really terrible at algorithms and leet code, but wrote quality code (that they tested!), had strong work ethics, were willing to be around off-hours for the occasional production outage, took ownership/accountability for their tasks, and were enjoyable people to work with/around. It's good to have at least one person on the team who is good with algorithms and stuff for those times when that comes up, but having that be a requirement for everyone on the team is an anti-pattern IMHO
joshstrange•6mo ago
I completely agree.

We've had almost every candidate (even some we _didn't_ hire) thank us for the interview process and how we went out of our way to make it stress-free and representative of the job. I'm not interested in tricking people or trying to show off "look how smart we are and how hard our test was"... puke. I want to work with solid performers who don't play games, why would I start that relationship off by playing a bunch of stupid games (silly interview tests/process)?

And here's the thing, even with our somewhat easier/less-stress testing process, we still got high quality hires and have had very few cases where we have later had to let someone go due to not being able to do the job. This isn't a "hire fast, fire fast" situation.

jghn•6mo ago
I agree with you that the real rubric is that I'd expect the candidate to be able to explain everything they did and why. For the most part I don't care how they got there.

Was thinking about this recently in a conversation about whether or not the candidate should have a full screen share so that interviewers could see what they're looking up, etc. I realized that I'd reached the point where I now trust my fellow interviewers less than I do the candidate at these bits. Personally I think it's nice to see how people are leveraging tools, but too often I see interviewers dig candidates anyways on the specifics.

I've now found this across multiple companies now has been that when coding exercises are "open book" like this that other interviewers *are* actively judging people based on what they're googling, and these days LLMing. If you're going to allow google but then fall back to "I can't believe they had to look *that* up!", that's not how it works.

sokoloff•6mo ago
I think “how to write for loop in $LANGUAGE” is something that it’s fair to judge a candidate on if they claim (or the job requires) recent experience in $LANGUAGE.
jghn•6mo ago
Maybe. I've been coding professionally for a few decades now. I have always looked stuff up constantly. It's a lot more seamless now due to first search engines and later the ever improving autocomplete. To be honest, I think half the time I just do it to do it. But for instance I bounce between languages so often that I can never remember the fine details of syntax. To use your example, things like for loops, matches, and the like are things that I tend to double check.

And that's the point. Everyone has their quirks, and if you're going to claim you want them to work they way they normally work, who cares? If they're able to turn out quality code at a reasonable clip, who cares what they need to look up.

zamalek•6mo ago
Makes sense. It's like we were prohibited from using calculators in school when I was growing up. If you're testing for on-the-job performance then you should have on-the-job tools.
lesuorac•6mo ago
Well, while I agree with the second half of your statement. It only applies to when you're testing on-the-job performance. If you want to test if somebody learned multiplication then you do need to take away a calculator otherwise you're just testing if they can multiply.

I'm not sure when an employer should really care about whether their cashier can add/subtract correctly vs just use the output from the machine. But in the education setting where you're making a claim that students will learn X you should be testing X.

psunavy03•6mo ago
The entertaining thing for at least Xennials and before is how our teachers used to tell us "out in the world, you won't just always have a calculator in your pocket!"

And yet here we are . . .

tracker1•6mo ago
I'm mixed on this one... In terms of schooling, it's about understanding how math works... not how to operate a calculator.

Similar to graphing calculators vs basic scientific calculators... if you're trying to demonstrate you understand an equation, having the calculator do that for you doesn't show you understand what it is doing.

Where in a job, you're probably not going to need to rebalance a B-Tree in any practical sense. For that matter, in my own experience, most actual development rarely has people optimizing their RDBMS usage in any meaningful way, for better or worse.

It's the difference between learning and comprehension and delivering value. Someone has to understand how the machine works to fix it when it breaks. And there are jobs at every level. Most of them aren't that constrained in practice though.

zamalek•6mo ago
> Similar to graphing calculators vs basic scientific calculators... if you're trying to demonstrate you understand an equation, having the calculator do that for you doesn't show you understand what it is doing.

That isn't entirely true and brings up an important nuance. I understood algebra and calculus extremely well but I routinely got half-score for fudging a calculation or somesuch because multiplication has a foundation of rote knowledge: you need to memorize you ten-times table. To this day (at 37) I continue to use algebra to do basic multiplication in my head (7*6 = 7*5+7 = 7/2*10+7 = 35+7 = 42).

Sure, using a graphing calculators solve the problem isn't demonstrating understanding, but 90s kids were simply asking for the basic calculator with 2 registers (display and M) and basic math and "sci" (sin, cos, sqrt, etc.) operators.

Preventing use of basic calculators does nothing to demonstrate knowledge of math at all, it actually hinders it.

tracker1•6mo ago
There's nothing wrong with breaking down 7*6 how you did, I do the same. It's not far from how common core, aka new math actually tries to teach it. It actually shows you understand the mechanics, where using a calculator to come up with the same answer does not... it only shows you know how to push buttons into a calculator.
zamalek•6mo ago
> using a calculator to come up with the same answer does not

I'm not sure why you're still talking about this. A basic "non-graphing" (90s "scientific calculator") is not capable of doing all the work for you. All it did was the basic things like add, mul, sub, sin, cos, etc. I am referring to one of these[1].

> It's not far from how common core, aka new math actually tries to teach it. It actually shows you understand the mechanics

I understand the mechanics of mathematics but still fared more poorly than I should have because I messed up a simple op like multiplication more often than I should have. My point is that had I had access to a basic calculator I would have scored significantly better in school. I went from Bs to As (80%, not sure what that is in GPA) in university by mere virtue of being able to use a simple calculator.

Again, not being able to use a [basic] calculator tests a signal that isn't actually important. Not being able to use Google/AI/whatever in your interview similarly tests an unimportant signal. The most important hire signal for a SDE is actually how they operate in a team and how they communicate with stakeholders, which these coding tests don't do.

[1]: https://upload.wikimedia.org/wikipedia/commons/4/4f/FX-77.JP...

tracker1•6mo ago
I'm actually against the leetcode style coding tests. They're exceedingly lopsided against more experienced developers further from school or younger learning where memorizing various memory structures is/was more emphasized.

I prefer a paired interview scenario with actual questions about approach to working through problems. As an interviewee I do horribly with the leetcode questions, there's no opportunity for clarifying questions, and some systems penalize you for looking away from the screen even. It's unnerving and I refuse to do them any more.

I also don't mind the "here's a task/challenge that should take a half hour or so." where you simply create a solution, commit to github, etc.

As an interviewer I prefer the walkthrough interviews as well... I usually don't get into pedantic exercises, or language/platform nuances unless the applicant uses "expert" on more than one or two subjects.

MichaelNolan•6mo ago
I'm curious what they will come up with. For smaller companies, leetcode style tests probably aren't the best, but for large companies that hire 10s of thousands of devs, leetcode style tests have a lot of good qualities. People who criticize leetcode usually prefer take home projects, pair programming with a current employee, "just a conversation" or something else, but these all have serious drawbacks at scale. Despite leetcode's flaws it has a lot of benefits:

* Objective and clear grading criteria.

* Unlimited supply of questions.

* Easily time bound to 30 min or 1 hr.

* Legal - i.e., no worries about disparate impact.

* Programming language/framework agnostic.

sp527•6mo ago
I can't even begin to imagine what sort of mind could observe the quality of Big Tech's software output and conclude that there's nothing wrong with their hiring process.
qwerpy•6mo ago
My problem with Big Tech software isn’t code quality, it’s the deliberate user-hostile decisions they make. Leetcode-style interviews appear to be doing fine at getting people who can write code.
paxys•6mo ago
Now look at their market cap.
saubeidl•6mo ago
It's almost as if capitalism wasn't a good system for deciding.. well, anything, really.
triceratops•6mo ago
Big Tech's software is faster and less buggy than the median software product.
artyom•6mo ago
Source?

Big tech software is successful and runs at scale.

I've got anecdotal experience in both worlds and no. Big tech software isn't faster (what you have is way more compute resources usually), and the claim about "less buggy" gives me goosebumps.

triceratops•6mo ago
> Source?

All the software I use. Netflix works perfectly every time. HBO Max is garbage. Amazon's website and app are pretty good, although the actual goods sold are trash. Costco is exactly the other way around.

artyom•6mo ago
Apples to oranges. You're not comparing big tech to "normal-size" tech (in the context of the original article), you're comparing tech companies with the biggest budgets in the world, with companies that aren't tech at all. In your example, one is a media company, the other is wholesale/retail one.
lienne•6mo ago
And still these engineers cannot build anything. If you were to ask if I would hire any of my Google colleagues as a startup founder, I would say never.
reverendsteveii•6mo ago
interviews should be about determining whether you can use the tools available to deliver the desired product, not some sort of purity test to determine whether you can build without tools. also I want to see you interact w AI because as of right now it's both an incredible and deeply flawed tool and your ability to recognize when it's about to walk you off a cliff is of increasing importance as we discover the limits of what it can do.
bgwalter•6mo ago
Zuckerberg has no clue about software development. He always wants young people, huge bullpen offices, moonshot projects like the Metaverse that fail.

Facebook open source software does not have great code quality. In some projects that always have been a huge mess they are now adding claude.md files to guide their beloved "AI". They did not add these files for humans before.

I think Facebook software is a lost case where it does not matter if you perform the weekly rewriting by LLM or by kLOC driven humans.

dangus•6mo ago
And yet they’re wildly successful, including the wide level of adoption of their open source projects like React.

Meta’s profit per employee is over $250,000, higher than Qualcomm. There is no Meta competitor in any of their verticals that has a larger customer base.

It seems to me that your definition of “software quality” and “lost cause” is factually wrong based on the metrics of success that matter.

And in any event, it is an engineer’s tunnel vision fallacy to believe that software quality is the most important factor to stakeholders. People will prefer buggy/slow software that solves their problem over fast/stable software that fails to solve their problem.

mepian•6mo ago
> higher than Qualcomm

That’s a pretty low bar for a software company.

dangus•6mo ago
Qualcomm basically has a monopoly in many of their product lines. They’re like the Nvidia of 5G modems and high end ARM chips. They’re also an insane patent licensing machine.
thefaux•6mo ago
And are almost never offered fast/stable software that solves their problem. When they are, it will come with aggressive fud by the competition or they will just lobby the government to mandate the incidental complexity of their own software so that the competition cannot even enter the market without being brought down to their level.
supportengineer•6mo ago
He is a business man. All business men make money by exploiting a resource.
bwfan123•6mo ago
Meta interviews in the past have been straight regurgitation of leetcode. They want to measure your flow and typing speed (sarcasm).

So, the interview can now be 2 leetcode hards in 20 min. Earlier, it was typing solution code from rote memory. Now it is furious vibe-coding copy-pasta from one browser window to another.

More seriously, what will the new questions look like ? In the age of LLMs how does one measure ability objectively. Is it with extreme underspecification of the probem ?

dangus•6mo ago
It’s going to be just like the Google/StackOverflow era where your value as an engineer is based on your resourcefulness.

The value of an employee who says “I don’t know how to do that” or “I’ll need to ask my coworkers for help” versus one that says “I am sure I can figure out just about anything by googling” is night and day, and I think the same is true with AI.

Half of the battle is going to be knowing what to ask for.

Lastly I’d like to point out that it makes general sense to test people on the real tools they’ll be using to get their work done. E.g., you wouldn’t get much value testing a bus driver on using a manual transmission if all your buses are automatic. Most corporate leaders are expecting and even demanding that their employees use AI tools to get their work done.

thefaux•6mo ago
I don't know how to do that is a great answer that I wish people would give more often.

> Most corporate leaders are expecting and even demanding that their employees use AI tools to get their work done.

Imagine for a second that I am an aspiring monopolist. Wouldn't a great way to achieve this be to make people believe that I am their agent, when I am really their adversary? Would I really offer them in good faith a tool that they could use to compete against me? Or would I try to get them to accept a trojan horse that destroys their profits from within. Once, I have siphoned enough profit to significantly damage the business, I can come in, borrow money to buy the company that I basically destroyed and then sell off the parts to an ignorant buyer who doesn't realize who badly I have wounded this once good business, or I just write off the loss for tax purposes.

garciasn•6mo ago
https://archive.is/9faUx
thr0w•6mo ago
This makes sense in principle, but then how do you do technical evaluation? I'm generally most interested in hearing the candidate think out loud about the problem, and explain their approach. LLMs do the "thinking", and also explain the approach.
add-sub-mul-div•6mo ago
It feels like a slow generational evolution beyond considering one employee different from another. They're biding their time until the AI can babysit itself.
ndriscoll•6mo ago
As always, it requires a knowledgeable interviewer. Is the candidate using it as a slightly advanced autocomplete? That's fine. Are they relying on it to design things for them/can't articulate out loud what they're hoping it will produce for them/can't identify and articulate when and how it does something stupid or wrong or differs from how they were thinking of doing it? Then they don't know what they're doing.

I do peer coding with people at work and we have copilot on. We'll discuss what we're trying to accomplish and make little comments out loud along the lines of "sure, almost" or "not quite, copilot" when it generates something. It's obvious whether someone is working in a way where they know what they want the tool to do.

tracker1•6mo ago
It's funny but I've been turning Github CoPilot off more often than on... it's absolutely been fantastic with some boilerplate SQL, and other tedious things, but any time I've tried doing complex things it seems to just get in the way more than help.

As to interviews, I'd be happier if whatever tool they're using actually had working API code sense working... so many of the tools I've had to use in interviews just had broken LSP results. I've also had issue when a given question requires specific domain knowledge over a third party library as part of the question.

marssaxman•6mo ago
For sure - I don't care about the code, I care about watching the applicant work, so I can decide whether I would want to work with them - that being the whole purpose of a job interview. If the "work" consists of delegating everything to a robot, I learn nothing and the interview is a waste of time.
gronglo•6mo ago
Why is Meta still interviewing SWEs, is my question. Surely, at this point, an agent can beat any human candidate at any task. What am I missing?
paxys•6mo ago
You are missing the fact that an agent cannot beat any human candidate at any task.
gronglo•6mo ago
Mark Zuckerberg himself predicted that AI would replace all mid-level engineers at Meta by the end of this year, with senior engineers continuing to work on tasks such as planning and architecture. With the release of Claude 4 it feels like, if anything, he was too conservative with his prediction — we're already there, and Mark must know it. So what's he playing at?
paxys•6mo ago
Look at Zuckerberg's actions, not what he says in podcasts to shill his AI tools and boost his company's stock price. Meta is currently hiring hundreds of software engineers every month and paying $200-300k/yr for entry level and $700k+ for senior roles. Why do you think that is?
saubeidl•6mo ago
AI snake oil salesman talks up AI snake oil, news at 11.
paxys•6mo ago
This is born out of necessity. I have seen some of the coding interview "cheating" tools in the market today and it is ridiculous just how good they are at helping a candidate fake not just the answer but their entire thought process. Leetcode style interviewing is basically dead, at least if done remotely.

The only way to fight this from the employer side is to embrace these tools and change your evaluation criteria.

supportengineer•6mo ago
This escalating war will only result in them hiring those most skillful at cheating.
msgodel•6mo ago
If you're good at interviewing I really don't think AI tools should make a difference tbh.
SkyPuncher•6mo ago
I do interviews for my company. We allow AI (and even encourage it). Use or lack of use of AI has zero correlation with candidate performance. In fact, I’d say there’s a slight negative trend against those who use AI.

We have candidates build a _very_ simple fullstack app. Think something like a TODO app, but only like 2 core functions. This is basic CRUD 101 stuff.

I’ve seen a boatload of candidates use AI only to flame out after the first round of prompting. They literally get stuck and can’t move forward.

The good candidates clearly know their fundamentals. They’re making intentional decisions, telling the AI scoped actions to take, and reviewing pretty much everything the AI generates. They reject code they don’t like. They explain why code is good or bad. They debug easily and quickly.

LLMs are very, very talented but context still matters. They can’t build proper understanding of the non-technical components. They can’t solve things in the simplest way possible. They can’t decide that to explain to the interviewer “if performance is a concern, I’d do X but it’s a time-bound interview so I’m going to Y. Happy to do X if you need”.

dekhn•6mo ago
Interesting. I interviewed at Meta a decade ago and was put off by the process (inexperienced devs who memorized leetcode medium problems). But I still get daily requests from their recruiters to try again (I am a senior leader in ML engineering). I would need to spend a fair amount of time thinking about what it would be like to do an interview with LLMs.

But honestly, I'd rather spend that time on figuring out how to use LLMs to interview people better (for example I already had an LLM write a collaboritive web editor with built-in code runner, so I don't need to license coderpad). I could see updating my prompt to have the coding agent generate a text box for entering prompts during the interview. Either way, I still expect candidates to be able to explain what a hash table is in their own words.