frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

SHARP, an approach to photorealistic view synthesis from a single image

https://apple.github.io/ml-sharp/
300•dvrp•6h ago•59 comments

Children with cancer scammed out of millions fundraised for their treatment

https://www.bbc.com/news/articles/ckgz318y8elo
203•1659447091•3h ago•145 comments

A linear-time alternative for Dimensionality Reduction and fast visualisation

https://medium.com/@roman.f/a-linear-time-alternative-to-t-sne-for-dimensionality-reduction-and-f...
56•romanfll•3h ago•14 comments

Quill OS: An open-source OS for Kobo's eReaders

https://quill-os.org/
273•Curiositry•9h ago•87 comments

The biggest heat pumps in the world

https://www.bbc.com/news/articles/c17p44w87rno
42•rayhaanj•1h ago•25 comments

Bonsai: A Voxel Engine, from scratch

https://github.com/scallyw4g/bonsai
50•jesse__•4h ago•6 comments

Erdős Problem #1026

https://terrytao.wordpress.com/2025/12/08/the-story-of-erdos-problem-126/
84•tzury•5h ago•8 comments

O'saasy License Agreement

https://osaasy.dev/
40•d3w1tt•3h ago•38 comments

A2UI: A Protocol for Agent-Driven Interfaces

https://a2ui.org/
4•makeramen•55m ago•0 comments

JetBlue flight averts mid-air collision with US Air Force jet

https://www.reuters.com/world/americas/jetblue-flight-averts-mid-air-collision-with-us-air-force-...
270•divbzero•11h ago•161 comments

Creating C closures from Lua closures

https://lowkpro.com/blog/creating-c-closures-from-lua-closures.html
31•publicdebates•4d ago•2 comments

Internal RFCs saved us months of wasted work

https://highimpactengineering.substack.com/p/the-illusion-of-shared-understanding
21•romannikolaev•5d ago•7 comments

8M users' AI conversations sold for profit by "privacy" extensions

https://www.koi.ai/blog/urban-vpn-browser-extension-ai-conversations-data-collection
480•takira•7h ago•151 comments

“Are you the one?” is free money

https://blog.owenlacey.dev/posts/are-you-the-one-is-free-money/
337•samwho•4d ago•65 comments

Native vs. emulation: World of Warcraft game performance on Snapdragon X Elite

https://rkblog.dev/posts/pc-hardware/pc-on-arm/x86_versus_arm_native_game/
77•geekman7473•10h ago•31 comments

7 Years, 2 Rebuilds, 40K+ Stars: Milvus Recap and Roadmap

https://milvus.io/blog/milvus-exceeds-40k-github-stars.md
17•Fendy•5d ago•5 comments

Economics of Orbital vs. Terrestrial Data Centers

https://andrewmccalip.com/space-datacenters
116•flinner•12h ago•167 comments

I'm a Tech Lead, and nobody listens to me. What should I do?

https://world.hey.com/joaoqalves/i-m-a-tech-lead-and-nobody-listens-to-me-what-should-i-do-e16e454d
12•joaoqalves•33m ago•1 comments

Show HN: I designed my own 3D printer motherboard

https://github.com/KaiPereira/Cheetah-MX4-Mini
65•kaipereira•1w ago•15 comments

Essential Semiconductor Physics [pdf]

https://nanohub.org/resources/43623/download/Essential_Semiconductor_Physics.pdf
186•akshatjiwan•2d ago•7 comments

Rollstack (YC W23) is hiring multiple software engineers (TypeScript) US/Canada

https://www.ycombinator.com/companies/rollstack-2/jobs/QPqpb1n-software-engineer-typescript-us-ca...
1•yjallouli•8h ago

Mark V Shaney

https://en.wikipedia.org/wiki/Mark_V._Shaney
13•djoldman•4d ago•1 comments

Light intensity steers molecular assemblies into 1D, 2D or 3D structures

https://phys.org/news/2025-11-intensity-molecular-1d-2d-3d.html
26•PaulHoule•5d ago•3 comments

In Defense of Matlab Code

https://runmat.org/blog/in-defense-of-matlab-whiteboard-style-code
123•finbarr1987•3d ago•128 comments

Chafa: Terminal Graphics for the 21st Century

https://hpjansson.org/chafa/
161•birdculture•15h ago•24 comments

Umbrel – Personal Cloud

https://umbrel.com
187•oldfuture•14h ago•100 comments

Secret Documents Show Pepsi and Walmart Colluded to Raise Food Prices

https://www.thebignewsletter.com/p/secret-documents-show-pepsi-and-walmart
421•connor11528•12h ago•101 comments

The appropriate amount of effort is zero

https://expandingawareness.org/blog/the-appropriate-amount-of-effort-is-zero/
126•gmays•14h ago•72 comments

A kernel bug froze my machine: Debugging an async-profiler deadlock

https://questdb.com/blog/async-profiler-kernel-bug/
95•bluestreak•13h ago•17 comments

Understanding carriage

https://seths.blog/2025/12/understanding-carriage/
50•herbertl•5d ago•12 comments
Open in hackernews

SHARP, an approach to photorealistic view synthesis from a single image

https://apple.github.io/ml-sharp/
299•dvrp•6h ago

Comments

brcmthrowaway•5h ago
So this is the secret sauce behind Cinematic mode. The fake bokeh insanity has reached its climax!
duskwuff•4h ago
As well as their "Spatial Scene" mode for lock screen images, which synthesizes a mild parallax effect as you move the phone.
Terretta•4h ago
It's available for everyday photos, portraits, everything, not just lock screens.
spike021•4h ago
you can also press the button while viewing a photo in the Photos app to see this.
calvinmorrison•4h ago
I understand AI for reasoning, knowledge, etc. I haven't figured out how anyone wants to spend money for this visual and video stuff. It just seems like a bad idea.
accurrent•4h ago
Simulation. It takes a lot of effort today to bring up simulations in various fields. 3 D programming is very nontrivial and asset development is extremely expensive. If I have a workspace I can take a photo of and just use it to generate a 3d scene I can then use it in simulations to test ideas out. This is particularly useful in robotics and industrial automation already.
re-thc•4h ago
Do people not spend on entertainment? Commercials? It's probably less of a bad idea than knowledge. AI giving a bad visual has less negatives than giving the wrong knowledge leading to the wrong decision.
rv3392•3h ago
This specific paper is pretty different to the kind of photo/video generation that has been hyped up in recent years. In this case, I think this might be what they're using for the iOS spatial wallpaper feature, which is arguably useless but is definitely an aesthetic differentiator to Android devices. So, it's indirectly making money.
yodon•4h ago
> photorealistic 3D representation from a single photograph in less than a second
arjie•4h ago
This is incredibly cool. It's interesting how it fails in the section where you need to in-paint. SVC seems to do that better than all the rest, though not anywhere close to the photorealism of this model.

Is there a similar flow but to transform either a video/photo/NeRF of a scene into a tighter, minimal polygon approximation of it. The reason I ask is that it would make some things really cool. To make my baby monitor mount I had to knock out the calipers and measure the pins and this and that, but if I could take a couple of photos and iterate in software that would be sick.

Geee•4h ago
This is great for turning a photo into a dynamic-IPD stereo pair + allows some head movement in VR.
SequoiaHope•4h ago
Ah and the dynamic IPD component preserves scale?
benatkin•4h ago
That is really impressive. However, it was a bit confusing at first because in the koala example at the top, the zoomed in area is only slightly bigger than the source area. I wonder why they didn't make it 2-3x as big in both axes like they did with the others.
yodon•4h ago
See also Spaitial[0] which announced today full 3D environment generation from a single image

[0]https://www.spaitial.ai/

andsoitis•4h ago
Why are all their examples of rooms?

Why no landscape or underwater scenes or something in space, etc.?

jaccola•4h ago
Constrained environments are much simpler.

I believe this company is doing image (or text) -> off the shelf image model to generate more views -> some variant of gaussian splatting.

So they aren't really "generating" the world as one might imagine.

boguscoder•4h ago
Requires email to view anything, that’s sad
dag11•4h ago
I'm confused, does it actually generate environments from photographs? I can't view the galleries since I didn't sign up for emails but all of the gallery thumbnails are AI, not photos.
jrflowers•2h ago
> I'm confused, does it actually generate environments from photographs?

It’s a website that collects people’s email addresses

avaer•2h ago
The best I've seen so far is Marble from World Labs, though that gives you a full 360 environment and takes several minutes to do so.
superfish•4h ago
"Unsplash > Gen3C > The fly video" is nightmare fuel. View at your own risk: https://apple.github.io/ml-sharp/video_selections/Unsplash/g...
ghurtado•4h ago
Seth Brundle has entered the chat.
Traubenfuchs•2h ago
Early AI „everything turns into dog heads“ vibes. Beautiful.
drcongo•55m ago
I miss those. Anyone know if it's still possible to get the models etc. needed to generate them?
Traubenfuchs•28m ago
I wish there was an archive of all those melty dreamscapes.

https://m.youtube.com/watch?v=DgPaCWJL7XI&t=1s&pp=2AEBkAIB0g...

https://www.youtube.com/watch?v=X0oSKFUnEXc

schneehertz•1h ago
san check, 1d10
harhargange•4h ago
TMPI looks just as good if not better.
jjcm•4h ago
Disagree - look at the sky in the seaweed shot. It doesn't quite get the depth right in anything, and the edges of things look off.
shwaj•3h ago
Agreed. The head of the fly also seems to have weird depth.
wfme•3h ago
Have a look through the rest of the images. TMPI has some pretty obvious shortcomings in a lot of them.

1. Sky looks jank 2. Blurry/warped behind the horse 3. The head seems to move a lot more than the body. You could argue that this one is desirable 4. Bit of warping and ghosting around the edges of the flowers. Particularly noticeable towards the top of the image. 5. Very minor but the flowers move as if they aren't attached to the wall

tartoran•4h ago
Impressive but something doesn't feel right to me.. Possibly too much sharpness, possibly a mix of cliches, all amplified at once.
remh•4h ago
Enhance! https://www.youtube.com/watch?v=LhF_56SxrGk
mvandermeulen•3h ago
I thought this was going to be the Super Troopers version
moondev•4h ago
cuda gpu only

https://github.com/apple/ml-sharp#rendering-trajectories-cud...

delis-thumbs-7e•3h ago
Interestingly Apple’s own models don’t work on MPS. Well, I guess you just have to wait for few years..
matthewmacleod•2h ago
This is specifically only for video rendering. The model itself works across GPU, CPU, and MPS.
diimdeep•1h ago
No, model works without CUDA then you have .ply that you can drop into gaussian splatter viewer like https://sparkjs.dev/examples/#editor

CUDA is needed to render side scrolling video, but there is many ways to do other things with result.

Leptonmaniac•4h ago
Can someone ELI5 what this does? I read the abstract and tried to find differences in the provided examples, but I don't understand (and don't see) what the "photorealistic" part is.
eloisius•3h ago
From a single picture it infers a hidden 3D representation, from which you can produce photorealistic images from slightly different vantage points (novel views).
avaer•3h ago
There's nothing "hidden" about the 3d represenation. It's a point cloud (in meters) with colors, and a guess at the the "camera" that produced it.

(I am oversimplifying).

eloisius•3h ago
Hidden in the sense of neural net layers. I mean intermediary representation.
avaer•3h ago
Right.

I just want to emphasize that this is not a NERF where the model magically produces an image from an angle and then you ask "ok but how did you get this?" and it throws up its hands and says "I dunno, I ran some math and I got this image" :D.

uh_uh•2h ago
"Hidden" or "latent" in a context like this just means variables that the algo is trying to infer because it doesn't have direct access to them.
ares623•3h ago
Takes a 2D image and allows you to simulate moving the angle of the camera with correct-ish parallax effect and proper subject isolation (seems to be able to handle multiple subjects in the same scene as well)

I guess this is what they use for the portrait mode effects.

p-e-w•3h ago
Agreed, this is a terrible presentation. The paper abstract is bordering on word salad, the demo images are meaningless and don’t show any clear difference to the previous SotA, the introduction talks about “nearby” views while the images appear to show zooming in, etc.
emsign•3h ago
Imagine history documentaries where they take an old photo and free objects from the background and move them round giving the illusion of parallax movement. This software does that in less than a second, creating a 3D model that can be accurately moved (or the camera for that matter) in your video editor. It's not new, but this one is fast and "sharp".

Gaussian splashing is pretty awesome.

kurtis_reed•2h ago
What are free objects?
ferriswil•2h ago
The "free" in this case is a verb. The objects are freed from the background.
Retr0id•2h ago
Until your comment I didn't realise I'd also read it wrong (despite getting the gist of it). Attempted rephrase of the original sentence:

Imagine history documentaries where they take an old photo, free objects from the background, and then move them round to give the illusion of parallax.

necovek•1h ago
I'd suggest a different verb like "detach" or "unlink".
tzot•1h ago
> Imagine history documentaries where they take an old photo, free objects from the background

Even using commas, if you leave the ambiguous “free” I suggest you prefix “objects” with “the” or “any”.

derleyici•3h ago
It turns a single photo into a rough 3D scene so you can slightly move the camera and see new, realistic views. "Photorealistic" means it preserves real textures and lighting instead of a flat depth effect. Similar behavior can be seen with Apple's Spatial Scene feature in the Photos app: https://files.catbox.moe/93w7rw.mov
avaer•3h ago
It makes your picture 3D. The "photorealistic" part is "it's better than these other ways".
carabiner•2h ago
Black Mirror episode portraying what this could do: https://youtu.be/XJIq_Dy--VA?t=14. If Apple ran SHARP on this photo and compared it to the show, that would be incredible.

Or if you prefer Blade Runner: https://youtu.be/qHepKd38pr0?t=107

avaer•4h ago
Is there a link with some sample gaussian splat files coming from this model? I couldn't find it.

Without that that it's hard to tell how cherry-picked the NVS video samples are.

EDIT: I did it myself, if anyone wants to check out the result (caveat, n=1): https://github.com/avaer/ml-sharp-example

derleyici•3h ago
Apple's Spatial Scene in the Photos app shows similar behavior, turning a single photo into a small 3D scene that you can view by tilting the phone. Demo here: https://files.catbox.moe/93w7rw.mov
Traubenfuchs•2h ago
It‘s awful and often creates a blurry mess in the imaginated space behind the object.

Photoshop content aware fill could do equally or better many years ago.

diimdeep•3h ago
Works great, model file is 2.8 GB, on M2 rendering took a few seconds, result is guassian .ply file but repo implementation requires CUDA card to render video, I have used one of webgl live renderers from here https://github.com/scier/MetalSplatter?tab=readme-ov-file#re...
Dumbledumb•51m ago
In Chapter D.7 they describe: "The complex reflection in water is interpreted by the network as a distant mountain, therefore the water surface is broken."

This is really interesting to me because the model would have to encode the reflection as both the depth of the reflecting surface (for texture, scattering etc) as well as the "real depth" of the reflected object. The examples in Figure 11 and 12 already look amazing.

Long tail problems indeed.