frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Live Translation on AirPods Expands to the EU (IE)

https://www.apple.com/ie/newsroom/2025/11/live-translation-on-airpods-expands-to-the-eu/
1•rbanffy•43s ago•0 comments

BBC on Gaza-Israel: One Story, Double Standards

https://cfmm.org.uk/bbc-on-gaza-israel-one-story-double-standards/
1•stuaxo•1m ago•0 comments

Nvidia's H100 GPU Takes AI Processing to Space

https://spectrum.ieee.org/nvidia-h100-space
1•rbanffy•3m ago•0 comments

UK outperforms US in creating unicorns from early stage VC investment

https://www.cityam.com/uk-outperforms-us-in-creating-unicorns-from-early-stage-vc-investment/
3•mmarian•3m ago•0 comments

ClusterMAX 2.0: The Industry Standard GPU Cloud Rating System

https://newsletter.semianalysis.com/p/clustermax-20-the-industry-standard
1•MasterScrat•3m ago•0 comments

How to lead products through layoff fear

https://www.mindtheproduct.com/scary-times-how-to-lead-through-layoff-fear/
1•mooreds•4m ago•0 comments

Self-Replicating Probes Could Be Operating in the Solar System

https://www.universetoday.com/articles/self-replicating-probes-could-be-operating-right-now-in-th...
2•rbanffy•8m ago•0 comments

Antropocene

https://feralatlas.org
1•RoyBean•11m ago•0 comments

II. Leaflet of the White Rose

https://www.weisse-rose-stiftung.de/white-rose-resistance-group/leaflets-of-the-white-rose/ii-lea...
1•breppp•11m ago•0 comments

One porn platform made millions suing its viewers

https://www.theguardian.com/society/ng-interactive/2025/nov/04/strike-3-porn-copyright-lawsuits
3•belter•11m ago•0 comments

NASA instrument arrives at ISS to demonstrate quantum entanglement

https://spaceandtelescope.com/nasa-instrument-arrives-at-iss-to-demonstrate-quantum-entanglement/
1•belter•12m ago•0 comments

How Tiles Works – Tiles Privacy

https://www.blog.tiles.run/p/how-tiles-works
1•_feynon•13m ago•0 comments

Ask HN: Can people please stop commenting on whether a submission is AI?

1•AndrewDucker•14m ago•2 comments

Creating a New Embedded Rust Projects for NXP LPC55S69

https://mcuoneclipse.com/2025/10/19/creating-an-new-embedded-rust-projects-for-nxp-lpc55s69/
1•hasheddan•14m ago•0 comments

Satisfying Bazel's relative paths requirement in C++ toolchains

https://pigweed.dev/blog/09-bazel-relative-toolchain-paths.html
1•kaycebasques•15m ago•0 comments

OpenAI Wants Federal Backstop for New Investments [video]

https://www.wsj.com/video/openai-wants-federal-backstop-for-new-investments/4F6C864C-7332-448B-A9...
2•mkhattab•16m ago•0 comments

Bombshell report exposes how Meta relied on scam ad profits to fund AI

https://arstechnica.com/tech-policy/2025/11/bombshell-report-exposes-how-meta-relied-on-scam-ad-p...
2•duxup•16m ago•1 comments

What Did Medieval Peasants Know? (2022)

https://www.theatlantic.com/health/archive/2022/05/medieval-history-peasant-life-work/629783/
1•thinkingemote•16m ago•0 comments

Cursor – Sixty days with the AI coding startup

https://joincolossus.com/article/inside-cursor/
1•davidgomes•16m ago•0 comments

JanitorBench: A new LLM benchmark for multi-turn chats

https://about.janitorai.com/
15•shep101•17m ago•2 comments

Lightstep is shutting down March 1, 2026

https://docs.lightstep.com/changelog/eol-notice
4•acid__•20m ago•1 comments

Ford Considers Scrapping Electric Version of F-150 Truck

https://www.wsj.com/business/autos/ford-150-lightning-ev-decision-89dc0d84
5•JumpCrisscross•22m ago•0 comments

Show HN: Deepcon – Get the most accurate context for coding agents

https://deepcon.ai
3•ethanpark•23m ago•1 comments

New court docs put Sam Altman's honesty in spotlight again

https://www.fastcompany.com/91436157/court-filing-sam-altman-openai-anthropic
2•teachrdan•24m ago•0 comments

Show HN: Stingray Security – In-browser AI checking for phishing and scams

https://chromewebstore.google.com/detail/stingray-block-phishing-c/mnlnnbihfaomfcapnmllagoomffobfoj
1•imack•25m ago•0 comments

Bloodhound/GriffonAD: exploit automatically bad configurations in AD

https://github.com/shellinvictus/GriffonAD
1•mikeharper•26m ago•0 comments

Most Frequent Applesoft Basic Tokens

https://jtauber.github.io/CARC/2025/11/06/most-frequent-applesoft-basic-tokens/
2•jtauber•28m ago•0 comments

Crown Office – The Gazette

https://www.thegazette.co.uk/notice/4992102
2•bifftastic•29m ago•1 comments

Evaluating Control Protocols for Untrusted AI Agents

https://arxiv.org/abs/2511.02997
1•timini•30m ago•1 comments

From silicon to softmax: Inside the Ironwood AI stack

https://cloud.google.com/blog/products/compute/inside-the-ironwood-tpu-codesigned-ai-stack
1•mariuz•30m ago•0 comments
Open in hackernews

Benchmarking the Most Reliable Document Parsing API

https://www.tensorlake.ai/blog/benchmarks
12•calavera•1h ago

Comments

serjester•1h ago
This is just a company advertisement, not even one that’s well done. They didn’t benchmark any of the real leaders in the space (reducto, extend, etc) and left Gemini out of the first two tests, presumably because it was the best performer (while also being multiple orders of magnitude cheaper).
JLO64•1h ago
Personally I use OpenAI models via the API for transcription of PDF files. Is there a big difference between them and Gemini models?
diptanu•43m ago
Hey! I am the founder of Tensorlake. We benchmarked the models that our customers consider using in enterprises or regulated industries where there is a big need for processing documents for various automation. Benchmarking takes a lot of time so we focussed on the ones that we get asked about.

On Gemini and other VLMs - we excluded these models because they don't do visual grounding - aka they don't provide page layouts, bounding boxes of elements on the pages. This is a table stakes feature for use-cases customers are building with Tensorlake. It wouldn't be possible to build citations without bounding boxes.

On pricing - we are probably the only company offer a pure on-demand pricing without any tiers. With Tensorlake, you can get back markdown from every page, summaries of figures, tables and charts, structured data, page classification, etc - in ONE api call. This means we are running a bunch of different models under the hood. If you add up the token count, and complexity of infrastructure to build a complex pipeline around Gemini, and other OCR/Layout detection model I bet the price you would end up with won't be any cheaper than what we provide :) Plus doing this at scale is very very complex - it requires building a lot of sophisticated infrastructure - another source of cost behind modern Document Ingestion services.

ianhawes•33m ago
I just tested a non-English document and it rendered English text. Does your model not support anything other than English?
diptanu•14m ago
It does, we have users in Europe and Asia using it with non English languages. Can you please send me a message at diptanu at tensorlake dot ai, would love to see why it didn’t work.
coderintherye•27m ago
Google's Vertex API for document processing absolutely does bounding boxes. In fact, some of the document processors are just a wrap around Google's product.
diptanu•15m ago
OP mentioned Gemini and not Google’s Vertex OCR API which has very different performance and accuracy characteristics than Gemini
hotpaper75•43m ago
Thanks for mentioning them, indeed their post seem to only surface a couple of names in the field and maybe not the most relevant ones.
karakanb•36m ago
I have been recently looking into extracting a bunch of details from a set of legacy invoice PDFs and had a subpar experience. Gemini was the best among the ones that I tried, but even that missed quite a bit. I'll definitely give this a look.

It seems like such a crowded space and there are many tools doing document extraction, I wonder if there's anything particular pulling more attention into the space?

recursive4•23m ago
Curious how it compares to https://github.com/datalab-to/chandra
diptanu•11m ago
We haven’t texted Chandra yet, because it’s very new. Under the hood Tensorlake is very similar to Marker - it’s a pipeline based OCR API, we do layout detection, Text Recognition and Detection, Table Structure Understanding, etc. We then use VLMs to enrich the results. Our models are much bigger than marker, and thus takes a little longer to parse documents. We optimized for accuracy. We will have a faster API soon.