frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Quick Look Source Code in Finder on macOS

https://anybox.ltd/source-code-preview
1•francisfeng•1m ago•0 comments

Against Vibes: When Is a Generative Model Useful

https://www.williamjbowman.com/blog/2026/03/05/against-vibes-when-is-a-generative-model-useful/
1•takira•2m ago•0 comments

Show HN: KaraMagic – automatic karaoke video maker

https://karamagic.com/
1•godot•3m ago•0 comments

What comes after agents? AI employees

https://www.ycombinator.com/launches/Pf7-beyond-agents-the-era-of-ai-employees
1•karissaho•3m ago•0 comments

Photocopier No More: The Reckoning with AI Creativity Has Arrived

https://reviews.ofb.biz/safari/article/1401.html
1•trbutler•4m ago•0 comments

Inverse Occam's Razor

https://arxiv.org/abs/2204.08284
1•jerlendds•5m ago•0 comments

Tell HN: Apple development certificate server seems down?

3•strongpigeon•5m ago•1 comments

Mother of All Grease Fires

https://milk.com/wall-o-shame/bucket.html
2•xk3•5m ago•0 comments

6-Axis Milling for Enhancing Quality of Fused Granular Fabrication Parts

https://www.mdpi.com/2073-4360/18/5/608
1•PaulHoule•6m ago•0 comments

Working to Decentralize FedCM

https://atproto.com/blog/working-to-decentralize-fedcm
1•sgoto•7m ago•0 comments

Agent-sync – sync between Claude Code and Codex configs

https://github.com/matanabudy/agent-sync
1•matanabudy•7m ago•0 comments

Helix 02 living room tidy

https://www.youtube.com/watch?v=CAdTjePDBfc
1•hheikinh•8m ago•0 comments

Don't let LLMs write for you

https://justismills.substack.com/p/dont-let-llms-write-for-you
1•c-oreills•9m ago•0 comments

Deep Learning: Our Year 1990-1991

https://people.idsia.ch/~juergen/deep-learning-miraculous-year-1990-1991.html
1•untilted•11m ago•0 comments

Ask HN: I built an AI-native codebase framework–could you evaluate it?

1•xodn348•15m ago•1 comments

The Slowest Viral Thing

https://pilgrima.ge/p/the-slowest-viral-thing
1•momentmaker•16m ago•0 comments

SoftBank eyes up to $40B loan to fund OpenAI investment

https://www.reuters.com/business/media-telecom/softbank-seeks-up-40-billion-loan-finance-openai-i...
4•devonnull•16m ago•0 comments

SEIA Solar Market Insight Report 2025 Year in Review

https://seia.org/research-resources/us-solar-market-insight/
1•toomuchtodo•17m ago•0 comments

A vertical tab companion app for aerospace window manager

https://github.com/raghavendra-talur/aeromux
1•rtalur•18m ago•1 comments

Uber rolls out women-only option in the US

https://www.bbc.com/news/articles/cx2gvrzwdr7o
2•alephnerd•18m ago•0 comments

Meta Is Buying Moltbook

https://lifehacker.com/tech/meta-is-buying-moltbook
1•umangsehgal93•18m ago•1 comments

GoT Timeline – a daily timeline game to test your Game of Thrones skills

https://www.got-timeline.com
1•onion92•19m ago•0 comments

Claude Code makes local LLMs 90% slower

https://unsloth.ai/docs/basics/claude-code
4•telotortium•22m ago•1 comments

Eventbrite Enters into Definitive Agreement to Be Acquired by Bending Spoons

https://www.businesswire.com/news/home/20251202408560/en/Eventbrite-Enters-into-Definitive-Agreem...
5•DocFeind•23m ago•1 comments

Why doesn't V8 fit on my microcontroller? (2021)

https://medium.com/the-toit-take/why-doesnt-v8-fit-on-my-microcontroller-71dc6e2d8f5c
1•tosh•24m ago•0 comments

Is there an MD5 Fixed Point where MD5(x) == x?

https://stackoverflow.com/questions/235785/is-there-an-md5-fixed-point-where-md5x-x
2•plaguna•25m ago•0 comments

GPT-4 leaks its own API internals through training data exposure

1•safteylayer•26m ago•0 comments

Andrew Tate Doesn't Get the Point of Books

https://www.theatlantic.com/ideas/2026/03/slow-reading-books-benefits/686266/
2•paulpauper•27m ago•2 comments

The Met Opera's Desperate Hunt for Money

https://www.nytimes.com/2026/03/08/arts/met-opera-peter-gelb-finances.html
3•paulpauper•28m ago•0 comments

Bubbles, Booms and Crashes in the US Stock Market 1792-2024

https://www.nber.org/papers/w34903
3•paulpauper•28m ago•0 comments
Open in hackernews

Show HN: Brainfuck to RISC-V JIT compiler written in Zig

https://github.com/evelance/brainiac
5•0x000xca0xfe•9mo ago
Hi everybody,

this was my project to learn Zig and RISC-V+x86_64 assembly.

Not sure if anybody is actually interested in yet another Brainfuck compiler, so I'll just write up some random things I learned while building it!

- A primitive assembly stitching compiler is 10x faster than the interpreter. Did not expect that.

- The generated x86 code is really bad (e.g. it always uses 6 or 7 byte sized instructions with 32-bit immediates when there are much smaller ones) but it doesn't really matter. Good code generated by GCC and clang for transpiled Brainfuck->C is not much faster as it's bottlenecked by memory accesses anyways.

- Zig is pretty far along actually. You can make serious projects with it!

- But the community seems to like self-punishment. Unused parameters and variables are hard errors and there is no way to disable that even for debug builds. Makes quickly commenting out part of the code a real PITA.

- I've had a miscompilation due to std.mem.span being broken and two source code breaks going from Zig 0.13 to 0.15 (std.mem.page_size got removed and ArrayList.popOrNull as well).

- But arbitrary size integers are fantastic! And well-defined two's complement behaviour!

Here is for example the code that encodes the c.beqz instruction:

  /// Branch if Equal to Zero (compressed): c.beqz rs1', offset -> beq rs1, x0, offset
  pub fn c_beqz(text: *std.ArrayList(u8), rs1: RV_X, offset: i9) !void {
      std.debug.assert(is3BitReg(rs1));
      std.debug.assert(@mod(offset, 2) == 0);
      const imm: u9 = @bitCast(offset);
      const RV_CB = packed struct(u16) {
          op: u2,
          offset5: u1,
          offset1_2: u2,
          offset6_7: u2,
          rsd_rs1_: u3,
          offset3_4: u2,
          offset8: u1,
          funct3: u3,
      };
      const ins = RV_CB {
          .op = 0x1,
          .offset5 = @truncate(imm >> 5),
          .offset1_2 = @truncate(imm >> 1),
          .offset6_7 = @truncate(imm >> 6),
          .rsd_rs1_ = @truncate(@intFromEnum(rs1) - 8),
          .offset3_4 = @truncate(imm >> 3),
          .offset8 = @truncate(imm >> 8),
          .funct3 = 0x6,
      };
      try appendInstruction(text, u16, @bitCast(ins));
  }
This is really nice as all the exotic integer sizes are actually checked, too.

- Zig support for Windows is good. Porting the project to Windows was very easy.

- When the RISC-V registers are carefully chosen, almost all instructions could be compressed in this projects.

- Compressed instructions and good branching code (using the branch instructions directly when the jump range is small enough instead of branching over a larger jump instruction) did not noticeably change performance on real hardware (OrangePi RV2).

- But somehow QEMU got a massive boost from that. Not sure why exactly.

So, that's about it!

I hope at least something was interesting...

Comments

sylware•9mo ago
thumbs up for this project (everything RISC-V is usually).

I write rv64 assembly (nearly core only, without memory reservation instructions) and run it on x86_64 with a very small (x86_64 assembly written) interpreter.

And your are right, I have had thoughts about a "RISC-V" x86_64 compiler (but it will probably require some runtime unfortunately).

Hopefully, rv22+ hardware with ultra-performant µ-architecture and with the latest silicon process will happen sooner than we expect. One less PI toxic lock and cleaner, _really standard_ assembly (the end game of much software).

0x000xca0xfe•9mo ago
Yeah I can't wait for a performant RISC-V core. Runtime code generation is so easy for RISC-V. I have many ideas or projects where I'd like to use it but it feels kinda pointless when JITed RISC-V machine code on current hardware gets destroyed by any half-decent x86 PC or Mac running naive C code.
sylware•9mo ago
Well, here are the tricks: interpreted rv64 assembly will be "slow"... actually "slower" than x86_64 native code... but in many execution contexts, for many pieces of software, here the first trick: the "slow" interpreted rv64 assembly machine code will be... "fast" enough... The 2nd trick: I have control on my rv64 machine interpreter, and I can write native x86_64 acceleration assembly along side of a rv64 reference implementation (I planned to do just that for my CPU renderer in my wayland compositor... actually I have already AVX2 code for some of that, even though the sweet spot is AVX512, but don't have the hardware for this, yet).

And once we have this rv64 shiny hardware, certainly won't be a drop-in, but the distance to code will be minimal.

One important SDK thing: I am careful at using the smallest number of rv64 machine instructions (we tend to forget 'R' in "RISC-V" means 'R'educed...), and I use basic, really basic, C preprocessors instead of the assembler preprocessor in order to decouple the assembly code from a specific assembler preprocessor. I don't even use assembler pseudo-instructions, or ABI register names, neither compressed machine instructions.

On top of that: I don't use ELF, I use a super minimal executable/system interface dynamic shared library format of my own, omega idiotically simple, which I wrap in ELF binaries for transparent support. People have to come to realize, ELF complexity, for a executable/system interface dynamic shared library is utterly and completely obsolete, even a liability once you are looking for binary stability in time (cf games), proven over more than the last decade.