My initial thoughts are that we would trade many of the conveniences we currently value for absolute, unambiguous clarity. For example:
- Would we get rid of most syntactic sugar? If there's only one, explicit way to write a `for` loop, the AI's output becomes more predictable and easier to review.
- Would we enforce extreme explicitness? Imagine a language where you must write `fn foo(none)` if there are no parameters, just to remove the ambiguity of `()`.
- How would we handle safety? Would features like mandatory visibility (`pub`/`priv`) and explicit ownership annotations for FFI calls become central to the language itself, providing guarantees the reviewer can see instantly?
- Would such a language even be usable by humans for day-to-day work, or would it purely be a compilation target for AI prompts?
What trade-offs would you be willing to make for a language that gave you higher confidence in the code an AI generates?
dtagames•6mo ago
Even after that, it will exhibit all the same problems as existing models and other languages. The unreliability of LLMs comes from the way they make predictions, rather than "retrieve" real answers, like a database would. Changing the content and context (your new language) won't change that.
baijum•6mo ago
But that makes me wonder if the goal should be reframed. Instead of trying to eliminate errors, what if we could change their nature?
The interesting hypothesis to explore, then, is whether a language's grammar can be designed to make an LLM's probabilistic errors fail loudly as obvious syntactic errors, rather than failing silently as subtle, hard-to-spot semantic bugs.
For instance, if a language demands extreme explicitness and has no default behaviors, an LLM's failure to generate the required explicit token becomes a simple compile-time error, not a runtime surprise.
So while we can't "fix" the LLM's core, maybe we can design a grammar that acts as a much safer "harness" for its output.
dtagames•6mo ago
The problem is that people don't like writing assembler, which is how we got Fortran in the first place.
The fundamental issue, then, is with the human language side of things, not the programming language side. The LLM is useful because it understands regular English, like "What is the difference between 'let' and 'const' in JS?," which is not something that can be expressed in a programming language.
To get the useful feature we want, natural language understanding, we have to accept the unreliable and predictive nature of the entire technique.
FloatArtifact•6mo ago
If it understands human language enough, it should be able to understand the logic laid out in the documentation mapped to symbols to construct code.
dtagames•6mo ago
You can also describe your own app's syntax, architecture, function signatures, etc. in markdown files or just in chat and Cursor will write code that conforms to your desired syntax, which definitely doesn't exist in the training set.
FloatArtifact•6mo ago
baijum•5mo ago