Its like a bootstrap css just dropped. People still giving "minimum effort" into their vibe code/eng projects but slap a domain on top. Is this to save token cost ?
Apparently Prof. Dr. style is the aesthetic name for pages with no custom styling of any kind, common at universities in the 01990s: https://contemporary-home-computing.org/prof-dr-style/
To give an example that annoys me to no end, Google's Antigravity insists on making everything "anthropomorphic", which gets interpreted as overtly rounded corners, way too much padding everywhere, and sometimes even text gradients (huge no-no in design). So, unless you instruct it otherwise, every webpage it creates looks like a lame attempt at this: https://m3.material.io/
These are the money lines:
NEVER use generic AI-generated aesthetics like overused font families
(Inter, Roboto, Arial, system fonts), cliched color schemes
(particularly purple gradients on white backgrounds), predictable
layouts and component patterns, and cookie-cutter design that lacks
context-specific character.
Interpret creatively and make unexpected choices that feel genuinely
designed for the context. No design should be the same. Vary between
light and dark themes, different fonts, different aesthetics. NEVER
converge on common choices (Space Grotesk, for example) across
generations.
[1]: https://github.com/anthropics/claude-code/blob/main/plugins/...Actually I would consider this setup to not be very user friendly. This makes a lot of assumptions about the data/format you have available already. Personally I would assume that anything operating on my bank transactions would be through some more turnkey/handsoff integration rather than a direct import.
Puzzle basically does this by hooking directly into my bank and gives me other tools where I can easily use the categorizations
Well, sure. Accounting solution. But in a post about scripting your way to financial fun and games (and I think there are a fair share of people here who are locked into their accounting platforms/apps for various reasons), what solution does the API call to the bank (unavailable to the small player) and then gives you an API endpoint to get cleaned data into whichever accounting solution you happen to be using? Puzzle.io ain't going to do it at any price.
This tools looks pretty nice, kudos for building it and putting it out there for others to try it
You can just install the tool and use it. It is a CLI with very verbose output. Verbose output is good for both humans and AI.
Here's Qwen3 Coder [1]
> claude
“There’s a CSV at @path please open it and classify my bank transactions”.
I actually just vibed a hyper-specific version of a similar tool for myself a couple weeks ago, mostly just for fun to see if I (or Claude) could. Took about an hour, and it's now able to automate the spreadsheet process my girlfriend and I use each month to split certain expenses. Saves us each ~15 minutes weekly.
I'm loving the ability LLMs provide to both build personal software so rapidly, as well as solve these kinds of fuzzier natural language problem spaces so relatively easily.
Side note: the state of consumer transaction reporting is absolute garbage. There should really be more metadata mandated by consumer protection regs or something. The fact that this is a hard problem in the first place feels very dumb.
I tried to use LLM::Function for category matching and, in my brief excursion that way found that submitting bank description strings to LLMs is pretty much the antithesis of what they are good at.
My solution does regex, then L-D lexical distance, then opens the Raku repl for manual fixes…
<<Catmap run tries to do an exact match to the description and auto apply a category. Failing that, it does a levenshtein / damarau lexical distance and proposes a best match category, prompting for [CR]. Or you can override and assign to an existing category. Go cats at the prompt to get a list of active categories.>>
It can be used by human or by AI agents.
I experiment the same with other mechanisms, and CLI are as effective - if not more effective - than MCP.
Granted, having access to AI I would use AI to run it. But nothing is stopping a manual, human centric, use.
I believe more tools should be written like that.
Y'all, please actually read the homepage before dunking on someone's project...
Fiveplus•22h ago
ezst•22h ago
nerdzoid•22h ago
minne•22h ago
pansa2•22h ago
throwaw12•22h ago
Why don't you accept it as, dude is experimenting and learning new tool, how cool is that, if this is possible, what else can I build with these tools?
supriyo-biswas•22h ago
throwaw12•21h ago
* simonw - Simon Willison, he could just continue building datasette or help Django, but he started exploring LLMs
* Armin Ronacher
* Steve Yegge
and many more
andy99•21h ago
yolo3000•20h ago
RevEng•18h ago
devsda•22h ago
throwaw12•21h ago
"leadership" or real leaders, want people to experiment a lot, so some of them will come up with novel ideas and either decide to build it on their own and get rich or build internally and make company rich.
Not always, but in many cases when someone becomes rich with innovation, it is probably because there was a benefit to a society (excluding gambling, porn, social media addictions)
stoneforger•20h ago
bitwize•22h ago
weird-eye-issue•22h ago
greatgib•21h ago
zwnow•21h ago
All they do is buy out companies and make a already finished product theirs.
makach•22h ago
senko•21h ago
I don't know about your bank transactions, but at least in my case the descriptions are highly irregular and basically need a hardcode for each and every damn pos location across each of the store location across each vendor.
I attempted that (with a Python script), gave up, and built myself a receipt tracker (photo + gemini based ocr) instead, which was way easier and is more reliable, even though - oh the horror! - it's using AI.
noufalibrahim•21h ago
There's a new feeling that I experience when using LLMs to do work. It's that every run, every commit has a financial cost (tokens). Claude code can write a nice commit message for me but it will cost me money. Alternatively, I can try to understand and write the message myself.
Perhaps the middle ground is to have the LLM write the classifier and then just use that on the exported bank statements.
yunohn•21h ago
ithkuil•20h ago
ozim•21h ago
That’s what I mostly do, I give it some examples ask to write code to handle stuff.
I don’t just dump data into LLM and ask for results, mostly because I don’t want to share all the data and I make up examples. But it also is much cheaper as once I have code to process data I don’t have to pay for processing besides what it costs to run that code on my machine.
mimimi31•20h ago
I think that's what GnuCash does by default. Even with years of past transaction data it still gets some very obvious matches wrong for me. In my experience it's about 90% accurate for the ones it really should be able to do based on the training data.
xtiansimon•15h ago
What's the pathway for the remaining 10%? Are they simply misclassified, and dropped into a queue for manual labeling? Do the outliers get managed by the GnuCash? Or do they get dumped into a misc 9000 account?
mimimi31•11h ago
conradev•18h ago
xtiansimon•18h ago
There was a time when I'd read this comment and then go looking for a tutorial on building a basic "Bayesian classifier". Invariably, I'd find several, which I'd line up like soldiers in tabs, and go through them until I find one that explained the what, why and how of it that spoke to my use (or near enough).
Of course, now ChatAI does all that for you in several Chat sessions. One does wonder though, if Chat is trained on text, and that was the history of what text was available, 10 years from now after everyone stopped writing 10 Blog posts about the same "Bayesian classifier", where's the ChatAI fodder coming from? I don't even know if this would be an outcome of fewer Blog posts [1]. It just strikes me as interesting because that would be _a very slow process_.
[1]: Not that this is necessarily true. People write blogs for all sorts of reasons, and having knockout quality competition from ChatAI does not KO all of them.
senko•18h ago
Sure, maybe, if I label enough data? At the number and variety of transactions I do, it wouldn't be be much better than hardcoding.
> It's that every run, every commit has a financial cost (tokens).
Ballpark figure for total cost for the Gemini OCRs for now (for me and a few hundred other people who have downloaded the app), for the past 6 or so months, is a few minutes of my hourly rate.
Absolutely not worth the manual grind for me.
andy99•21h ago
This is a hard, for all intents and purposes non deterministic problem.
Now if you’ll excuse me I have to draft another post admonishing the use of
hrimfaxi•21h ago
ManuelKiessling•21h ago
dizhn•21h ago
I had a coding agent write this for me last week. :D
It takes an excel export of my transactions that I have to download obviously since no bank is giving out API access to their customers. It uses some python pandas and excel stuff and streamlit to classify the transactions and "tally" results and show on the screen as color coded tabular data. (Streamlit seems really nice but super super limited in what it can do.) It also creates an excel file (with same color coding) so I can actually open it up and check if necessary. This excel file has dropdowns to reclassify rows. The final excel file also has formulas in place to update live. Code can also compare its own programmatic calculations with the result from formulas from excel. Why not? My little coding sweatshop never complains. (All with free models and clis by the way. I haven't had a reason to try Claude yet.)
xtiansimon•15h ago
I think Citi bank has an API (https://developer.citi.com/). Not that it's public for account holders, but third-parties can build on that. I'm looking at Plaid.com
One thing about Plaid--I've not been happy when encountering Plaid in the wild. For example, when a vendor directs me to use Plaid to validate a new bank account integration. I'd much rather wait a few days and use the 0.01 deposit route. Like using a katana for shaving.
But signing up to use Plaid for bank transaction ingestion via API is a whole different matter.
zwnow•21h ago
alwillis•6h ago
dimitri-vs•20h ago
I think the smarter play is having an agent take the first crack at it, and build up a high confidence regex rule set. And then from there handle things that don't match and do periodic spot checks to maintain the rule set.
xtiansimon•18h ago
Curious, what's the inputs for an agent when handling your dataset? What can you feed the agent so it can later "learn" from your _manual labeling_?
init•19h ago
Regex works well if you have a very limited set of sender and recipient accounts that don't change often
Bayesian or DNN classifiers work well when you have labeled data.
LLMs work well when you have a lot of data from lots of accounts.
You can even combine these approaches for higher accuracy
conradev•18h ago
Just like manually editing Sieve/Gmail filters: I want full determinism, but managing all of that determinism can be annoying…
qaboutthat•18h ago