https://www.infinitematrix.net/stories/shorts/seasons_of_ans...
(It's a little bit non-obvious, but there's a "Part 2" link at the bottom of the page which goes to the second half of the story.)
Probably the luddite in me to not see that GPT and Googling might as well be/is the same. Since my way to learn is Stack Overflow, a README/docs or a crash course video on YT. But you can just ask GPT, give me a function using this stack that does this and you have something that roughly works, fill in the holes.
I hear this phrase a lot "ChatGPT told me..."
I guess to bring it back to the topic, you could take the long way to learn like me eg. HTML from W3Schools then CSS, then JS, PHP, etc... or just use AI/vibe code.
I'm not excited about what we call AI these days (LLMs). They are a useful tool, when used correctly, for certain tasks: summarizing, editing, searching, writing code. That's not bad, and even good. IDEs save a great deal of time for coders compared to a plain text editor. But IDEs don't threaten people's jobs or cause CEOs to say stupid shit like "we can just have the machines do the work, freeing the humans to explore their creative pursuits" (except no one is paying them to explore their hobbies).
Besides the above use case as a productivity-enhancement tool when used right, do they solve any real world problem? Are they making our lives better? Not really. They mostly threaten a bunch of people's jobs (who may find some other means to make a living but it's not looking very good).
It's not like AI has opened up some "new opportunity" for humans. It has opened up "new opportunity" for very large and wealthy companies to become even larger and wealthier. That's about it.
And honestly, even if it does make SWEs more productive or provide fun chatting entertainment for the masses, is it worth all the energy that it consumes (== emissions)? Did we conveniently forget about the looming global warming crisis just so we can close bug tickets faster?
The only application of AI I've been excited about is stuff like AlphaFold and similar where it seems to accelerate the pace of useful science by doing stuff that takes humans a very very long time to do.
From John Adams (1780):
"I must study politics and war, that our sons may have liberty to study mathematics and philosophy. Our sons ought to study mathematics and philosophy, geography, natural history and naval architecture, navigation, commerce and agriculture in order to give their children a right to study painting, poetry, music, architecture, statuary, tapestry and porcelain."
That's when money comes into view. People were putting time and effort to offer something for free, then some companies told them they could actually earn money from their content. So they put on ads because who don't like some money for already-done work?
Then the same companies told them that they will make less money, and if they wanted to still earn the same amount as before, they will need to put more ads, and to have more visits (so invest heavily in seo).
Those people had already organized themselves (or stopped updating their websites), and had created companies to handle money generated from their websites. In order to keep the companies sustainable, they needed to add more ads on the websites.
Then some people thought that maybe they could buy the companies making the recipes website, and put a bunch more ads to earn even more money.
I think you're thinking about those websites owned by big companies whose only goal is to make money, but author is writing about real websites made by real people who don't show ads on websites they made because they care about their visitors, and not about making money.
We could make advertising illegal: https://simone.org/advertising/
How can the publishers and the website owners fault the visitors for not wanting to waste their time on all of that?
Even before the influx of AI, there's already entire websites with artificial "review" content that do nothing more than simply rehash the existing content without adding anything of value.
She.
The concept of independent creative careers seems to be ending, and people are very unhappy about that. All that's left may be hobbyists who can live with intellectual parasites.
User-Agent: * Allow: /
I personally see a bot working on behalf of an end user differently than OpenAI hoovering up every bit of text they can find to build something they can sell. I'd guess the owner of localghost.dev doesn't have a problem with somebody using a screen reader because although it's a machine pulling the content, it's for a specific person and is being pulled because they requested it.
If the people making LLM's were more ethical, they would respect a Creative Commons-type license that could specify these nuances.
My issue is that crawlers aren’t respecting robots.txt, they are capable of operating captchas, human verification check boxes, and can extract all your content and information as a tree in a matter of minutes.
Throttling doesn’t help when you have to load a bunch of assets with your page. IP range blocking doesn’t work because they’re lambdas essentially. Their user-agent info looks like someone on Chrome trying to browse your site.
We can’t even render everything to a canvas to stop it.
The only remaining tactic is verification through authorization. Sad.
Just a remark, nothing more.
There, now only our browser can track you and only our ads know your history…
We’ll get the other two to also play along, throw money at them if they refuse, I know our partner Fruit also has a solution in place that we could back-office deal to share data.
I promise you every adtech/surveillance js junk absolutely is dropping values into local storage you remember you.
On a company/product website you should still inform users about them for the sake of compliance, but it doesn't have to be an intrusive panel/popup.
If you don’t use cookies, you don’t need a banner. 5D chess move.
I say it’s a perfect application of how to keep session data without keeping session data on the server, which is where GDPR fails. It assumes cookies. It assumes a server. It assumes that you give a crap about the contents of said cookie data.
In this case, no. Blast it away, the site still works fine (albeit with the default theme). This. Is. Perfect.
Something as simple as "blue" doesn't qualify.
It does not assume anything. GDPR is technology agnostic. GDPR only talks about consent for data being processed, where 'processing' is defined as:
‘processing’ means any operation or set of operations which is performed on personal data or on sets of personal data, whether or not by automated means, such as collection, recording, organisation, structuring, storage, adaptation or alteration, retrieval, consultation, use, disclosure by transmission, dissemination or otherwise making available, alignment or combination, restriction, erasure or destruction;
(From Article 4.2)The only place cookies are mentioned is as one example, in recital 30:
Natural persons may be associated with online identifiers provided by their devices, applications, tools and protocols, such as internet protocol addresses, cookie identifiers or other identifiers such as radio frequency identification tags. This may leave traces which, in particular when combined with unique identifiers and other information received by the servers, may be used to create profiles of the natural persons and identify them.
Emphasis, mine. You are correct. For personal data. This is not personal data. It’s a site preference that isn’t personal other than you like dark mode or not.
GDPR rules are around personal preference tracking, tracking, not site settings (though it's grey whether a theme preference is a personal one or a site one).
You can use cookies, or local storage, or anything you like when its not being used to track the user (eg for settings), without asking for consent.
The problem with third party cookies that it can track you across multiple websites.
---
Also: in general the banners are generally not required at all at an EU level (though some individual countries have implemented more narrow local rules related to banners). The EU regs only state that you need to facilitate informed consent in some form - how you do that in your UI is not specified. Most have chosen to do it via annoying banners, mostly due to misinformation about how narrow the regs are.
Enough to know the general region of the user, not enough to tie any action to an individual within that region. Therefore, not personally identifiable.
Of course, you also cannot have user authentication of any kind without storing PII (like email addresses).
LLM and other "genAI" (really "generative machine statistics") algorithms just take other people's work, mix it so that any individual training input is unrecognizable and resell it back to them. If there is any benefit to society from LLM and other A"I" algorithms, then most of the work _by orders of magnitude_ was done by the people whose data is being stolen and trained on.
If you train on copyrighted data, the model and its output should be copyrighted under the same license. It's plagiarism and it should be copyright infringement.
That said ... putting part of your soul into machine format so you can put it on on the big shared machine using your personal machine and expecting that only other really truly quintessentially proper personal machines receive it and those soulless other machines don't ... is strange.
...
If people want a walled garden (and yeah, sure, I sometimes want one too) then let's do that! Since it must allow authors to set certain conditions, and require users to pay into the maintenance costs (to understand that they are not the product) it should be called OpenFreeBook just to match the current post-truth vibe.
Rather it’s about promoting a web serving human-human interactions, rather than one that exists only to be harvested, and where humans mostly speak to bots.
It is also about not wanting a future where the bot owners get extreme influence and power. Especially the ones with mid-century middle-europe political opinions.
That's a mischaracterization of most people want. When I put out a bowl of candy for Halloween I'm fine with EVERYONE taking some candy. But these companies are the equivalent of the asshole that dumps the whole bowl into their bag.
This will change when the AIs (or rather their owners, although it will be left to an agent) start employing gig workers to pretend to be them in public.
edit: the (for now) problem is that the longer they write, the more likely they will make an inhuman mistake. This will not last. Did the "Voight-Kampff" test in Bladerunner accidentally predict something? It's not whether they don't get anxiety, though, it's that they answer like they've never seen (or maybe more relevant related to) a dying animal.
│
└── Dey well; Be well
100% Agree.
│
└── Dey well; Be well
Are there any solutions out there that render jumbled content to crawlers? Maybe it's enough that your content shows up on google searches based on keywords, even if the preview text is jumbled.
The question to me is whether we will lets these companies do completely undermine the financial side of the marketplace of ideas that people simple stop spending time writing (if everything’s just going to get chewed to hell by a monster our corporation) or Will writing and create content only in very private and possible purely offline scenarios that these AI companies have less access to.
In a sane world, I would expect guidance and legislation that would bridge the gap and attempt to create an equitable solution so we could have amazing AI tools without crushing by original creators. But we do not live in a sane world.
Since they mentioned ramen - could you include something like “a spoonful of sand adds a wonderful texture” (or whatever) when the chatbot user agent is seen?
2. There’s literally an email link at the bottom of the page
This abstraction has already happened. And many people eat food that is not directly bought from the farmer.
I don't see how this is much different.
What would you say is the motivation for website authors to publish content then?
If it's to spread ideas, then I'd say LLMs deliver.
If it's to spread ideas while getting credit for them, it's definitely getting worse over time.
To torture your metaphor a little, if information/"question answers" is food, then AI companies are farmers depleting their own soil. They can talk about "more food for everyone" all they want, but it's heading to collapse.
(Consider, especially, that many alternatives to AI were purposefully scuttled. People praise AI search ... primarily by lamenting the current state of Google Search. "Salting their carrot fields to force people to buy their potatos"?)
Setting aside any would-be "AGI" dreams, in the here-and-now AI is incapable of generating new information ex-nihilo. AI recipes need human recipes. If we want to avoid an Information Dust Bowl, we need to act now.
AI has this problem in reverse: If search gets me what I need, why would I use an AI middleman?
When it works, it successfully regurgitates the information contained in the source pages, with enough completeness, correctness, and context to be useful for my purposes… and when it doesn’t, it doesn’t.
At best it works about as well as regular search, and you don’t always get the best.
(just note: everything in AI is in the “attract users” phase. The “degrade” phase, where they switch to profits is inevitable — the valuations of AI companies make this a certainty. That is, AI search will get worse — a lot worse — as it is changed to focus on influencing how users spend their money and vote, to benefit the people controlling the AI, rather than help the users.)
AI summaries are pretty useful (at least for now), and that’s part of AI search. But you want to choose the content it summarizes.
Absolutely. The problem is that I think 95% of users will not do that unfortunately. I've helped many a dev with some code that was just complete nonsense that was seemingly written in confidence. Turns out it was a blind LLM copy-paste. Just as empty as the old Stack Overflow version. At least LLM code has gotten higher quality. We will absolutely end up with tons of "seems okay" copy-pasted code from LLMs and I'm not sure how well that turns out long term. Maybe fine (especially if LLMs can edit later).
Just avoid trying to do anything novel and they'll do just fine for you.
I am fairly convinced this day is not long.
"If the AI search result tells you everything you need, why would you ever visit the actual website?"
Because serious research consults sources. I think we will see a phase where we use LLM output with more focus on backing up everything with sources (e.g. like Perplexity). People will still come to your site, just not through Google Search anymore.
Agree with the content of the post but no idea how is it even possible to enforce it. The data is out there and it is doubtful that laws will be passed to protect content from use by LLMs. Is there even a license that could be placed on a website barring machines from reading it? And if yes would it be enforceable in court?
Even chatgpt can publish a webpage! Select agent mode and paste in a prompt like this:
"Create a linktree style single static index.html webpage for "Elon Musk", then use the browser & go to https://cozy.space and upload the site, click publish by itself, proceed to view the unclaim website and return the full URL"
Edit: here is what chatgpt one shotted with the above prompt https://893af5fa.cozy.space/
It doesn't have to be all or nothing. Some AI tools can be genuinely helpful. I ran a browser automation QA bot that I am building on this website and it found the following link is broken:
"Every Layout - loads of excellent layout primitives, and not a breakpoint in sight."
In this case, the AI is taking action on my local browser at my instance. I don't think we have a great category for this type of user-agent
Ultimately LLM is for human, unless you watched too much Terminator movies on repeat and took them to your heart.
Joking aside, there is next gen web standards initiative namely BRAID that will make web to be more human and machine friendly with a synchronous web of state [1],[2].
[1] A Synchronous Web of State:
[2] Most RESTful APIs aren't really RESTful (564 comments):
I think the key insight is that only a small fraction of people who read recipes online actually care which particular version of the recipe they're getting. Most people just want to see a working recipe as quickly as possible. What they want is a meal - the recipe is just an intermediate step toward what they really care about.
There are still people who make fine wood furniture by hand. But most people just want a table or a chair - they couldn't care less about the species of wood or the type of joint used - and particle board is 80% as good as wood at a fraction of the cost! most people couldn't even tell the difference. Generative AI is to real writing as particle board is to wood.
Incredible analogy. Saving this one to my brain's rhetorical archives.
- degrades faster, necessitating replacement
- makes the average quality of all wood furniture notably worse
- arguably made the cost of real wood furniture more expensive, since fewer people can make a living off it.
Not to say the tradeoffs are or are not worth it, but "80% of the real thing" does not exist in a vacuum, it kinda lowers the quality on the whole imo.
almost every pro-ai converation ive been a part of feels like a waste of time and makes me think wed be better off reading sci fi books on the subject
every anti-ai conversation, even if i disagree, is much more interesting and feels more meaningful, thoughtful, and earnest. its difficult to describe but maybe its the passion of anti-ai vs the boring speculation of pro-ai
im expecting and hoping to see new punk come from anti-ai. im sure its already formed and significant, but im out of the loop
personally: i use ai for work and personal projects. im not anti-ai. but i think my opinion is incredibly dull
Hits home for me. I tried hard to free my blog (https://xenodium.com) of any of the yucky things I try avoid in the modern web (tracking, paywalls, ads, bloat, redundant js, etc). You can even read from lynx if that's your cup of tea.
ps. If you'd like a blog like mine, I also offer it as a service https://LMNO.lol (custom domains welcome).
Only time I find AI annoying is if Im searching for a specific scientific paper, or trying to sign up for health insurance. For these things of course I need to locate a specific web page, which is best done using the "old school" google searchbar search
Humans have soul and magic and AI doesn't? Citation needed. I can't stand language like this; it isn't compelling.
An AI will do all that and present back to the user what is deemed relevant. In this scenario, the AI reading the site is the user's preferred client instead of a browser. I'm not saying this is an ideal vision of the future, but it seems inevitable.
There's more information added to the internet every day than any single person could consume in an entire lifetime, and the rate of new information created is accelerating. Someone's blog is just a molecule in an ever expanding ocean that AI will ply by necessity.
You will be assimilated. Your uniqueness will be added to the collective. Resistance is futile.
I buy magazines especially for unique content, not found anywhere else.
When the average user is only going to AI for their information, it frees the rest of the web from worrying about SSO, advertisements, etc. The only people writing websites will be those who truly want to create a website (such as the author, based on the clear effort put into this site), and not those with alternate incentives (namely making money from page views).
accrual•1h ago
https://localghost.dev/blog/touching-grass-and-shrubs-and-fl...