> Idiotic piece
That's unnecessary; please don't do that here. Weird that you created an account just to post an unsubstantive comment.
https://github.com/ipfs/kubo/issues/10327
https://discuss.ipfs.tech/t/moved-ipfs-node-result-netscan-d...
>This happens with Hetzner all the time because they have no VLANs and all customers are on a single LAN and IPFS tries to discover other nodes in the same LAN by default.
> running and managing numerous DBs, k8s clusters, ci/cd pipelines and stateless container systems is simply impossible with a team of 1-2 people.
That's a medium to large homelab worth of stuff, which means it can be run by a single nerd in their spare time.
The gulf between these two insofar as what approach, technologies, and due-diligences are necessary is vast.
But I think for many (most?) businesses, one nine is just fine. That's perfectly doable by one person, even if you want, say, >=96% uptime, which allows for 350 hours of downtime per year. Even two nines allows for ~88 hours of downtime per year, and one person could manage that without much trouble.
Most businesses aren't global. Downtime outside regular business hours for your timezone (and perhaps one or two zones to the west and east of you) is usually not much of a problem, especially if you're running a small B2B service.
For a small business that runs on 1-3 servers (probably very common!), keeping a hot spare for each server (or perhaps a single server that runs all services in a lower-supported-traffic mode) can be a simple way to keep your uptime high without having to spend too much time or money. And people don't have to completely opt out of the cloud; there are affordable options for e.g. managed RDBMS hosting that can make maintenance and incident response significantly easier and might be a good choice, depending on your needs.
(Source: I'm building a small one-person business that is going to work this way, and I've been doing my research and gaming it out.)
> The whole debate of “is this still the cloud or not” is nonsense to me. You’re just getting lost in naming conventions. VPS, bare metal, on-prem, colo, who cares what you call it. You need to put your servers somewhere. Sure, have a computer running in your mom’s basement if that makes you feel like you’re exiting the cloud more, I’ll have mine in a datacenter and both will be happy.
If you can't drive to the location where your stuff is running, and then enter the building blindfolded, yet put your hands on the correct machine, then it's cloud.
The "is this cloud or not" debate in the piece makes perfect sense. Who cares whether Hetzner is defined as "the cloud" or not? The point is, he left AWS without going to Azure or some other obvious cloud vendor. He took a step towards more hands on management. And he saved a ton of money.
However, one situation where I think the cloud might be useful is for archive storage. I did a comparison between AWS Glacier Deep Storage and local many-hard-drive boxes, for storing PB-scale backups, and AWS just squeaked in as slightly cheaper, but only because you only pay for the amount you use, whereas if you buy a box then you have to pay for the unused space. And it's off-site, which is a resilience advantage. And the defrosting/downloading charge was acceptable at effectively 2.5 months worth of storage. However, at smaller scales you would probably win with a small NAS, and at larger scales you'd be able to set up a tape library and fairly comprehensively beat AWS for price.
That being said, the cloud does have a lot of advantages:
- You're getting a lot of services readily available. Need offsite backups? A few clicks. Managed database? A few clicks. Multiple AZs? Available in seconds.
- You're not paying up-front costs (vs. investing hundreds of dollars for buying server hardware) and everything is available right now [0]
- Peak-heavy loads can be a lot cheaper. Mostly irrelevant for you average compute load, but things are quite different if you need to train an LLM
- Many services are already certified according to all kinds of standards, which can be very useful depending on your customers
Also, engineering time and time in general can be expensive. If you are a solo entrepreneur or a slow growth company, you have a lot of engineering time for basically free. But in a quick growth or prototyping phase, not to speak of venture funding, things can be quite different. Buying engineering time for >150€/hour can quickly offset a lot of saving [1].
Does this apply to most companies? No. Obviously not. But the cloud is not too expensive - you're paying for stuff you don't need. That's an entirely different kind of error.
[0] Compared to the rack hosting setup described in the post. Hetzner, Linode, etc. do provide multiple AZs with dedicated servers.
[1] Just to be fair, debugging cloud errors can be time consuming, too, and experienced AWS engineers will not be cheaper. But an RDS instance with solid backups-equivalent will usually not amortize quickly, if you need to pay someone to set it up.
...
...
"P.S. follow me on Twitter"
So uh, not everything
Strawman arguments, ad hominem attacks and Spongebob mocking memes, and the casual venturing into conspiracy theories and malicious intentions...
> Why do all these people care if I save more money or not? ... If they’re wrong, and if I and more people like me manage to convince enough people that they’re wrong, they may be out of a job soon.
I have a feeling him leaving AWS with the three people that he convinces will not make it into their quarterly earnings report.
> You will hear a bunch of crap from people that have literally never tried the alternative. People with no real hands-on experience managing servers for their own projects for any sustained period of time.
This is more of a rant than a thoughtful technical article. I don't know what I was expecting, because I clicked on the title knowing it was clickbait, so shame on me, I guess...
Is this what I'm missing by not having Twitter?
This means that teams must make an up-front architectural decision to develop apps in a server-agnostic manner, and developers must stay disciplined to keep components portable from day one, but you can get a lot of mileage out of free credits without burning dollars on any infrastructure. The biggest challenge becomes finding the time to perform these migrations among other competing priorities, such as new feature development, especially if you're growing fast.
Our startup is mostly built on Google Cloud, but I don't think our sales rep is very happy with how little we spend or that we're unwilling to "commit" to spending. The ability to move off of the cloud, or even just to another cloud, provides a lot of leverage in the negotiating seat.
Cloud vendors can also lead to an easier risk/SLA conversation for downstream customers. Depending on your business, enterprise users like to see SLAs and data privacy laws respected around the globe, and cloud providers make it easy to say "not my problem" if things are structured correctly.
* at least I assume what this post is; I’m still waiting for it to load.
On AWS an Aurora RDS is not cheap. But I don't have to spend time or money on an admin.
Is the cost justified? Because that's what cloud is. Not even talking about the level of compliance I get from having every layer encrypted when my hosted box is just a screwdriver away from data getting out the old school way.
When I'm small enough or big enough, self managed makes sense and probably is cheaper. But when getting the right people with enough redundancy and knowledge is getting the expensive part...
But actually - I've never seen this in any if these arguments so far. Probably because actual time required to manage a db server is really unpredictable.
This, and also startups are quite heterogeneous. If you have an engineer on your team with experience in hosting their own servers (or at least a homelab-person), setting up that service with sufficient resiliency for your average startup will be done within one relaxed afternoon. If your team consists of designers and engineers who hardly ever used a command line, setting up a shaky version of the same thing will cost you days - and so will any issue that comes up.
* off-site db backups
* a guaranteed db restore process
* auditable access to servers
* log persistence and integrity
* timely security patching
* intrusion detection
so that my employer can save money.
He could have summed up with "AWS is expensive, host your own server instead".
I'd be more interested to understand (from folk who were there) what the conditions were that made AWS et al such a runaway hit. What did folks gain, and have those conditions meaningfully changed in some way that makes it less of a slam dunk?
My recollection from working at a tech company in the early 2010s is that renting rack space and building servers was expensive and time consuming, estimating what the right hardware configuration would be for your business was tricky, and scaling different services independently was impossible. also having multi regional redundancy was rare (remember when squarespace was manually carrying buckets of petrol for generators up many flights of stairs to keeps servers online post sandy?[1]).
AWS fixed much of that. But maybe things have changed, and there's a new middleground that meaningfully improves on the experience?
[1] https://www.squarespace.com/press-coverage/2012-11-1-after-s...
vcpu, iops, transfer fees, storage -- they are all resources going into a pool .
If Hetzner is giving you 10TB for $100 , then host your static files/images there and save $800.
Apps are very modular. You have services, asyncs, LBs, static files . Just put the compute where it is most cost effective.
You don't have to close your AWS account to stick it to the man. Like any utility, just move your resources to where they are most affordable.
I once got into an argument with a lead architect about it and it's really easy to twist the conversation into "don't you think we'll reach that scale?" To justify complexity.
The bottom line is for better or worse, the cloud and micro services are keeping a lot of jobs relevant and there's no benefit in convincing people otherwise
“It is difficult to get a man to understand something, when his salary depends on his not understanding it.”
― Upton Sinclair
why? why be so obnoxious to other people who you claim are being obnoxious to you. no need to read your blog post now/
ranger_danger•51m ago
If you need a lot of, well, anything, be it compute, memory, storage, bandwidth etc., of course cloud stuff is going to be more expensive... but if you don't need that, then IMO $3/mo on-demand pricing really can't be beat when I don't have to maintain any equipment myself. Oracle also offers perpetually free VM instances if you don't mind the glow.
kelnos•17m ago
I can certainly see a use for that small amount of compute & RAM, but it's not clear that your level of needs is common. I've been paying for a $16/mo VPS (not on AWS) for about 15 years. It started out at $9/mo, but I've upgraded it since then as my needs have grown. It's not super beefy with 2 vCPUs, 5GiB of RAM, and 60GiB of disk space (with free data ingress/egress), but it does the job, even if I could probably find it cheaper elsewhere.
But not at Amazon. Closest match is probably a t3.medium, with 2 vCPUs and 4GiB RAM. Add a 60GiB gp2 EBS volume, and it costs around $35/mo, and that's not including data transfer.
The point that you're missing is we're not looking for the cheapest thing ever, we're looking for the cheapest thing that meets requirements. For many (most?) applications, you're going to overpay (sometimes by orders of magnitude) for AWS.
You say "if you need a lot", but "lot" is doing a bit of work there. My needs are super modest, certainly not "a lot", and AWS is by far not the cheapest option.