• 0 Posts
  • 24 Comments
Joined 6 months ago
cake
Cake day: February 6th, 2025

help-circle
  • Multi-cloud is far from trivial, which is why most companies… don’t.

    Even if you are multi-cloud, you will be egressing data from one platform to another and racking up large bills (imagine putting CloudFront in front of a GCS endpoint lmao), you are incentivized to stick on a single platform. I don’t blame anyone for being single-cloud with the barriers they put up, and how difficult maintaining your own infrastructure is.

    Once you get large enough to afford tape libraries then yeah having your own offsite for large backups makes a lot of sense, but otherwise the convenience and reliability (when AWS isn’t nuking your account) of managed storage is hard to beat — cold HDDs are not great, and m-disc is pricey.


  • In this guy’s specific case, it may be financially feasible to back up onto other cloud solutions, for the reasons you stated.

    However public cloud is used for a ton of different things. If you have 4TiB of data in Glacier, you will be paying through the absolute nose pulling that data down into another cloud; highway robbery prices.

    Further as soon as you talk about something more than just code (say: UGC, assets, databases) the amount of data needing to be “egressed” from the cloud balloons, as does the price.



  • The recent boom in neural net research will have real applicable results that are genuine progress: signal processing (e.g. noise removal), optical character recognition, transcription, and more.

    However the biggest hype areas with what I see as the smallest real return is in the huge model LLM space, which basically try to portray AGI as just around the corner. LLMs will have real applications in summarization, but largely otherwise they just generate asymptotically plausible babble, very good for filling the Internet with slop, not actually useful to replace all the positions OAI, et al, need it to (for their funding to be justified).



  • Writing tests is a good example. It’s not great at writing tests, but it is definitely better than the average developer when you take the probability of them writing tests in the first place into account.

    Outside of everything else discussed here, this is something I disagree with on a fundamental level, flawed tests are worse than no tests, IMO.
    Not to get too deep in to the very contentious space of testing in development, but when it comes to automated testing, I think we’re better off with more rigorous[1] testing instead of just chasing test coverage metrics.


    1. Validating tests through chaos/mutagen testing; or model verification (e.g. Kani) ↩︎



  • It solves a pretty hard problem that is a self-hosted video platform; a lot of places use YouTube to host videos, even if they aren’t doing so to make money through adsense, this is for their own site material, posting to groupchats, and similar purposes.

    Issue is that otherwise you rely on platform owners like Google, who can decide to unperson you, your business, or an employee. It effectively happened to me, YT terminated my channel for unsubstantiated reasons, and hosting my own peertube is likely in the future to replace where I host my decades of video content.

    Further, ideologically, we should be collectively moving away from “platforms” for what should be obvious reasons to those of us on the fediverse.


  • I think it’s “the algorithm”, people basically just want to be force-fed “content” – look how successful TikTok is, largely because it has an algorithm that very quickly narrows down user habits and provides endless distraction.

    Mastodon and fediverse alternatives by comparison have very simple feeds and ways to surface content, it simply doesn’t “hook” people the same way, and that’s competition.

    On one hand we should probably be doing away with “the algorithm” for reasons not enumerated here for brevity, but on the other hand maybe the fediverse should build something to accommodate this demand, otherwise the non-fedi sites will.



  • I don’t think it’s hyperbole to say a significant percentage of Git activity happens on GitHub (and other “foundries”) – which are themselves a far cry from efficient.

    My ultimate takeaway on the topic is that we’re stuck with Git’s very counterintuitive porcelain, and only satisfactory plumbing, regardless of performance/efficiency; but if Mercurial had won out, we’d still have its better interface (and IMO workflow), and any performance problems could’ve been addressed by a rewrite in C (or the Rust one that is so very slowly happening).




  • They have smallweb and news indexing, but other than that AFAICT they rely completely on other providers. Which is a shame, Google allows submitting sites for indexing and notifies if they can’t.

    Running a scraper doesn’t need to cover everything since they have access to other indexes, but they really should be developing that ability instead of relying on Bing and other providers to provide good results, or results at all.



  • I’ve been using Kagi for the last year+.

    Personally, I wish they’d tone down the AI stuff that ruined Google, but at least you can turn most of it off.

    Their results are okay, a little better than Bing, but obviously they’re limited by their existing index providers, I wish they’d run their own spiders and crawl for their own data, since I think Bing fails on a lot of coverage of obscure websites.

    In general I find the weighting of modern indexes to be subpar, though the SEO industry has made it a hard problem to tackle, I wish more small websites and forums were higher ranked, and AI slop significantly de rated.

    TW: Self harm

    Also not a huge fan of the company and a lot of it’s ardent customers, who heavily protested a suicide prevention popup if you used it to searched for how to kill yourself.