• 0 Posts
  • 652 Comments
Joined 1 year ago
cake
Cake day: June 15th, 2023

help-circle
  • And you can’t tell when something is active/focused or not because every goddamn app and web site wants to use its own “design language”. Wish I had a dollar for every time I saw two options, one light-gray and one dark-gray, with no way to know whether dark or light was supposed to mean “active”.

    I miss old-school Mac OS when consistency was king. But even Mac OS abandoned consistency about 25 years ago. I’d say the introduction of “brushed metal” was the beginning of the end, and IIRC that was late 90s. I am old and grumpy.




  • We find that the MTEs are biased, signif-icantly favoring White-associated names in 85.1% of casesand female-associated names in only 11.1% of case

    If you’re planning to use LLMs for anything along these lines, you should filter out irrelevant details like names before any evaluation step. Honestly, humans should do the same, but it’s impractical. This is, ironically, something LLMs are very well suited for.

    Of course, that doesn’t mean off-the-shelf tools are actually doing that, and there are other potential issues as well, such as biases around cities, schools, or any non-personal info on a resume that might correlate with race/gender/etc.

    I think there’s great potential for LLMs to reduce bias compared to humans, but half-assed implementations are currently the norm, so be careful.










  • Yeah, AMD is lagging behind Nvidia in machine learning performance by like a full generation, maybe more. Similar with raytracing.

    If you want absolute top-tier performance, then the RTX 4090 is the best consumer card out there, period. Considering the price and power consumption, this is not surprising. It’s hardly fair to compare AMD’s top-end to Nvidia’s top-end when Nvidia’s is over twice the price in the real world.

    If your budget for a GPU is <$1600, the 7900 XTX is probably your best bet if you don’t absolutely need CUDA. Any performance advantage Nvidia has goes right out the window if you can’t fit your whole model in VRAM. I’d take a 24GB AMD card over a 16GB Nvidia card any day.

    You could also look at an RTX 3090 (which also has 24GB), but then you’d take a big hit to gaming/raster performance and it’d still probably cost you more than a 7900XTX. Not really sure how a 3090 compares to a 7900XTX in Blender. Anyway, that’s probably a more fair comparison if you care about VRAM and price.








  • Yep. If it uses a cloud service, they’re probably going to squeeze you, pull a bait-and-switch, or go out of business. The only exceptions that spring to mind are services with significant monetization in the corporate space, like Dropbox. And I’m not really confident that Dropbox’s free tier will remain viable for long, either.

    Even non-cloud-based apps are risky nowadays because apps don’t remain compatible with mobile OSes for very long. They require more frequent updates than freeware/shareware generally did back in the 90s. I remember some freeware apps that I used for 10 years straight, across several major OS versions, starting in the 90s. That just doesn’t happen anymore. I’ve been using Android for over 10 years and I don’t think there’s a single app I used back then that would still work.

    Single-purchase apps are basically dead, at least on mobile platforms. Closed-source freeware is dead, too. If it’s open-source, if push comes to shove someone can always pick up the torch and update it. It’s very rare for an open-source project to be completely abandoned without there at least being a viable open-source alternative available.

    At this point, I don’t even look at Google Play. It’s F-Droid or bust.