AI isn’t a shortcut for thinking. In her guide for skeptics, Hilary Gridley reframes AI as a collaborator—not a replacement. Use it like spellcheck for your thoughts. Don’t fear it—iterate with it. Insight improves, speed follows. Full post: https://hils.substack.com/p/the-ai-skeptics-guide-to-ai-collaboration
But it is still a giant inefficient use of resources and energy 🤣
@prologic@twtxt.net Since you have to check and double check everything it spits out (without providing sources), I don’t find any of this helpful. It’s like someone’s in the room with you and that person is saying random stuff that might or might not be correct. At best, it might spark some new idea in your head and then you follow that idea the traditional way.
Information published on the internet (or anywhere, for that matter) was never guaranteed to be correct. But at least you had a “frame of reference”: “Ah, I read this information about Linux on a blog that usually posts about Windows, so this one single Linux post might not necessarily be correct.” That is completely lost with LLMs. It’s literally all mushed together. 🤷
@movq@www.uninformativ.de Yeah I couldn’t agree more. The utility of using it in any way to form “truths” or to do anything that require a high degree of “accuracy” is utterly pointless.
There are other tasks LLM(s) are far better suited for, which are also its downsides, and gawd so expensive and unrealistic to run yourself 🤦♂️ Do you know what one of these NVIDIA H100’s cost? 💲 That’s right! 🤣 > $50k USD 😱 And many of the models out there require 8 of these suckers 🤣 Each one consumes around ~400W of power each (not including the machine that houses them!)
@prologic@twtxt.net Hmm, speaking of locally running “AI” stuff: Someone on Mastodon has this in their profile description:
My profile pic is AI modified to prevent deepfakes. I used local Stable Diffusion on my solar powered 7900XTX to average a few selfies.
That sounds like a fun thing to do. Do I have a chance of doing that on my old box from 2013 without a dedicated GPU? 😂