• 1 Post
  • 981 Comments
Joined 3 years ago
cake
Cake day: July 9th, 2023

help-circle


  • Good anecdote but this is just hegemonic propaganda. Social media has also revealed the reality behind the hegemonic narrative. That’s what they’re actually afraid of.

    It’s not propaganda, it’s a fact. The rise of conspiracy theories becoming mainstream, the rise of fascist groups that are currently undermining global peace and stability, the ability for long-debunked pseudoscience to be treated as equal with science: all of that is facilitated by social media giving an equal platform to people that do not deserve one, particularly the platforms run by capitalists. Social media has indeed done some good, but my argument was never that social media is wholly bad, just that it’s a net negative.

    I agree that “they” are afraid of The People organizing and seeing through all the bullshit, but that’s not something unique that social media is able to facilitate, and it’s not something social media has been particularly effective at doing. People of the past were able to see through the bullshit without social media, and if we all lost the internet tomorrow, people would still manage to communicate and share ideas. We did it for decades through books, newspapers, speaking events, zines, etc.

    We don’t need social media to progress, and I would argue that recent history seems to indicate the contrary.

    It’s not true. What about the people in charge of this platform? The bulk of the issues arise from capitalism and this type of censorship is designed to abolish its criticism.

    There are no people “in charge” of this platform. If you wanted to, you could spin up your own instance with the sole member being you. You could fork the code and start your own Lemmy v2.0. We are collectively responsible for the operation of this federation of services, and even here, you still find the tolerance of bad actors and the spread of rotten ideas.

    Has the Fediverse been a net positive? Maybe. But we are small fish compared to the fat cats that are Facebook, Instagram, TikTok, YouTube, Xitter, etc., and there’s no dispute that their influence has reached far and the ideas they’ve allowed to fester for profit have been destructive, to say the least.

    Social media doesn’t exist in a vacuum; it’s within the context of a global society run by greed, and the fact that it sometimes does good doesn’t outweigh the capitalists who weaponize it against us.


  • I mean, it has enabled every goober and bad actor with an opinion to essentially have a megaphone and build platforms and movements. I’d argue that’s a net negative. Even the Fediverse isn’t immune to propaganda and conspiracy theories.

    I think putting a warning on the tin is appropriate, especially for platforms run by billionaires whose explicit goal is to get people hooked and keep them feeding the machine by any means necessary.

    It’s true that the bulk of the issue arises from the people in charge of the platforms, but nobody currently in power is going to do anything about the billionaire problem. This is at least a vague gesture acknowledging that a problem exists. Also, it’s just a sign. When have warning signs stopped people from doing things that are unhealthy?









  • Human digital interfaces aren’t a secret, but other things like remote-viewing, etc. have been known about for a long time, and they were failures. There’s even a whole movie about it called Men Who Stare At Goats. Pointing to a few examples of actual conspiracies or weird projects doesn’t mean every claim has validity. It just means the government is generally untrustworthy, but that also means you need to take each claim individually, in practice. You can’t just generalize and say that “government untrustworthy, therefore believe the opposite of anything they say.” That’s being reactive, not skeptical.

    That’s not to say that there’s not scary tech out there (it’s been demonstrated that they can not only see but hear conversations through walls by interpolating Wi-Fi signals), but it’s all very much within the realm of science, not the paranormal.


  • It’s not just the money. It’s the knowledge and expertise needed to use the algorithms, at all…Not everyone has the time, energy, and attention to learn that stuff.

    I agree. That does not mean that LLMs are leveling the playing field with people who can’t/won’t get an education regarding computer science (and let’s not forget that most algorithms don’t just appear; they’re crafted over time). LLMs are easy, but they are not better or even remotely equivalent. It’s like saying, “Finally, the masses can tell a robot to build them a table,” and saying that the expertise of those “elite” woodworkers is no longer needed.

    …damn if I am tired of having to rely on “Zillow and a prayer” if I want to get a house or apartment.

    And this isn’t a problem LLMs can solve. I feel for you, I do. We’re all feeling this shit, but this is a capitalism problem. Until the ultracapitalists who are making these LLMs (OpenAI, Google, Meta, xAI, Anthropic, Palantir, etc.) are no longer the drivers of machine learning, and until the ultracapitalist companies stop using AI or algorithms to decide who gets what prices/loans/rental rates/healthcare/etc., we will not see any kind of level playing field you or the author are wishing for.

    You’re looking at AI, ascribing it features and achievements it doesn’t deserve, then wishing against all the evidence that it’s solving capitalism. It’s very much not, and if anything, it’s only exacerbating the problems caused by it.

    I applaud your optimism—I was optimistic about it once, too—but it has shown, time and again, that it won’t lead to a society not governed by the endless chasing of profits at the expense of everyone else; it won’t lead to a society where the billionaires and the rest of us compete on equal footing. What we regular folk have gotten from them will not be their undoing.

    If you want a better society where you don’t have to claw the most meager of scraps from the hand of the wealthy, it won’t be found here.


  • Okay. Claims are not evidence. “I read it somewhere” is not even close to substantial, because anyone can write anything they want on the Internet. Without evidence or even consensus amongst experts, it just sounds like a conspiracy theory.

    The CIA is often the bogeyman, because they do lots in secret, and the government is inherently untrustworthy. That doesn’t mean they have wireless brain interfaces, however.


  • No they’re not. That’s just the claptrap the billionaire Tech Bros want you to believe in. “Ooo, AGI is just around the corner! Buy in now to get it first! Ooo!”

    They just have access to militarized versions through specialized LoRAs and no restraints. It’s not anything beyond what’s possible for regular people right now, it’s just that regular people will never get access to the kind of training data needed to achieve the same results (not that the government should be able to, either).


  • Then along comes the language model. Suddenly, you just talk to the computer the way you’d talk to another human, and you get what you ask for.

    That’s not at all how LLMs work, and that’s why people are saying this whole premise is a bad take. Not only do LLMs get things wrong, they do it in such a way that it completely fabricates answers at times; they do this, because they’re pattern generation engines, not database parsers. Algorithms don’t do that, because they digest a set of information and return a subset of that information.

    Also, so what if algorithms cost a lot of money? That’s not really an argument for why LLMs level the playing field. They’re not analogous to each other, and the LLMs being foisted on the unassuming public by the billionaires are certainly not some kind of power leveler.

    Furthermore, it takes a fuckton more processing resources to run an LLM than it does an algorithm, and I’m just talking about cycles. If we went beyond just cycles, the relative power needed to solve the same problem using an LLM versus an algorithm is not even close. There’s an entire branch of mathematics dedicated to algorithm analysis and optimization, but you’ll find no such thing for LLMs, because they’re not remotely the same.

    No, all we have are fancy chatbots at the end of the day that hallucinate basic facts, not especially different from the annoying Virtual Assistants of a few years ago.