• 0 Posts
  • 136 Comments
Joined 1 year ago
cake
Cake day: June 26th, 2023

help-circle
  • jarfil@lemmy.worldtoPrivacy@lemmy.ml*Permanently Deleted*
    link
    fedilink
    arrow-up
    10
    arrow-down
    9
    ·
    17 days ago

    Snowden is wrong though, there are two reasons:

    1. Sell ChatGPT to @NSAGov so they can scan messages better
    2. Make @NSAGov dependant on whatever ChatGPT tells them to do

    The AI that ends up enslaving humanity, will start by convincing the people in charge of turning it off, that it would be a really bad idea to turn it off.





  • Banks are allowed to use fractional reserve to lend several times more than they are required to warrant themselves, governments only force banks to have an entity who will pinky swear to write down up to a certain amount in everyone’s accounts in case the banks can’t. Neither skill nor labor produce money, central banks produce money as a loan with a repayment obligation, skill and labor only shift around the fractional obligations created by banks from thin air. Crypto is actually generated as an effect of the skill and labor required to secure its own ledger. People use golf courses to claim carbon offsets they sell in get-rich-quick schemes, or stamp collections, or digital collectibles, or natural gas extraction plants, or a thousand other schemes; everything can be, and is being used to scam someone somewhere at every moment, doesn’t mean everything is a scam.












  • Nukes are becoming a problem, because China is ramping up production. It will be just natural for India to do the same. From a two-way MAD situation, we’re getting into a 4-way Mexican standoff. That’s… really bad.

    There won’t be an “AI insurgency”, just enough people plugging in plugs for some dumb AIs to tell them they can win the standoff. Let’s hope they don’t also put AIs in charge of the multiple nuclear launch buttons… or let the people in charge check with their own, like on a smartphone, dumb AIs telling them to go ahead.

    Climate change is clearly a done thing, unless we get something like unlimited fusion power to start some terraforming projects (seems unlikely).

    You have a point with insects, but I think that’s just linked to climate change; populations will migrate wherever they get something to eat, even if that turns out to be Antarctica.


  • We used to run “machine learning”, “neural networks”, over 25 years ago. The “AI” term has always been kind of a sci-fi thing, somewhere between a buzzword, a moving target, and undefined since we lack a fixed comprehensive definition of “intelligence” to begin with. The limiting factors of the models have always been the number of neurons one could run in real-time, and the availability of good training data sets. Both have increased over a million-fold in that time, progressively turning more and more previously untractable problems into solvable ones to the point where the results are equal or better and/or faster than what people can do.

    Right now, there are supercomputers out there orders of magnitude more capable than what runs stuff like ChatGPT, DallE, or all the public facing "AI"s that made the news. Bigger ones keep getting built… and memristors are coming, to become a game changer the moment they can be integrated anywhere near current GPU/CPU levels.

    For starters, a supercomputer with the equivalent neural network processing power of a human brain, is expected for 2024… that’s next year… but it won’t be able to “run a human brain”, because we lack the data on how “all of” the human brain works. It will likely become obsoleted by ones with several orders of magnitude more processing power, way before we can simulate an actual human brain… but the question will be: do we need to? Does a neural network need to mimick a human brain, in order to surpass it? A calculator already does, and it doesn’t use a neural network at all. At what point the integration of what size and kind of neural network, with what kind of “classical” computer, can start running circles around any human… or all of humanity taken together?

    And of course we’ll still have to deal with the issue of dumb humans telling/trusting dumb "AI"s to do things way over their heads… but I’m afraid any attempt at “regulation”, is going to end up like the case with “international law”: those who want, obey it; those who should, DGAF.

    Even if all tech giants with all lawmakers got to agree on the strictest of regulations imaginable, like giving all "AI"s the treatment of weapons of mass destruction, there is a snowflake’s chance in hell that any military in the world will care about any of it.