unfaithful-functor

END my suffering

  • 2 Posts
  • 9 Comments
Joined 1 year ago
cake
Cake day: July 27th, 2023

help-circle





  • The next comment is so peak tech hubris to me.

    It’s “just” predicting the next token so it means nothing

    This form of argument should raise red flags for everyone. It is an argument against the possibility of emergence, that a sufficient number of simple systems cannot give rise to more complex ones. Human beings are “just” a collection of cells. Calculators are “just” a stupid electric circuit.

    The fact is, putting basic components together is the only way we know how to make things. We can use those smaller component to make a more complex thing to accomplish a more complex task. And emergence is everywhere in nature as well.

    This is the part of the AGI Discourse I hate because anyone can approach this with aesthetic and analogy from any field at all to make any argument about AI and its just mind-grating.

    This form of argument should raise red flags for everyone. It is an argument against the possibility of emergence, that a sufficient number of simple systems cannot give rise to more complex ones. Human beings are “just” a collection of cells. Calculators are “just” a stupid electric circuit.

    I’ve never seen a non-sequitur more non. The argument is that predicting the next term is categorically not what language is. That is, it’s not that there is nothing emerging, but that what is emerging is just straight up not language.

    The fact is, putting basic components together is the only way we know how to make things. We can use those smaller component to make a more complex thing to accomplish a more complex task. And emergence is everywhere in nature as well.

    “Look! This person thinks predicting the next token is not consciousness. I bet they must also not believe that humans are made of cells, or that many small things can make complex thing. I bet they also believe the soul exists and lives in the pineal gland just like old NON-SCIENCE PEOPLE.”




  • I’m just as clueless. I think there are three syllogisms that tech brains orbit around.

    • AI will improve society, a better society would face climate change more effectively, so AI will help us face climate change.
      • Basically an extension of the milder claims with regards to AI improving education, health care, research, the economy, etc.
      • Very vague and feel-good and I think more a reaction to distrust of AI than an assertion of anything.
    • AI will help us better understand complex systems like climate, understanding the complexity of climate change helps us, so AI will help us face climate change.
      • Stemming from skepticism of current climate science methodology that doesn’t fit with what they think science should be.
      • Secretly hope that AI will show us climate change isn’t actually even real and the fact we think it is is some byproduct of our feeble minds trying to understand something so dynamic and complex.
    • There’s some magic bullet technological solution to climate change that is outside of current human ability space to invent. AI can potentially eclipse these limits and invent things we can’t. So AI can invent this magic solution.
      • Hardcore AI singularity takeoff yadda yadda folks. Goes hand in hand with the ideas AI will invent microorganisms or nanobots that will take over the entire biosphere or thinking it will find a new theory of physics that lets it teleport places or shit like that.
      • In this POV climate is even a non-issue since AI could easily solve it but we can’t easily solve how to not make this AI kill us.