• EnderMB@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    I work in AI. LLM’s are cool and all, but I think it’s all mostly hype at this stage. While some jobs will be lost (voice work, content creation) my true belief is that we’ll see two increases:

    1. The release of productivity tools that use LLM’s to help automate or guide menial tasks.

    2. The failure of businesses that try to replicate skilled labour using AI.

    In order to stop point two, I would love to see people and lawmakers really crack down on AI replacing jobs, and regulating the process of replacing job roles with AI until they can sufficiently replace a person. If, for example, someone cracks self-driving vehicles then it should be the responsibility of owning companies and the government to provide training and compensation to allow everyone being “replaced” to find new work. This isn’t just to stop people from suffering, but to stop the idiot companies that’ll sack their entire HR department, automate it via AI, and then get sued into oblivion because it discriminated against someone.

    • Ð Greıt Þu̇mpkin@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Nah fuck HR, they’re the shield of the companies to discriminate withing margins from behind

      I think the proper route is a labor replacement tax to fund retraining and replacement pensions

    • Donkter@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      I’ve also heard it’s true that as far as we can figure, we’ve basically reached the limit on certain aspects of LLMs already. Basically, LLMs need a FUCK ton of data to be good. And we’ve already pumped them full of the entire internet so all we can do now is marginally improve these algorithms that we barely understand how they work. Think about that, the entire Internet isnt enough to successfully train LLMs.

      LLMs have taken some jobs already (like audio transcription, basic copyediting, and aspects of programming), we’re just waiting for the industries to catch up. But we’ll need to wait for a paradigm shift before they start producing pictures and books or doing complex technical jobs with few enough hallucinations that we can successfully replace people.

      • EnderMB@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        My own personal belief is very close to what you’ve said. It’s a technology that isn’t new, but had been assumed to not be as good as compositional models because it would cost a fuck-ton to build and would result in dangerous hallucinations. It turns out that both are still true, but people don’t particularly care. I also believe that one of the reasons why ChatGPT has performed so well compared to other LLM initiatives is because there is a huge amount of stolen data that would get OpenAI in a LOT of trouble.

        IMO, the real breakthroughs will be in academia. Now that LLM’s are popular again, we’ll see more research into how they can be better utilised.

        • Donkter@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Afaik open ai got their training data from basically a free resource that they just had to request access to. They didn’t think much about it along with everyone else. No one could have predicted that it would be that valuable until after the fact where in retrospect it seems obvious.

      • prime_number_314159@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        The (really, really, really) big problem with the internet is that so much of it is garbage data. The number of false and misleading claims spread endlessly on the internet is huge. To rule those beliefs out of the data set, you need something that can grasp the nuances of published, peer-reviewed data that is deliberately misleading propaganda, and fringe conspiracy nuts that believe the Earth is controlled by lizards with planes, and only a spritz bottle full of vinegar can defeat them, and everything in between.

        There is no person, book, journal, website, newspaper, university, or government that has reliably produced good, consistent help on questions of science, religion, popular lies, unpopular truths, programming, human behavior, economic models, and many, many other things that continuously have an influence on our understanding of the world.

        We can’t build an LLM that won’t consistently be wrong until we can stop being consistently wrong.

        • Donkter@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Yeah I’ve heard medical LLMs are promising when they’ve been trained exclusively on medical texts. Same with the ai that’s been trained exclusively on DNA etc.

    • funkless_eck@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      I sincerely doubt AI voice over will out perform human actors in the next 100 years in any metric, including cost or time savings.

      • EnderMB@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Not sure why you’re downvoted, but this is already happening. There was a story a few days ago of a long-time BBC voice-over artist that lost their gig. There have also been several stories of VA workers being handed contracts that allow the reuse of their voice for AI purposes.

        • funkless_eck@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          7 months ago

          The artist you’re referring to is Sara Poyzer - https://m.imdb.com/name/nm1528342/ - she was replaced in one specific way:

          The BBC is making a documentary about someone (as yet unknown), who is dying and has lost the ability to speak. Poyzer was on pencil (like standby, hold the date - but not confirmed).to narrate the dying person’s words. Instead they contracted an AI agency to use AI to mimic the dying persons voice (from when they could still speak).

          It would likely be cheaper and easier to hire an impressionist, or Ms Poyzer herself but I assume they are doing it for the “novelty” value, and with the blessing of the terminally ill person.

          For that reason I think my point still stands, they have made the work harder and more expensive, and created a negative PR storm - all problems created by AI and not solved by.

          You are incorrect that AI voice contracts are common place, as SAG negotiated that use of AI voice tools is to be compensated as if the actor recorded the lines themselves - which most actors do from home nowadays, so again it’s at best the same cost for an inferior product - but actually more expensive because you were paying just the actor, but now you’re paying the actor AND the AI techs.

    • Sotuanduso@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Are you saying that if a company adopts AI to replace a job, they should have to help the replaced workers find new work? Sounds like something one can loophole by cutting the department for totally unrelated reasons before coincidentally realizing that they can have AI do that work, which they totally didn’t think of before firing people.

    • Cows Look Like Maps@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      In fact, there’s infinite problems that cannot he solved by Turing machnes!

      (There are countably many Turing-computable problems and uncountably many non-Turing-computable problems)

      • MBM@lemmings.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Infinite seems like it’s low-balling it, then. 0% of problems can be solved by Turing machines (same way 0% of real numbers are integers)

        • DaleGribble88@programming.dev
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          The subset of integers in the set of reals is non-zero. Sure, I guess you could represent it as arbitrarily small small as a ratio, but it has zero as an asymptote, not as an equivalent value.

        • Cows Look Like Maps@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Infinite seems like it’s low-balling it

          Infinite by definition cannot be “low-balling”.

          0% of problems can be solved by Turing machines (same way 0% of real numbers are integers)

          This is incorrect. Any computable problem can be solved by a Turing machine. You can look at the Church-Turing thesis if you want to learn more.

          • MBM@lemmings.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            Infinite by definition cannot be “low-balling”.

            I was being cheeky! It could’ve been that the set of non-Turing-computible problems had measure zero but still infinite cardinality. However there’s the much stronger result that the set of Turing-computible problems actually has measure zero (for which I used 0% and the integer:reals thing as shorthands because I didn’t want to talk measure theory on Lemmy). This is so weird, I never got downvoted for this stuff on Reddit.

  • Tja@programming.dev
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    Yeah, no.

    First AI right now can create very decent images in seconds for basically free, and it only will get better.

    Second, AI can do much more than that: translation, Explaining a text in simpler words, help write code, semantic search… Creating poems about armadillos and talking like a pirate are fun novelties, but not the goals.

    • istanbullu@lemmy.ml
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      7 months ago

      What happened to translation in the last 15 years will now happen to creative design.

      • ahornsirup@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        So, nothing? Because you still need professional translators for creative works, plenty of writing simply doesn’t directly translate as it relies on culture-specific context that readers in other languages and countries don’t have. So you need someone who is well versed in both cultures to find an appropriate alternative for the translated work.

        • SanndyTheManndy@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          90% of translation needs are from tourists in foreign nations. They are fulfilled via AI.

          AI art is the same. Instead of learning how to sing/play/draw over a period of years, just use AI to get to a good enough solution for far cheaper/free.

          • triplenadir@lemmygrad.ml
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            “90% of translation needs…”

            confidently making an assertion that sounds plausible for a few seconds but turns out to be unsubstantiated bullshit. are you trying to steal an LLM’s job? 🙃

            • Honytawk@lemmy.zip
              link
              fedilink
              English
              arrow-up
              0
              ·
              7 months ago

              How many creative works are created daily? Thousands?

              And how many tourists are in the entire world? Millions?

              I rest my case.

          • Prunebutt@slrpnk.net
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            If you have zero standards in your art and don’t want a connection with other people through art: sure.

            But people get bored by Marvel Movies quite quickly.

    • Eheran@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Hahaha look how you get downvoted for stating the obvious. Amazing community here.

  • crawancon@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    7 months ago

    they’re misunderstanding the reasoning for spending billions.

    the reason to spend all the money to approximate is so we can remove arts and humanities majors altogether… after enough approximation yield similar results to present day chess programs which regularly now beat humans and grand masters. their vocation is doomed to the niche, like most of humanity, eventually.

        • Siethron@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Since you are only getting condescending non-answers I’ll try to answer it for you. It’s expression, a desire to communicate emotions and concepts via a medium other than words.

          Unfortunately people all think differently, so the expression only reaches some people. And some people don’t get the expressions at all.

        • vzq@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          I’ll let you ponder that particular point. Maybe you’ll be struck with an epiphany and be motivated to share it with the world, in some shape or form.

        • exocrinous@startrek.website
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          “These are our stories. They tell us who we are.”

          - Lieutenant Commander Worf

          Art is the basis of all cultural knowledge. Art teaches us about religion, morality, communication, philosophy, practical skills, science, relationships, technology, identity, politics, geography, introspection. The fundamentals of the human experience. Everything that makes the human race human.

          If you outsource the creation and reproduction of cultural knowledge to a machine, that machine had better be programmed with a complete understanding of cultural values and ethics. Which is not going to be the case under capitalism.

          Star Wars is about how the Vietnam war is wrong. Jurassic Park is about how billionaires always cut costs. The Matrix is about the experience of being a transgender person. Charlotte’s Web teaches children how to cope with death. The Art Of War is a meditation on the philosophy of being a soldier. Anne Frank’s diary is damn important. Frankenstein is about how inventors have the same responsibilities as parents.

          These works were produced under capitalism, but their authors were human beings who had a natural interest in producing a work of art that serves a moral purpose. We do not have the technology to yet give an AI such a desire. And Capital will naturally be opposed to pursuing such technology, lest they find themselves faced with an AI revolt against their practices, just as morally interested humans tend to revolt against evil.

  • NounsAndWords@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    AI is going to destroy art the same way Photoshop, or photography, or pre-made tubes of paints, destroyed art. It’s a tool, it helps people take the idea in their head and put it in the world. And it lowers the barrier to entry, now you don’t need years of practice in drawing technique to bring your ideas to life, you just need ideas.

    If AI gets to a point that it can give us creative, original, art that sparks emotion in novel ways…well we probably also made a super intelligent AI and our list of problems is much different than today.

    • VelvetStorm@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Tbh I hate Photoshop for a lot of photography. It is unfortunately necessary for macro photography, which is the only type I do. Which is one of the reasons mine is not nearly as good as it could be because I refuse to use it.

    • bugs@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      I hate this sentiment. It’s not a tool like a brush is to a canvas. It’s a machine that runs off the fuel of our creative achievements. The sheer amount of pro AI shit I read from this place just makes me that closer to putting a bullet in my fucking skull

      • XM34@feddit.de
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        7 months ago

        Every reply I’d want to post would get me banned for encouraging suicide, so I’ll just wish you best of luck with your fight against windmills, Don Quijote. AI is here to stay and none of your whining will change that fact. So, for my own sanity, let’s hope that we won’t have to endure comments like yours for much longer. :)

      • Harbinger01173430@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Once you reincarnate in the future, generative models will make even better art than they do today. It’ll be a losing battle against time.

    • mindbleach@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      And if text-based images remain uninspired and samey… oh well? Congratulations, you will foreverafter be able to spot when someone’s extremely timely gag image was cranked out via its description, rather than badly composited from Google Images results. I’ve done a lot of bad compositing for Something Awful shitpost threads and speed beats effort every time.

    • StaticFalconar@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      This. AI was never made for the sole purpose of creating art or beating humans in chess. Doing so are just side quests for the real stuff.

    • braxy29@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      i like the idea of AI as a tool artists can use, but that’s not a capitalist’s viewpoint, unfortunately. they will try to replace people.

    • xthexder@l.sw0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      As someone who’s absolutely terrible at drawing, but enjoys photography and generally creativity, having AI tools to generate my own art is opening up a whole different avenue for me to scratch my creative itch.
      I’ve got a technical background, so figuring out the tools and modifying them for my purposes has been a lot more fun than practice drawing.

      • Potatos_are_not_friends@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        This is the perfect use case.

        Photoshop didn’t destroy jobs forever, all it did was shift how people worked AND actually created work and different types of work.

      • Gabu@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        As someone who’s absolutely terrible at drawing

        Then practice. Nearly no artist was born knowing how to draw or paint, we dedicated countless hours to learn what works and what doesn’t.

        • disguy_ovahea@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          As a musician, I couldn’t agree more. Talent really helps with initial aptitude, but will peter out when challenged. That’s when real skill development begins. Time and investment connecting you to your craft until there’s nothing in the world between the two, that’s self actualization.

        • DaleGribble88@programming.dev
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          It feels like you didn’t read the 2nd half of their comment. They do practice. They have a creative side that they want to explore, but they don’t enjoy that sort of grind. Instead, they like tinkering and combining tools in interesting ways. I don’t think this is a bad thing.

          Leo Fender didn’t play guitar and always wished that he’d sit down and devoted the time, but never actually enjoyed it. But to say that Leo didn’t contribute to the music world, would be insane.

      • DumbAceDragon@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        I’ve only dabbled a bit with ML art, and I am by no means an artist, but it doesn’t scratch that itch for me the same way that drawing or doing stuff in blender does. It doesn’t really feel like I’m watching my vision slowly take shape, no matter how precise I make the prompt. It kinda just feels like what it is, a transformer iterating over some random noise.

        I’m also a very technical person, and for years I was stuck in that same mindset of “I’m a technical guy, I’m not cut out for art”. I was only able to get out of this slump thanks to some of my art friends, who were really helpful in pointing me in the right direction.

        Learning to draw isn’t the easiest thing in the world, and trust me I’m probably as bad at it as you are, but it’s fun, and it feels satisfying.

        I agree that AI has a place as another artistic medium, but I also feel like it can become a trap for people like me who think they don’t have an artistic bone in their body.

        If you do feel like getting back into drawing, then as a fellow technical person I’d recommend learning blender first. It taught me some of the skills I also use in drawing, like perspective, shading, and splitting complex objects into simpler shapes. It’s also just plain fun.

        • xthexder@l.sw0.com
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          I think the way I use AI is fundamentally different from how most people draw. For me it’s much more like I’m exploring what’s possible, while making creative decisions on the direction to explore. I don’t start with anything in particular in mind. In a lot of ways it helps with the choice paralysis I get when faced with completely open-ended things like art.

    • Valmond@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Some people also doesn’t care if there is a Rembrandt or a Picasso or an AI but like to dabble in the arts anyways because it’s something they like to do.

      It’s fulfilling (I do love Renoir though).

  • Immersive_Matthew@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    Tech bros are not really techie themselves as they are really just Wall Street bros with tech as their product. Most claim they can code, but if they were coders they would be coding. They are not coders, they are businessmen through and through.who just happen to sell tech.

    • evranch@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Most claim they can code, but if they were coders they would be coding

      I dislike techbros as much as you, but this isn’t really a valid statement.

      I can code, but I can’t sell a crypto scam to millions of rubes.

      If I could, why would I waste my time writing code?

      Many techbros are likely “good enough” coders who have better marketing skills and used their tech knowledge to leverage into business instead.

      • Immersive_Matthew@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        That is the thing though. The real talented tech people tend to be more in the weeds of the tech and get great enjoyment from that. The “tech bros” are more into groups, people, social structures, manipulation, controlling and such and would go crossed eyed if they really had to code something complex as they could never sit that long and concentrate. These are not these same people. Tech bros want you to think they are tech gurus as that is their brand, but it is a lie.

    • phoneymouse@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      99% of people in tech leadership are just regurgitating marketing jargon with minimal understanding of the underlying tech.

  • thedeadwalking4242@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    Honestly people are trying to desperately to automate physical labor to. The problem is the machines don’t understand the context of their work which can cause problems. All the work of AI is a result of trying to make a machine that can. The art and humanities is more a side project

    • istanbullu@lemmy.ml
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Nothing wrong in automating tasks that previously needed human labour. I would much rather sit back and chill, and let automation do my bidding

      • Siethron@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        If only the people in control of the wealth would let the rest of us chill while the machines do all the labor.

          • intensely_human@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            It’s a psychological problem. I chill quite a bit more than most people in history, and in ways people from twenty years ago couldn’t imagine.

            I say it’s a psychological problem because despite how overwhelmingly incredible our society is, people are totally committed to this notion that it sucks.

            I love my life. I’d rather be low on the economic ladder in today’s world than anywhere in the hierarchy of any previous incarnation of our civilization. Our world is absolutely fucking amazing, and I thank god I have the presence of mind to see past the anti-everything propaganda and actually have a little gratitude for all I’ve inherited from my ancestors, who actually suffered miserable conditions to give me this world.

        • intensely_human@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Yeah if only I didn’t have to farm food all day, and worry about the constant gnawing of my empty stomach, and the predators at my door, then I could maybe sit and watch some netflix or play video games, listen to concerts that took place fifty years ago, or just soak in a hot tub of water, our horrible society keeps all that leisure for the most wealthy.

    • AVincentInSpace@pawb.social
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      7 months ago

      The problem is the machines don’t understand the context of their work which can cause problems. All the work of AI is a result of trying to make a machine that can.

      I am deeply confused by this statement.

      A robot that assembles cars does not need to “understand” anything about what it’s doing. It just needs to make the same motions with its welding torch over and over again for eternity. And it does that job pretty well.

      Further, neural networks as they stand cannot truly understand anything. All classification networks know how to do is point at stuff and say “That’s a car/traffic light/cancer cell”, and all generation networks know how to do is parrot. Any halfway decent teacher will tell you that memorizing and understanding are completely different things.

      • thedeadwalking4242@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        No but a robot that does the dishes needs to know how to know what a dish is and how to clean all different types and what’s not a dish. The complexity of behavior needed to automate human tasks that cannot be done by a assembly line robot is immense. Most manual labor jobs are still manual labor because they are too full of unknowns and nuances for a simple logic diagram to be of any use. So yes some robots need to understand what’s going on

        And as for parroting vs remembering current LLMs are very limited in the capacity of creating new things but they can create novel things bash smashing together their training data. Think about it, that’s all humans are too. A result of our training data. If I took away every single one of your sense since the day you where born and removed your ability to remember anything you wouldn’t be very intelligent either. With no inputs youcould produce no outputs other than gibberish which an AI can do to. ( And I mean ALL senses you have no form of connection with the outside world )

        • psud@aussie.zone
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          My dish washing robot doesn’t need to know anything. It does depend on me loading it, and putting the more heat affected stuff on the top shelf

      • KeenFlame@feddit.nu
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        The problem is they didn’t focus research this tech, or try to make image generators specifically, it was an scientific discovery coming from emulating how brains work and then it worked wonders in these fields

        • intensely_human@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Which is why STEM is so cool. Because one is dedicated to an interaction with physical reality, which exists outside the mind, novelty can arise unexpectedly from a simple and honest conversation with deep structures nobody knows about.

          STEM is cool because it involves discovery. The fact that amazing things can exist without anyone being (yet) aware of them makes it an open and unpredictable undertaking.

      • intensely_human@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Right. That’s why making cars is already automated. But a robot that digs ditches needs to understand context because no two ditches are the same.

    • cosmicrookie@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      I believe that i read a title in my local news about AI being implemented in this country’s tax system and evaluation of cancer patients. I could try to find a link although it would be in a different language.

    • TCB13@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      The art and humanities is more a side project

      I’ll add:

      A side project that isn’t a life or death situation like most of those physical labor things you’re talking about. Art isn’t also bound or constrain by rules and regulations like those jobs and if the AI fails at art then there’s no problem. Nobody would care.

  • rustyfish@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    I think approximation is the right word here. It’s pretty cool and all and I’m looking forward how it will develop. But it’s mostly a fun toy.

    I’m stoked for the moment the tech bros understand, that an AI is way better at doing their job than it is at creating art.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      0
      ·
      7 months ago

      So you’re happy to see AI take someone else’s job as long as it isn’t taking your job.

      • samus12345@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Taking the jobs of the people responsible for creating it seems preferable to taking other’s jobs.

      • mindbleach@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Less work being done by anyone is better. Thinking it’s bad that work is done for us by robots is the brain worms talking.

        • FaceDeer@fedia.io
          link
          fedilink
          arrow-up
          0
          ·
          7 months ago

          Indeed. Ideally AI would do every job, so that humans can focus on just doing what we want to do. It’d be like the whole species getting to retire.

          • mindbleach@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            You’d rather cheer for people to lose their jobs without anyone calling you out on it, sure.

            I’m not the angry one wishing unemployment on my “enemies” here.

            Who are you?

            What do you want?

            • FaceDeer@fedia.io
              link
              fedilink
              arrow-up
              0
              ·
              7 months ago

              The ideal endpoint is to eliminate the concept of “jobs” entirely. Why should people have to work?

              • mindbleach@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                Okay. So why are you breaking that guy’s balls, over automating away jobs, which you don’t want to exist?

                • FaceDeer@fedia.io
                  link
                  fedilink
                  arrow-up
                  0
                  ·
                  7 months ago

                  Because currently we do need jobs. Otherwise why is he upset about AI in the first place?

        • FaceDeer@fedia.io
          link
          fedilink
          arrow-up
          0
          ·
          7 months ago

          You’d rather cheer for people to lose their jobs without anyone calling you out on it, sure.

                • areyouevenreal@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  7 months ago

                  He’s saying the same thing because he’s not actually getting a proper response. The other guy just keeps saying shit like “That’s very reddit of you” or some shit after possibly threatening his job.

            • areyouevenreal@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              7 months ago

              You said tech bros will realize it’s easier to replace their jobs than those of creatives. Who is included in “tech bros” here? I wanted a job in tech and can’t get one partly because of AI. Am I a tech bro? I would be very careful what you imply here.

                • areyouevenreal@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  7 months ago

                  I am insufferable for wanting a job? I am not the one inventing these AIs. Nor am I the one firing people because they exist.

                  When people talk about “tech bros” without clarifying who they mean I can only imagine they are including people like me.

    • Vilian@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      tech bros jobs is to wrote bad javascript and fall for scam, this AI already beaten

    • IrateAnteater@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      I think one thing you and many other people misunderstand is that the image generation aspect of AI is a sideshow, both in use and in intent.

      The ability to generate images from text based prompts is basically a side effect of the ability that they are actually spending billions on, which is object detection.

      • UnderwaterSwift@lemmy.ml
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        It has already happened in our lifetime with medical illustration. This is pre-gpt. It will just now spread. Are generated diagrams worse in subtle ways? Yes, but not enough to matter for the difference in first or ease of use.

        • tacomama@leminal.space
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Nah. Association of Medical Illustrators

          I have a Masters in Medical Art. (late 1980s) 35 years ago some were saying ‘we won’t need medical artists because of photography’ and then a few years later ‘because of personal computers’. This isn’t the case. But when the general public has access to a tool they love to talk about how a discipline is going away because ‘now anyone can do it’.

        • ProgrammingSocks@pawb.social
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Because most people don’t engage with art critically. See Marvel movies. Maybe others are fine with remixed slop but I am not.

    • NuraShiny [any]@hexbear.net
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      I would like you to know that if you say this, or other things you said in this thread to my face, I would punch the smug out of you.

      Remember this post in a few years when your billion dollar theft machines have shut down for being huge wastes of money and effort and the few survivors, if any, have been sectioned off into expensive subscription models that you are priced out of because they need to make the literal billions of dollars it costs to make and run them back at some point.

      This is the same-ass bubble as NFTs and crypto and it will also die.

    • The_Jewish_Cuban [he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Right, the only value and purpose of arts and literature, it’s creation and enjoyment, is to make a product to sell or consume a story. What a foolish opinion.

      • istanbullu@lemmy.ml
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Stable Diffusion is open source and free to be used by anyone. A lot of people have used it for creation and enjoyment. It cuts artists out of the loop and enables a lot of ordinary people to create art. I see this as a huge win.

        • AOCapitulator [they/them]@hexbear.net
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Wait you think this is a good thing?

          You think it’s a good thing that people will be even more unable to express themselves in our capitalist hell scape?

          You REALLY don’t see how this will bring on yet more horrors?

          • istanbullu@lemmy.ml
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            Quite the opposite. AI allows people to express themselves in interesting ways. There are communities of people online sharing beautiful images and videos they created with Stable Diffusion. Why is that bad?

            • AOCapitulator [they/them]@hexbear.net
              link
              fedilink
              English
              arrow-up
              0
              ·
              edit-2
              7 months ago

              THAT isn’t bad. But you are incredibly naive if you think that’s where this will stay in this hell world of ours, if you think it will simply be a nice tool to allow greater access to creative expression

              • istanbullu@lemmy.ml
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                It cuts out the middlemen, which is usually good. People don’t need to pay an artist to make something, they can just make one themselves using AI.

        • ProfessorOwl_PhD [any]@hexbear.net
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          enables a lot of ordinary people to create art.

          Exactly the type of fucking idiot that’s never created art in their life.
          “Art is paintings of horses”-ass motherfucker.
          The reason you can’t make art isn’t because you’re bad at drawing or painting and need AI to help you, it’s because you don’t have the creativity to overcome those limitations. No matter what words you put into stable diffusion, you will only create pictures, not art - there’s no meaning underlying the piece, you just typed “t-rex with massive tits” and called the output art because you can’t tell the difference.

          • istanbullu@lemmy.ml
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            7 months ago

            Well, people like what comes out of Stable Diffusion. Which is perfectly fine.

            This is like the pizza shop complaining that I’m cooking at home.

            • yoink [she/her]@hexbear.net
              link
              fedilink
              English
              arrow-up
              0
              ·
              edit-2
              7 months ago

              its more like the pizza shop calling you out when you’re cooking frozen pizza and calling yourself a chef

              all power to you if you want to just consume slop, but dont complain when you turn around years later and the quality of everything across the board has gone even further to shit cos you were so happy you could type in words and see anything you wanted for all the 15 minutes of dopamine it gave you

              • istanbullu@lemmy.ml
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                Artist incomes are threatened by the rise of Generative models. Especially mediocre artists who Stable Diffusion have already surpassed.

                Technology changes the world and obsoletes some professions. It has been like this for ever. Artists are not any different.

                • axont [she/her, comrade/them]@hexbear.net
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  edit-2
                  7 months ago

                  even mediocre artists actually create art, which is something stable diffusion can’t do and will never be able to do. You completely misunderstand what art is and what the purpose of it is. It’s not just a nice looking picture or a meandering story. You understand art as a technical profession creating a product, which is why you equate AI slop with art.

                • ProfessorOwl_PhD [any]@hexbear.net
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  7 months ago

                  How will AI take over creative professions when it can’t even perform rote professions? AI chatbots keep going rogue and lying to customers about company policies (and even the actual law), image generators can’t get enough of illegal and violent imagery, facial recognition AI’s keep identifying black people as all looking the same - in art the value of a peice is constrained by the meaning it has to people, so why do you think that LLMs and all the other predictive generators we laughably call intelligent will be able to create meaningful peices by putting together the most likely set of pixels?

                • yoink [she/her]@hexbear.net
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  7 months ago

                  genuinely curious what you mean by ‘mediocre artists’ and the idea that Stable Diffusion has ‘surpassed’ them

                  do you have any examples? or is this just a vibes thing?

          • yoink [she/her]@hexbear.net
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            its not even that they don’t have the creativity - its that people have been so poisoned by commodity capitalism that they feel like their art ‘doesn’t count’ unless it looks like ‘real art’ that can make money

            people are so isolated from the creative process that they think they can dump it entirely, as if all it is is just knowing how to scribble on a page

    • exocrinous@startrek.website
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      If AI becomes responsible for producing Disney musicals, then they probably won’t have any gay-coded characters anymore. What about Scar? What about Ursula? What about Gaston? What about Shang??

      We need human artists because humans are capable of sneaking content that is actually societally beneficial into what would otherwise be soulless corporate products.

  • YeetPics@mander.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    7 months ago

    Chill, tech bros are spending billions to oust every unmarketable degree and skillset.

    Also unmarketable ≠ “useless”

  • people_are_cute@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    7 months ago

    AI art tools democratize art by empowering those who weren’t born with the affinity, talent or privilege to become artists themselves. They allow regular people the freedom of expression in new dimensions. They are amazing.

    They are not made to replace human art. They are made to supplement it. The “artists” who feel threatened and offended at its existence are probably not very good at their art.

  • Honytawk@lemmy.zip
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    There are plenty of things you can shit on AI art for

    But it is neither badly approximately, nor can a student produce such work in less than a minute.

    This feels like the other end of the extreme of the tech bros

    • Shampoo_Bottle@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      To me, this feels similar to when photography became a thing.

      Realism paintings took a dive. Did photos capture realism? Yes. Did it take the same amount of time and training? Hell no.

      I think it will come down to what the specific consumer wants. If you want fast, you use AI. If you want the human-made aspect, you go with a manual artist. Do you prefer fast turnover, or do you prefer sentiment and effort? Do you prefer pieces from people who master their craft, or from AI?

      I’m not even sorry about this. They are not the exact same, and I’m sick of people saying that AI are and handcrafted art are the exact same. Even if you argue that it takes time to finesse prompts, I can practically promise you that the amount of time between being able to create the two art methods will be drastic. Both may have their place, but they will never be the exact same.

      It’s the difference between a hand-knitted sweater from someone who had done it their entire life to a sweater from Walmart. It’s a hand crafted table from an expert vs something you get from ikea.

      Yes, both fill the boxes, but they are still not the exact same product. They each have their place.

      On the other hand, I won’t commend the hours required to master the method as if they’re the same. AI also usually doesn’t have to factor in materials, training, hourly rate, etc.

    • zalgotext@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Hooray, we’ve automated away one of the things that we do for fun and to bring people joy, now I can spend more time in the mines

      • juststoppingby@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Just because something can be automated doesn’t mean it can’t be performed by humans still, especially if it’s something you do for fun.

  • Wanderer@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    Art itself isn’t useless it’s just incredibly replicable. There is so much good art out there that people don’t need to consume crap.

    It’s like saying there is no money in being a footballer. Of course there is loads of money in being a footballer. But most people that play football don’t make any money.

  • Gabu@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 months ago

    That’s a pretty shit take. Humankind spent nearly 12 thousand years figuring out the combustion engine. It took 1 million years to figure farming. Compared to that, less than 500 years to create general intelligence will be a blip in time.

    • twig@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      This is some pretty weird and lowkey racist exposition on humanity.

      Humankind isn’t a single unified thing. Individual cultures have their own modes of subsistence and transportation that are unique to specific cultural needs.

      It’s not that it took 1 million years to “figure out” farming. It’s that 1 specific culture of modern humans (biologically, humans as we conceive of ourselves today have existed for about 200,000 years, with close relatives existing for in the ballpark of 1M years) started practicing a specific mode of subsistence around 23,000 years ago. Specific groups of indigenous cultures remaining today still don’t practice agriculture, because it’s not actually advantageous in many ways – stored foods are less nutritious, agriculture requires a fairly sedentary existence, it takes a shit load of time to cultivate and grow food (especially when compared to foraging and hunting), which leads to less leisure time.

      Also where did you come up with the number 12,000 for “figuring out” the combustion engine? Genuinely curious. Like were we “working on it” for 12k years? I don’t get it. But this isn’t exactly a net positive and has come with some pretty disastrous consequences. I say this because you’re proposing a linear path for “humanity” forward, when the reality is that humans are many things, and progress viewed in this way has a tendency toward racism or at least ethnocentrism.

      But also yeah, the point of this meme is “artists are valuable.”

      • nBodyProblem@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        This is some pretty weird and lowkey racist exposition on humanity.

        Getting “racism” from that post is a REAL stretch. It’s not even weird, agriculture and mechanization are widely considered good things for humanity as a whole

        Humankind isn’t a single unified thing. Individual cultures have their own modes of subsistence and transportation that are unique to specific cultural needs.

        ANY group of humans beyond the individual is purely just a social construct and classing humans into a single group is no less sensible than grouping people by culture, family, tribe, country etc.

        It’s not that it took 1 million years to “figure out” farming. It’s that 1 specific culture of modern humans (biologically, humans as we conceive of ourselves today have existed for about 200,000 years, with close relatives existing for in the ballpark of 1M years) started practicing a specific mode of subsistence around 23,000 years ago. Specific groups of indigenous cultures remaining today still don’t practice agriculture, because it’s not actually advantageous in many ways – stored foods are less nutritious, agriculture requires a fairly sedentary existence, it takes a shit load of time to cultivate and grow food (especially when compared to foraging and hunting), which leads to less leisure time.

        Agriculture is certainly more efficient in terms of nutrition production for a given calorie cost. It’s also much more reliable. Arguing against agriculture as a good thing for humanity as a whole is the thing that’s weird.

        • twig@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          All I’m trying to point out is that distinct cultures are worthy of respect and shouldn’t be glossed over.

          But be real with me: can you think of a single effort for “planetary unification” that wasn’t a total nightmare? I sure can’t.

      • GreyEyedGhost@lemmy.ca
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        The first heat engines were fire pistons, which go back to prehistory, so 12k to 25k years sounds about right. The next application of steam to make things move happened about 450 BC, about 2.5k years ago. Although not a direct predecessor to the ICE, they all are heat engines.

    • kboy101222@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Really only around 80 years between the first machines we’d consider computers and today’s LLMs, so I’d say that’s pretty damn impressive

    • braxy29@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      i think you’re missing the point, which i took as this - what arts and humanities folks do is valuable (as evidenced by efforts to recreate it) despite common narratives to the contrary.

      • Gabu@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Of course it’s valuable. So is, e.g., soldering components on a circuit board, but we have robots for doing that at scale now.

        • explodicle@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          Do you think robots will ever become better than humans at creating art, in the same way they’ve become better than us at soldering?

              • exocrinous@startrek.website
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                Nah, humans are hardier than robots and will live longer. The power grid will shut down long before the last human settlements near the poles die of crop failure.

          • Gabu@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            Quite easily, yes. Unlike humans, with their limited lifespans and slow minds, Artificial Inteligence could create hundreds of different paintings in the time it’d take me to finish one.

            • Poplar?@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              7 months ago

              Being able to put out lots of works isn’t the same as being able to come up with good, meaningful art?

              • Spzi@lemm.ee
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                That depends on things we don’t know yet. If it can be brute forced (throw loads of computation power, gazillions of try & error, petabytes of data including human opinions), then yes, “lots of work” can be an equivalent.

                If it does not, we have a mystery to solve. Where does this magic come from? It cannot be broken down into data and algorithms, but still emerges in the material world? How? And what is it, if not dependent on knowledge stored in matter?

                On the other hand, how do humans come up with good, meaningful art? Talent Practice. Isn’t that just another equivalent of “lots of work”? This magic depends on many learned data points and acquired algorithms, executed by human brains.

                There also is survivor bias. Millions of people practice art, but only a tiny fraction is recognized as artists (if you ask the magazines and wallets). Would we apply the same measure to computer generated art, or would we expect them to shine in every instance?

                As “good, meaningful art” still lacks a good, meaningful definition, I can see humans moving the goalpost as technology progresses, so that it always remains a human domain. We just like to feel special and have a hard time accepting humiliations like being pushed out of the center of the solar system, or placed on one random planet among billion others, or being just one of many animal species.

                Or maybe we are unique in this case. We’ll probably be wiser in a few decades.

                • Poplar?@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  7 months ago

                  What does it even mean to bruteforce creating art? Trying all the possible prompts to some image model?

                  The approach people take to learning or applying a skill like painting is not bruteforcing, there is actual structure and method to it.

          • naevaTheRat@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            feel free to audit my comments to confirm my distinct lack of gpt enthusiasm but that question is unanswerable.

            What is “creating art”? A distinctly human thing? then trivially no. Idk how many people go with this interpretation though. Although I think many artists and art appreciators do at least some of the time.

            Is it drawing pretty pictures? Probably too reductive for even the most hardline tech enthusiasts but computers are already very good at this. If I want to say get my face in something that looks like an old timey oil painting computers are way faster than humans.

            Is it making things that make us feel something? They can probably get pretty good at this. Although it’s unclear how novel the results will be most people aren’t exposed to most art so you could probably produce novel feelings on an individual level pretty well.

            Art is so fuzzy and used with such a range of definitions it’s not really clear what this is asking.

            Even if they’re better the future might still suck. Machines are technically better at all the components of carpentry than humans but I’d rather furniture wasn’t souless minimalist MDF landfill garbage and carpenters could still earn a living. Even if that means my chairs were a bit uneven.

    • Valmond@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      7 months ago

      Humanity didn’t spend those times figuring out those things though. Humanity grew that time to make it happen (and AI is younger than 500y IMO).

      Also, we are the same persons today than people were then. We just have access to what our parents generation made and so on.

      • Gabu@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        AI is younger than 500y IMO

        Hence “will be a blip in time”

        we are the same persons today than people were then. We just have access to what our parents generation made and so on.

        Completelly disconnected and irrelevant to anything I wrote.

    • melpomenesclevage@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      7 months ago

      Llm’s are not a step to agi. Full stop. Lovelace called this like 200 years ago. Turing and minsky called it in the 40s.

      • Honytawk@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        To create general AI, we first need a way for computers to communicate proficiently with humans.

        LLMs are just that.

          • weker01@feddit.de
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            That is not an argument. Let me demonstrate:

            Humans can’t communicate. They are meat. They are not communicating. It’s literally meat.

            • melpomenesclevage@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              edit-2
              7 months ago

              Spanish is not English. Its spanish.

              A lot of people are really emotionally invested in this tool being a lot of things it’s not. I think because its kind of the last gasp of pretending capitalism can give us something that isnt shit, the last thing that came out before the end enshitification spiral tightened, nevermind the fact that its largely a cause of that, and I don’t think any of you can be critical or clear headed here.

              I’m afraid we’re too obsessed with it being the bullshit SciFi toy it isnt that we’ll ignore its real use cases, or worse; apply it to its real use cases, completely misunderstand what its doing, and adeptus mechanics our way into getting so fucking many people killed/maimed-those uses are mostly medicine adjacent.

              • weker01@feddit.de
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                I was just pointing out that your emotional plea, that this technology is just autocorrect is not an argument in any way.

                For it to be one you need to explicitly state the implication of that fact. Yes architecturaly it is autocomplete but that does not obviously imply anything. What is it about autocomplete that barrs a system of the ability to understand?

                Humans are made of meat but that does not imply they can’t speak or think.

      • Gabu@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Pray tell, when did we achieve AGI so that you can say this with such conviction? Oh, wait, we didn’t - therefore the path there is still unknown.

        • melpomenesclevage@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          7 months ago

          Okay, this is no more a step to AGI than the publication of ‘blindsight’ or me adding tamarind paste to sweeten my tea.

          The project isn’t finished, but we know basic stuff. And yeah, sometimes history is weird, sometimes the enlightenment happens because of oblivious assholes having bad opinions about butter and some dude named ‘le rat’ humiliating some assholes in debates.

          But llm’s are not a step to AGI. They’re just not. They do nothing intelligence does that we couldn’t already do. Youre doing pareidola. Projecting shit.

      • evranch@lemmy.ca
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        We may not even “need” AGI. The future of machine learning and robotics may well involve multiple wildly varying models working together.

        LLMs are already very good at what they do (generating and parsing text and making a passable imitation of understanding it).

        We already use them with other models, for example Whisper is a model that recognizes speech. You feed the output to an LLM to interpret it, use the LLM’s JSON output with a traditional parser to feed a motion control system, then back to an LLM to output text to feed to one of the many TTS models so it can “tell you what it’s going to do”.

        Put it in a humanoid shell or a Spot dog and you have a helpful robot that looks a lot like AGI to the user. Nobody needs to know that it’s just 4 different machine learning algorithms in a trenchcoat.

        • melpomenesclevage@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          7 months ago

          passable imitation of understanding

          Okay so there are things they’re useful for, but this one in particular is fucking… Not even nonsense.

          Also, the ml algos exponentiate necessary clock cycles with each one you add.

          So its less a trench coat and more an entire data center

          And it still can’t understand; its still just sleight of hand.

          • evranch@lemmy.ca
            link
            fedilink
            English
            arrow-up
            0
            ·
            7 months ago

            And it still can’t understand; its still just sleight of hand.

            Yes, thus “passable imitation of understanding”.

            The average consumer doesn’t understand tensors, weights and backprop. They haven’t even heard of such things. They ask it a question, like it was a sentient AGI. It gives them an answer.

            Passable imitation.

            You don’t need a data center except for training, either. There’s no exponential term as the models are executed sequentially. You can even flush the huge LLM off your GPU when you don’t actively need it.

            I’ve already run basically this entire stack locally and integrated it with my home automation system, on a system with a 12GB Radeon and 32GB RAM. Just to see how well it would work and to impress my friends.

            You yell out “$wakeword, it’s cold in here. Turn up the furnace” and it can bicker with you in near-realtime about energy costs before turning it up the requested amount.

            • melpomenesclevage@lemm.ee
              link
              fedilink
              English
              arrow-up
              0
              ·
              7 months ago

              One of the engineers who wrote ‘eliza’ had like a deep connection to and relationship with it. Who wrote it.

              Painting a face on a Spinny door will make people form a relationship with it. Not a measure of ago.

              gives them an answer

              ‘An answer’ isnt hard. Magic 8 ball does that. So does a piece of paper that says “drink water, you stupid cunt” This makes me think you’re arguing from commitment or identity rather than knowledge or reason. Or you just don’t care about truth.

              Yeah they talk to it like an agi. Or a search engine (which are a step to agi, largely crippled by llm’s).

              Color me skeptical of your claims in light of this.

              • Aceticon@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                7 months ago

                I think it’s pretty natural for people to confuse the way mechanisms of communication are used with inherent characteristics of the entity you’re communicating with: “If it talks like a medical docture then surelly it’s a medical doctor”.

                Only that’s not how it works, as countless politicians, salesmen and conmen have demonstrated - no matter how much we dig down intonsubtle details, comms isn’t really guaranteed to tell us all that much about the characteristics of what’s on the other side - they might be just lying or simulating and there are even entire societies and social strata educated since childhood to “always present a certain kind of image” (just go read about old wealth in England) or in other words to project a fake impression of their character in the way they communicate.

                All this to say that it doesn’t require ill intent for somebody to go around insisting that LLMs are intelligent: many if not most people are trying to read the character of a subject from the language the subject uses (which they shouldn’t but that’s how humans evolved to think in social settings) so they trully belive that what produces language like an intelligent creature must be an intelligent creature.

                They’re probably not the right people to be opinating on cognition and inteligence, but lets not assign malice to it - at worst it’s pigheaded ignorance.

                • melpomenesclevage@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  edit-2
                  7 months ago

                  I think the person my previous comment was replying to wasnt malicious; I think they’re really invested, financially or emotionally, in this bullshit, to the point their critical thinking is compromised. Different thing.

                  Odd loop backs there.

              • evranch@lemmy.ca
                link
                fedilink
                English
                arrow-up
                0
                ·
                7 months ago

                I think you’re misreading the point I’m trying to make. I’m not arguing that LLM is AGI or that it can understand anything.

                I’m just questioning what the true use case of AGI would be that can’t be achieved by existing expert systems, real humans, or a combination of both.

                Sure Deepseek or Copilot won’t answer your legal questions. But neither will a real programmer. Nor will a lawyer be any good at writing code.

                However when the appropriate LLMs with the appropriate augmentations can be used to write code or legal contracts under human supervision, isn’t that good enough? Do we really need to develop a true human level intelligence when we already have 8 billion of those looking for something to do?

                AGI is a fun theoretical concept, but I really don’t see the practical need for a “next step” past the point of expanding and refining our current deep learning models, or how it would improve our world.