• 1 Post
  • 51 Comments
Joined 1 年前
cake
Cake day: 2023年7月3日

help-circle
  • I’m in the same boat. Markov chains are a lot of fun, but LLMs are way too formulaic. It’s one of those things where AI bros will go, “Look, it’s so good at poetry!!” but they have no taste and can’t even tell that it sucks; LLMs just generate ABAB poems and getting anything else is like pulling teeth. It’s a little more garbled and broken, but the output from a MCG is a lot more interesting in my experience. Interesting content that’s a little rough around the edges always wins over smooth, featureless AI slop in my book.


    slight tangent: I was interested in seeing how they’d work for open-ended text adventures a few years ago (back around GPT2 and when AI Dungeon was launched), but the mystique did not last very long. Their output is awfully formulaic, and that has not changed at all in the years since. (of course, the tech optimist-goodthink way of thinking about this is “small LLMs are really good at creative writing for their size!”)

    I don’t think most people can even tell the difference between a lot of these models. There was a snake oil LLM (more snake oil than usual) called Reflection 70b, and people could not tell it was a placebo. They thought it was higher quality and invented reasons why that had to be true.

    Orange site example:

    Like other comments, I was also initially surprised. But I think the gains are both real and easy to understand where the improvements are coming from. [ . . . ]

    I had a similar idea, interesting to see that it actually works. [ . . . ]

    Reddit:

    I think that’s cool, if you use a regular system prompt it behaves like regular llama-70b. (??!!!)

    It’s the first time I’ve used a local model and did [not] just say wow this is neat, or that was impressive, but rather, wow, this is finally good enough for business settings (at least for my needs). I’m very excited to keep pushing on it. Llama 3.1 failed miserably, as did any other model I tried.

    For story telling or creative writing, I would rather have the more interesting broken english output of a Markov chain generator, or maybe a tarot deck or D100 table. Markov chains are also genuinely great for random name generators. I’ve actually laughed at Markov chains before with friends when we throw a group chat into one and see what comes out. I can’t imagine ever getting something like that from an LLM.



  • Friends don’t let friends OSINT

    i can stop any time I want I swear

    The youtube page you found is less talked about, though a reddit comment on one of them said “anyone else thinking burntbabylon is Luigi?”. I will point out that the rest of his online presence doesn’t really paint him as “anti tech” overall, but who can say.

    apparently there was an imposter youtube channel too I missed

    https://www.cnn.com/us/live-news/brian-thompson-unitedhealthcare-death-investigation-12-9-24#cm4hp9zyk000m3b6nobdudr15

    not sure what his official instagram is, but I saw a mention of the instagram account @nickakritas_ around the beginning of his channel (assuming it’s his). didn’t appear in the internet archive though.

    also saw these twitter & telegram links to promote his channel, the twitter one was deleted or nuked (I use telegram to talk with friends who have it but the lack of content removal + terrible encryption means I don’t touch unknown telegram links with a 10ft pole, so I have no idea what’s in there):

    I missed a couple videos which survived on the internet archive but I couldn’t make it through 5 seconds of any of them. one of them (“How Humans Are Becoming Dumber”) cites that tech priest guy Gwern Branwen and “Anti-Tech” was gone from the channel name by then. he changed the channel name a lot so maybe he veered away from it being an anti-tech channel?

    • Nick Akritas (around January 2023)
    • Anti-Tech Cabin
    • AntiTechCabin, but he changed the handle to @Cabin_Club during a short transitory period
    • Cabin Club (around March 18th 2023)
    • Cabin Productions
    • Laconian (lasted until November 2023)





  • Time for another round of Rothschild nutso’s to come around now that ChatGPT can’t say one of their names.

    At first I was thinking, you know, if this was because of the GDPR’s right to be forgotten laws or something that might be a nice precedent. I would love to see a bunch of people hit AI companies with GDPR complaints and have them actually do something instead of denying their consent-violator-at-scale machine has any PII in it.

    But honestly it’s probably just because he has money

    I think Sam Altman’s sister accused him of doing this to her name awhile ago too (semi-recent example). I don’t think she was on a “don’t generate these words ever” blacklist, but it seemed like she was erased from the training data and would only come up after a web search.










  • I know this shouldn’t be surprising, but I still cannot believe people really bounce questions off LLMs like they’re talking to a real person. https://ai.stackexchange.com/questions/47183/are-llms-unlikely-to-be-useful-to-generate-any-scientific-discovery

    I have just read this paper: Ziwei Xu, Sanjay Jain, Mohan Kankanhalli, “Hallucination is Inevitable: An Innate Limitation of Large Language Models”, submitted on 22 Jan 2024.

    It says there is a ground truth ideal function that gives every possible true output/fact to any given input/question, and no matter how you train your model, there is always space for misapproximations coming from missing data to formulate, and the more complex the data, the larger the space for the model to hallucinate.

    Then he immediately follows up with:

    Then I started to discuss with o1. [ . . . ] It says yes.

    Then I asked o1 [ . . . ], to which o1 says yes [ . . . ]. Then it says [ . . . ].

    Then I asked o1 [ . . . ], to which it says yes too.

    I’m not a teacher but I feel like my brain would explode if a student asked me to answer a question they arrived at after an LLM misled them on like 10 of their previous questions.




  • I feel like the Internet Archive is a prime target for techfashy groups. Both for the amount of culture you can destroy, and because backed up webpages often make people with an ego the size of the sun look stupid.

    Also, I can’t remember but didn’t Yudkowsky or someone else pretty plainly admit to taking a bunch of money during the FTX scandal? I swear he let slip that the funds were mostly dried up. I don’t think it was ever deleted, but that’s the sort of thing you might want to delete and could get really angry about being backed up in the Internet Archive. I think Siskind has edited a couple articles until all the fashy points were rounded off and that could fall in a similar boat. Maybe not him specifically, but there’s content like that that people would rather not be remembered and the Internet Archive falling apart would be good news to them.

    Also (again), it scares me a little that their servers are on public tours. Like it’d take one crazy person to do serious damage to it. I don’t know but I’m hoping their >100PB of storage is including backups, even if it’s not 3-2-1. I’m only mildly paranoid about it lol.