• 0 Posts
  • 14 Comments
Joined 1 year ago
cake
Cake day: June 9th, 2023

help-circle

  • Not necessarily. For a game like this that only functions online, you could presumably determine all the possible server calls and point them to a server you own. You could do this purely via clever network settings without modifying the game at all. If you could do that, the game would run fine and you could even use the original authentication server to ensure the user holds a valid license.

    At that point, you “just” need to implement and run a server for the game. This also doesn’t involve modifying the game, but could run afoul of potential laws against reverse engineering if not done in a clean room manner (I’m not a lawyer so there could be other things too since unfortunately US law tends to not favor the end user).

    Regardless of any of that, it always feels silly to me when companies fight tooth-and-nail against people not only performing free work and hosting for a dead game but ALSO trying to ensure people actually own the game before playing on their private server. Of course they could just use 🏴‍☠️ versions and black-hole the authentication server. All the company does by withdrawing licenses is ensure they have to skip authentication so the company loses out.




  • Huh, go figure. Thanks for the info! I honestly never would have found that myself.

    I still think it should be possible to use in:channel on the channel-specific search though. One less button press and it can’t be that confusing UX-wise since you have clear intent when doing it (if anything, the fact that the two searches work differently has to be more confusing UX-wise).





  • Copilot, yes. You can find some reasonable alternatives out there but I don’t know if I would use the word “great”.

    GPT-4… not really. Unless you’ve got serious technical knowledge, serious hardware, and lots of time to experiment you’re not going to find anything even remotely close to GPT-4. Probably the best the “average” person can do is run quantized Llama-2 on an M1 (or better) Macbook making use of the unified memory. Lack of GPU VRAM makes running even the “basic” models a challenge. And, for the record, this will still perform substantially worse than GPT-4.

    If you’re willing to pony up, you can get some hardware on the usual cloud providers but it will not be cheap and it will still require some serious effort since you’re basically going to have to fine-tune your own LLM to get anywhere in the same ballpark as GPT-4.



  • isildun@sh.itjust.workstoMemes@lemmy.mlSure it is
    link
    fedilink
    English
    arrow-up
    27
    ·
    edit-2
    1 year ago

    Definitely AI generated. Look at the bottom-right of the Confederate flag. It’s all messed up, classic generative AI “artifacting” for lack of a better word for it.

    Edit: lower down in the thread the original was posted. This was upscaled (very poorly) by AI.