• 1 Post
  • 90 Comments
Joined 1 year ago
cake
Cake day: July 14th, 2023

help-circle


  • Code mods are great, maps and assets are in there but not officially, so compatibility going forward probably isn’t great for those. Full modding support is being worked on and is one of their highest priorities, so I’m not surprised there wasn’t much discussion about it. Asset mod support is “before summer” so they’ve got another month according to their last statement on it. PDX Mods has some bugs but overall it’s actually pretty slick and functional, and they’ve made a few highly requested improvements to it already.






  • Let’s just take NYT for example. Subscription costs $325/year. Why would I ever pay that much? It’s not 1954. I’m not sitting down with my morning coffee and reading the damn thing front to back. I’m reading maybe one article a week from 15 different sources. Am I supposed to pay $5000/year just to cover my bases?

    As with everything else in [CURRENT YEAR] the value proposition is so absurdly out of step with reality that fixing it basically relies on rolling out the guillotines.






  • Bought a Win 11 laptop recently (it’s now a linux device), and on first boot up there was no way to gracefully decline using a Microsoft account to sign in. Luckily, I was in a hotel and couldn’t connect to the wifi without going through a login page, so the lack of internet connection allowed me to set up a local account. In any case, if you’re forced to log in with an MS account, OneDrive starts syncing right away. You can disable it, but maybe not before it’s already done some damage.







  • I guess I’m wondering if there’s some way to bake the contextual understanding into the model instead of keeping it all in vram. Like if you’re talking to a person and you refer to something that happened a year ago, you might have to provide a little context and it might take them a minute, but eventually, they’ll usually remember. Same with AI, you could say, “hey remember when we talked about [x]?” and then it would recontextualize by bringing that conversation back into vram.

    Seems like more or less what people do with Stable Diffusion by training custom models, or LORAs, or embeddings. It would just be interesting if it was a more automatic process as part of interacting with the AI - the model is always being updated with information about your preferences instead of having to be told explicitly.

    But mostly it was just a joke.