So do they load all those matrices (totalling to 175b params in this case) to available GPUs for every token of every user?
So do they load all those matrices (totalling to 175b params in this case) to available GPUs for every token of every user?
That’s how llms work. When they say 175 billion parameters, it means at least that many calculations per token it generates
I don’t get it, how is it possible that so many people all over the world use this concurrently, doing all kinds of lengthy chats, problem solving, codegeneration, image generation and so on?
Should be “starting your own instance”, because otherwise you still have to conform to the rules of the instance you create your community/sub on.
It is what it feels like, but it’s not really 100% this way (yet). It is a bad self-reinforcing cognitive bias: we think “forums are dead, that’s why we stick to the sitename” instead of actually finding dozens of still alive forums and going there, in turn sitename gets more populated while forums feel more dead. But there are still plenty alive. Also, there are relatively new kinds of forums which sometimes work very well for their niche, like Discourse communities for example.
After you use ChatGPT for a bit, you will start recognizing its style of writing in posts and comments. I’ve seen dozens of obviously ChatGPT generated posts or replies on Reddit and Lemmy. Usually there will be a person who already replied to them something like “Thanks, ChatGPT”, because it is that obvious. This only happens with naive prompts though, if you ask ChatGPT to present its answer to your prompt in a different style (for example, mimic some famous writer, or being cheerful/angry/excited and avoid overly safe language), it will immediately start writing differently and there’s likely no limit on variety of writing styles you can pull out of it with enough effort of just asking it to write this or that way.
Yeah, got a bit carried away yesterday. Ultimately, there can’t be right or wrong, since the whole discussion is simply about individual understanding or even preferences for the term “slop”, nothing more. Some people will try to be as empirical as possible and choose a meaning based on how they’ve seen it being used in the wild, others will try to push the meaning they want for the term, it’s all good and subjective.
This is where the fundamental difference in attribution of connotations lies. From what you say, you perceive the term “slop” as a direct synonym to “low quality”, without any extras. I perceive it as something more of a synonym to “repetitive” but with extra connotations, the most accurate common divisor of which is “repetitive content produced at speeds suggesting low effort”.
I understand your point of view. But would you call something complex, high-quality, but repetitive, slop? And the same question, but if the person who produces it, does it extremely fast.
How would you use that term? Would you call “slop” something that was just mindlessly generated using AI in a single prompt and non-“slop” something that uses AI in more sophisticated/deliberate ways? What is the threshold of something being “slop” ultimately? Is this just result not looking decent enough or amount of effort combined with amount of knowledge and experience that was used to create that? I’m personally conflicted on this, because sometimes even mindless prompt may give great result, and sometimes a lot of manual effort may give shit result. I guess with “slop” I tend to gravitate towards “amount of effort combined with amount of knowledge and experience that was used to create” and perhaps also the amount of content that particular person produces and speed of its production. So if someone is really good with some tools (not necessarily AI) and figured some overpowered shortcuts that allow to produce results very fast with little effort, it also can be called “slop” just for the rate of production alone.
FWIW, the thing with killswitch it not due to Bazzite, nor KDE. There’s a f*ck load of user reports all over the internet with different systems that have experienced the same thing; e.g. this one by a GNOME user on Pop!_OS.
My bad, so it’s probably ProtonVPN client doing tricky hidden things that can break.
As for your criticism on kdewallet, I was also bothered by it the last few times I engaged with KDE Plasma.
I also got a kdewallet problem with flatpak VS Code authenticating to github, but that one is so widely known, they even included guidelines in docs on how to solve it.
The only real issue I’ve had was that the btrfs partition sometimes shits itself and requires some CLI commands in emergency mode to fix it.
This sounds scary, not sure I’d be able to fix that. Hopefully, with some search if that happens to me.
IIRC there was a widget for setting preferred GPU in the taskbar?
Couldn’t find this one, and in general couldn’t find any UI for configuring GPUs systemwide. It’s possible to set preferred GPU in Lutris settings, but it didn’t work for some reason. I installed most heavy games via Steam, and Lutris doesn’t see my Steam games and setting preferred GPU in its Steam category doesn’t affect games in question.
Cool idea, I almost forgot this feature even exists. I think I dismissed it the past when I realized it’s probably not going to be easy to switch VPN servers this way.
It does hurt, your VPN should support proper port-forwarding for soulseek to work well. In most cases, you will only be able to download files, but your shares will be inaccessible. It doesn’t seem to work with ProtonVPN for example, even when you built-in port-forwarding feature. And even if it did work, you would need to reconfigure and restart soulseek every time you reconnect the VPN, because their port-forwarding is randomizing the ports and there’s no way to turn that off.
Let me explain how Honkai Star Rail handles gearing. Every single character has six relic slots: head, hands, body, feet, planar orb, and planar ornament. These relics go from level 0 to level 15, and four of them have a randomized primary stat. They all feature four randomized secondary stats, and every three levels a random one of those secondary stats gets a bonus. Each relic also belongs to a set of relics, and characters benefit from having two or four pieces of a given relic set. That means for every character in your party, you need to get the right items at the maximum rarity, the right primary stats, the right secondary stats, and the right level-ups for those secondary stats.
This is min-maxer mindset and I would hope randomized systems like this will prevent it but unfortunately no: even here some people think they actually need to roll every dice exactly the right way. I don’t think it’s true that this is really necessary. And no, it is not necessary to do top 10 world parses; you can just beat endgame content on modest, casual difficulty and call it a day, rather than try hard to set a record.
Even co-op in gacha games doesn’t qualify as MMO, because for that you need hundreds or thousands of players being simultaneously in the same persistent world. This is the same reason why games like Dota, League of Legends or Counter Strike aren’t considered MMO.
Yes, the cheapest ones might have some risks, I mostly presented it as an example of what the opposite extremity looks like. There is a lot in-between, something a bit more expensive is even more guaranteed win. For example last time I used Hetzner, I had a server with 64gb RAM, 2TB SSD, and 16 cores Ryzen for something like €34/month. Hetzner support is very decent and they’re very well known, have decent reputation and been providing their services for a long time.
I’m talking about 3d software one, and author obviously talks about that one too.
Maybe the problem is that they are using ridiculously overpriced enterprise services like AWS or Azure, which provide their own solutions for a lot of common things like backups, replicas, logging, etc, but cost 100x more than what you can get with DIY on some cheap VPS if you’re fine with spending 1.25x more time.
Also, given that the instance is called “infosec.exchange”, you can be sure that he is not running this on some cheap VPS.
Why not, though.
I wonder why it needs so much money for infra? Last time I rented a VPS it was €7/month for 8 Core Xeon E5 V4, 12 GB DDR4 RAM, 150 GB SSD/NVME, Unlimited Traffic, 1 Gbps Port.
I also asked ChatGPT itself, and it listed a number of approaches, and one that sounded good to me is to pin layers to GPUs, for example we have 500 GPUs: cards 1-100 have permanently loaded layers 1-30 of AI, cards 101-200 have permanently loaded layers 31-60 and so on, this way no need to frequently load huge matrices itself as they stay in GPUs permanently, just basically pipeline user prompt through appropriate sequence of GPUs.