• 32 Posts
  • 233 Comments
Joined 2 years ago
cake
Cake day: June 9th, 2023

help-circle
  • j4k3@lemmy.worldtoFediverse memes@feddit.ukIn a nutshell.
    link
    fedilink
    English
    arrow-up
    3
    ·
    16 minutes ago

    We need some kind of subletting system. Package a stupid simple Raspberry Pi image with the minimal amount of configuration required for me to self host, and set up a fediverse based DNS and certificate authority. Make it stupid simple for a user like me to self host my traffic without needing to pay for a domain and sort out all of the requirements. Like I technically have money but don’t control my finances and am heavily subsidized by family due to my situation and disability. I could probably set it up if I was super motivated but networking complexity is a rabbit hole that feels intentionally convoluted and a pain in the ass every time I mess with it. Locking down a Linux image with an immutable base and well configured PAM, SELinux, and automatic updating is equally daunting when a person has no experience or professional contacts familiar with the setup to ground one’s understanding.

    I don’t see why we need to rely on the traditional internet infrastructure. We should be able to make something like fediverse.lemmy.user.j4k3 where we collectivise the address fediverse.lemmy, use it as a DNS server, and then distribute the traffic to individual servers.


  • No, not a conspiracy but, it is not actually "emotional issues" either. It is dogma, and an emotional solution will not work directly either.

    Emotional solutions may seem to help in some cases but not all. The core underlying issue is the human tribal scope. Dogma exists specifically at the level of tribalism. Tribalism is why logic and reasoning are ineffective. Any information from humans outside of the tribe is invalidated based upon membership alone. No information from outsiders is considered valid.

    So if you can reach someone with an emotional connection, you are really convincing them that you are part of their tribe. However, if they accept you while the rest of the tribe has not, or they are of low social hierarchical rank in the tribe, they risk becoming an outcast other too, and preventing propagation of reasoning logic.

    This dynamic of tribalism is why religious leaders are so powerful and why secular authorities engage with them. If a leader or high ranking members of a tribe endorse such a secular authority, all of the tribe must follow blindly because there is no logic in dogma, only tribalism. It can sound rather absurd, but the reinforcement mechanism is social network isolation of the tribe. Most people don’t actually believe the tribal mythos, but if they leave, they lose access to their social network they were born within because this group is mutually exclusive. This social network isolated reinforcement mechanism is why there are regional faiths, and these do not compete on merit of logic or reasoning. If you were born into a region of Jewish, Christian, Buddhist, Zoroastrian, or Muslim faith tribes, you will be a part of one of these. The issue is not the validity of the tribe. Battling the mythos or ethos of the tribe is only capable of provoking the most combative and least broad-scope self aware members of the group. The only way to effectively change people is to offer them a better tribe and social network than what they have and all that they have ever known. For most humans, leaving their tribe is a death sentence in the same primitive instincts that are at play with dogma.

    So, if you follow this logic, first off, welcome to the tribe, because we are doing it too! Tribalism is inescapable at this point in human evolution. Second, it should be rather obvious that any logic or argument with dogma results in the exact opposite effect of what you are trying to accomplish. If you truly want to change a dogmatic person, you must welcome them into your tribe openly. If this action seems difficult, it is because your own tribal scope is not what you believe it to be, and in a sense you were coming to a battle. Third and finally, a High Machiavellian type person with broad stroke abstractive skills can see this type of dynamic like playing with Lego bricks to make a small box; it is trivial. The skill doesn’t have anything to do with sensing or emotions. It is like watching a cutaway of an engine assembly turning and intuitively seeing how all the pieces work together well enough to understand the mechanism. And this is why a person that regularly visited Epstein Island, and solicited a porn star for sex because she looked like his daughter, or an open polygamist oligarch is able to buy the heads of dogmatic tribes and get blindly accepted. These people are high Machs too. We are rare relatively speaking in terms of functional thought. High Machiavellian is not necessarily bad. It can be used for good, or like myself – to be one of the few people dumb enough to abandon their isolated socially exclusive network of dogmatism. It also enables raising awareness of those that are dangerously wielding the skill, though only at smaller scales of within a tribe, unless one makes a goal of broad scope influence but that is a challenge as well. Like, I can understand collective motivations like a machine, but I do not understand emotions of influence and popularity well. The same applies to others like Trump and Musk. If they could understand these elements, they would displace the leaders of the dogmas. Instead they must still work to appease these intermediaries.




  • I haven’t looked into the issue of PCIe lanes and the GPU.

    I don’t think it should matter with a smaller PCIe bus, in theory, if I understand correctly (unlikely). The only time a lot of data is transferred is when the model layers are initially loaded. Like with Oobabooga when I load a model, most of the time my desktop RAM monitor widget does not even have the time to refresh and tell me how much memory was used on the CPU side. What is loaded in the GPU is around 90% static. I have a script that monitors this so that I can tune the maximum number of layers. I leave overhead room for the context to build up over time but there are no major changes happening aside from initial loading. One just sets the number of layers to offload on the GPU and loads the model. However many seconds that takes is irrelevant startup delay that only happens once when initiating the server.

    So assuming the kernel modules and hardware support the more narrow bandwidth, it should work… I think. There are laptops that have options for an external FireWire GPU too, so I don’t think the PCIe bus is too baked in.



  • j4k3@lemmy.worldtoSelfhosted@lemmy.worldConsumer GPUs to run LLMs
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    1
    ·
    edit-2
    2 days ago
    Anything under 16 is a no go. Your number of CPU cores are important. Use Oobabooga Textgen for an advanced llama.cpp setup that splits between the CPU and GPU. You'll need at least 64 GB of RAM or be willing to offload layers using the NVME with deepspeed. I can run up to a 72b model with 4 bit quantization in GGUF with a 12700 laptop with a mobile 3080Ti which has 16GB of VRAM (mobile is like that).

    I prefer to run a 8×7b mixture of experts model because only 2 of the 8 are ever running at the same time. I am running that in 4 bit quantized GGUF and it takes 56 GB total to load. Once loaded it is about like a 13b model for speed but is ~90% of the capabilities of a 70b. The streaming speed is faster than my fastest reading pace.

    A 70b model streams at my slowest tenable reading pace.

    Both of these options are exponentially more capable than any of the smaller model sizes even if you screw around with training. Unfortunately, this streaming speed is still pretty slow for most advanced agentic stuff. Maybe if I had 24 to 48gb it would be different, I cannot say. If I was building now, I would be looking at what hardware options have the largest L1 cache, the most cores that include the most advanced AVX instructions. Generally, anything with efficiency cores are removing AVX and because the CPU schedulers in kernels are usually unable to handle this asymmetry consumer junk has poor AVX support. It is quite likely that all the problems Intel has had in recent years has been due to how they tried to block consumer stuff from accessing the advanced P-core instructions that were only blocked in microcode. It requires disabling the e-cores or setting up a CPU set isolation in Linux or BSD distros.

    You need good Linux support even if you run windows. Most good and advanced stuff with AI will be done with WSL if you haven’t ditched doz for whatever reason. Use https://linux-hardware.org/ to see support for devices.

    The reason I mentioned avoid consumer e-cores is because there have been some articles popping up lately about all p-core hardware.

    The main constraint for the CPU is the L2 to L1 cache bus width. Researching this deeply may be beneficial.

    Splitting the load between multiple GPUs may be an option too. As of a year ago, the cheapest option for a 16 GB GPU in a machine was a second hand 12th gen Intel laptop with a 3080Ti by a considerable margin when all of it is added up. It is noisy, gets hot, and I hate it many times, wishing I had gotten a server like setup for AI, but I have something and that is what matters.


  • I like to write, but have never done so professionally. I disagree that it hurts writers. I think people reacted poorly to AI because of the direct and indirect information campaign Altmann funded to try and make himself a monopoly. AI is just a tool. It is fun to play with in unique areas, but these often require very large models and/or advanced frameworks. In my science fiction universe I must go to extreme lengths to get the model to play along with several aspects like a restructure of politics, economics, and social hierarchy. I use several predictions I imagine about the distant future that plausibly make the present world seem primitive in several ways and with good reasons. This restructuring of society violates both some of our cultural norms in the present and is deep within areas of politics that are blocked by alignment. I tell a story where humans are the potentially volatile monsters to be feared. That is not the plot, but convincing a present model to collaborate on such a story ends up in the gutter a lot. My grammar and thought stream is not great and that is the main thing I use a model to clean up, but it is still collaborative to some extent.

    I feel like there is an enormous range of stories to tell and that AI only makes these more accessible. I have gone off on tangents many times exploring parts of my universe because of directions the LLM took. Like I limit the model to generate a sentence at a time and I’m writing half or more of every sentence for the first 10k tokens. Then it picks up on my style so much that I can start the sentence with a word or change one word in a sentence and let it continue with great effect. It is most entertaining to me because it is almost as fast as me telling a story as fast as I can make it up. I don’t see anything remotely bad about that. No one makes a career in the real world by copying someone else’s writing. There are tons of fan works but those do not make anyone real money and they only increase the reach of the original author.

    No, I think all the writers and artists hype was all about Altmann’s plan for a monopoly that got derailed when Yann LeCunn covertly leaked the Llama weights after Altmann went against the founding principles of OpenAI and made GPT3 proprietary.

    People got all upset about digital tools too back when they first came on the scene; about how they would destroy the artists. Sure it ended the era of hand painted cartoon cell animation, but it created stuff like Pixar.

    All of AI is a tool. The only thing to hate is this culture of reductionism where people are given free money in the form of great efficiency gains and they choose to do the same things with less people and cash out the free money instead of using the opportunity to offer more, expand, and do something new. A few people could get a great tool chain together and create a franchise greater, better planned, and more rich than anything corporations have ever done to date. The only thing to hate are these little regressive stupid people without vision, without motivation, and far too conservatively timid to take risks and create the future. We live in an age of cowards worthy of loathing. That is the only problem I see.


  • I use the term myth loosely in abstraction. Generalization of the tools of industry is still a mythos in an abstract sense. Someone with a new lathe they bought to bore the journals of an engine block has absolutely no connection or intentions related to class, workers, or society. That abstraction and assignment of meaning like a category or entity or class is simply the evolution of a divine mythos in the more complex humans of today.

    Stories about Skynet or The Matrix are about a similar struggle of the human class against machine gods. These have no relationship to the actual AI alignment problem and are instead a battle with more literal machine gods. Point is that the new thing is always the boogie man. Evolution must be deeply conservative most of the time. People display a similar trajectory of conservative aversion to change. In this light, the reasons for such resistance are largely irrelevant. It is a big change and will certainly get a lot of push back from conservative elements that collectively ensure change is not harmful. Those elements get cut off in the long term as the change propagates.

    You need a 16 GB or better GPU from the 30 series or higher, but then run Oobabooga text gen with the API and an 8×7b or like a 34b or 70b coder in a GGUF quantized model. Those are larger than most machines can run but Oobabooga can pull it off by splitting the model between CPU and GPU. You’ll just need the ram to initially load the thing or deepspeed to load it from NVME.

    Use a model with a long context and add a bunch of your chats into the prompt. Then ask for your user profile and start asking it questions about you that seem unrelated to any of your previous conversations in the context. You might be surprised by the results. Inference works both directions. You’re giving a lot of information that is specifically related to the ongoing interchanges and language choices. If you add a bunch of your social media posts, it is totally different in what the model will make up about you in a user profile. There is information of some sort that the model is capable of deciphering. It is not absolute or like some kind of conspiracy or trained behavior (I think), but the accuracy seemed uncanny to me. It spat out surprising information across multiple unrelated sessions when I tried it a year ago.



  • When tech changes quickly, some people always resist exponentially in the opposite vector. The bigger and more sudden the disruption, the bigger the push back.

    If you read some of Karl Marx stuff, it was the fear of the machines. Humans always make up a mythos of divine origin. Even atheists of the present are doing it. Almost all of the stories about AI are much the same stories of god machines that Marx was fearful of. There are many reasons why. Lemmy has several squeaky wheel users on this front. It is not a very good platform for sharing stuff about AI unfortunately.

    There are many reasons why AI is not a super effective solution and overused in many applications. Exploring uses and applications is the smart thing to be doing in the present. I play with it daily, but I will gatekeep over the use of any cloud based service. The information that can be gleaned from any interaction with an AI prompt is exponentially greater than any datamining stalkerware that existed prior. The real depth of this privacy evasive potential is only possible with a large number of individual interactions. So I expect all applications to interact with my self hosted OpenAI compatible server.

    The real frontier is in agentic workflows and developing effective niche focused momentum. Any addition of AI into general use type stuff is massively over used.

    Also people tend to make assumptions about code as if all devs are equal or capable. In some sense I am a dev, but not really. I’m more of a script kiddie that dabbles in assembly at times. I use AI more like stack exchange to good effect.





  • The only fundamental issue with the CPU and tensors is the L2 to L1 cache bus width. This cannot be altered and maintain the speed. This is not a real issue in the grand scheme of things. It is only an issue with the total design cycle. Don’t get sucked into the little world of marketing nonsense surrounding specific fab nodes and whatever spin nonsense the sales fools are pedaling. Real hardware takes 10 years from initial concept to first market availability. Nvidia was lucky because their plans happened to align with the AI boom. They could adjust a few minor packaging tweaks to tailor the existing designs in the pipeline to the present market, but they had no prescient genius about how AI would explode like the last two years. Such a premise assumes they began the 40 series knowing about the AI boom in 2012, nearly 4 years before OpenAI was founded.

    The FPGA does not work for AI. It does not scale like you assume and the power required is untenable. You can find information about Intel/Altera AI researchers that were well funded and traversed this path before the constraints were discovered. You need simpler architecture with a lower transistor count. This is like the issue with static RAM versus DRAM. Static is functionally superior in nearly every way, but it simply can’t scale due to power and space requirements.

    With tensors all that is needed is throughput. That is a solvable problem. Single thread speeds in CPUs is a sales gimmick and nothing more. Your brain is a much more powerful biological computer and operates on 3 main clocks the fastest of which is only around 100 Hz. Parallelism can be used to create an even faster and more rich user experience than the present. This is the future. The dual processor paradigm has been done before in the x286 - x386 era and it failed because data centers rejected such a dual processor paradigm in favor of slightly better hardware that was nearly good enough. This is the reality of the present too. Any hardware that is good enough to do both workloads will be adopted by data centers and therefore the market. This is where the real design edge is made and all consumer products are derived.

    None of Nvidia’s products are relevant 8 years from now. They are a temporary hack. This is why they must use their enormous capital to buy a new future beyond the GPU, and they will.




  • j4k3@lemmy.worldtoPrivacy@lemmy.mlPrivacy Recommendations for a Young Teen
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    1
    ·
    edit-2
    5 days ago

    I think authoritarianism is a giant mistake and only creates duplicitous behavior. In my opinion tracking is ridiculous. None of us existed like this and ended up fine. In my opinion, all of this nonsense is acting as a stand in for relationships and real parenting. Humans make decisions and develop ethics based upon trust and autonomy. By stealing that factor of trust and autonomy, and replacing it with authoritarianism a parent is stunting the child’s growth of independent ethics and character. Make compelling discussions of why they should do whatever thing, but let them decide their own path. The lack of compelling discussions and real trust that requires risk is a major factor in the problems that exist in the present world.

    The one time you actually need to know where your kid is at because something has happened, you will not know because you have taught them that the only path to independence is to turn off the device and put it into a Faraday cage like pouch, or someone else will do so. If you have a fundamentally trusting relationship with open dialog and respect for their autonomy, they will tell you openly exactly where they are going and any potential for danger. If you can handle that information without allowing anxiety to overwhelm reasoning skills, you will be in a far better position to help them if something bad happens.

    The most long term valuable aspect of schooling is the development of one’s social network and connections, along with the habits and ethics. The actual information learned is rather limited in valuable application in the end. Who one knows and how one appears to others is of far more value than what one knows. For these reasons, there may be value in corporate social media. Simply teach the kid to understand how these places are both a trap and a tool. A trap, in that many of the smartest humans are manipulating users in ways that are nearly impossible for the users to escape. Never invest emotions into such a trap. Use the tool if needed for external social benefits, but use it as a manipulation tool with a layer of disconnect from who you really are. Teach them to use a work profile to isolate any apps from their device. That is just how I look at the issue.


  • My neighbor spent around 3 hours calling around to track down a source today. The wasted time and money that represents is the only crime I see. Don’t go talking about it or selling it, but whatever a person does in their kitchen for their own needs is none of my business. The wasted time is theft and therefore criminal actions against common people with no recourse. There are likely millions of victims of this crime.




  • j4k3@lemmy.worldtoOpen Source@lemmy.mlAccessability of github for europe
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    1
    ·
    6 days ago

    Most key stuff is not on GitHub or GitHub is just a mirror. The heir apparent to Linux is Hartman and he moved to Europe a long time ago.

    No mobile devices are safe. Those are all proprietary black boxes for hardware. If the shit hits the fan, it is back to dumb phones and x86 computers. Digital doomsday prepers are not sounding all that crazy right now IMO.

    I have gotten weird interactions with rate limiting through GitHub because I will not whitelist their stalkerware collector server. They also pushed 2 factor to stalk and exploit through the only documented path they wanted people to take. I quit because of it.