• 1 Post
  • 27 Comments
Joined 2 years ago
cake
Cake day: August 11th, 2023

help-circle

  • There is a manual pre-installed on your machine for most commands available. You just type man and the name of the thing you want the manual for. Many commands also have a --help option that will give you a list of basic options.

    I should point out this isn’t Linux specific either. Many of these commands come from Unix or from other systems entirely. macOS has a similar command line system actually. It’s more that Linux users tend to use and recommend the command line more. Normally because it’s the way of doing things that works across the largest number of distributions and setups, but also because lots of technical users prefer command line anyway. Hence why people complain about Windows command lines being annoying. I say command lines because they actually have two of them for some odd reason. Anyway I hope this helped explain why things are the way they are.















  • There is a lot that can be discussed in a philosophical debate. However, any 8 years old would be able to count how many letters are in a word. LLMs can’t reliably do that by virtue of how they work. This suggests me that it’s not just a model/training difference. Also evolution over million of years improved the “hardware” and the genetic material. Neither of this is compares to computing power or amount of data which is used to train LLMs.

    Actually humans have more computing power than is required to run an LLM. You have this backwards. LLMs are comparably a lot more efficient given how little computing power they need to run by comparison. Human brains as a piece of hardware are insanely high performance and energy efficient. I mean they include their own internal combustion engines and maintenance and security crew for fuck’s sake. Give me a human built computer that has that.

    Anyway, time will tell. Personally I think it’s possible to reach a general AI eventually, I simply don’t think the LLMs approach is the one leading there.

    I agree here. I do think though that LLMs are closer than you think. They do in fact have both attention and working memory, which is a large step forward. The fact they can only process one medium (only text) is a serious limitation though. Presumably a general purpose AI would ideally have the ability to process visual input, auditory input, text, and some other stuff like various sensor types. There are other model types though, some of which take in multi-modal input to make decisions like a self-driving car.

    I think a lot of people romanticize what humans are capable of while dismissing what machines can do. Especially with the processing power and efficiency limitations that come with the simple silicon based processors that current machines are made from.





  • Yeah it’s not always that simple. You haven’t been around long enough to see the stuff that can go wrong with installing Windows. For example I recently had Windows refuse to see both SSDs in a machine. All because of something called Intel VMD. Took me a handful of attempts before I found the problem.

    When Windows installs work they are fairly simple if long, but when they don’t work oh boy.

    The unplugging of internet to get a local account?

    Also they disabled that for Windows Home.

    Some Lemmy users are actually just wankers. I would like it if you all stopped. It’s especially great when I have people like you who probably aren’t even experienced in tech.