

Rust coreutils has 17,000 commits and is 12 years old.
Admin of the Bestiverse


Rust coreutils has 17,000 commits and is 12 years old.


Quit rewriting history, those were absolutely not lowball milestones.
If Mr. Musk were somehow to increase the value of Tesla to $650 billion — a figure many experts would contend is laughably impossible and would make Tesla one of the five largest companies in the United States, based on current valuations — his stock award could be worth as much as $55 billion
That is a quote from this contemporary article: https://www.cnbc.com/2018/01/23/teslas-pay-deal-to-keep-elon-musk-all-or-nothing.html


The reason his last pay package of 50 billion was awarded to him is because he met the milestones for that. It was a similar deal to this one where they set top end milestones that everyone felt were ridiculous and they’d never hit them. Mostly stock targets IIRC.


I too fixed performance problems in that repo a few years back and did a write up on it - https://jackson.dev/post/rust-coreutils-dd/
I’m glad this project is getting some more attention, maybe even getting funding from Ubuntu since they’re using it? Last time I touched it most of the code was still pretty clearly written by Rust beginners and non-systems programmers so it likely had/has many such issues to uncover. Ubuntu putting it into their distro should hopefully get more experienced (and actually paid!) devs taking a closer look.


So you’re also contributing to the declining viewership that these channels are complaining about. They make no money off of your watch time so to them it is the same as no view at all.


And use what instead?


My AI Skeptic Friends Are All Nuts - https://fly.io/blog/youre-all-nuts/
You still think it’s sketchy?
I’ve explained that it’s perfectly normal, that it’s just someone who wants to use Unicode in their domain name (in this case because they probably speak a non-ascii based language), and most good web clients should be showing that link as the Unicode characters. Firefox for example shows that as the proper Unicode directly.
It literally is just a way for non-english speakers to have a domain name in their native language.
Are you talking about the “xn—“ domain name? Because FYI that’s just a punycode domain. It’s pretty commonly used for non-ascii domains. https://en.wikipedia.org/wiki/Punycode
The article itself is only available over Tor or I2P anyways though.


How do you make it illegible for LLMs?


So you prefer ad supported content?


130ms is perceivable but still quite small, and you’d only hit it once per domain (per TTL). If you care enough to intentionally use it then I wouldn’t worry about it. You’ll rarely notice the difference.
There are a few other services with similar ethos that you may want to check out as alternatives. Quad9 is the one I remember off the top of my head.
Some things do charge different amounts though. YouTube Premium for example is more expensive if you subscribe in iOS but maybe that’s just because it’s Google.
They also could have just not let anyone subscribe through the iOS app. Lots of things do that.


Japan already passed a law that explicitly allows training on copyrighted material. And many other countries just wouldn’t care. So if it becomes a real problem the companies will just move.
I think they need to figure out a middle ground where we can extract value from the for profit AI companies but not actually restrict the competition.


I don’t think they’re wrong in saying that if they aren’t allowed to train on copyrighted works then they will fall behind. Maybe I missed it in the article, but Japan for example has that exact law (use of copyright to train generative AI is allowed).
Personally I think we need to give them somewhat of an out by letting them do it but then taxing the fuck out of the resulting product. “You can use copyrighted works for training but then 50% of your profits are taxed”. Basically a recognition that the sum of all copyrighted works is a societal good and not just an individual copyright holders.
I can’t help, just chiming in to say that I’ve also had that experience with Immich. It’s the one service I’ve used that has somehow managed to break itself multiple times like this.
No idea how it happens, I don’t do anything weird with the setup and it just breaks. I’d heard that feedback from other people too but didn’t believe it until it happened to me. It’s been a few months so maybe I’ll try again, I’m just not too happy importing hundreds of gigs of photos multiple times.
So yea just… you’re not alone, good luck.


~15k lines of actual Rust code.
@ ❯ git clone https://github.com/torvalds/linux && cd linux && tokei
Cloning into 'linux'...
remote: Enumerating objects: 10655741, done.
remote: Counting objects: 100% (1067/1067), done.
remote: Compressing objects: 100% (208/208), done.
remote: Total 10655741 (delta 961), reused 859 (delta 859), pack-reused 10654674 (from 3)
Receiving objects: 100% (10655741/10655741), 5.13 GiB | 13.37 MiB/s, done.
Resolving deltas: 100% (8681589/8681589), done.
Updating files: 100% (87840/87840), done.
===============================================================================
Language Files Lines Code Comments Blanks
===============================================================================
Alex 2 222 180 0 42
ASN.1 15 656 441 87 128
Assembly 10 5226 4764 0 462
GNU Style Assembly 1336 372898 271937 56600 44361
Autoconf 5 433 377 26 30
Automake 3 31 23 3 5
BASH 59 2029 1368 352 309
C 34961 24854959 18510957 2766479 3577523
C Header 25450 10090846 7834037 1503620 753189
C++ 7 2267 1946 81 240
C++ Header 2 125 59 55 11
CSS 3 295 172 69 54
Device Tree 5582 1744314 1430810 83215 230289
Gherkin (Cucumber) 1 333 199 97 37
Happy 10 6049 5332 0 717
HEX 2 173 173 0 0
INI 2 13 6 5 2
JSON 894 542554 542552 0 2
LD Script 8 377 289 29 59
Makefile 3062 81226 55970 12993 12263
Module-Definition 2 128 113 0 15
Objective-C 1 89 72 0 17
Perl 61 43843 34461 3909 5473
Python 280 84204 66996 5198 12010
RPM Specfile 1 131 111 2 18
ReStructuredText 3672 761388 577410 0 183978
Ruby 1 29 25 0 4
Shell 957 187353 130476 23721 33156
SVG 79 52122 50727 1303 92
SWIG 1 252 154 27 71
TeX 1 234 155 73 6
Plain Text 1455 134747 0 110453 24294
TOML 3 47 28 12 7
Unreal Script 5 671 415 158 98
Apache Velocity 1 15 15 0 0
Vim script 1 42 33 6 3
XSL 10 200 122 52 26
XML 24 22177 19862 1349 966
YAML 4545 512759 417504 19285 75970
-------------------------------------------------------------------------------
HTML 2 28 22 3 3
|- JavaScript 1 7 7 0 0
(Total) 35 29 3 3
-------------------------------------------------------------------------------
Markdown 1 248 0 177 71
|- BASH 1 2 2 0 0
|- C 1 20 12 6 2
(Total) 270 14 183 73
-------------------------------------------------------------------------------
Rust 91 15207 11065 2248 1894
|- Markdown 85 7773 747 5253 1773
(Total) 22980 11812 7501 3667
===============================================================================
Total 82608 39520940 29971358 4591687 4957895
===============================================================================


Because most people do not understand what this technology is, and attribute far too much control over the generated text to the creators. If Copilot generates the text “Trans people don’t exist”, and Microsoft doesn’t immediately address it, a huge portion of people will understand that to mean “Microsoft doesn’t think trans people exist”.
Insert whatever other politically incorrect or harmful statement you prefer.
Those sorts of problems aren’t easily fixable without manual blocks. You can train the models with a “value” system where they censor themselves but that still will be imperfect and they can still generate politically incorrect text.
IIRC some providers support 2 separate endpoints where one is raw access to the model without filtering and one is with filtering and censoring. Copilot, as a heavily branded end user product, obviously needs to be filtered.


I understand why they need to implement these blocks, but they seem to always be implemented without any way to workaround them. I hit a similar breakage using Cody (another AI assistant) which made a couple of my repositories unusable with it. https://jackson.dev/post/cody-hates-reset/
I’m working on something that allows for custom CRDTs since I agree no one CRDT strategy is best for any given app. There are several others I know of but they only use a single type. I think Automerge is the most popular current one but I don’t know if it has many actual users.
Mine is Eidetica, still very much experimental but I’m making progress https://github.com/arcuru/eidetica