

I don’t know how to tell you this but… Every body gives a shit. We’re born shitters.
Father, Hacker (Information Security Professional), Open Source Software Developer, Inventor, and 3D printing enthusiast


I don’t know how to tell you this but… Every body gives a shit. We’re born shitters.


The real problem here is that Xitter isn’t supposed to be a porn site (even though it’s hosted loads of porn since before Musk bought it). They basically deeply integrated a porn generator into their very publicly-accessible “short text posts” website. Anyone can ask it to generate porn inside of any post and it’ll happily do so.
It’s like showing up at Walmart and seeing everyone naked (and many fucking), all over the store. That’s not why you’re there (though: Why TF are you still using that shithole of a site‽).
The solution is simple: Everyone everywhere needs to classify Xitter as a porn site. It’ll get blocked by businesses and schools and the world will be a better place.


“To solve this puzzle, you have to get your dog to poop in the circle…”


Yep. Stadia also had a feature like this (that no one ever used).
Just another example of why software patents should not exist.


It’s cold outside all year round and there’s abundant geothermal energy. Basically, it’s the perfect place to build data centers.


Working on (some) AI stuff professionally, the open source models are the only models that allow you to change the system prompt. Basically, that means that only open source models are acceptable for a whole lot of business logic.
Another thing to consider: There’s models that are designed for processing: It’s hard to explain but stuff like Qwen 3 “embedding” is made for in/out usage in automation situations:
https://huggingface.co/Qwen/Qwen3-Embedding-8B
You can’t do that effectively with the big AI models (as much as Anthropic would argue otherwise… It’s too expensive and risky to send all your data to a cloud provider in most automation situations).


This doesn’t make sense when you look at it from the perspective of open source models. They exist and they’re fantastic. They also get better just as quickly as the big AI company services.
IMHO, the open source models will ultimately what pops the big AI bubble.


I assume you missed “scrolling” because of all the alcohol?


Shit. I just realized I went straight to the comments!


It’s close enough. The key is that it’s not something that was just regurgitated based on a single keyword. It’s unique.
I could’ve generated hundreds and I bet a few would look a lot more like a banana.


Hard disagree. You just have to describe the shape and colors of the banana and maybe give it some dimensions. Here’s an example:
A hyper-realistic studio photograph of a single, elongated organic object resting on a wooden surface. The object is curved into a gentle crescent arc and features a smooth, waxy, vibrant yellow skin. It has distinct longitudinal ridges running its length, giving it a soft-edged pentagonal cross-section. The bottom end tapers to a small, dark, organic nub, while the top end extends into a thick, fibrous, greenish-brown stalk that appears to have been cut from a larger cluster. The yellow surface has minute brown speckles indicating ripeness.
It’s a lot of description but you’ve got 4096 tokens to play with so why not?
Remember: AI is just a method for giving instructions to a computer. If you give it enough details, it can do the thing at least some of the time (also remember that at the heart of every gen AI model is a RNG).

Note: That was the first try and I didn’t even use the word “banana”.


It’s more like this: If you give a machine instructions to construct or do something, is the end result a creative work?
If I design a vase (using nothing but code) that’s meant to be 3D printed, does that count as a creative work?
That vase was made using code (literally just text) I wrote in OpenSCAD. The model file is the result of the code I wrote and the physical object is the output of the 3D printer that I built. The pretty filament was store-bought, however.
If giving a machine instructions doesn’t count as a creative process then programming doesn’t count either. Because that’s all you’re doing when you feed a prompt to an AI: Giving it instructions. It’s just the latest tech for giving instructions to machines.


What that Afghanistan girl image demonstrates is simply a lack of diversity in Midjourney’s training data. They probably only had a single image categorized as “Afghanistan girl”. So the prompt ended up with an extreme bias towards that particular set of training values.
Having said that, Midjourney’s model is entirely proprietary so I don’t know if it works the same way as other image models.
It’s all about statistics. For example, there were so many quotes and literal copies of the first Harry Potter book in OpenAI’s training set that you could get ChatGPT to spit out something like 70% of the book with a lot of very, very specific prompts.
At the heart of every AI is a random number generator. If you ask it to generate an image of an Afghan girl—and it was only ever trained on a single image—it’s going to output something similar to that one image every single time.
On the other hand, if it had thousands of images of Afghan girls you’d get more varied and original results.
For reference, finding flaws in training data like that “Afghanistan girl” is one of the methods security researchers use to break large language models.
Flaws like this are easy to fix once they’re found. So it’s likely that over time, image models will improve and we’ll see fewer issues like this.
The “creativity” isn’t in the AI model itself, it’s in its use.


I’ve seen original sources reproduced that show exactly what an AI copied to make images.
Show me. I’d honestly like to see it because it means that something very, very strange is taking place within the model that could be a vulnerability (I work insecurity).
The closest thing to that I’ve seen is false watermarks: If the model was trained on a lot of similar images with watermarks (e.g. all images of a particular kind of fungus might have come from a handful of images that were all watermarked), the output will often have a nonsense watermark that sort of resembles the original one. This usually only happens with super specific things like when you put the latin name of a plant or tree in your prompt.
Another thing that can commonly happen is hallucinated signatures: On any given image that’s supposed to look like a painting/drawing, image models will sometimes put a signature-looking thing in the lower right corner (because that’s where most artist signatures are placed).
The reason why this happens isn’t because the image was directly copied from someone’s work, it’s because there’s a statistical chance that the model (when trained) associated the keywords in your prompt with some images that had such signatures. The training of models is getting better at preventing this from happening though, as they apply better bounding box filtering to the images as a pretraining step. E.g. a public domain Audibon drawing of a pelican would only use the bird itself and not the entire image (which would include the artist signature somewhere).
The reason why the signature should not be included is because the resulting image would not be drawn by that artist. That would be tantamount to fraud (bad). Instead, what image models do (except OpenAI with ChatGPT/DALL-E) is tell the public exactly what their images were trained on. For example, they’ll usually disclose that they used ImageNET (which you yourself can download here: https://www.image-net.org/download.php ).
Note: I’m pretty sure the full ImageNET database is also on Huggingface somewhere if you don’t want to create an account with them.
Also note: ImageNET doesn’t actually contain images! It’s just a database of image metadata that includes bounding boxes. Volunteers—for over a decade—spent a lot of time drawing bounding boxes with labels/descriptions on public images that are available for anyone to download for free (with open licenses!). This means that if you want to train a model with ImageNET, you have to walk the database and download all the image URLs it contains.
If anything was “stolen”, it was the time of those volunteers that created the classification system/DB in order for things like OpenCV to work so that your doorbell/security camera can tell the difference between a human and a cat.
especially the ones made over the injections of workers
Well there’s the problem! As good as it sounds, you actually lose a lot of the nutrition when employees are processed into injectable paste. Ultra processed workers are bad for you.
Eat them raw as capitalism intended!


If someone has never seen a banana they wouldn’t be able to draw it either.
Also, AIs aren’t stealing anything. When you steal something you have deprived the original owner of that thing. If anything, AIs are copying things but even that isn’t accurate.
When an image AI is trained, it reads though millions upon millions of images that live on the public Internet and for any given image it will increment a floating point value by like 0.01. That’s it. That’s all they do.
For some reason people have this idea in their heads that every AI-generated image can be traced back to some specific image that it somehow copied exactly then modified slightly and combined together for a final output. That’s not how the tech works at all.
You can steal a car. You can’t steal an image.


All creative works are built on top of works that came before them. AI is no different.
If you tell AI to generate an image of a fat, blue, lizardperson with a huge dick, wearing a fedora, holding way too many groceries… It’ll do that. And you just used it to make something original that (probably) never existed before.
It will have literally created something new. Thanks to the instructions of a creative human.
Saying AI is incapable of creating something new is like saying that programmers can’t create anything new because they’re just writing instructions for a computer to follow. AI is just the latest technology for doing that.


Nothing says, “our culture/political system is the best” more than a 2,000-question ideological test that attempts to make sure it will not be questioned.


Big AI is a bubble but AI in general is not.
If anything, the DRAM shortages will apply pressure on researchers to come up with more efficient AI models rather than more efficient (normal) software overall.
I suspect that as more software gets AI-assisted development we’ll actually see less efficient software but eventually, more efficient as adoption of AI coding assist becomes more mature (and probably more formalized/automated).
I say this because of experience: If you ask an LLM to write something for you it often does a terrible job with efficiency. However, if you ask it to analyze an existing code base to make it more efficient, it often does a great job. The dichotomy is due to the nature of AI prompting: It works best if you only give it one thing to do at a time.
In theory, if AI code assist becomes more mature and formalized, the “optimize this” step will likely be built-in, rather than something the developer has to ask for after the fact.
I don’t know, man… Have you even seen Amber? It might be worth an alert 🤷