• 0 Posts
  • 408 Comments
Joined 2 years ago
cake
Cake day: August 27th, 2023

help-circle
  • The bubble continuing ensures the current paradigm soldiers on, meaning hideously expensive projects shove local models into people’s hands for free, because everyone else is doing that.

    And once it bursts, there’s gonna be an insulating layer of dipshits repeating “guess it was nothing!” over the next decade of incremental wizardry. For now, tolerating the techbro cult’s grand promises of obvious bullshit means the unwashed masses are interpersonally receptive to cool things happening.

    Already the big boys are pivoted toward efficiency instead of raw speed at all costs. The closer they get toward a toaster matching current tech with a model trained for five bucks, the better. I’d love for VCs to burn money on experimentation instead of scale.



  • This is the real future of neural networks. Trained on supercomputers - runs on a Game Boy. Even in comically large models, the majority of weights are negligible, and local video generation will eventually be taken for granted.

    Probably after the crash. Let’s not pretend that’s far off. The big players in this industry have frankly silly expectations. Ballooning these projects to the largest sizes money can buy has been illustrative, but DeepSeek already proved LLMs can be dirt cheap. Video’s more demanding… but what you get out of ten billion weights nowadays is drastically different from a six months ago. A year to date ago, video models barely existed. A year to date from now, the push toward training on less and running on less will presumably be a lot more pressing.



  • Due to some disagreements—some recent; some tolerated for close to 2 decades—with how collaboration should work, we’ve decided that the best course of action was to fork the project

    Okay, that was always allowed!

    Programming is the weirdest place for kneejerk opposition to anything labeled AI, because we’ve been trying to automate our jobs for most of a century. Artists will juke from ‘the quality is bad!’ to ‘the quality doesn’t matter!’ the moment their field becomes legitimately vulnerable. Most programmers would love if the robot did the thing we wanted. That’s like 90% of what we’re looking for in the first place. If writing ‘is Linux in dark mode?’ counted as code, we’d gladly use that, instead of doing some arcane low-level bullshit. I say this as someone who has recently read through IBM’s CGA documentation to puzzle out low-level bullshit.

    You have to check if it works. But if it works… what is anyone bitching about?





  • Stubborn prick misidentified parts of a computer forty years ago, and clung to relevance by never admitting programs don’t need the processor to understand a damn thing. He invented bigotry against robots. His one hit is a thought experiment on par with Descarte saying ‘the dog only acts like it feels pain,’ as he slides the dagger deeper.

    A computer could be indistinguishable from a human - it could simulate a real person’s brain scan at a subatomic level - and John Searle would insist it’s no different from Eliza. May we never hear his bullshit touted again.



  • I would be shocked if any diffusion model could do that based on a description. Most can’t overfill a wine glass.

    Rendering over someone demonstrating the movement, as video-to-video, is obviously easier than firing up Blender. But: that’s distant from any dream of treating the program like an actress. Each model’s understanding is shallow and opinionated. You cannot rely on text instructions.

    The practical magic from video models, for the immediate future, is that your video input can be real half-assed. Two stand-ins can play a whole cast, one interaction at a time. Or a blurry pre-vis in Blender can go straight to a finished shot. At no point will current technologies be more than loose control of a cartoon character, because to these models, everything is a cartoon character. It doesn’t know the difference between an actor and a render. It just knows shinier examples with pinchier proportions move faster.










  • I fully endorse photorealistic cartoons. Characters can look like anything, without having to find a specific guy and hope fits the role.

    Getting the cartoon onscreen can still involve an actual actor. Diffusion turns whatever you have into whatever you describe. Turning a guy into another guy is not a big ask. It’s how some of this mess started, with Nicolas Cage deepfaked as Superman, and defictionalizing the Stallone version of The Terminator. The target face does not need to be a real person. Three actors can stage nearly any script.

    Same goes for voice-acting. VAs are understandably concerned about being cloned. Nobody’s talking about the opposite: making up what characters sounds like, so any actor can play anybody. Or everybody. You can even substitute, when a scene needs extra oomph - like a band featuring a guitarist for a solo. Same sound… distinct performance.