• 1 Post
  • 310 Comments
Joined 6 months ago
cake
Cake day: April 27th, 2025

help-circle




  • Let’s not forget these systems were deployed as a (deliberately escalatory) response to GWB withdrawing from the ABM treaty and placing interceptor batteries on shore in Eastern Europe.

    Interceptor batteries are hard to distinguish from IRBMs without inspecting the canister contents because they follow similar trajectories to the missiles they need to intercept and have therefore similar physical characteristics.

    This is not to excuse the Russians (fuck those guys) but to point out that this is the result of a strategic calculus that EVERYONE in the field warned the Bush administration about for literally years in the 2001–2008 timeframe.

    To be clear: by everyone I mean everyone. The arms control people were like “don’t do this, it will suck for everyone”, and the military types were like “start building IRBMs now so when the russkis start escalating we’ll have more”, but EVERYONE saw it coming.




  • Well, sort of. There is a difference between models that eat text and output images (diffusion models like Dalle and stable diffusion) and the models that eat images and text and output text (vision llms like qwen3-vl), but the way they both know what things look like is based on contrastive learning, based on an older model called CLIP and its descendants.

    Basically you feed a model both images and descriptions of images and train it to produce the same output vectors in both cases. Essentially it learns what a car looks like, and what the image of a car is called in whatever languages it’s trained in. If you only train a model on “acceptable” image/descriptions it literally never learns the words for “unacceptable” things and acts.

    Diffusion models are often fine tuned on specific types of porn (either full parameter or QLoRa), often with great effect. The same is much more work for llms though. Even if you remove the censorship (eg through abliteration, modifying the weights to inhibit outright denials), the models that’s left will not know the words it needs to express the concepts in the images.