Leminal Space
  • Communities
  • Create Post
  • Create Community
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
David Gerard@awful.systemsM to TechTakes@awful.systemsEnglish · 5 days ago

AI image generators have just 12 generic templates

pivot-to-ai.com

external-link
message-square
15
link
fedilink
34
external-link

AI image generators have just 12 generic templates

pivot-to-ai.com

David Gerard@awful.systemsM to TechTakes@awful.systemsEnglish · 5 days ago
message-square
15
link
fedilink
There’s a new paper: “Autonomous language-image generation loops converge to generic visual motifs”  — diffusion models have just 12 standard templates. [Cell; Cell, with supplements, PDF; press re…

how many three-window rooms can one AI make

https://www.youtube.com/watch?v=khysGsyK9Qo&list=UU9rJrMVgcXTfa8xuMnbhAEA - video
https://pivottoai.libsyn.com/20251222-ai-image-generators-have-just-12-templates - podcast

time: 6 min 46 sec

  • David Gerard@awful.systemsOPM
    link
    fedilink
    English
    arrow-up
    7
    ·
    5 days ago

    oh yeah, Suno does the same, it has about 12 songs

    • flaviat@awful.systems
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      1
      ·
      5 days ago

      clanker’s dozen

    • Soyweiser@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      ·
      5 days ago

      Wonder if this is some sort of pre model collapse sign.

      • corbin@awful.systems
        link
        fedilink
        English
        arrow-up
        3
        ·
        10 hours ago

        Nah, it’s more to do with stationary distributions. Most tokens tend to move towards it; only very surprising tokens can move away. (Insert physics metaphor here.) Most LLM architectures are Markov, so once they get near that distribution they cannot escape on their own. There can easily be hundreds of thousands of orbits near the stationary distribution, each fixated on a simple token sequence and unable to deviate. Moreover, since most LLM architectures have some sort of meta-learning (e.g. attention) they can simulate situations where part of a simulation can get stuck while the rest of it continues, e.g. only one chat participant is stationary and the others are not.

        • Soyweiser@awful.systems
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 hours ago

          Thanks!

    • Witless Protection Program@mastodon.me.uk
      link
      fedilink
      arrow-up
      3
      ·
      5 days ago

      @dgerard @Soyweiser the Randy Newman record?

TechTakes@awful.systems

techtakes@awful.systems

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !techtakes@awful.systems

Big brain tech dude got yet another clueless take over at HackerNews etc? Here’s the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 53 users / day
  • 433 users / week
  • 1.01K users / month
  • 3.97K users / 6 months
  • 6 local subscribers
  • 2.34K subscribers
  • 1.03K Posts
  • 29.2K Comments
  • Modlog
  • mods:
  • David Gerard@awful.systems
  • BE: 0.19.14
  • Modlog
  • Legal
  • Instances
  • Docs
  • Code
  • join-lemmy.org