This isn’t always true.
This isn’t always true.
I spotted that too. It has to be.
Art isn’t work, it’s speech. It’s part of the human condition. Art is useless, said Wilde. Art is for art’s sake—that is, for beauty’s sake.
I do not make art, I just post it here on lemmy. I’d be OK with that. People freely create, copy, and iterate on memes, and they are the greatest cultural touchstones we have. First and foremost, people create because they have something to say.
People already make memes and mods for free. Humans are a social species and will continue to create and share things until the end of time. Making money off of creation is a privilege for only a tiny few.
Have you tried Flux?
It keeps making me look for a hidden message.
The humor is just from someone shitposting Lorem ipsum at absurd length.
How much weed is that, to scale?
You keep moving the goal posts and putting words in my mouth. I never said you can do new things out of nothing. Nothing I mentioned is approaching, equaling, or exceeding the effort of training a model.
You haven’t answered a single one of my questions, and you are not arguing in good faith. We’re done here. I can’t say it’s been a pleasure.
Do you have any examples of how they fail? There are plenty of ways to explain new concepts to models.
https://arxiv.org/abs/2404.19427 https://arxiv.org/abs/2406.11643 https://arxiv.org/abs/2403.12962 https://arxiv.org/abs/2404.06425 https://arxiv.org/abs/2403.18922 https://arxiv.org/abs/2406.01300
What kind of creativity are you talking about then? I’ve also never heard of a bloated model. Which models are bloated?
But at what point does that guidance just become the dataset you removed from the training data?
The whole point is that it didn’t know the concepts beforehand, and no it doesn’t become the dataset. Observations made of the training data are added to the model’s weights after training, the dataset is never relevant again as the model’s weights are locked in.
To get it to run Doom, they used Doom.
To realize a new genre, you’ll “just” have to make that game the old fashion way, first.
Or you could train a more general model. These things happen in steps, research is a process.
There are more forms of guidance than just raw words. Just off the top of my head, there’s inpainting, outpainting, controlnets, prompt editing, and embeddings. The researchers who pulled this off definitely didn’t do it with text prompts.
I mean, you’ve never seen a purple elephant with a tennis racket. None of that exists in the data set since elephants are neither purple nor tennis players. Exposure to all the individual elements allows for generation of concepts outside the existing data, even though they don’t exit in reality or in the data set.
Page returns 404 for me.
You should read these two articles from Cory Doctorow. I think they’ll help clear up some thing for you.
https://pluralistic.net/2024/05/13/spooky-action-at-a-close-up/#invisible-hand
What the hell?
You’re good to go as long as it’s original enough. If it isn’t then that’s just copyright infringement. You might as well have right-clicked “save as”.