@Roundcat This is one way to think of it, but my take is that all the arguments people make against leaving Twitter are just proxy arguments. The real arguments people don't want to say out loud are:
"I am attached to my follower count and don't want to rebuild"
and
"I like fighting with people, and this place encourages my brand of toxicity"
On the latter front, Bluesky is booming with bullies and scolds.
Small correction: Stable Diffusion has 2.3B 256x256 images, and 170M 512x512 images that it is trained on.
But still, that comes out to less than one byte of model data per image in the training set. There's literally no way to argue that a 256x256 image got compressed down to less than a byte.
It is worth noting to those who are in opposition to AI that the "models trained on my work contain my work" argument isn't technically or legally sound. Better arguments need to be made. If we could cram 1B+ 512x512 images into a single 2GB model, it would represent the single greatest breakthrough in data compression technology in human history. It turns out, this isn't a compression breakthrough, it's just proof that the original data isn't in the model.
I agree that some degree of mimicry isn't infringement - fair use includes some degree of derivation - but I really don't think that training and using a model to deliberately imitate another artist's work represents fair use, especially if it can be found to harm the artist.
The line shouldn't be drawn at "is it an exact copy?", it should be drawn at "would it confuse a reasonable person into thinking this was original art by another artist?".
Yesterday's lawsuits against Generative AI companies like Stable Diffusion and Midjourney were mostly dismissed, and I generally agree with that. There is one case still ongoing between Sarah Andersen and StabilityAI that I expect Sarah to lose on the grounds that her accusation isn't sound (her artwork isn't actually stored in the model).
But I am disappointed that they ruled that copying styles isn't infringement unless it is an exact copy of a work. I feel like this isn't a nuanced ruling...
@stux I really want to learn how they do it. I have tried a lot of experiments with Stable Diffusion, and it does a terrible job with a 360 view. I'm amazed that Skybox manages to get all the shadows pointing the right way!
@apLundell @jalefkowit No worse than most of the new houses here in Seattle. Lots of 3+ storey townhomes with about 300 square feet of space per floor. Something like 20% of the floorplan is stairwell.
I was legitimately searching for housing a year ago, and gave up when the agent asked me to make an offer the same day without an inspection on a townhome that had the kitchen on the third floor.
@apLundell @jalefkowit If they built this thing, I would never want to live in it, but I would feign interest just to get a tour. It wouldn't be the first time I pretended to have enough money to afford real estate just to tour a stupid looking house.
The worst I ever saw was a house with three different second stories that didn't connect to each other, and they each had a separate stairwell for access...
@apLundell @jalefkowit I typed "McMansion" and "McMansion Indoors" into Stable Diffusion, and gave it way too much canvas to work with, and I think you nailed it.
@esvrld@octodon.social There are also some holes in my philosophy as I consider this. There are a few models out there are unabashedly designed to copy a specific style. I have a Pixar-theme model that I have played with that is undoubtedly trained on Pixar's IP.
But I feel substantially less conflicted in infringing upon a $7.4B company, and should be more nuanced about how such training directed at smaller artists' works could be damaging to those artists.
@esvrld@octodon.social I think your proposal that "ai isn't actually capable of copying either the style or the subject matter of artists in the way that is alleged" is a philosophically interesting idea, but as a cautious AI advocate I would argue that, in practical terms, AI is very capable of copying an artist's style to the level of infringement, and would argue that responsible and ethically informed use of generative models is of high importance to anyone exploring this tech.
@esvrld@octodon.social this is interesting input. It definitely would be an issue that certain niche art styles in the corpus could end up zeroing in non-trivially on a specific artist's style. I hadn't fully considered this, and have been assuming that in the absence of an artist's name in the prompt that the art would be more generic. This immediately has me feeling more skeptical of third party models that don't share their datasets and could be weighted in such a way to be infringing by default.
And done
Get your free 'NFTs' or whatever now and do whatever you like with 'em
Live: http://stux.stuxnet.ai/f-nft
Code: https://gitcat.org/stux/f-nft
@stux These are really fun! I'll make sure to clone the git repo so I too can own one of these "NFTs".
Just curious, are the backgrounds generated from Skybox AI (Blockade Labs), or sourced from somewhere else?
@doug FWIW, The USSR had copyright laws that lasted 15 years. If we had similar laws, training a model on only public domain materials would be extremely easy.
But in the USA, thanks to corporate lobbying, a copyrights lasts 70 years, and a trademark can be renewed indefinitely. Narrowing datasets down to the public domain is impossibly complicated.
What choice do AI engineers have but to ask for forgiveness rather than permission in these matters?
@doug I also don't see how relaxed views on intellectual property are "anti-socialist". I am deeply opposed to the US system of IP law, especially with recent attacks on fair use. I don't think that artists should be able to retain complete and permanent ownership over their works once they are available to the public. I believe that information should be free, and that expression, even controversial expression, should not be restricted by anyone's financial interests.
@doug Well, I am not a billionaire investor, and I am not using AI to benefit financially. I feel like you are in opposition to the right things, but are a little too broad with your criticism.
I primarily use Stable Diffusion which is MIT licensed FOSS software, so I haven't even paid the creators of these models. I am entirely a private user. My biggest uses are experimentation, and creating desktop backgrounds for private use.
And if I do share, I do so without any form of license.
@doug On the other side of this, if someone set up an Etsy store selling prints of AI art that used prompts like "Painting by [living artist]", and were deliberately emulating that artist or artists for the purpose of profiting off of their reputation, I would absolutely call that an infringement. That would be deeply unethical.
I think that there is some nuance to be afforded here.
Personally, I hope to see artist names removed from AI data sets so that their styles cannot be directly copied.
#Netsec Professional. Whitehat #Hacker. #Demoscene spectator. Nerd.
I'm a fan of #Linux, #FOSS, #Decentralization (not Crypto), Crypto (as in #Cryptography), and #Socialism. Always #Antifascist & #Antiwar.
Seattle, WA