An image of an archeologist adventurer who wears a hat and uses a bullwhip
647 comments
·April 3, 2025ionwake
District5524
I asked Sora to turn a random image of my friend and myself into Italian plumbers. Nothing more, just the two words "Italian plumbers". The created picture was not shown to me because it was in violation of OpenAI's content policy. I asked then just to turn the guys on the picture into plumbers, but I asked this in the Italian language. Without me asking for it, Sora put me in an overall and gave me a baseball cap, and my friend another baseball cap. If I asked Sora to put mustache on us, one of us received a red shirt as well, without being asked to. Starting with the same pic, if I asked to put one letter on the baseball caps each - guess, the letters chosen were M and L. These extra guardrails are not really useful with such a strong, built-in bias towards copyright infringement of these image creation tools. Should it mean that with time, Dutch pictures will have to include tulips, Italian plumbers will have to have a uniform with baseball caps with L and M, etc. just not to confuse AI tools?
weinzierl
Many years ago I tried to order a t-shirt with the postscript tiger on the front from Spreadshirt.
It was removed on Copyright claims before I could order one item myself. After some back and forth they restored it for a day and let me buy one item for personal use.
My point is: Doesn't have to be Sony, doesn't have to be a snitch - overzealous anticipatory obedience by the shop might have been enough.
Xmd5a
>After some back and forth they restored it for a day and let me buy one item for personal use.
I used Spreadshirt to print a panel from the Tintin comic on a T-shirt, and I had no problem ordering it (it shows Captain Haddock moving through the jungle, swatting away the mosquitoes harassing him, giving himself a big slap on the face, and saying, 'Take that, you filthy beasts!').
beardyw
I bought Tintin T-shirts 40 years ago in Thailand (the "branded" choices were amazing). They were actually really good, still got them!
moffkalast
> my whole shop was taken offline
I think the problem there was being dependent on someone who is a complete pushover, doesn't bother to check for false positives and can kill your business with a single thought.
lukan
How was your shop taken down?
Usually there are lawyers letters involved first?
apgwoz
Print in demands definitely have terms of service allowing them to take whatever down. You’re playing by their rules, and your $2 revenue / tshirt and very few overall sales is not worth the potentially millions in legal fees to fight for you.
djoldman
I don't condone or endorse breaking any laws.
That said, trademark laws like life of the author + 95 years are absolutely absurd. The ONLY reason to have any law prohibiting unlicensed copying of intangible property is to incentivize the creation of intangible property. The reasoning being that if you don't allow people to exclude 3rd party copying, then the primary party will assumedly not receive compensation for their creation and they'll never create.
Even in the case where the above is assumed true, the length of time that a protection should be afforded should be no more than the length of time necessary to ensure that creators create.
There are approximately zero people who decide they'll create something if they're protected for 95 years after their death but won't if it's 94 years. I wouldn't be surprised if it was the same for 1 year past death.
For that matter, this argument extends to other criminal penalties, but that's a whole other subject.
csallen
> The ONLY reason to have any law prohibiting unlicensed copying of intangible property is to incentivize the creation of intangible property.
That was the original purpose. It has since been coopted by people and corporations whose incentives are to make as much money as possible by monopolizing valuable intangible "property" for as long as they can.
And the chief strategic move these people have made is to convince the average person that ideas are in fact property. That the first person to think something and write it down rightfully "owns" that thought, and that others who express it or share it are not merely infringing copyright, they are "stealing."
This plan has largely worked, and now the average person speaks and thinks in these terms, and feels it in their bones.
specproc
It's been a US-led project for the benefit of American corporations.
If I was running the trade emergency room in any European state right now, I'd have "stop enforcing US copyright" up there next to "reciprocal tarrifs".
TeMPOraL
Unfortunately we have a bunch of copyright-friendly groups in EU, so this would only work in the "stop enforcing US copyright in retaliation" sense, but not likely in the "stop enforcing copyright because on the net, it's a scam" sense.
null
InDubioProRubio
Worked for china
the_other
Philosophers were waaaay ahead of this game.
null
jl6
> There are approximately zero people who decide they'll create something if they're protected for 95 years after their death but won't if it's 94 years.
I’m sure you’re right for individual authors who are driven by a creative spark, but for, say, movies made by large studios, the length of copyright is directly tied to the value of the movie as an asset.
If that asset generates revenue for 120 years, then it’s slightly more valuable than an asset that generates revenue for 119 years, and considerably more valuable than an asset that generates revenue for 20 years.
The value of the asset is in turn directly linked to how much the studio is willing to pay for that asset. They will invest more money in a film they can milk for 120 years than one that goes public domain after 20.
Would studios be willing to invest $200m+ in movie projects if their revenue was curtailed by a shorter copyright term? I don’t know. Probably yes, if we were talking about 120->70. But 120->20? Maybe not.
A dramatic shortening of copyright terms is something of a referendum on whether we want big-budget IP to exist.
In a world of 20 year copyright, we would probably still have the LOTR books, but we probably wouldn’t have the LOTR movies.
AnthonyMouse
> If that asset generates revenue for 120 years, then it’s slightly more valuable than an asset that generates revenue for 119 years, and considerably more valuable than an asset that generates revenue for 20 years.
Not so, because of net present value.
The return from investing in normal stocks is ~10%/year, which is to say ~670% over 20 years, because of compounding interest. Another way of saying this is that $1 in 20 years is worth ~$0.15 today. A dollar in 30 years is worth ~$0.05 today. A dollar in 40 years is worth ~$0.02 today. As a result, if a thing generates the same number of dollars every year, the net present value of the first 20 years is significantly more than the net present value of all the years from 20-120 combined, because money now or soon from now is worth so much more than money a long time from now. And that's assuming the revenue generated would be the same every year forever, when in practice it declines over time.
The reason corporations lobby for copyright term extensions isn't that they care one bit about extended terms for new works. It's because they don't want the works from decades ago to enter the public domain now, and they're lobbying to make the terms longer retroactively. But all of those works were already created and the original terms were sufficient incentive to cause them to be.
fashion-at-cost
You analysis misses the incredibly important caveat that revenue rises with inflation.
jl6
> And that's assuming the revenue generated would be the same every year forever, when in practice it declines over time.
For the crown jewel IP that the studios are most interested in protecting, the opposite of this assumption is true. Star Wars, for example, is making more money than ever. Streaming revenues will probably invalidate that assumption for an even wider pool of back catalog properties.
m000
> I’m sure you’re right for individual authors who are driven by a creative spark, but for, say, movies made by large studios, the length of copyright is directly tied to the value of the movie as an asset.
That would be fine, if the studios didn't want to have it both ways. They want to retain full copyright control over their "asset", but they also use Hollywood Accounting [1] to both avoid paying taxes and cheat contributors that have profit-sharing agreements.
If studios declare that they made a loss on producing and releasing something to get a tax break, the copyright term for that work should be reduced to 10 years tops.
johanvts
For movies in particular the tail is very thin. Only very few 50 year old movies are ever watched. Was any commercial movie ever financed without a view to making a profit in the box office/initial release?
techpression
According to Matt Damon (in one of many interviews) a lot of movies were produced with the second revenue stream (vhs/dvd) being part of the calculations, that is why we now get a lot less variety and alternative movies made, that second revenue stream doesn’t exist any more (I assume streaming pays very little in comparison).
lupusreal
With books it's even worse. Movies might get a trickle of revenue from TV licensing but once a book is out of print (which usually happens very quickly, and most never go into print again), that's it. No more revenue from that book, it continues to circulate in libraries and used bookstores but the author and publisher gets nothing from that.
benwad
The Fellowship of the Ring, the first of Peter Jackson's LOTR movies released in 2001, made $887 million in its original theatrical run (on a $93 million budget). It would absolutely still have been made if copyright was only 20 years. And now it would be in the public domain!
jl6
The success that we can now measure through hindsight wasn’t assured at the time of greenlighting the film. They took a huge gamble:
https://variety.com/2021/film/news/lord-of-the-rings-peter-j...
It would have been an even bigger gamble if they weren’t able to bank on any long term revenue (I’m certain Netflix continues to pay for the rights to stream the trilogy after 2021).
londons_explore
> If that asset generates revenue for 120 years, then it’s slightly more valuable than an asset that generates revenue for 119 years, and considerably more valuable than an asset that generates revenue for 20 years
Due to the fairy high cost of capital right now, pretty much anything more than 5 years away is irrelevant. 10 years max, even for insanely high returns on investment.
aucisson_masque
> If that asset generates revenue for 120 years, then it’s slightly more valuable than an asset that generates revenue for 119 years
Movies for instance make most of their revenue in the 2 week following their release in theater. Beyond, peoole who wanted to see it had already seen it and the others don't care.
I'd argue it's similar for other art form, even for music. The gain at the very end of the copyright lifetime is extremely marginal and doesn't influence spending decision, which is mostly measured on a return basis of at most 10 years.
pegasus
Seems to me most of that inflated budget is needed for the entertainment role of films, not the art in them, which a low budget can often stimulate rather than inhibit. In which case nothing of importance would be lost by a drastic shortening of copyright terms.
jpc0
You are putting a pretty blatant cap on what constitutes "art"...
legulere
This line of reasoning doesn't make sense for retroactive lengthening of copyright though, as the author is not gaining anything from that anymore.
RataNova
While I think the laws are broken, I also get why companies fight so hard to defend their IP: it is valuable, and they've built empires around it. But at some point, we have to ask: are we preserving culture or just hoarding it?
jpc0
You are missing a bunch of edge cases, and the law is all about edge cases.
An artist who works professionally has family members, family members who are dependent on them.
If they pass young, become popular just before they pass and their extremely popular works are now public domain. Their family sees nothing from their work, that is absolutely being commercialized ( publishing and creation generally spawns two seperate copyrights).
TeMPOraL
GP's not missing those edge cases; GP recognizes those edge cases are themselves a product of IP laws.
Those laws are effectively attempting to make information behave as physical objects, by giving them simulated "mass" through a rent-seeking structure. The case you describe is where this simulated physical substrate stops behaving like physical substrate, and choice was made to paper over that with extra rules, so that family can inherit and profit from IP of a dead creator, much like they would inherit physical products of a dead craftsman and profit from selling them.
It's a valid question whether or not this is taking things too far, just for the sake of making information conform to rules of markets for physical goods.
aziaziazi
Il not sure IP should be used as a life insurance, there’s already many public and private ideas tools for that.
Also it seems you assume inheritance is a good think. Most people do think the same on a personal level, however when we observes the effect on a society the outcome is concentration of wealth on a minority and barriers for wealth hand change === barriers for "American dream".
Matumio
You seem to talk about fairness. Copyright law isn't supposed to be fair, it's supposed to benefit society. On one side you have the interest of the public to make use of already created work. On the other side is the financial incentive to create such work in the first place.
So the question to ask is whether the artist would have created the work and published it, even knowing that it isn't an insurance to their family in case of their early death.
null
anhner
If copyright law is reduced to say, 20 years from the date of creation (PLENTY of time for the author to make money), then it's irrelevant if he dies young or lives until 100.
IanCal
The weirdness there is tying it to someone's life.
noduerme
You're conflating trademark with copyright.
Regardless, it's not just copyright laws that are at issue here. This is reproducing human likenesses - like Harrison Ford's - and integrating them into new works.
So if I want to make an ad for a soap company, and I get an AI to reproduce a likeness of Harrison Ford, does that mean I can use that likeness in my soap commercials without paying him? I can imagine any court asking "how is this not simply laundering someone's likeness through a third party which claims to not have an image / filter / app / artist reproducing my client's likeness?"
All seemingly complicated scams come down to a very basic, obvious, even primitive grift. Someone somewhere in a regulatory capacity is either fooled or paid into accepting that no crime was committed. It's just that simple. This, however, is so glaring that even a child could understand the illegality of it. I'm looking forward to all of Hollywood joining the cause against the rampant abuse of IP by Silicon Valley. I think there are legal grounds here to force all of these models to be taken offline.
Additionally, "guardrails" that prevent 1:1 copies of film stills from being reprinted are clearly not only insufficient, they are evidence that the pirates in this case seek to obscure the nature of their piracy. They are the evidence that generative AI is not much more than a copyright laundering scheme, and the obsession with these guardrails is evidence of conspiracy, not some kind of public good.
planb
> So if I want to make an ad for a soap company, and I get an AI to reproduce a likeness of Harrison Ford, does that mean I can use that likeness in my soap commercials without paying him?
No, you can't! But it shouldn't be the tool that prohibits this. You are not allowed to use existing images of Harrison Ford for your commercial and you also will be sued into oblivion by Disney if you paint a picture of Mickey Mouse advertising your soap, so why should it be any different if an AI painted this for you?
noduerme
Well, precisely. What then is the AI company's justification for charging money to paint a picture of Harrison Ford to its users?
The justification so far seems to have been loosely based on the idea that derivative artworks are protected as free expression. That argument loses currency if these are not considered derivative but more like highly compressed images in a novel, obfuscated compression format. Layers and layers of neurons holding a copy of Harrison Ford's face is novel, but it's hard to see why it's any different legally than running a JPEG of it through some filters and encoding it in base64. You can't just decode it and use it without attribution.
adrianmsmith
> you also will be sued into oblivion by Disney if you paint a picture of Mickey Mouse advertising your soap, so why should it be any different if an AI painted this for you?
If the AI prompt was "produce a picture of Micky Mouse", I'd agree with you.
The creators of AI claim their product produces computer-generated images, i.e. generated/created by the computer. Instead it's producing a picture of a real actual person.
If I contract an artist to produce a picture of a person from their imagination, i.e. not a real person, and they produce a picture of Harrison Ford, then yeah I'd say that's on the artist.
AnthonyMouse
> This is reproducing human likenesses - like Harrison Ford's - and integrating them into new works.
The thing is though, there is also a human requesting that. The prompt was chosen specifically to get that result on purpose.
The corporate systems are trying to prevent this, but if you use any of the local models, you don't even have to be coy. Ask it for "photo of Harrison Ford as Indiana Jones" and what do you expect? That's what it's supposed to do. It does what you tell it to do. If you turn your steering wheel to the left, the car goes to the left. It's just a machine. The driver is the one choosing where to go.
IanCal
> This, however, is so glaring that even a child could understand the illegality of it
If you have to explain "laundering someone's likeness" to them maybe not, I think it's a frankly bizarre phrase.
FeepingCreature
Human appearance does not have enough dimensions to make likeness a viable thing to protect; I don't see how you could do that without say banning Elvis impersonators.
That said:
> I'm looking forward to all of Hollywood joining the cause against the rampant abuse of IP by Silicon Valley.
If you're framing the sides like that, it's pretty clear which I'm on. :)
noduerme
Interesting you should bring that up:
https://www.calcalistech.com/ctechnews/article/1517ldjmv
Loads of lawsuits have been filed by celebrities and their estates over the unauthorized use of their likeness. And in fact, in 2022, Las Vegas banned Elvis impersonators from performing weddings after a threat from the Presley estate's licensing company:
https://www.dailymail.co.uk/news/article-10872855/Elvis-imag...
But there are also a couple key differences between putting on a costume and acting like Elvis, and using a picture of Elvis to sell soap.
One is that a personal artistic performance could be construed as pastiche or parody. But even more importantly, if there's a financial incentive involved in doing that performance, the financial incentive has to be aligned more with the parody than with drawing an association to the original. In other words, dressing up as Elvis as a joke or an act, or even to sing a song and get paid to perform a wedding is one thing if it's a prank, it's another thing if it's a profession, and yet another thing if it's a mass-marketing advertisement that intends for people to seriously believe that Elvis endorsed this soap.
ghaff
You’re thinking of copyright law, not trademark law. Which serves a different function. If you’re going to critique something it’s useful to get your facts right.
southernplaces7
>I don't condone or endorse breaking any laws.
Really? Because there are a lot of very stupid laws out there that should absolutely be broken, regularly, flagrantly, by as many people as possible for the sake of making enforcement completely null and pointless. Why write in neutered corporate-speak while commenting on a casual comment thread while also (correctly) pointing out the absurdity of certain laws.
kazinator
New rule: you get to keep your belongings for 20 years and that's it. Then they are taken away. Every dollar you made twenty or more years ago, every asset you acquired, gone.
That oughtta be enough to incentivize people to work and build their wealth.
Anything more than that is unnecessary.
lazyasciiart
I think you mean "20 years after you die". That seems like a perfectly rational way to deal with people who want to be buried in their gold and jewels in a pyramid.
FeepingCreature
We should never have accepted the term "intellectual property" at all, if this is the mindset it leads to.
ars
Intellectual property does not belong to you. The entire concept of "owning" intellectual property is a very recent thing.
So how about we go back to how it used to be and just remove this entire concept.
You can own things you can't own an idea.
hackable_sand
I disagree that you can own things, but I certainly concede IP as a good starting point.
KronisLV
I think the cat is out of the bag when it comes to generative AI, the same way how various LLMs for programming have been trained even on codebases that they had no business using, yet nobody hasn’t and won’t stop them. It’s the same as what’s going to happen with deepfakes and such, as the technology inevitably gets better.
> Hayao Miyazaki’s Japanese animation company, Studio Ghibli, produces beautiful and famously labor intensive movies, with one 4 second sequence purportedly taking over a year to make.
It makes me wonder though - whether it’s more valuable to spend a year on a scene that most people won’t pay that much attention to (artists will understand and appreciate, maybe pause and rewind and replay and examine the details, the casual viewer just enjoy at a glance) or use tools in addition to your own skills to knock it out of the park in a month and make more great things.
A bit how digital art has clear advantages over paper, while many revere the traditional art a lot, despite it taking longer and being harder. The same way how someone who uses those AI assisted programming tools can improve their productivity by getting rid of some of the boilerplate or automate some refactoring and such.
AI will definitely cheapen the art of doing things the old way, but that’s the reality of it, no matter how much the artists dislike it. Some will probably adapt and employ new workflows, others stick to tradition.
M95D
It's a very clear difference between a cheap animation and Ghibli. Anyone can see it.
In the first case, there's only one static image for an entire scene, scrolled and zoomed, and if they feel generous, there would be an overlay with another static image that slides over the first at a constant speed and direction. It feels dead.
In the second case, each frame is different. There's chaotic motions such as wind and there's character movement with a purpose, even in the background, there's always something happening in the animation, there's life.
paulluuk
There is a huge middle ground between "static image with another sliding static image" and "1 year of drawing per 4 second Ghibli masterpiece". From your comment is almost looks like you're suggesting that you have to choose either one or the other, but that is of course not true.
I bet that a good animator could make a really impressive 4-second scene if they were given a month, instead of a year. Possibly even if they were given a day.
So if we assume that there is not a binary "cheap animation vs masterpiece" but rather a sort of spectrum between the two, then the question is: at what point do enough people stop seeing the difference, that it makes economic sense to stay at that level, if the goal is to create as much high-quality content as possible?
zipmapfoldright
anyone _can_ see it, but _most_ people don't (and don't care)
To be clear, I am not saying it's not valuable, only that to the vast majority, it's not.
soneca
I wonder if really great stuff are always for a minority. You have to have listened a lot of classical music to notice a great interpretation of Mozart from a good one. To realize how great was a chess move, how magical was a soccer play, how deep was the writing of a philosopher. Not only for stuff that requires previous effort, but also the subjectiveness of art. Picasso will be really moving for a minority of people. The Godfather. Even Shakespeare.
Social media and generative AI may be good business because the capture the attention of the majority, but maybe they are not valuable to anyone.
umko21
I think you’re right that most people don’t notice, but without the extra effort, it would’ve ended up as just another mediocre animation. And standing out from mediocrity is what made it appealing to many people.
whywhywhywhy
> but _most_ people don't (and don't care)
Perhaps it's not for everyone.
M95D
Who cares if it's valuable for the majority? What do you think this is? Stock market for slop?
This is art.
balamatom
So what?
IanCal
Fundamentally I think this comes down to answering the question of "why are you creating this?".
There are many valid answers.
Maybe you want to create it to tell a story, and you have an overflowing list of stories you're desperate to tell. The animation may be a means to an end, and tools that help you get there sooner mean telling more stories.
Maybe you're pretty good at making things people like and you're in it for the money. That's fine, there are worse ways to provide for your family than making things people enjoy but aren't a deep thing for you.
Maybe you're in it because you love the act of creating it. Selling it is almost incidental, and the joy you get from it comes down to spending huge amounts of time obsessing over tiny details. If you had a source of income and nobody ever saw your creations, you'd still be there making them.
These are all valid in my mind, and suggest different reasons to use or not to use tools. Same as many walks of life.
I'd get the weeds gone in my front lawn quickly if I paid someone to do it, but I quite enjoy pottering around on a sunny day pulling them up and looking back at the end to see what I've achieved. I bake worse bread than I could buy, and could buy more and better bread I'm sure if I used the time to do contracting instead. But I enjoy it.
On the other hand, there are things I just want done and so use tools or get others to do it for me.
One positive view of AI tools is that it widens the group of people who are able to achieve a particular quality, so it opens up the door for people who want to tell the story or build the app or whatever.
A negative side is the economics where it may be beneficial to have a worse result just because it's so much cheaper.
mytailorisrich
> It makes me wonder though - whether it’s more valuable to spend a year on a scene that most people won’t pay that much attention to
In this case, yes it is.
People do pay attention to the result overall. Studio Ghibli has got famous because people notice what they produce.
Now people might not notice every single detail but I believe that it is this overall mindset and culture that enables the whole unique final product.
xandrius
I think most like the vibes, not the fact it took ages to make.
null
Qualitionion
Its the quality or level of detail.
Which might indicate an environment were quality is above quantity
happyraul
To me the question of what activity/method is more "valuable" in the context of art is kind of missing the point of art.
MgB2
Idk, the models generating what are basically 1:1 copies of the training data from pretty generic descriptions feels like a severe case of overfitting to me. What use is a generational model that just regurgitates the input?
I feel like the less advanced generations, maybe even because of their limitations in terms of size, were better at coming up with something that at least feels new.
In the end, other than for copyright-washing, why wouldn't I just use the original movie still/photo in the first place?
yk
Tried Flux.dev with the same prompts [0] and it seems actually to be a GPT problem. Could be that in GPT the text encoder understands the prompt better and just generates the implied IP, or could be that a diffusion model is just inherently less prone to overfitting than a multimodal transformer model.
[0] https://imgur.com/a/wqrBGRF Image captions are the impled IP, I copied the prompts from the blog post.
jsemrau
DALL-E 3 already uses a model that trained on synthetic data that take the prompt and augments it. This might lead to the overfitting. It could also be, and might be the simpler explanation, that its just looks up the right file from a RAG.
vjerancrnjak
If it overfits on the whole internet then it’s like a search engine that returns really relevant results with some lossy side effect.
Recent benchmark on unseen 2025 Math Olympiad shows none of the models can problem solve . They all accidentally or on purpose had prior solutions in the training set.
jks
You probably mean the USAMO 2025 paper. They updated their comparison with Gemini 2.5 Pro, which did get a nontrivial score. That Gemini version was released five days after USAMO, so while it's not entirely impossible for the data to be in its training set, it would seem kind of unlikely.
jsemrau
The same timing is actually suspicious. And it would not be the first time something like this happened.
iamacyborg
I was noodling with Gemini 2.5 Pro a couple days ago and it was convinced Donald Trump didn’t win the 2024 election and that he conceded to Kamala Harris so I’m not entirely sure how much weight I’d put behind it.
gertlex
What if the word "generic" were added to a lot of these image prompts? "generic image of an intergalactic bounty hunter from space" etc.
Certainly there's an aspect of people using the chat interface like they use google: describe xyz to try to surface the name of a movie. Just in this case, we're doing the (less common?) query of: find me the picture I can vaguely describe; but it's a query to a image /generating/ service, not an image search service.
squeaky-clean
Generic doesn't help. I was using the new image generator to try and make images for my Mutants and Masterminds game (it's basically D&D with superheroes instead of high fantasy), and it refuses to make most things citing that they are too close to existing IP, or that the ideas are dangerous.
So I asked it to make 4 random and generic superheroes. It created Batman, Supergirl, Green Lantern, and Wonder Woman. Then at about 90% finished it deleted the image and said I was violating copyright.
I doubt the model you interact with actually knows why the babysitter model rejects images, but it claims to know why and leads to some funny responses. Here is it's response to me asking for a superhero with a dark bodysuit, a purple cape, a mouse logo on their chest, and a spooky mouse mask on their face.
> I couldn't generate the image you requested because the prompt involved content that may violate policy regarding realistic human-animal hybrid masks in a serious context.
gertlex
Yeah... so much for that hope on my end! Thanks for testing.
(hard to formulate why I was too lazy to test myself :) )
jeroenhd
People like what they already know. When they prompt something and get a realistic looking Indiana Jones, they're probably happy about it.
To me, this article is further proof that LLMs are a form of lossy storage. People attribute special quality to the loss (the image isn't wrong, it's just got different "features" that got inserted) but at this point there's not a lot distinguishing a seed+prompt file+model from a lossy archive of media, be it text or images, and in the future likely video as well.
The craziest thing is that AI seems to have gathered some kind of special status that earlier forms of digital reproduction didn't have (even though those 64kbps MP3s from napster were far from perfect reproductions), probably because now it's done by large corporations rather than individuals.
If we're accepting AI-washing of copyright, we might as well accept pirated movies, as those are re-encoded from original high-resolution originals as well.
balamatom
Probably the majority of people in the world already "accept pirated movies". It's just that, as ever, nobody asks people what they actually want. Much easier to tell them what to want, anyway.
To a viewer, a human-made work and an AI-generated one both amount to a series of stimuli that someone else made and you have no control over; and when people pay to see a movie, generally they don't do it with the intent to finance the movie company to make more movies -- they do it because they're offered the option to spend a couple hours watching something enjoyable. Who cares where it comes from -- if it reached us, it must be good, right?
The "special status" you speak of is due to AI's constrained ability to recombine familiar elements in novel ways. 64k MP3 artifacts aren't interesting to listen to; while a high-novelty experience such as learning a new culture or a new discipline isn't accessible (and also comes with expectations that passive consumption doesn't have.)
Either way, I wish the world gave people more interesting things to do with their brains than make a money, watch a movies, or some mix of the two with more steps. (But there isn't much of that left -- hence the concept of a "personal life" as reduced to breaking one's own and others' cognitive functioning then spending lifetimes routing around the damage. Positively fascinating /s)
RataNova
Yeah, I've been feeling the same. When a model spits out something that looks exactly like a frame from a movie just because I typed a generic prompt, it stops feeling like “generative” AI and more like "copy-paste but with vibes."
FeepingCreature
To my knowledge this happens when that single frame is overrepresented in its training data. For instance, variations of the same movie poster or screenshot may appear hundreds of times. Then the AI concludes that this is just a unique human cultural artifact, like the Mona Lisa (which I would expect many human artists could also reproduce from memory).
tshaddox
Idk, a couple of the examples might be generic enough that you wouldn't expect a very specific movie character. But most of the prompts make it extremely clear which movie character you would expect to see, and I would argue that the chat bot is working as expected by providing that.
grotorea
Even if I'm thinking of an Indiana Jones-like character doesn't mean I want literally Indiana Jones. If I wanted Indiana Jones I could just grab a scene from the movie.
infthi
if someone gets Indiana on the suspiciously detailed request the author provided and it appears they wanted something else, they can clarify that to the chat bot, e.g. by copying this your comment.
I have a strong suspicion that many human artists would behave in a way the chat bot did (unless they start asking clarifying questions. Which chatbots should learn to do as well)
m000
Good luck grabbing that Ghibli movie scene, where Indiana Jones arm-wrestles Lara Croft in a dive-bar, with Brian Flanagan serving a cocktail to Allan Quatermain in the background.
stevage
Why? For fan content, using the original characters in new situations, mashups, new styles etc.
Lerc
I'm not sure if this is a problem with overfitting. I'm ok with the model knowing what Indiana Jones or the Predator looks like with well remembered details, it just seems that it's generating images from that knowledge in cases where that isn't appropriate.
I wonder if it's a fine tuning issue where people have overly provided archetypes of the thing that they were training towards. That would be the fastest way for the model to learn the idea but it may also mean the model has implicitly learned to provide not just an instance of a thing but a known archetype of a thing. I'm guessing in most RLHF tests archetypes (regardless of IP status) score quite highly.
masswerk
What I'm kind of concerned about is that these images will persist and will be reinforced by positive feedback. Meaning, an adventurous archeologist will be the same very image, forever. We're entering the epitome of dogmatic ages. (And it will be the same corporate images and narratives, over and over again.)
duskwuff
And it's worth considering that this issue isn't unique to image generation, either.
baq
Welcome to the great age of slop feedback loops.
vkou
> I'm ok with the model knowing what Indiana Jones or the Predator looks like with well remembered details,
ClosedAI doesn't seem to be OK with it, because they are explicitly censoring characters of more popular IPs. Presumably as a fig leaf against accusations of theft.
red75prime
If you define feeding of copyrighted material into a non-human learning machine as theft, then sure. Anything that mitigates legal consequences will be a fig leaf.
The question is "should we define it as such?"
jauntywundrkind
Obviously a horrible hideous theft machine.
One thing I would say, it's interesting to consider what would make this not so obviously bad.
Like, we could ask AI to assess the physical attributes of the characters it generated. Then ask it to permute some of those attributes. Generate some random tweaks: ok but brawy, short, and a different descent. Do similarly on some clothing colors. Change the game. Hit the "random character" button on the physical attributes a couple times.
There was an equally shatteringly-awful less-IP-theft (and as someone who thinks IP is itself incredibly ripping off humanity & should be vastly scoped down, it's important to me to not rest my arguments on IP violations).... An equally shattering recent incident for me. Having trouble finding it, don't remember the right keywords, but an article about how AI has a "default guy" type that it uses everywhere, a super generic personage, that it would use repeatedly. It was so distasteful.
The nature of 'AI as compression', as giving you the most median answer is horrific. Maybe maybe maybe we can escape some of this trap by iterating to different permutations, by injecting deliberate exploration of the state spaces. But I still fear AI, worry horribly when anyone relies on it for decision making, as it is anti-intelligent, uncreative in extreme, requiring human ingenuity to budge off its rock of oppressive hypernormality that it regurgitates.
areoform
Theft from whom and how?
Are you telling me that our culture should be deprived of the idea of Indiana Jones and the feelings that character inspires in all of us forever just because a corporation owns the asset?
Indiana Jones is 44 years old. When are we allowed to remix, recreate and expand on this like humanity has done since humans first started sitting down next to a fire and telling stories?
edit: this reminds of this iconic scene from Dr. Strangelove, https://www.youtube.com/watch?v=RZ9B7owHxMQ
Mandrake: Colonel... that Coca-Cola machine. I want you to shoot the lock off it. There may be some change in there.
Guano: That's private property.
Mandrake: Colonel! Can you possibly imagine what is going to happen to you, your frame, outlook, way of life, and everything, when they learn that you have obstructed a telephone call to the President of the United States? Can you imagine? Shoot it off! Shoot! With a gun! That's what the bullets are for, you twit!
Guano: Okay. I'm gonna get your money for ya. But if you don't get the President of the United States on that phone, you know what's gonna happen to you?
Mandrake: What?
Guano: You're gonna have to answer to the Coca-Cola company.
I guess we all have to answer to the Walt Disney company.calmbell
"idea of Indiana Jones and the feelings that character inspires in all of us forever just because a corporation owns the asset" is very different from the almost exact image of Indiana Jones.
GolfPopper
And a reason people are getting ticked at the AI companies is the hypocrisy. They're near-universally arguing that it's okay for them to treat copyright in a way that it is illegal for us to, apparently on the basis of, "we've got a billions in investment capital, and applying the law equally will make it hard for us to get a return on that investment".
chongli
Exactly. The idea of Indiana Jones, the adventurer archaeologist more at home throwing a punch than reading a book, is neither owned by nor unique to Lucasfilm (Disney). There is a ton of media out there featuring this trope character [1]. Yes, the trope is overwhelmingly associated with the image of Harrison Ford in a fedora within the public consciousness, but copyright does not apply to abstract ideas such as tropes.
Some great video games to feature adventurer archaeologists:
* NetHack (One of the best roles in the game)
* Tomb Raider series (Lara Croft is a bona fide archaeologist)
* Uncharted series (Nathan Drake is more of a treasure hunter but he becomes an archaeologist when he retires from adventuring)
* Professor Layton series
* La-Mulana series (very obviously inspired by Indiana Jones, but not derivative)
* Spelunky (inspired by La-Mulana)
[1] https://tvtropes.org/pmwiki/pmwiki.php/Main/AdventurerArchae...
_ph_
Not forever. But 75 years after the death of the creator by current international agreement. I definitely think that the exact terms of copyright should be revisited - a lot of usages should be allowed like 50 years of publishing a piece of work. But that needs to be agreed upon and converted into law. Till then, one should expect everyone, especially large corporations, to stick to the law.
saulpw
When Mickey Mouse was created (1928), copyright was 28 years that could be reupped once for an additional 28 years. So according to those terms, Mickey Mouse would have ascended to the public domain in 1984.
IMO any change to copyright law should not be applied retroactively. Make copyright law to be what is best for society and creators as a whole, not for lobbyists representing already copyrighted material.
fullstop
I mean, at least shouldn't we wait until Harrison Ford has passed?
jauntywundrkind
Its kind of funny that everyone is harping this way or that way about IP.
This is a kind of strange comment for me to read. Because imby tone it sounds like a rebuttal? But by content, it agrees with a core thing I said about myself:
> and as someone who thinks IP is itself incredibly ripping off humanity & should be vastly scoped down, it's important to me to not rest my arguments on IP violations
What's just such a nightmare to me is that the tech is so normative. So horribly normative. This article shows that AI again and again reproduced only the known, only the already imagined. Its not that it's IP theft that rubs me so so wrong, it's that it's entirely bankrupt & uncreative, so very stuck. All this power! And yet!
You speak at what disgusts me yourself!
> When are we allowed to remix, recreate and expand on this like humanity has done
The machine could be imagining all kinds of Indianas. Of all different remixed recreated expanded forms. But this pictures are 100% anything but that. They're Indiana frozen in Carbonite. They are the driest saddest prison of the past. And call into question the validity of AI entirely, show something greviously missing.
sothatsit
> All this power! And yet!
You are completely ignoring the fact that you can provide so much more information to the LLMs to get what you want. If you truly want novel images, ChatGPT can absolutely provide them, but you have to provide a better starting point than "An image of an archeologist adventurer who wears a hat and uses a bullwhip".
If you just provide a teensy bit more information, the results dramatically change. Try out "An image of an Indian female archeologist adventurer who wears a hat and uses a bullwhip". Or give it an input image to work with.
From just adding a couple words, ChatGPT produces an entirely new character. It's so easy to get it to produce novel images. It is so easy in fact, that it makes a lot of posts like this one feel like strawmen, intentionally providing so little information to the LLMs that the generic character is the only obvious output that you would expect.
Now, would it be better if it didn't default to these common movie tropes? Sure. But the fact that it can follow these tropes doesn't mean that it cannot also be used to produce entirely new images filled with your imagination as well. You just have to actually ask it for that.
dcow
It strikes me that perhaps the prompts are not expansive or expressive enough. If you look at some of the prompts our new wave of prompt artists use to generate images in communities like midjourney, a single sentence doesn't cut it.
If AI is just compression, then decompressing a generic pop-culture-seeking prompt will yield a generic uninspired image.
lupusreal
I have given detailed descriptions of my own novel ideas to these image generators and they have faithfully implemented my ideas. I don't need the bot to be creative, I can do that myself. The bot is a paint brush. Give it to somebody who isn't creative and you won't get anything creative out of it. That isn't the tool's fault, it's merely an inadequacy of the user.
littlecranky67
But I can hire an artist and ask him to draw me a picture of Indiana Jones, he creates a perfect copy and I hang it on my fridge. Where did I (or the artist) violate any copyright (or other) laws? It is the artist that is replaced by the AI, not the copyrighted IP.
rdtsc
> But I can hire an artist and ask him to draw me a picture of Indiana Jones,
Sure, assuming the artist has the proper license and franchise rights to make and distribute copies. You can go buy a picture of Indy today that may not be printed by Walt Disney Studios but by some other outfit or artists.
Or, you mean if the artist doesn't have a license to produce and distribute Indiana Jones images? Well they'll be in trouble legally. They are making "copies" of things they don't own and profiting from it.
Another question is whether that's practically enforceable.
> Where did I (or the artist) violate any copyright (or other) laws?
When they took payment and profited from making unauthorized copies.
> It is the artist that is replaced by the AI, not the copyrighted IP.
Exactly, that's why LLMs and the companies which create them are called "theft machines" -- they are reproducing copyrighted material. Especially the ones charging for "tokens". You pay them, they make money and produce unauthorized copies. Show that picture of Indy to a jury and I think it's a good chance of convincing them.
I am not saying this is good or bad, I just see this having a legal "bite" so to speak, at least in my pedestrian view of copyright law.
saaaaaam
The likeness of Indiana Jones is not protected in any way - as far as I know - that would stop a human artist creating, rendering and selling a work of art representing their creative vision of Indiana Jones. And even more so in a private context. Even if the likeness is protected (“archaeologist, adventurer, whip, hat”) then this protection would only be in certain jurisdictions and that protection is more akin to a design right where the likeness would need to be articulated AND registered. Many jurisdictions don’t require copyright registration and do not offer that sort of technical likeness registration.
If they traced a photo they might be violating the copyright of the photographer.
But if they are drawing an archaeologist adventurer with a whip and a hat based on their consumption and memory of Indiana Jones imagery there is very little anyone could do.
If that image was then printed on an industrial scale or printed onto t-shirt there is a (albeit somewhat theoretical) chance that in some jurisdictions sale of those products may be able to be restricted based on rights to the likeness. But that would be a stretch.
planb
> Or, you mean if the artist doesn't have a license to produce and distribute Indiana Jones images? Well they'll be in trouble legally. They are making "copies" of things they don't own and profiting from it.
Ok, my sister can draw, and she gifts me an image of my favorite Marvel hero she painted to hang on my wall. Should that be illegal?
alphan0n
That’s not how copyright law works.
Commissioned work is owned by the commissioner unless otherwise agreed upon by contract.
So long as the work is not distributed, exhibited, performed, etc, as in the example of keeping the artwork on their refrigerator in their home, then no infringement has taken place.
Velorivox
That does infringe copyright...you're just unlikely to get in trouble for it. You might get a cease and desist if the owner of the IP finds out and can spare a moment for you.
RajT88
Totally agree. LLM's are just automating that infringement process.
If you make money off it, it's no longer fair use; it's infringement. Even if you don't make money off it, it's not automatically fair use.
My own favorite crazy story about copyright violations:
Metallica sued Green Jello for parodying Enter Sandman (including a lyric where it says "It's not Metallica"):
https://en.wikipedia.org/wiki/Electric_Harley_House_(of_Love...
They lost that case. The kicker? Metallica were guest vocalists on that album.
ryandrake
This doesn't make any sense to me. No media is getting copied, unless the drawing is exactly the same as an existing drawing. Shouldn't "copy"right apply to specific, tangible artistic works? I guess I don't understand how the fantasy of "IP" works.
What if the drawing is of Indiana Jones but he's carrying a bow and arrow instead of a whip? Is it infringement?
What if it's a really bad drawing of Indiana Jones, so bad that you can't really tell that it's the character? Is that infringement?
What if the drawing is of Indiana Jones, but in the style of abstract expressionism, so doesn't even contain a human shape? Is it infringement?
What if it's a good drawing that looks very much like Indiana Jones, but it's not! The person's name is actually Iowa Jim. Is that infringement?
What if it's just an image of an archeologist adventurer who wears a hat and uses a bullwhip, but otherwise doesn't look anything like Indiana Jones? Is it infringement?
piyh
Presumably the artist is a human who directly or indirectly paid money to view a film containing an archaeologist with the whip.
I don't think this is about reproduction as much as how you got enough data for that reproduction. The riaa sent people to jail and ruined their lives for pirating. Now these companies are doing it and being valued for hundreds of billions of dollars.
RataNova
You're right, it's not just about reproduction, it's about how the data was collected
the_af
Plus the scale of it all.
A human friend can get tired and there's so many request he/she can fulfill and at a max rate. Even a team of human artists have a relatively low limit.
But Gen AI has very high limits and speeds, and it never gets tired. It seems unfair to me.
layer8
The artist is violating copyright by selling you that picture. You can’t just start creating and distributing pictures of a copyrighted property. You need a license from the copyright holder.
You also can’t sell a machine that outputs such material. And that’s how the story with GenAI becomes problematic. If GenAI can create the next Indiana Jones or Star Wars sequel for you (possibly a better one than Disney makes, it has become a low bar of sorts), I think the issue becomes obvious.
mcmcmc
Depends on if you paid him or not. If he sold it to you, then he is infringing on Disney’s IP and depriving them of that revenue.
eb0la
If you're paying for tokens used to generate that...
null
the_af
I think framing this as "IP theft" is a mistake.
Nobody can prevent you from drawing a photo realistic picture of Indy, or taking a photo of him from the internet and hanging it on your fridge. Or asking a friend to do it for you. And let's be honest -- because nobody is looking -- said friend could even charge you a modest sum to draw a realistic picture of Indy for you to hang on your fridge; yes, it's "illegal" but nobody is looking for this kind of small potatos infringement.
I think the problem is when people start making a business out of this. A game developer could think "hey, I can make a game with artwork that looks just like Ghibli!", where before he wouldn't have anyone with the skills or patience to do this (see: the 4-second scene that took a year to make), now he can just ask the Gen AI to make it for them.
Is it "copyright infringement"? I dunno. Hard to tell, to be honest. But from an ethical point of view, it seems odd. And before you actually required someone to take the time and effort to copy the source material, now it's an automated and scalable process that does this, and can do this and much more, faster and without getting tired. "Theft at scale", maybe not so small potatos anymore.
--
edit: nice, downvotes. And in the other thread people were arguing HN is such a nice place for dissenting opinions.
eb0la
I believe when we talk about this there's a big misunderstanding between Copyright, Trademarks, and Fair use.
Indy, with its logo, whiplash, and hat, is a trademark from Disney. I don't know the specific stuff; but if you sell a t-shirt with Indiana Jones, or you put the logo there... you might be sued due to trademark violation.
If you make copies of anything developed, sold, or licensed by Disney (movies, comics, books, etc) you'll have a copyright violation.
The issue we have with AI and LLM is that: - The models compress information and can make a lot of copies of it very cheaply. - Artist wages are quite low. Higher that what you'd pay OpenAI, but not enough to make a living even unless you're hired by a big company (like Marvel or DC) and they give you regular work ($100-120 for a cover, $50-80/page interior work. One page needs about one day to draw.) - AI used a lot of images from the internet to train models. Most of them were pirated. - And, of course, it is replacing low-paying jobs for artist.
Also, do not forget it might make verbatim copies of copyrighted art if the model just memorized the picture / text.
kridsdale3
I would argue that countless games have already been made by top tier professional artists that IP-Steal the Ghibli theme.
Breath of the Wild, and Tears of the Kingdom should be included there.
airstrike
Can we not call it "theft"? It's such a loaded term and doesn't really mean the same thing when we're talking about bits and bytes.
moolcool
OK, but then we need a common standard. If Facebook is allowed to use libgen, I should also be allowed.
whywhywhywhy
How is it different from fan art, which is legal.
Pet_Ant
> Obviously a horrible hideous theft machine.
I hate how it is common to advance a position to just state a conclusion as if it were a fact. You keep repeating the same thing over and over until it seems like a concensus has been reached instead of an actual argument reasoned from first principle.
This is no theft here. Any copyright would be flimsier than software patents. I love Studio Ghibli (including $500/seat festival tickets) but it's the heart and the detail that make them what they are. You cannot clone that. Just some surface similarity. If that's all you like about the movies... you really missed the point.
Imagine if in early cinema someone had tried to claim mustachioed villian, ditsy blonde, or dumb jock? These are just tropes and styles. Quality work goes much much much deeper, and that cannot be synthesised. I can AI generate a million engagement rings, but I cannot pick the perfect one that fits you and your partners love story.
PS- the best work they did was "When Marnie was There". Just fitted together perfectly.
dvsfish
The engagement ring is a good example object, but I feel it serves the opposite argument better.
If engagement rings were as ubiquitous and easy to generate as Ghibli images have become, they would lose their value very quickly -- not even just in the monetary sense, but the sentimental value across the market would crash for this particular trinket. It wouldn't be about picking the right one anymore, it would be finding some other thing that better conveys status or love through scarcity.
If you have a 3d printer you'd know this feeling where abundance diminishes the value of something directly. Any pure plastic items you have are reduced to junk very quickly once you know you can basically have anything on a whim (exceptions for things with utility, however these are still printable). If I could print 30 rings a day, my partner wouldn't want any of them as a show of my undying love. Something more special and rare and thoughtful would have to take its place.
This isn't meant to come across as shallow in any way, its just classic supply and demand relating to non monetary value.
planb
>If I could print 30 rings a day, my partner wouldn't want any of them as a show of my undying love. Something more special and rare and thoughtful would have to take its place.
And now I think this serves the opposite argument better. Downloading some random ring from the internet would not show your undying love. Designing a custom ring just for your partner, even if it is made from plastic, and even if you use AI as a tool in the process, is where the value is generated.
asadotzler
>it's the heart and the detail that make them what they are. You cannot clone that
You absolutely can and these theft machines are proving that, literally cloning those details with very high precision and fidelity.
RataNova
I don't think AI is doomed to be uncreative but it definitely needs human weirdness and unpredictability to steer it
shadowgovt
> Obviously a horrible hideous theft machine.
I mean... If I go to Google right now and do an image search for "archeologist adventurer who wears a hat and uses a bullwhip," the first picture is a not-even-changed image of Indiana Jones. Which I will then copy and paste into whatever project I'm working on without clicking through to the source page (usually because the source page is an ad-ridden mess).
Perhaps the Internet itself is the hideous theft machine, and AI is just the most efficient permutation of user interface onto it.
(Incidentally, if you do that search, you will also, hilariously, turn up images of an older gentleman dressed in a brown coat and hat who is clearly meant to be "The Indiana Jones you got on Wish" from a photo-licensing site. The entire exercise of trying to extract wealth via exclusive access to memetic constructs is a fraught one).
WhyOhWhyQ
Your position cannot distinguish stealing somebody's likeness and looking at them.
rurp
The key difference is that the google example is clearly copying someone elses work and there are plenty of laws and norms that non-billionaires need to follow. If you made a business reselling the image you copied you would expect to get in trouble and have to stop. But AI companies are doing essentially the same thing in many cases and being rewarded for it.
The hypocrisy is much of the problem. If we're going to have IP laws that severely punish people and smaller companies for reselling the creative works of others without any compensation or permission then those rules should apply to powerful well-connected companies as well.
m3kw9
Do google pay anyone when I use image search and the results are straight from their website?
AkBKukU
This was decided in court[1] over two decades as acceptable fair-use and that thumbnail images do not constitute a copyright violation.
[1] https://scholar.google.com/scholar_case?case=137674209419772...
neomantra
> Maybe Studio Ghibli making it through the seemingly deterministic GPT guardrails was an OpenAI slip up, a mistake,
The author is so generous... but Sam Altman literally has a Ghibli-fied Social profile and in response to all this said OpenAI chooses its demos very carefully. His primary concern is that Ghibli-fying prompts are over-consuming their GPU resources, degrading the service by preventing other ChatGPT tasks.
flessner
Everyone is talking about theft - I get it, but there's a more subtler point being made here.
Current generation of AI models can't think of anything truly new. Everything is simply a blend of prior work. I am not saying that this doesn't have economic value, but it means these AI models are closer to lossy compression algorithms than they are to AGI.
The following quote by Sam Altman from about 5 years ago is interesting.
"We have made a soft promise to investors that once we build this sort-of generally intelligent system, basically we will ask it to figure out a way to generate an investment return."
That's a statement I wouldn't even dream about making today.
nearbuy
> Current generation of AI models can't think of anything truly new.
How could you possibly know this?
Is this falsifiable? Is there anything we could ask it to draw where you wouldn't just claim it must be copying some image in its training data?
mjburgess
Novelty in one medium arises from novelty in others, shifts to the external environment.
We got brass bands with brass instruments, synth music from synths.
We know therefore, necessarily, that they can be nothing novel from an LLM -- it has no live access to novel developments in the broader environment. If synths were invented after its training, it could never produce synth music (and so on).
The claim here is trivially falsifiable, and so obviously so that credulous fans of this technology bake it in to their misunderstanding of novelty itself: have an LLM produce content on developments which had yet to take place at the time of its training. It obviously cannot do this.
Yet an artist which paints with a new kind of black pigment can, trivially so.
jedimastert
The problem with generating genuinely new art is it requires "inputs" that aren't art. It's requires life experiences.
Davidzheng
I beseech you, in the bowels of Christ, think it possible that you may be mistaken.
kubanczyk
Oliver Cromwell, a letter to the General Assembly of the Church of Scotland, 3 August 1650
bbor
Disregarding the (common!) assumption that AGI will consist of one monolithic LLM instead of dozens of specialized ones, I think your comment fails to invoke an accurate, consistent picture of creativity/"truly new" cognition.
To borrow Chomsky's framework: what makes humans unique and special is our ability to produce an infinite range of outputs that nonetheless conform to a set of linguistic rules. When viewed in this light, human creativity necessarily depends on the "linguistic rules" part of that; without a framework of meaning to work within, we would just be generating entropy, not meaningful expressions.
Obviously this applies most directly to external language, but I hope it's clear how it indirectly applies to internal cognition and--as we're discussing here--visual art.
TL;DR: LLMs are definitely creative, otherwise they wouldn't be able to produce semantically-meaningful, context-appropriate language in the first place. For a more empirical argument, just ask yourself how a machine that can generate a poem or illustration depicting [CHARACTER_X] in [PLACE_Y] doing [ACTIVITY_Z] in [STYLE_S] without being creative!
[1] Covered in the famous Chomsky v. Foucault debate, for the curious: https://www.youtube.com/watch?v=3wfNl2L0Gf8
burnished
Oooh those guardrails make me angry. I get why they are there (dont poke the bear) but it doesn't make me overlook the self serving hypocrisy involved.
Though I am also generally opposed to the notion of intellectual property whatsoever on the basis that it doesn't seem to serve its intended purpose and what good could be salvaged from its various systems can already be well represented with other existing legal concepts, i.e deceptive behaviors being prosecuted as forms of fraud.
teddyh
The problem is people at large companies creating these AI models, wanting the freedom to copy artists’ works when using it, but these large companies also want to keep copyright protection intact, for their regular business activities. They want to eat the cake and have it too. And they are arguing for essentially eliminating copyright for their specific purpose and convenience, when copyright has virtually never been loosened for the public’s convenience, even when the exceptions the public asks for are often minor and laudable. If these companies were to argue that copyright should be eliminated because of this new technology, I might not object. But now that they come and ask… no, they pretend to already have, a copyright exception for their specific use, I will happily turn around and use their own copyright maximalist arguments against them.
(Copied from a comment of mine written more than three years ago: <https://news.ycombinator.com/item?id=33582047>)
ToValueFunfetti
I don't care for this line of argument. It's like saying you can't hold a position that trespassing should be illegal while also holding that commercial businesses should be legally required to have public restrooms. Yes, both of these positions are related to land rights and the former is pro- while the latter is anti-, but it's a perfectly coherent set of positions. OpenAI can absolutely be anti-copyright in the sense of whether you can train an an NN on copyrighted data and pro-copyright in the sense of whether you can make an exact replica of some data and sell it as your own without making it into hypocrisy territory. It does suggest they're self-interested, but you have to climb a mountain in Tibet to find anybody who isn't.
Arguments that make a case that NN training is copyright violation are much more compelling to me than this.
belorn
The example you gave with public restroom do not work because of two main concept: They are usually getting paid for it by the government, and operating a company usually holds benefits given by the government. Industry regulations as a concept is generally justified in that industry are getting "something" from society, and thus society can put in requirements in return.
A regulation that require restaurants to have a public bathroom is more akin to regulation that also require restaurants to check id when selling alcohol to young customers. Neither requirement has any relation with land rights, but is related to the right of operating a company that sell food to the public.
TremendousJudge
No, the exception they are asking for (we can train on copyrighted material and the image produced is non-copyright infringing) is copyright infringing in the most basic sense.
I'll prove it by induction: Imagine that I have a service where I "train" a model on a single image of Indiana Jones. Now you prompt it, and my model "generates" the same image. I sell you this service, and no money goes to the copyright holder of the original image. This is obviously infringment.
There's no reason why training on a billion images is any different, besides the fact that the lines are blurred by the model weights not being parseable
jofla_net
I guess the best explanation for what we're witnessing is the notion that 'Money Talks', and sadly nothing more. To think thats all that fair use activists lacked in years passed..
theshrike79
It's not just the guardrails, but the ham-fisted implementation.
Grok is supposed to be "uncensored", but there are very specific words you just can't use when asking it to generate images. It'll just flat out refuse or give an error message during generation.
But, again, if you go in a roundabout way and avoid the specific terms you can still get what you want. So why bother?
Is it about not wanting bad PR or avoiding litigation?
mrweasel
The implementation is what gets to me too. Fair enough that a company doesn't want their LLM used in a certain way. That's their choice, even if it's just to avoid getting sued.
How they then go about implementing those guardrails is pretty telling about their understand and control over what they've build and their line of thinking. Clearly, at no point before releasing their LLMs onto the world did anyone stop and ask: Hey, how do we deal with these things generating unwanted content?
Resorting to blocking certain terms in the prompts is like searching for keywords in spam emails. "Hey Jim, I got another spam email from that Chinese tire place" - "No worry boss, I've configured the mail server to just delete any email containing the words China or tire".
Some journalist should go to a few of these AI companies and start asking questions about the long term effectiveness and viability of just blocking keywords in prompts.
coderenegade
I don't see why this is an issue? The prompts imply obvious and well-known characters, and don't make it clear that they want an original answer. Most humans would probably give you similar answers if you didn't add an additional qualifier like "not Indiana Jones". The only difference is that a human can't exactly reproduce the likeness of a famous character without significant time and effort.
The real issue here is that there's a whole host of implied context in human languages. On the one hand, we expect the machine to not spit out copyrighted or trademarked material, but on the other hand, there's a whole lot of cultural context and implied context that gets baked into these things during training.
dgunay
I think the point is that for a lot of them there are endless possible alternatives to the character design, but it still generates one with the exact same design. Why can't, for example, the image of Tomb Raider have a different colored tank top? Why is she wearing a tank top and not a shirt? Why does she have to have a gun? Why is she a busty, attractive brunette? These are all things that could be different but the dominance of Lara Croft's image and strong association with the words "tomb raider" in popular culture clearly influences the model's output.
coderenegade
Because it's not clear that that's what you want. What's the context? Are we playing a game where I guess a character? Is it a design session for a new character based on a well known one, maybe a sidekick? Is it a new take on an old character? Are you just trying to remember what a well-known character looks like, and giving a brief prompt?
It's not clear what the asker wants, and the obvious answer is probably the culturally relevant one. Hell, I'd give you the same answers as the AI did here if I had the ability to spit out perfect replicas.
echoangle
And how is that bad or surprising? It’s actually what I would expect from how AI works.
SV_BubbleTime
Exactly. We designed systems that work on attention and inference… and then surprised that it returns popular results?
mvieira38
It's an IP theft machine. Humans wouldn't be allowed to publish these pictures for profit, but OpenAI is allowed to "generate" them?
victorbjorklund
I would 100% be allowed to draw an image of Indiana Jones in illustrator. There is no law against me drawing his likeness.
pier25
No, you aren't allowed to monetize an image of Indiana Jones even if you made it yourself.
otabdeveloper4
You 100% wouldn't be allowed to sell your Indiana Jones drawing services.
echoangle
But would you be allowed to publish it in the same way the AI companies do?
why_at
I'm honestly trying to wrap my head around the law here because copyright is often very confusing.
If I ask an artist to draw me a picture of Indiana Jones and they do it would that be copyright infringement? Even if it's just for my personal use?
bawolff
Probably that would be a derrivative work. Which means the original owner would have some copyright in it.
It may or may not be fair use, which is a complicated question (ianal).
Avicebron
IANAL, but if OpenAI makes any money/commercial gains from producing a Ghibli-esque image when you ask, say you pay a subscription to OpenAI. What percentage of that subscription is owed to Ghibli for running Ghibli art through OpenAI's gristmill and providing the ability to create that image with that "vibe/style" etc. How long into perpetuity is OpenAI allowed to re-use that original art whenever their model produces said similar image. That seems to be the question.
xboxnolifes
I would think yes. Consider the alternate variation where the artist proactively draws Indiana Jones, in all his likeness, and attempts to market and sell it. The same exchange is ultimately happening, but this clearly is copyright infringement.
pwarner
To me a lot has to do with what a human does with them one the tool generates them no?
Smithalicious
Won't somebody think of the billionaire IP holders? The horror.
asadotzler
And the small up and coming artists whose work is also stolen, AI-washed, and sold to consumers for a monthly fee, destroying the market for those up and coming artists to sell original works. You don't get to pretend this is only going to hurt big players when there are already small players whose livelihoods have been ruined.
divCraftsman
[dead]
jmull
Normally (well, if you're ethical) credit is given.
Also, there are IP limits of various sorts (e.g. copyright, trademark) for various purposes (some arguably good, some arguably bad), and some freedoms (e.g., fair use). There's no issue if this follows the rules... but I don't see where that's implemented here.
It looks like they may be selling IP they don't own the right to.
runarberg
Overfitting is generally a sign of a useless model
mlsu
I was really hoping that the conversation around AI art would at least be partially centered on the perhaps now dated "2008 pirate party" idea that intellectual property, the royalty system, the draconian copyright laws that we have today are deeply silly, rooted in a fiction, and used over and over again, primarily by the rich and powerful, to stifle original ideas and hold back cultural innovation.
Unfortunately, it's just the opposite. It seems most people have fully assimilated the idea that information itself must be entirely subsumed into an oppressive, proprietary, commercial apparatus. That Disney Corp can prevent you from viewing some collection of pixels, because THEY own it, and they know better than you do about the culture and communication that you are and are not allowed to experience.
It's just baffling. If they could, Disney would scan your brain to charge you a nickel every time you thought of Mickey Mouse.
kokanee
The idea of open sourcing everything and nullifying patents would benefit corporations like Disney and OpenAI vastly more than it would benefit the people. The first thing that would happen is that BigCorp would eat up every interesting or useful piece of art, technology, and culture that has ever been created and monetize the life out of it.
These legal protections are needed by the people. To the Pirate Party's credit, undoing corporate personhood would be a good first step, so that we can focus on enforcing protections for the works of humans. Still, attributing those works to CEOs instead of corporations wouldn't result in much change.
pixl97
>The first thing that would happen is that BigCorp would eat up every interesting or useful piece of art, technology, and culture that has ever been created and monetize the life out of it.
Wait, I'm still trying to figure out the difference between your imaginary world and the world we live in now?
Lerc
I think the main difference is if everything were freely available they may attempt to monetize the life out of it, but they will fail if they can't actually provide something people actually want. There's no more "You want a thing so you're going to buy our thing because we are the exclusive providers of it. That means we don't even have to make it very good"
If anyone in the world could make a Star Wars movie, the average Star Wars movie would be much worse, but the best 10 Star Wars movies might be better that what we currently have.
dragontamer
Thor would have red hair in the imaginary world, rather than being a Blonde man which was made to be a somewhat distinguished comic book character.
The Disney or otherwise copyrighted versions allow for unique spins on these old characters to be re-copyrighted. This Thor from Disney/Marvel is distinguished from Thor from God of War.
dcow
How do restaurants work, then? You can’t copyright a recipe. Instructions can’t generally be copyrighted, otherwise someone would own the fastest route from A to B and charge every person who used it. The whole idea of intellectual property gets really weird when you try to pinpoint what exactly is being owned.
I do not agree with your conjecture that big corps would win by default. Ask why would people need protection from having their work stolen when the only ones welding weaponized copyright are the corporations. People need the freedom to wield culture without restriction, not protection from someone having the same idea as them and manifesting it.
singleshot_
It’s more reasonable to say that the idea of intellectual property is challenging for nonlawyers because of the difficulty in understanding ownership not as one thing, but as a bundle of various elements of control, exclusion, obligation, or entitlement, even some of which spring into existence out of nowhere.
In other words, the challenge is not to understand “what exactly is being owned,” and instead, to understand “what exactly being owned is.”
hammock
> How do restaurants work, then? You can’t copyright a recipe.
They barely work. Recipes are trade secrets, and the cooks who use them are either paid very well, given NDAs or given only part of the most guarded recipes
ipsento606
> How do restaurants work, then?
Primarily because recipe creation is not one of the biggest cost centers for restaurants?
api
A restaurant is a small manufacturing facility that produces a physical product. It’s not the same at all.
awesome_dude
Closed source - when was the last time your restaurant told you what was in, and how to make, your favourite dish?
What's in Coca Cola?
What are the 11 herbs and spices in Kentucky Fried Chicken?
How do I make the sauce in a Big Mac?
dragonwriter
> The idea of open sourcing everything and nullifying patents would benefit corporations like Disney and OpenAI vastly more than it would benefit the people.
Disney would be among the largest beneficiaries if the right to train AI models on content was viewed as an exclusive right of the copyright holder; they absolutely do not benefit from AI training being considered fair use.
julianeon
Maybe now, post-AI.
But if you'd asked this question in 2015 or earlier, everyone would have said Disney -> pro-patent, average people & indie devs -> anti-patent. Microsoft was famously pro-patent, as were a number of nuisance companies that earned the label "patent troll."
Honestly, this idea of "patents to protect the people" would've come across as a corporate lawyer trick pre-2015.
csallen
This is the exact opposite of the truth.
Look at YouTube. Look at SoundCloud. Look at all the fan fiction sites out there, internet mangas and manwhas and webtoons, all the podcasts, all the influencers on X and Instagram and TikTok and even OnlyFans, etc etc. Look at all the uniquely tiny distribution channels that small companies and even individuals are able to build in connection with their fans and customers.
There is endless demand for the endless variety of creativity and content that's created by normal people who aren't Disney, and endless ways to get it into people's hands. It is literally impossible for any one company to hoover all of it up and somehow keep it from the people.
In fact, the ONLY thing that makes it possible for them to come close to doing that is copyright.
And the only reason we have such a huge variety of creativity online is because people either (a) blatantly violate copyright law, or (b) work around gaps in copyright law that allow them to be creative without being sued.
The idea that we need copyrights to protect us from big companies is exactly wrong. It's the opposite. Big companies need copyright to protect their profits from the endless creativity and remixing of the people.
echelon
The original claim is false,
> intellectual property [...] used over and over again, primarily by the rich and powerful, to stifle original ideas and hold back cultural innovation.
There's nothing about IP which prevents you from creating your own. There are, in fact, a near infinite number of things you can create. More things than there exist stars in our galaxy.
The problem with ideas is that they have to be good. They have to be refined. They have to hit the cultural zeitgeist, solve a particular problem, or just be useful. That's the hard part that takes the investment of time and money.
In the old world before Gen AI, this was the hard thing that kept companies in power. That world is going away fast, and now creation will be (relatively) easy. More taste makers will be slinging content and we'll wind up in a land of abundance. We won't need Disney to give us their opinion on Star Wars - we can make our own.
The new problem is distributing that content.
> The idea of open sourcing everything and nullifying patents would benefit corporations like Disney and OpenAI vastly more than it would benefit the people. The first thing that would happen is that BigCorp would eat up every interesting or useful piece of art, technology, and culture that has ever been created and monetize the life out of it.
Unless the masses can create and share on equal footing, you're 100% right.
If it turns out, however, that we don't need Google, OpenAI, or big tech to make our own sci-fi epics, share them with a ton of people, and interact with friends and audiences, then the corporations won't be able to profit off of it.
If social networks were replaced with common carriers and protocols.
If Gen AI could run at the edge without proprietary models or expensive compute.
If the data of YouTube, Reddit, Twitter, Instagram didn't require hyperscaler infra to store, search, and serve.
Unfortunately, there are too many technical reasons why the giants will win. And network effects will favor the few versus many. Unless those parameters change, we'll be stuck with big tech distribution.
Even if the laws around IP change, the hard tech challenges keep the gatekeepers in power. The power accrues to those who can dominate creation (if creation is unilateral), or even more so, to the distributors of that content.
codedokode
> The problem with ideas is that they have to be good.
No they don't, look at music popular in social networks.
> and now creation will be (relatively) easy. More taste makers will be slinging content and we'll wind up in a land of abundance.
Even before the generative AI, I think we live in the era where there are more creators than ever in history: everybody today can publish their music or art without any large investments (except for instruments: they are expensive as always). I would prefer we have cheaper pianos, samples and microphones instead of worthless music-copying models.
dcow
> We won't need Disney to give us their opinion on Star Wars - we can make our own.
Disney would say that you can’t. And in the current copyright regime, it’s not unlikely that they’d convince the court that they’re right.
api
This is the same argument we made in the 90s about what the web was going to do. What ended up happening was the growth of aggregators and silos like Facebook that baited everyone with ease of use into putting everything into their walled garden and then monetized it. The creators, namely the posters of the content, got nothing.
The same is happening already with AI creations. Doing it yourself is work and takes some technical skill, so most people use hosted AI services. Guess who makes all the money?
You will be able to create and share your own spin on Star Wars. You won’t see anything for that except maybe cred or some upvotes. The company that hosts it and provides the gateway and controls the algorithms that show it to people will get everything.
eaglelamp
If we are going to have a general discussion about copyright reform at a national level, I'm all for it. If we are going to let billion dollar corporations break the law to make even more money and invent legal fictions after the fact to protect them, I'm completely against it.
Training a model is not equivalent to training a human. Freedom of information for a mountain of graphics cards in a privately owned data center is not the same as freedom of information for flesh and blood human beings.
r3trohack3r
You’re setting court precedent that will apply equally to OpenAI as it does to the llama.cpp and stable diffusion models running on your own graphics card.
munificent
SGTM.
Honestly, seriously. Imagine some weird Thanos showed up, snapped his fingers and every single bit of generative AI software/models/papers/etc. were wiped from the Earth forever.
Would that world be measurably worse in any way in terms of meaningful satisfying lives for people? Yes, you might have to hand draw (poorly) your D&D character.
But if you wanted to read a story, or look at an image, you'd have to actually connect with a human who made that thing. That human would in turn have an audience for people to experience the thing they made.
Was that world so bad?
photonthug
I don’t know about that, we seem to be so deeply into double standards for this stuff that we’ve forgotten they are double standards. If I aggressively scrape content from anywhere and everywhere ignoring robots.txt and any other terms and conditions, then I’ll probably be punished. Corporate crawlers that are feeding the beast just do this on a massive scale and laugh off all of the complaints, including those from smaller corporations who hire lawyers..
codedokode
Can stable diffusion be created without using copyrighted content? Maybe we should have some exemption for non-commercial research but definitely not for commercial exploitation or generating copyrighted images using open-source models.
robocat
> invent legal fictions after the fact
You're reading into the situation...
For the US getting legislators to do anything is impossible: even the powerful fail.
When a legal system is totally roadblocked, what other choice is there? The reason all startups ask forgiveness is that permission is not available.
(edit). Shit. I guess that could be a political statement. Sorry
tastyface
A different way of looking at it: AI, by design, defaults to regurgitating the poppiest of pop culture content. Every whip-wielding archaeologist is now Harrison Ford. Every suave British spy is now Daniel Craig. With the power of AI, creativity is dead and buried.
slg
This is what was often missed in the previous round of AI discourse that criticized these companies for forcing diversity into their systems after the fact. Every suave spy being Daniel Craig is just the apolitical version of every nurse being a woman or every criminal being Black. Converging everything to the internet's most popular result represents an inaccurate and a dumped down version of the world. You don't have to value diversity as a concept at all to recognize this as a systemic flaw of AI, it is as easy as recognizing that Daniel Craig isn't the only James Bond let alone the only "suave English spy".
dcow
It’s only a flaw insofar as it’s used in ways in which the property of the tool is problematic. Stereotypes are use for good and bad all the time, let’s not pretend that we have to attack every problem with a funky shaped hammer because we can’t admit that it’s okay to have specialized tools in the tool belt.
card_zero
The backlash against AI compels creative types to be more original, maybe. It could be that AI improves culture by reflecting it in insipid parody, with the implicit message "stop phoning it in".
darioush
don't you think it is empowering and aspiring for artists? they can try several drafts of their work instantaneously, checking out various compositions etc before even starting the manual art process.
they could even input/train it on their own work. I don't think someone can use AI to copy your art better than the original artist.
Plus art is about provenance. If we could find a scrap piece of paper with some scribbles from Picasso, it would be art.
Kim_Bruning
This does seem to work for writing. Feed your own writing back in and try variations / quickly sketch out alternate plots, that sort of thing.
Then go back and refine.
Treat it the same as programming. Don't tell the AI to just make something and hope it magically does it as a one-shot. Iterate, combine with other techniques, make something that is truly your own.
autoexec
> A different way of looking at it: AI, by design, defaults to regurgitating the poppiest of pop culture content.
That's the whole problem with AI. It's not creative. There's no "I" in AI. There's just what we feed it and it's a whole lot of "garbage in, garbage out". The more the world is flooded with derivative AI slop the less there will be of anything else to train AI on and eventually we're left with increasingly homogenized and uncreative content drowning out what little originality is still being made without AI.
xorcist
I think what you observe is more like a natural blowback to the prevailing idea that this is somehow beyond critique because it will fundamentally change culture and civilization forever.
There's a bit of irony here too. The intellectual discourse around intellectural property, a diverse and lively one from an academic standpoint, the whole free and open source software movements, software patents, the piracy movement and so on have analyzed the history, underlying ideas and values in detail for the past thirty years. Most people know roughly what is at stake, where they stand, and can defend their position in an honest way.
Then comes new technology, everyone and their mother gets excited about it, and steamrolls all those lofty ideas into "oh look at all the shiny things it can produce!". Be careful what you wish for.
achierius
Let's be clear. You can be for free software, against copyright, etc., and STILL be in favor of these firms being punished for violating copyright as they have. Because frankly, we -- normal people -- have always known that we would be punished if we did anything close to this: so many people have been thrown in jail, even killed themselves, because they distributed some film or hosted some books. But now, when a big corporation does it, and in doing so seeks to replace and impoverish thousands, millions of hard-working, law-abiding people, now is when we should expect the government to finally say -- oh, that copyright thing was silly all along? No. Perhaps if the deal was that the whole system would go away entirely -- that we, too, could do what these firms have done. But that's not what's being proposed. That will not happen. They want the laws to be for them, not for us, and I will always be opposed to attempts at actualizing that injustice.
FeepingCreature
IMO the natural effect of this will be to massively devalue any individual cultural artifact, and that this will also achieve the benefit of defanging the big copyright holders. Is it the right way to go about it? No. Is it an insult to anyone who ever got nabbed for piracy? Sure. But tbh as a pirate voter I'll still very much take it.
jason_pomerleau
> If they could, Disney would scan your brain to charge you a nickel every time you thought of Mickey Mouse.
This reminds me of Tom Scott’s “Welcome to Life: The Singularity, Ruined by Lawyers” [1]
II2II
> That Disney Corp can prevent you from viewing some collection of pixels, because THEY own it
A world without copyright is just as problematic as a world with copyright. With copyright, you run into the problem of excessive control. This wasn't too much of a problem in the past. If you bought a book, record, or video recording, you owned that particular copy. You could run into disagreeable situations because you didn't own the rights, but it was difficult to prevent anyone from from viewing a work once it had been published. (Of course, modern copyright laws and digitial distribution has changed that.)
On the flip side, without copyright, it would be far easier for others to exploit (or even take credit) for the work of another person without compensation or recourse. Just look at those AI "generated" images, or any website that blatently rips off the content created by another person. There is no compensation. Heck, there isn't even credit. Worse yet, the parties misrepresenting the content are doing their best to monetize it. Even authors who are more than willing to give their work away have every right to feel exploited under those circumstances. And all of that is happening with copyright laws, where there is the opportunity for recourse if you have the means and the will.
singpolyma3
To reply to the parenthetical, copyright has nothing to do with credit. Taking credit for someone else's work is banned in some places in some contexts (they call this a moral rights regime) but not the same thing as what is being talked about when people say copyright (which is about copying and performing)
dcow
You don’t need credit to talk about pop culture. I don’t need to credit the Indian Jones copyright holder when I paint a stunning likeness of Ford in a kaki outfit with a whip, even if the holder might try to sue me over it. Copyright and credit aren’t the same.
paulryanrogers
There are also trademark protections. I heard Ford actually trademarked his likeness to ensure he got a piece of the merchandise action.
idiotsecant
The idea that someone can't use ideas without someone else making money from it is a really, really, radically weird idea and is very new in the history of human society.
ryandrake
Not just some particular collection of pixels, but an infinite number of combinations of collections of pixels, any of which remotely invoke a shadow of similarity to hundreds of "properties" that Disney lays claim to.
codedokode
But why do you want to make a collection of pixels that resembles existing characters and not create your own?
chimpanzee
Essentially: “information wants to be free”.
I agree.
But this must include the dissolution of patents. Otherwise corporations and the owners of the infrastructure will simply control everything, including the easily replicable works of individuals.
j-bos
At least patents only last 20 years as opposed to nearly over a century for copyright.
paulryanrogers
In practice it's often longer. Drug companies queue up minor tweaks to their formulas and can threaten to sue anyone even close to the new way, even carbon copies of the now expired patent. Few can afford to win a lawsuit.
We need more courts and judges to speed the process, to make justice more accessible, and universal SLAPP protections to weed out frivolous abuse.
codedokode
I am against dissolution of patents if the technology took lot of research. In this case the patent protects from others copying the result of research.
However, obvious patents like "a computer system with a display displaying a product and a button to order it" should not be allowed. Also, software patents should not exist (copyright is enough).
wsintra2022
What if all that research led to some incredible world changing for the better idea/concept/product in an open society that would benefit everyone, in the closed society only those allowed to use the patent benefit
zparky
Here's what Ars got out of GPT-4 a year ago: https://arstechnica.com/tech-policy/2024/02/why-the-new-york...
dcow
Either, (1) LLMs are just super lossy compress/decompress machines and we humans find fascination in the loss that happens at decompression time, at times ascribing creativity and agency to it. Status quo copyright is a concern as we reduce the amount of lossiness, because at some point someone can claim that an output is close enough to the original to constitute infringement. AI companies should probably license all their training data until we sort the mess out.
Or, (2) LLMs are creative and do have agency, and feeding them bland prompts doesn't get their juices flowing. Copyright isn't a concern, the model just regurgitated a cheap likeness of Indiana Jones as Harrison Ford the world has seen ad nauseam. You'd probably do the same thing if someone prompted you the same way, you lazy energy conserving organism you.
In any case, perhaps the idea "cheap prompts yield cheap outputs" holds true. You're asking the model respond to the entirely uninspired phrase: "an image of an archeologist adventurer who wears a hat and uses a bullwhip". It's not surprising to me that the model outputs a generic pop-culture-shaped image that looks uncannily like the most iconic and popular rendition of the idea: Harrison Ford.
If you look at the type of prompts our new generation of prompt artists are using over in communities like Midjourney, a cheap generic sentence doesn't cut it.
sothatsit
You don't even need to add much more to the prompts. Just a few words, and it changes the characters you get. It won't always produce something good, but at least we have a lot of control over what it produces. Examples:
"An image of an Indian female archeologist adventurer who wears a hat and uses a bullwhip" (https://sora.com/g/gen_01jqzet1p8fjaa808bmqnvf7rk)
"An image of a fat Russian archeologist adventurer who wears a hat and uses a bullwhip" (https://sora.com/g/gen_01jqzfk727erer98a6yexafe70)
"An image of a skeletal archeologist adventurer who wears a hat and uses a bullwhip" (https://sora.com/g/gen_01jqzfnaz6fgqvgwqw8w4ntf6p)
Or, give ChatGPT a starting image. (https://sora.com/g/gen_01jqzf7vdweg4v5198aqfynjym)
And by further remixing the images ChatGPT produces, you can get your images to be even more unique. (https://sora.com/g/gen_01jqzfzmbze0wa310m42f8j5yw)
GrantMoyer
All four of those are dressed like Indiana Jones. They look like different versions of Indiana Jones you'd see in super hero multi-verse story.
sothatsit
So... ask it to dress them differently. You can just ask it to make whatever changes you want.
"An image of an archeologist adventurer who wears a hat and uses a bullwhip. He is wearing a top hat, a scarf, a knit jumper, and pink khaki pants. He is not wearing a bag" (https://sora.com/g/gen_01jqzkh4z2fqctzr9k1jsfnrhy)
Want to get rid of the pose? Add that the archeologist is "fun and joyous" to the prompt. (https://sora.com/g/gen_01jqzksmjgfppbv5p51hw0xrzn)
You have so much control, it is up to you to ask for something that is not a trope.
jcheng
Those are great, I would watch any one of those movies. Maybe even the "Across the Indiana-Verse" one where they are all pulled into a single dimension.
otabdeveloper4
Archeologists don't actually wear fedora hats.
And the stereotypical meme "archeologist hat" is the pith helmet.
sothatsit
Here, I asked ChatGPT to generate an image using a pith helmet for you: https://sora.com/g/gen_01jqzmab6hfxxtrt3atd0jgpg7
You can just ask for whatever changes you want.
snowwrestler
This is the opposite of how people have thought about creativity for centuries, though.
The most creative person is someone who generates original, compelling work with no prompting at all. A very creative person will give you something amazing and compelling from a very small prompt. A so-so creative person will require more specific direction to produce something good. All the way down to the new intern who need paragraphs of specs and multiple rounds of revision to produce something usable. Which is about where the multi-billion-dollar AI seems to be?
toddmorey
"Prompt artist" makes me sigh out loud
Not sure if anyone is interested in this story, but I remember at the height of the PokemonGo craze I noticed there were no shirts for the different factions in the game, cant rememebr what they were called but something like Teamread or something. I setup an online shop to just to sell a red shirt with the word on it. The next day my whole shop was taken offline for potential copyright infringement.
What I found surprising is I didnt even have one sale. Somehow someone had notified Nintendo AND my shop had been taken down, to sell merch that didn't even exist for the market and if I remember correctly - also it didnt even have any imagery on it or anything trademarkable - even if it was clearly meant for pokmeonGo fans.
Im not bitter I just found it interesting how quick and ruthless they were. Like bros I didn't even get a chance to make a sale. ( yes and also I dont think I infringed anything).