MicrostockGroup Sponsors


Author Topic: Should Adobe share the revenue earned by "fast lane" generative fill with us?  (Read 1532 times)

0 Members and 1 Guest are viewing this topic.

« on: October 22, 2023, 09:14 »
+4
Recently there was a livestream with Mat Hayward and Terry White (https://www.behance.net/videos/eb2dab18-ad9b-4535-a64e-127037055b8e/New-Creative-Cloud-features-for-Adobe-Stock-Contributors-with-Mat-Hayward-and-Terry-White) where it was revealed that customers have a certain number of "credits" they can spend on generative fill features. Namely, they called this "fast lane", which puts your request in a separate queue and you get your output in 10-15 seconds. Otherwise, you go into the other lane, which is slower, and which could take a lot longer to do the generative fill.

When customers used up their credits, they can buy new ones, and use it to put their requests in the "fast lane". Now - I might be biased, but it seems pretty obvious with me that a part of this revenue should be shared with us - the contributors who helped train the underlying model. Yes - we got that one-time payment to use our assets to create the model, but that is separate from continuously reaping the rewards of our work, and not sharing any revenue with us. So, I think a percfentage of the revenue should be shared with us.

I asked Mat about that in the topic about the livestream (https://www.microstockgroup.com/fotolia-com/adobe-stock-livestream-today-with-mat-hayward-and-terry-white/msg594091/#msg594091), but although the questions posed after mine were anwered, mine was ignored. I'm no mind reader, but it's probably because there isn't any revenue to share with us.

Do you think that we, the contributors whose content trained their model, should be fairly compensated for the continuous revenue that Adobe will get from selling these (but not limited to!) fast lane credits? Is a one-time payment enough to keep you quiet? Or will you raise your voices only when AI completely takes over and you lose the entierty of your revenue? Let me know in the comments <3


« Reply #1 on: October 22, 2023, 11:24 »
+2
Totally 100% agree contributors should get PERPETUAL RECURRING revenue - EVERY SINGLE TIME an asset is "created" that is referenced.

It is 100% possible to do with the "AI" model - contrary to what anyone says. They might not "want" to do it (because they idea is "greed" and trying to "take it all") - but extremely feasible.

See my post here:
https://www.microstockgroup.com/general-stock-discussion/since-'ai-tools'-get-perpetual-recurring-revenue-contributors-should-too/

In essence - the current algorithm (for stealing people's content) would be revised slightly, basically:
a) Every time an asset is stolen, er, "trained" - the contributor whose asset was trained is "tagged" with an ID.
b) When an AI "image" generates an image - it does reference essentially computer models. However - it is totally easy to say for say a "car" - 55 contributors were "tagged" in that computer representation of a "car".
c) When the asset is generated, contributors get micropayments (i.e., say the "ai" image was "worth" $0.10 to the company, and the revenue split was 50-50. So 55 contributors each get $0.05 / 55 ~ 0.0005 cents. May not seem like a "lot" - but with the millions of images generated daily - quickly adds up. (So say each image was like that, 55 contributors for different 'models' - and say 100000 images were generated referencing their input in a month, then 0.0005 * 100000 = $50)
d) The contributor then gets the $50 for their REGULARLY RE-USED contribution. If that same amount continued every month - then the contributor gets that same payment for the rest of their life EVERY SINGLE MONTH, as their asset is referenced.

VERY EASY TO DO. REQUIRES RE-PROGRAMMING the current algorithm. PUSH for that. It is simply a matter of DOING it.

« Reply #2 on: October 22, 2023, 12:42 »
0
....
b) When an AI "image" generates an image - it does reference essentially computer models. However - it is totally easy to say for say a "car" - 55 contributors were "tagged" in that computer representation of a "car".
c) When the asset is generated, contributors get micropayments (i.e., say the "ai" image was "worth" $0.10 to the company, and the revenue split was 50-50. So 55 contributors each get $0.05 / 55 ~ 0.0005 cents. May not seem like a "lot" - but with the millions of images generated daily - quickly adds up. (So say each image was like that, 55 contributors for different 'models' - and say 100000 images were generated referencing their input in a month, then 0.0005 * 100000 = $50)...

besides the fact that there is no, way to trace which images were used & worse even if your 'easy to do' way of marking were possible, for most images (maybe close to 0%) there is no way to identify who the artist is for the billions of images used - many have no names assoc'd and those that do lack verification and an address to pay to). how would your revised training know who (& how) to make payments

but you've stacked the odds in your favor - you claim there would be millions of images generated using only 50 originals, when when many thousands of images would be likely ( a conservative estimate), so your estimate of payment dueis off by several orders of magnitude and certainly not millions of images generated daily

« Reply #3 on: October 22, 2023, 15:02 »
+2
Let's please not derail this topic to whether it is possible to identify which assets have been used to create a particular variation of the generative fill, as it is not pertinent to this discussion - discuss that topic elsewhere please.

Let's assume it isn't possible. Revenue share could be done in a 65/35% split, with the 35% of the revenue going into a "contributor's fund", and then paid out as a frequency of the amount of assets the contributor has provided for training. This number is known, as we already received a one-time payment for the initial Firefly training.

So let's say the whole model has been trained on 50 million assets, and contributor A has provided 5000 assets to that training database. His "exposure" would be 5000/50 million = 0.01%

If the total revenue of "fast lane" credits is 2 million per month, 35% (for the contributor fund) is 700000$. 0.01% of that is 70$. Contributor A would get 70$ that month.

This is feasible, realistic and, in my opinion, fair.

Why is this not being incorporated and why aren't we voicing our concerns? Do you really not care?

« Reply #4 on: October 22, 2023, 17:13 »
0
....
b) When an AI "image" generates an image - it does reference essentially computer models. However - it is totally easy to say for say a "car" - 55 contributors were "tagged" in that computer representation of a "car".
c) When the asset is generated, contributors get micropayments (i.e., say the "ai" image was "worth" $0.10 to the company, and the revenue split was 50-50. So 55 contributors each get $0.05 / 55 ~ 0.0005 cents. May not seem like a "lot" - but with the millions of images generated daily - quickly adds up. (So say each image was like that, 55 contributors for different 'models' - and say 100000 images were generated referencing their input in a month, then 0.0005 * 100000 = $50)...

besides the fact that there is no, way to trace which images were used & worse even if your 'easy to do' way of marking were possible, for most images (maybe close to 0%) there is no way to identify who the artist is for the billions of images used - many have no names assoc'd and those that do lack verification and an address to pay to). how would your revised training know who (& how) to make payments

but you've stacked the odds in your favor - you claim there would be millions of images generated using only 50 originals, when when many thousands of images would be likely ( a conservative estimate), so your estimate of payment dueis off by several orders of magnitude and certainly not millions of images generated daily

You obviously don't have any computer background, or very little. Do you have ANY programming experience whatsoever, let alone large datasets? Yes, it is VERY possible, and VERY doable.

The illustration/example is designed to keep it simple, so you can understand. Obviously programming would be a little more sophisticated than that. But it is VERY VERY easy to do - it is simply a matter of DOING it.

« Reply #5 on: October 22, 2023, 17:21 »
0
Let's please not derail this topic to whether it is possible to identify which assets have been used to create a particular variation of the generative fill, as it is not pertinent to this discussion - discuss that topic elsewhere please.

Let's assume it isn't possible. Revenue share could be done in a 65/35% split, with the 35% of the revenue going into a "contributor's fund", and then paid out as a frequency of the amount of assets the contributor has provided for training. This number is known, as we already received a one-time payment for the initial Firefly training.

So let's say the whole model has been trained on 50 million assets, and contributor A has provided 5000 assets to that training database. His "exposure" would be 5000/50 million = 0.01%

If the total revenue of "fast lane" credits is 2 million per month, 35% (for the contributor fund) is 700000$. 0.01% of that is 70$. Contributor A would get 70$ that month.

This is feasible, realistic and, in my opinion, fair.

Why is this not being incorporated and why aren't we voicing our concerns? Do you really not care?

Actually - it's not derailing - it's actually very relevant, and pertinent to the discussion - simply because the agencies are acting on the premise that they have never considered that as an option (plus, would like people to think it wasn't an option) - simply because "they" in general want the lions share of $$$$. Most - if they could - would probably try to get rid of none-computer generated assets out of pure greed.

Fact is - the "tools" being built are built on the hard work of others - and - going forward - perpetually - contributors whose assets are referenced should be compensated fairly. The "tagging" model (pseudo-code of course for simplicity), accomplishes this.

In terms of a 1-time payment - your model - while good in principle - is too simplistic - and does not compensate users based on usage. I.e., "one" asset could be referenced millions of times, but only get a "1 asset" payment, while another (i.e., the people who spam 50,000 pencils) would get "50,000" asset payments. As well - usage would not necessarily be fairly compensated (i.e., evergreen content, versus say editorial/time specific (i.e, "the virus") over the last couple years).

The perpetual income based on tagging assets used for "ai" image generation is very fair, reasonable, and doable. It is simply a matter of reprogramming the current algorithm to tag assets, and when an "ai image" is generated - compensating every contributor whose assets were used as part of that computer model.


« Reply #6 on: October 23, 2023, 02:22 »
+3


besides the fact that there is no, way to trace which images were used & worse even if your 'easy to do' way of marking were possible, for most images (maybe close to 0%) there is no way to identify who the artist is for the billions of images used - many have no names assoc'd and those that do lack verification and an address to pay to). how would your revised training know who (& how) to make payments


We're talking about Adobe generative fill. Adobe knows very well where to find the artists who's photos were used and how to pay them as they used images from their own database.
« Last Edit: October 23, 2023, 12:34 by Her Ugliness »

« Reply #7 on: October 23, 2023, 06:29 »
+2
yes of course they should share the revenue, the logic is pretty simple.

« Reply #8 on: October 23, 2023, 13:09 »
0
....
b) When an AI "image" generates an image - it does reference essentially computer models. However - it is totally easy to say for say a "car" - 55 contributors were "tagged" in that computer representation of a "car".
c) When the asset is generated, contributors get micropayments (i.e., say the "ai" image was "worth" $0.10 to the company, and the revenue split was 50-50. So 55 contributors each get $0.05 / 55 ~ 0.0005 cents. May not seem like a "lot" - but with the millions of images generated daily - quickly adds up. (So say each image was like that, 55 contributors for different 'models' - and say 100000 images were generated referencing their input in a month, then 0.0005 * 100000 = $50)...

besides the fact that there is no, way to trace which images were used & worse even if your 'easy to do' way of marking were possible, for most images (maybe close to 0%) there is no way to identify who the artist is for the billions of images used - many have no names assoc'd and those that do lack verification and an address to pay to). how would your revised training know who (& how) to make payments

but you've stacked the odds in your favor - you claim there would be millions of images generated using only 50 originals, when when many thousands of images would be likely ( a conservative estimate), so your estimate of payment dueis off by several orders of magnitude and certainly not millions of images generated daily

You obviously don't have any computer background, or very little. Do you have ANY programming experience whatsoever, let alone large datasets? Yes, it is VERY possible, and VERY doable.

The illustration/example is designed to keep it simple, so you can understand. Obviously programming would be a little more sophisticated than that. But it is VERY VERY easy to do - it is simply a matter of DOING it.

you know NOTHING about my computer experience.

you don't address the biggest problem - how do you identify who made the image & how to contact & pay them?  that information isnt available in most cases

and now you've changed the goal posts, saying funds should be distributed to everyone, not based on where their images were used -  you know little/nothing about ML it seems and that is relevant as  the major fallacy in your proposal assumes you can track where an image is used

« Reply #9 on: October 23, 2023, 13:13 »
0

besides the fact that there is no, way to trace which images were used & worse even if your 'easy to do' way of marking were possible, for most images (maybe close to 0%) there is no way to identify who the artist is for the billions of images used - many have no names assoc'd and those that do lack verification and an address to pay to). how would your revised training know who (& how) to make payments


We're talking about Adobe generative fill. Adobe knows very well where to find the artists who's photos were used and how to pay them as they used images from their own database.
yes, it's about AS specifically but i was responding to the comments that were not limited to AS.

again, how does AS know whose images were used for each creation since ML eliminates any way to track that, even if identifiers were attached initially.  each image is translated into thousands of datapoints and millions/billions of operations are performed to generate each new image.
« Last Edit: October 23, 2023, 13:17 by cascoly »

« Reply #10 on: October 24, 2023, 01:32 »
+2
you know NOTHING about my computer experience.

you don't address the biggest problem - how do you identify who made the image & how to contact & pay them?  that information isnt available in most cases

and now you've changed the goal posts, saying funds should be distributed to everyone, not based on where their images were used -  you know little/nothing about ML it seems and that is relevant as  the major fallacy in your proposal assumes you can track where an image is used

You are correct, I don't know your computer experience, which is why I was asking. Likewise, you know absolutely nothing about my experience, and are quite arrogant and accusatory in your statements, which would lead me to believe you know very little, if anything related to computer science/computer engineering/programming/etc - and especially - actual "AI" versus "ML" algorithms. Fact is, I actually do know what I am talking about. Have you EVER dealt with large datasets, scraped data, creating actual databases from scratch, written image algorithms, ANY of that? Have you actually ever even created your own "AI" algorithm? If you had - you'd probably know what I am talking about is very feasible, and simply a matter of doing it. It does require "work", and it does seem agencies are trying to figure out how to cut out artists while appearing to be noble (in many ways from a purely greedy standpoint). when fact is, they should be held fully accountable for any theft of artist assets, and compensate accordingly - in the same perpetual/recurring revenue model they so desperately desire.

Answering your questions:

a) For compensation - depends on where/how the data/images/etc were scraped. It would make most sense to simply begin with the major agencies whose data was scraped (i.e., DT/SS/AS/P5/etc). Super easy to figure out who to issue payments to. For other agencies, would be a matter of writing more sophisticated algorithms...

And - as a "tongue in cheek" statement - since "everyone" seems to think actual thinking "AI" exists (most people don't actually understand what actual "ai", and believe it is "thinking", versus the sophisticated theft & pattern re-arrangement being called "ai") - but if "ai" actually existed - it would be super simple - just ask the "ai" to figure out who to issue payments to.

b) it is SUPER easy to track/source image usage. If you don't realize this, it seems you've never scraped data before?

« Last Edit: October 24, 2023, 01:56 by SuperPhoto »

« Reply #11 on: October 24, 2023, 01:36 »
0

yes, it's about AS specifically but i was responding to the comments that were not limited to AS.

again, how does AS know whose images were used for each creation since ML eliminates any way to track that, even if identifiers were attached initially.  each image is translated into thousands of datapoints and millions/billions of operations are performed to generate each new image.

If you are using an out-of-the box "ML" solution, without ANY kind of revision whatsoever that had no built in tracking/etc, then yes, you would be correct.

But if you REVISE the algorithm (let's say a simple "reinforcement" model) - and assign weights with "ids" of original source to the original inputs - it becomes very easy to "track" sources. It does require revising the generic algorithms taught in most computer science textbooks.

« Reply #12 on: October 24, 2023, 01:50 »
0

besides the fact that there is no, way to trace which images were used & worse even if your 'easy to do' way of marking were possible, for most images (maybe close to 0%) there is no way to identify who the artist is for the billions of images used - many have no names assoc'd and those that do lack verification and an address to pay to). how would your revised training know who (& how) to make payments


We're talking about Adobe generative fill. Adobe knows very well where to find the artists who's photos were used and how to pay them as they used images from their own database.
yes, it's about AS specifically but i was responding to the comments that were not limited to AS.

again, how does AS know whose images were used for each creation since ML eliminates any way to track that, even if identifiers were attached initially.  each image is translated into thousands of datapoints and millions/billions of operations are performed to generate each new image.

Okay - you aren't quite thinking correctly here.

It requires revising the machine learning algorith to incorporate identifiers and then attribute those identifiers to outputted information.

Keeping things simple.

Lets say you have 3 contributors, named "|A|" and "|B|" and "|C|".
|A| has images of an apple, and a pear
|B| has an image of an apple
|C| has an image of a pear, and an orange

Let's say the "machine" ("AI") version of an apple is "ML-APPLE" and "ML-PEAR".

The "AI" (ML/machine learning algorithm) creates a "representation" of what it believes an "apple" to be by scraping |A| + |B|'s image.
It then does the same for a pear, by scraping |A| + |C|'s image.

In it's internal representation, it would look like:

[ML-APPLE]:{|A|,|B|} (simply meaning what I stated above - the "ai" version of an apple references |A| + |B|'s image
[ML-PEAR]:{|A|,|C|} (simply meaning what I stated above - the "ai" version of a pears references |A| + |C|'s image
[ML-ORANGE]:{|C|} (simply meaning what I stated above - the "ai" version of an orange references |C|'s image

Let's say you then have a customer that generates images. Let's say they pay $1/image (simplicity), & its a 50-50 share between agency (the "ML" image) + the source contributors.

They decide to make a picture of a "pear". Since the "ML-PEAR" references |A|+|C|'s image - |A|+|C| would be compensated for the use of that image generation.
I.e., $1 = $.50 agency, $.50 to contributors. Since two contributors (A+C) made the "pear" image, the revenue for contributors would be issued to them).

Now let's say you made an image of an orange.
Since |C| was the only source document referenced, |C| would get full credit for this image. (I.e., $1 => $0.50 to agency, $.50 to |C|).

That is a super basic illustration of what I am talking about.

Of course, the pseudocode above is an extremely simplistic concept - it is simply designed to illustrate how it would be done on a most basic level, and some of the requirements to revise the agorithm to attribute source images.

Of course, actual code would be much more sophisticated, and one could then decide whether to attach weights to "how much" of the model was used (i.e., was it a "tiny" pear in the image, or a "big" pear in the image, and should they be compensated accordingly?) As well as how "much" of the "pear" was attributed to a specific contributor. (I.e., did |C| say have 50 images of pears, and |A| only 1 image of pear, that was used in the model/representation of what a 'pear' was - such that |C| should get 50x the 'credit' for the pear image?) Of course - that is a little more in depth, and this example was simply used for illustration purposes.

Fact is - it IS super easy to properly attribute source images, AND - it is ALSO super easy to properly CREDIT source images - on a PERPETUAL RECURRING BASIS.

It is simply a matter of taken the time to revise the algorithm to do so.
« Last Edit: October 24, 2023, 01:58 by SuperPhoto »

« Reply #13 on: October 24, 2023, 07:25 »
+1
PPS,

I should add - what do you think the words are used to describe in creating an image? I.e., "pear growing in a tree in a field".
"pear,grow[ing],tree,field"

Those are tags. The "machine learning" already DOES associate with tags.
It extracts that information (i.e., "keywords") associated with the "image" - and then associates that with the model.

So it is SUPER easy to simply add "contributor-id" (which can be the name/URL/etc or an actual number that contains all that information). And then SUPER easy to associate WHICH contributors file(s) were used in creating a "composite" image (i.e., an "ai" generated image).

SUPER SUPER EASY. Just a matter of doing it, then fairly compensating contributors with the SAME RECURRING PERPETUAL INCOME REVENUE model that the agencies so desperately and greedily want for themselves, and trying to convince contributors that anything else is "fair" (which of course, it's not). Sharing the recurrnig revenue model, with opt-in/opt-out features so at ANY time the contributor can opt-out if they don't like the terms  - and assets going forward do NOT reference the input items - is fair.

« Reply #14 on: October 24, 2023, 15:14 »
0
...
So it is SUPER easy to simply add "contributor-id" (which can be the name/URL/etc or an actual number that contains all that information). And then SUPER easy to associate WHICH contributors file(s) were used in creating a "composite" image (i.e., an "ai" generated image).

SUPER SUPER EASY. Just a matter of doing it, then fairly compensating contributors with the SAME RECURRING PERPETUAL INCOME REVENUE model that the agencies so desperately and greedily want for themselves, and trying to convince contributors that anything else is "fair" (which of course, it's not). Sharing the recurrnig revenue model, with opt-in/opt-out features so at ANY time the contributor can opt-out if they don't like the terms  - and assets going forward do NOT reference the input items - is fair.

i agree, starting from scratch, adding contributor id is the easy part, but given the number of possible contributors to each piece of an ai-gen image, the book-keeping becomes expensive - many thousands of entries for each image.

but the much larger problem you don't address is actually finding out who the author is and how to pay them as scraping only works if the image has the artist's verifiable info (and how to verify they're who they claim to be when opting out).  true, when agencies create datasets thery have info for payment, when they license these datasets who tracks payments? and will agencies actually sell data sets with the private info of their contributors?

your examples are great for explaining your proposal, but they don't scale up - the amount payable to any artist is 2-3 orders of magnitude less 


Uncle Pete

  • Great Place by a Great Lake - My Home Port
« Reply #15 on: October 25, 2023, 13:02 »
+2
...
So it is SUPER easy to simply add "contributor-id" (which can be the name/URL/etc or an actual number that contains all that information). And then SUPER easy to associate WHICH contributors file(s) were used in creating a "composite" image (i.e., an "ai" generated image).

SUPER SUPER EASY. Just a matter of doing it, then fairly compensating contributors with the SAME RECURRING PERPETUAL INCOME REVENUE model that the agencies so desperately and greedily want for themselves, and trying to convince contributors that anything else is "fair" (which of course, it's not). Sharing the recurrnig revenue model, with opt-in/opt-out features so at ANY time the contributor can opt-out if they don't like the terms  - and assets going forward do NOT reference the input items - is fair.

i agree, starting from scratch, adding contributor id is the easy part, but given the number of possible contributors to each piece of an ai-gen image, the book-keeping becomes expensive - many thousands of entries for each image.

but the much larger problem you don't address is actually finding out who the author is and how to pay them as scraping only works if the image has the artist's verifiable info (and how to verify they're who they claim to be when opting out).  true, when agencies create datasets thery have info for payment, when they license these datasets who tracks payments? and will agencies actually sell data sets with the private info of their contributors?

your examples are great for explaining your proposal, but they don't scale up - the amount payable to any artist is 2-3 orders of magnitude less

Just a question, since I'm spinning with different views and claims.

Once the AI has used the image to train, it is not using the image again, when it generates a new image from a request. It's just using what it learned. I mean no direct access to the past or the training images.

Comparable thought. I write a reference book, and I'm paid for each sale. The buyer looks at the book and learns something. The author is not paid, every time the buyer opens the book again and reads something else, or the same facts over again.

I'm not against being paid more or making more for the use of my images, but I'm just trying to understand, if they are used a second time, after the training?

And if I was paid for reference, to my image, why would they have to pay every time the AI looks back at it?


 

Related Topics

  Subject / Started by Replies Last post
17 Replies
5416 Views
Last post July 20, 2023, 10:35
by spike
3 Replies
4563 Views
Last post September 18, 2023, 11:27
by Injustice for all
6 Replies
818 Views
Last post February 12, 2024, 06:35
by Faustvasea
1 Replies
849 Views
Last post February 18, 2024, 20:09
by Jo Ann Snover
15 Replies
474 Views
Last post March 17, 2024, 12:11
by stoker2014

Sponsors

Mega Bundle of 5,900+ Professional Lightroom Presets

Microstock Poll Results

Sponsors