One evening in early September, T, a 28-year-old actor who asked to be identified by his first initial, took his seat in a rented Hollywood studio space in front of three cameras, a director, and a producer for a somewhat unusual gig.
The two-hour shoot produced footage that was not meant to be viewed by the publicâat least, not a human public.
Rather, Tâs voice, face, movements, and expressions would be fed into an AI database âto better understand and express human emotions.â That database would then help train âvirtual avatarsâ for Meta, as well as algorithms for a London-based emotion AI company called Realeyes. (Realeyes was running the project; participants only learned about Metaâs involvement once they arrived on site.)
The âemotion studyâ ran from July through September, specifically recruiting actors. The project coincided with Hollywoodâs historic dual strikes by the Writers Guild of America and the Screen Actors Guild (SAG-AFTRA). With the industry at a standstill, the larger-than-usual number of out-of-work actors may have been a boon for Meta and Realeyes: here was a new pool of âtrainersââand data pointsâperfectly suited to teaching their AI to appear more human.
For actors like T, it was a great opportunity too: a way to make good, easy money on the side, without having to cross the picket line.
âThere arenât really clear rules right now.”
âThis is fully a research-based project,â the job posting said. It offered $150 per hour for at least two hours of work, and asserted that âyour individual likeness will not be used for any commercial purposes.â
The actors may have assumed this meant that their faces and performances wouldnât turn up in a TV show or movie, but the broad nature of what they signed makes it impossible to know for sure the full implications of what theyâve given away. In fact, in order to participate, they had to sign away certain rights âin perpetuityâ for technologies and use cases that may not yet exist.
And while the job posting insisted that the project âdoes not qualify as struck workâ (that is, work produced by employers against whom the union is striking), it nevertheless speaks to some of the strikeâs core issues: how actorsâ likenesses can be used, how actors should be compensated for that use, and what informed consent should look like in the age of AI.
âThis isnât a contract battle between a union and a company,â said Duncan Crabtree-Ireland, SAG-AFTRAâs chief negotiator, at a panel on AI in entertainment at San Diego Comic-Con this summer. âItâs existential.â
Many actors across the industry, particularly background actors (also known as extras), worry that AIâmuch like the models described in the emotion studyâcould be used to replace them, whether or not their exact faces are copied. And in this case, by providing the facial expressions that will teach AI to appear more human, study participants may in fact have been the ones inadvertently training their own potential replacements.
âOur studies have nothing to do with the strike,â Max Kalehoff, Realeyesâs vice president for growth and marketing, said in an email. âThe vast majority of our work is in evaluating the effectiveness of advertising for clientsâwhich has nothing to do with actors and the entertainment industry except to gauge audience reaction.â The timing, he added, was âan unfortunate coincidence.â Meta did not respond to multiple requests for comment.
Given how technological advancements so often build upon one another, not to mention how quickly the field of artificial intelligence is evolving, experts point out that thereâs only so much these companies can truly promise.
In addition to the job posting, MIT Technology Review has obtained and reviewed a copy of the data license agreement, and its potential implications are indeed vast. To put it bluntly: whether the actors who participated knew it or not, for as little as $300, they appear to have authorized Realeyes, Meta, and other parties of the two companiesâ choosing to access and use not just their faces but also their expressions, and anything derived from them, almost however and whenever they wantâas long as they do not reproduce any individual likenesses.
Some actors, like Jessica, who asked to be identified by just her first name, felt there was something âexploitativeâ about the projectâboth in the financial incentives for out-of-work actors and in the fight over AI and the use of an actorâs image.
Jessica, a New Yorkâbased background actor, says she has seen a growing number of listings for AI jobs over the past few years. âThere arenât really clear rules right now,â she says, âso I donât know. Maybe ⊠their intention [is] to get these images before the union signs a contract and sets them.â
Do you have any tips related to how AI is being used in the entertainment industry? Please reach out at tips@technologyreview.com or securely on Signal at 626.765.5489.
All this leaves actors, struggling after three months of limited to no work, primed to accept the terms from Realeyes and Metaâand, intentionally or not, to affect all actors, whether or not they personally choose to engage with AI.
âItâs hurt now or hurt later,â says Maurice Compte, an actor and SAG-AFTRA member who has had principal roles on shows like Narcos and Breaking Bad. After reviewing the job posting, he couldnât help but see nefarious intent. Yes, he said, of course itâs beneficial to have work, but he sees it as beneficial âin the way that the Native Americans did when they took blankets from white settlers,â adding: âThey were getting blankets out of it in a time of cold.â
Humans as data
Artificial intelligence is powered by data, and data, in turn, is provided by humans.
It is human labor that prepares, cleans, and annotates data to make it more understandable to machines; as MIT Technology Review has reported, for example, robot vacuums know to avoid running over dog poop because human data labelers have first clicked through and identified millions of images of pet wasteâand other objectsâinside homes.
When it comes to facial recognition, other biometric analysis, or generative AI models that aim to generate humans or human-like avatars, it is human faces, movements, and voices that serve as the data.
Initially, these models were powered by data scraped off the internetâincluding, on several occasions, private surveillance camera footage that was shared or sold without the knowledge of anyone being captured.
But as the need for higher-quality data has grown, alongside concerns about whether data is collected ethically and with proper consent, tech companies have progressed from âscraping data from publicly available sourcesâ to âbuilding data sets with professionals,â explains Julian Posada, an assistant professor at Yale University who studies platforms and labor. Or, at the very least, âwith people who have been recruited, compensated, [and] signed [consent] forms.â
But the need for human data, especially in the entertainment industry, runs up against a significant concern in Hollywood: publicity rights, or âthe right to control your use of your name and likeness,â according to Corynne McSherry, the legal director of the Electronic Frontier Foundation (EFF), a digital rights group.
This was an issue long before AI, but AI has amplified the concern. Generative AI in particular makes it easy to create realistic replicas of anyone by training algorithms on existing data, like photos and videos of the person. The more data that is available, the easier it is to create a realistic image. This has a particularly large effect on performers.
He believes itâs that improvisation requirement that explains why Realeyes and Meta were specifically recruiting actors.
Some actors have been able to monetize the characteristics that make them unique. James Earl Jones, the voice of Darth Vader, signed off on the use of archived recordings of his voice so that AI could continue to generate it for future Star Wars films. Meanwhile, de-aging AI has allowed Harrison Ford, Tom Hanks, and Robin Wright to portray younger versions of themselves on screen. Metaphysic AI, the company behind the de-aging technology, recently signed a deal with Creative Artists Agency to put generative AI to use for its artists.
But many deepfakes, or images of fake events created with deep-learning AI, are generated without consent. Earlier this month, Hanks posted on Instagram that an ad purporting to show him promoting a dental plan was not actually him.
The AI landscape is different for noncelebrities. Background actors are increasingly being asked to undergo digital body scans on set, where they have little power to push back or even get clarity on how those scans will be used in the future. Studios say that scans are used primarily to augment crowd scenes, which they have been doing with other technology in postproduction for yearsâbut according to SAG representatives, once the studios have captured actorsâ likenesses, they reserve the rights to use them forever. (There have already been multiple reports from voice actors that their voices have appeared in video games other than the ones they were hired for.)
In the case of the Realeyes and Meta study, it might be âstudy dataâ rather than body scans, but actors are dealing with the same uncertainty as to how else their digital likenesses could one day be used.
Teaching AI to appear more human
At $150 per hour, the Realeyes study paid far more than the roughly $200 daily rate in the current Screen Actors Guild contract (nonunion jobs pay even less).
This made the gig an attractive proposition for young actors like T, just starting out in Hollywoodâa notoriously challenging environment even had he not arrived just before the SAG-AFTRA strike started. (T has not worked enough union jobs to officially join the union, though he hopes to one day.)
In fact, even more than a standard acting job, T described performing for Realeyes as âlike an acting workshop where ⊠you get a chance to work on your acting chops, which I thought helped me a little bit.â
For two hours, T responded to prompts like âTell us something that makes you angry,â âShare a sad story,â or âDo a scary scene where youâre scared,â improvising an appropriate story or scene for each one. He believes itâs that improvisation requirement that explains why Realeyes and Meta were specifically recruiting actors.
In addition to wanting the pay, T participated in the study because, as he understood it, no one would see the results publicly. Rather, it was research for Meta, as he learned when he arrived at the studio space and signed a data license agreement with the company that he only skimmed through. It was the first heâd heard that Meta was even connected with the project. (He had previously signed a separate contract with Realeyes covering the terms of the job.)
The data license agreement says that Realeyes is the sole owner of the data and has full rights to âlicense, distribute, reproduce, modify, or otherwise create and use derivative worksâ generated from it, âirrevocably and in all formats and media existing now or in the future.â
This kind of legalese can be hard to parse, particularly when it deals with technology that is changing at such a rapid pace. But what it essentially means is that âyou may be giving away things you didnât realize ⊠because those things didnât exist yet,â says Emily Poler, a litigator who represents clients in disputes at the intersection of media, technology, and intellectual property.
âIf I was a lawyer for an actor here, I would definitely be looking into whether one can knowingly waive rights where things donât even exist yet,â she adds.
As Jessica argues, âOnce they have your image, they can use it whenever and however.â She thinks that actorsâ likenesses could be used in the same way that other artistsâ works, like paintings, songs, and poetry, have been used to train generative AI, and she worries that the AI could just âcreate a composite that looks âhuman,â like believable as human,â but âit wouldnât be recognizable as you, so you canât potentially sue themââeven if that AI-generated human was based on you.
This feels especially plausible to Jessica given her experience as an Asian-American background actor in an industry where representation often amounts to being the token minority. Now, she fears, anyone who hires actors could ârecruit a few Asian peopleâ and scan them to create âan Asian avatarâ that they could use instead of âhiring one of you to be in a commercial.â
Itâs not just images that actors should be worried about, says Adam Harvey, an applied researcher who focuses on computer vision, privacy, and surveillance and is one of the co-creators of Exposing.AI, which catalogs the data sets used to train facial recognition systems.
What constitutes âlikeness,â he says, is changing. While the word is now understood primarily to mean a photographic likeness, musicians are challenging that definition to include vocal likenesses. Eventually, he believes, âit will also ⊠be challenged on the emotional frontierââthat is, actors could argue that their microexpressions are unique and should be protected.
Realeyesâs Kalehoff did not say what specifically the company would be using the study results for, though he elaborated in an email that there could be âa variety of use cases, such as building better digital media experiences, in medical diagnoses (i.e. skin/muscle conditions), safety alertness detection, or robotic tools to support medical disorders related to recognition of facial expressions (like autism).â
Now, she fears, anyone who hires actors could ârecruit a few Asian peopleâ and scan them to create âan Asian avatarâ that they could use instead of âhiring one of you to be in a commercial.â
When asked how Realeyes defined âlikeness,â he replied that the company used that termâas well as âcommercial,â another word for which there are assumed but no universally agreed-upon definitionsâin a manner that is âthe same for us as [a] general business.â He added, âWe do not have a specific definition different from standard usage.â
But for T, and for other actors, âcommercialâ would typically mean appearing in some sort of advertisement or a TV spotââsomething,â T says, âthatâs directly sold to the consumer.â
Outside of the narrow understanding in the entertainment industry, the EFFâs McSherry questions what the company means: âItâs a commercial company doing commercial things.â
Kalehoff also said, âIf a client would ask us to use such images [from the study], we would insist on 100% consent, fair pay for participants, and transparency. However, that is not our work or what we do.â
Yet this statement does not align with the language of the data license agreement, which stipulates that while Realeyes is the owner of the intellectual property stemming from the study data, Meta and âMeta parties acting on behalf of Metaâ have broad rights to the dataâincluding the rights to share and sell it. This means that, ultimately, how itâs used may be out of Realeyesâs hands.
As explained in the agreement, the rights of Meta and parties acting on its behalf also include:
- Asserting certain rights to the participantsâ identities (âIdentifying or recognizing you ⊠creating a unique template of your face and/or voice ⊠and/or protecting against impersonation and identity misuseâ)
- Allowing other researchers to conduct future research, using the study data however they see fit (âconducting future research studies and activities ⊠in collaboration with third party researchers, who may further use the Study Data beyond the control of Metaâ)
- Creating derivative works from the study data for any kind of use at any time (âusing, distributing, reproducing, publicly performing, publicly displaying, disclosing, and modifying or otherwise creating derivative works from the Study Data, worldwide, irrevocably and in perpetuity, and in all formats and media existing now or in the futureâ)
The only limit on use was that Meta and parties would ânot use Study Data to develop machine learning models that generate your specific face or voice in any Meta productâ (emphasis added). Still, the variety of possible use casesâand usersâis sweeping. And the agreement does little to quell actorsâ specific anxieties that âdown the line, that database is used to generate a work and that work ends up seeming a lot like [someoneâs] performance,â as McSherry puts it.
When I asked Kalehoff about the apparent gap between his comments and the agreement, he denied any discrepancy: âWe believe there are no contradictions in any agreements, and we stand by our commitment to actors as stated in all of our agreements to fully protect their image and their privacy.â Kalehoff declined to comment on Realeyesâs work with clients, or to confirm that the study was in collaboration with Meta.
Meanwhile, Meta has been building photorealistic 3D âCodec avatars,â which go far beyond the cartoonish images in Horizon Worlds and require human training data to perfect. CEO Mark Zuckerberg recently described these avatars on the popular podcast from AI researcher Lex Fridman as core to his vision of the futureâwhere physical, virtual, and augmented reality all coexist. He envisions the avatars âdelivering a sense of presence as if youâre there together, no matter where you actually are in the world.â
Despite multiple requests for comment, Meta did not respond to any questions from MIT Technology Review, so we cannot confirm what it would use the data for, or who it means by âparties acting on its behalf.â
Individual choice, collective impact
Throughout the strikes by writers and actors, there has been a palpable sense that Hollywood is charging into a new frontier that will shape how weâall of usâengage with artificial intelligence. Usually, that frontier is described with reference to workersâ rights; the idea is that whatever happens here will affect workers in other industries who are grappling with what AI will mean for their own livelihoods.
Already, the gains won by the Writers Guild have provided a model for how to regulate AIâs impact on creative work. The unionâs new contract with studios limits the use of AI in writersâ rooms and stipulates that only human authors can be credited on stories, which prevents studios from copyrighting AI-generated work and further serves as a major disincentive to use AI to write scripts.
In early October, the actorsâ union and the studios also returned to the bargaining table, hoping to provide similar guidance for actors. But talks quickly broke down because âit is clear that the gap between the AMPTP [Alliance of Motion Picture and Television Producers] and SAG-AFTRA is too great,â as the studio alliance put it in a press release. Generative AIâspecifically, how and when background actors should be expected to consent to body scanningâwas reportedly one of the sticking points.
Whatever final agreement they come to wonât forbid the use of AI by studiosâthat was never the point. Even the actors who took issue with the AI training projects have more nuanced views about the use of the technology. âWeâre not going to fully cut out AI,â acknowledges Compte, the Breaking Bad actor. Rather, we âjust have to find ways that are going to benefit the larger picture⊠[It] is really about living wages.â
But a future agreement, which is specifically between the studios and SAG, will not be applicable to tech companies conducting âresearchâ projects, like Meta and Realeyes. Technological advances created for one purposeâperhaps those that come out of a âresearchâ studyâwill also have broader applications, in film and beyond.
âThe likelihood that the technology that is developed is only used for that [audience engagement or Codec avatars] is vanishingly small. Thatâs not how it works,â says the EFFâs McSherry. For instance, while the data agreement for the emotion study does not explicitly mention using the results for facial recognition AI, McSherry believes that they could be used to improve any kind of AI involving human faces or expressions.
(Besides, emotion detection algorithms are themselves controversial, whether or not they even work the way developers say they do. Do we really want âour faces to be judged all the time [based] on whatever products weâre looking at?â asks Posada, the Yale professor.)
This all makes consent for these broad research studies even trickier: thereâs no way for a participant to opt in or out of specific use cases. T, for one, would be happy if his participation meant better avatar options for virtual worlds, like those he uses with his Oculusâthough he isnât agreeing to that specifically.
But what are individual study participantsâwho may need the incomeâto do? What power do they really have in this situation? And what power do other peopleâeven people who declined to participateâhave to ensure that they are not affected? The decision to train AI may be an individual one, but the impact is not; itâs collective.
âOnce they feed your image and ⊠a certain amount of peopleâs images, they can create an endless variety of similar-looking people,â says Jessica. âItâs not infringing on your face, per se.â But maybe thatâs the point: âTheyâre using your image without ⊠being held liable for it.â
T has considered the possibility that, one day, the research he has contributed to could very well replace actors.
But at least for now, itâs a hypothetical.
âIâd be upset,â he acknowledges, âbut at the same time, if it wasnât me doing it, theyâd probably figure out a different wayâa sneakier way, without getting peopleâs consent.â Besides, T adds, âthey paid really well.â
Do you have any tips related to how AI is being used in the entertainment industry? Please reach out at tips@technologyreview.com or securely on Signal at 626.765.5489.
Article Source link and Credit

