OpenAI’s innovative forays into the world of artificial intelligence are both exciting and bring with them certain uncertainties. The recently announced Sora model is at the center of these uncertainties. The Wall Street Journal’s interview with Mira Murati is filled with significant question marks regarding the sources of Sora’s training data. OpenAI’s Chief Technology Officer Murati avoided giving clear answers to questions about Sora’s data sources and skirted the issue with vague statements.
Mira Murati’s Statements Cause Confusion
Ambiguity, especially regarding where Sora’s training data comes from, is concerning for those seeking clarity. Murati’s statement “we used public data and licensed data” actually raises many questions. What public data was used? From which sources were the licensed data obtained, and what kind of data does it include?
Murati gave vague answers when asked about the data source for the Sora model, which can generate videos from text instructions. Murati replied to the question of how the company developed its new model, valued at $80 billion, with “We used public data and licensed data.”
Joanna Stern from the Journal asked whether Sora was trained with data from social media platforms such as YouTube, Instagram, or Facebook. Murati responded, “Actually, I’m not sure,” and added:
“You know, if they were public – available for public use. But I’m not sure. I’m not sure about this myself.”
The question of whether Sora was trained with data from social media platforms remains completely unanswered. Murati’s “I’m not sure” response frankly reflects a stance far from transparency. At this point, clarifying the source of the data used and the privacy, fair use, and safety of these data in terms of society is extremely important.
Another notable point is the information about OpenAI’s partnership with Shutterstock. Unfortunately, Murati’s explanations about the sources of the data are not entirely satisfactory. While it is important whether the data are public or licensed, how and for what purposes these data are used should also be clarified.
Situations That Could Lead to Serious Concerns
Artificial intelligence models are built on training data, and the accuracy, reliability, and source of these data are critical factors determining the model’s success and societal impact. Therefore, the uncertainty of the training data for an important artificial intelligence model like Sora can lead to serious concerns both within the industry and in society.
OpenAI emphasizing transparency and accountability principles in innovations like the Sora model could increase trust and reduce concerns. Ultimately, while advancing in the world of artificial intelligence is important, this progress must be in harmony with the interests and values of society.
When you step inside OpenAI, you feel as if you have opened the door to the future. However, lately, this door has been filled with a series of legal problems obstructing the company’s path. Recent developments related to Murati’s management of some of OpenAI’s key projects have both elevated and questioned the company’s name.
In November of last year, the company was shaken by the dismissal of OpenAI’s board chairman Sam Altman, leading to a temporary leadership vacuum. During this critical period, Murati took the helm as the interim CEO. However, the change in management attracted attention not only within the company but also from the outside world.
Serious Criticisms Arise
OpenAI’s artificial intelligence models have faced serious criticism regarding the accuracy and use of training data. In July 2023, well-known authors such as Sarah Silverman, Richard Kadrey, and Christopher Golden filed a lawsuit accusing ChatGPT of copyright infringement. According to their claims, the AI was creating summaries based on copyrighted content, using the authors’ works.
This lawsuit was not limited to OpenAI’s legal turmoil. In December, The New York Times filed a similar lawsuit against Microsoft and OpenAI, accusing the companies of using the newspaper’s content without permission to train AI chatbots. Additionally, in California, another class-action lawsuit was filed alleging that OpenAI collected private user information from the internet without permission to train ChatGPT.
The allegations raised in these lawsuits show that OpenAI must consider not only technological advancement but also ethical and legal responsibilities. As a giant at the forefront of the artificial intelligence world, how OpenAI will resolve these challenging issues is eagerly awaited. However, one thing is certain: as artificial intelligence shapes the future, progressing on this journey without the guidance of the law seems impossible.
As a writer, my comment on these negative developments is that they have the potential to seriously affect artificial intelligence coins. Any negative approach towards a giant company like OpenAI could impact other AI coins, including WLD.