The Limited Times

Now you can see non-English news...

Open AI launches Sora, a revolutionary video tool with artificial intelligence

2024-02-15T23:39:38.387Z

Highlights: Open AI launches Sora, a revolutionary video tool with artificial intelligence. Sora is capable of creating high-quality images from text explanations. The model understands not only what the user has asked for in the request, but also how those things exist in the physical world, according to OpenAI. The company is also giving access to a range of visual artists, designers and filmmakers to give feedback on how to improve the model to make it more useful to creative professionals. For now, the tool has been made available to the so-called red teams.


Sora is capable of creating high-quality images from text explanations


While the fascination with ChatGPT and generative artificial intelligence language models has not yet passed, OpenAI has just introduced a dazzling and revolutionary video creation tool called Sora.

With it, just give a description of what you want to see on the screen and there it is, created by artificial intelligence.

Some are more accomplished than others, sometimes they have that video game style that differentiates them from reality, but they are all surprising.

The CEO of OpenAI, Sam Altman, made the announcement of the launch on the social network X, which was immediately flooded with the new creations.

Realistic, futuristic, crazy, cartoon videos... The videos include all kinds of automatic creations created with generative artificial intelligence.

Sora is capable of generating entire videos in one go or expanding the generated videos to make them longer.

Sora is capable of generating complex scenes with multiple characters, specific types of movement, and precise subject and background details.

The model understands not only what the user has asked for in the request, but also how those things exist in the physical world, according to OpenAI.

The model has a deep understanding of language, allowing it to accurately interpret cues and generate compelling characters that express vibrant emotions, the company explains.

“Here is Sora, our video generation model,” Altman wrote.

“We offer access to a limited number of creators,” he added, before asking his followers to make suggestions for creating new videos, in addition to the samples that he had already offered on his website.

The instructions may be more or less detailed.

One of the examples offered by OpenAI responds to the following description: “An elegant woman walks down a Tokyo street full of warm bright neon and lively urban signage.

She is wearing a black leather jacket, a long red dress, black boots and a black bag.

She is wearing sunglasses and red lipstick.

She walks with confidence and carelessness.

The street is wet and reflective, creating a mirror effect of the colored lights.

Many pedestrians walk around.”

And the result is surprising.

Another notes: “Trailer for a movie about the adventures of the 30-year-old spaceman wearing a red wool knitted motorcycle helmet, blue sky, salt desert, cinematic style, shot on 35mm film, bright colors.” .

In addition to being able to generate a video solely from text instructions, the model is capable of taking an existing still image and generating a video from it, animating the image content with precision and attention to small details.

The model can also take an existing video and enlarge it or fill in missing frames.

You can ask for content, style and give all kinds of instructions.

Altman has been publishing new videos requested by tweeters, proving that the results are immediate.

Sora can also create multiple shots within a single generated video while accurately maintaining characters and visual style.

“We are teaching AI to understand and simulate the physical world in motion, with the goal of training models that help people solve problems that require interaction in the real world,” explains OpenAI when presenting the new text conversion tool. to video.

“Sora can generate videos up to one minute long while maintaining visual quality and fidelity to user input,” she adds.

For now, the tool has been made available to the so-called red teams.

The members of these teams try to question a product or service, push it to the limit, put it to the test and find its faults as if they were enemies of the company.

Here they have the specific mission of evaluating critical areas for potential damage or risks.

Among them are experts in areas such as misinformation, hateful content and bias.

Open AI is also giving access to a range of visual artists, designers and filmmakers to give feedback on how to improve the model to make it more useful to creative professionals.

“We are sharing our research progress in advance to begin working with people outside of OpenAI and receive their feedback, as well as to give the public an idea of ​​​​the AI ​​​​capabilities that are on the horizon,” explains the company.

Defects to be polished

The artificial intelligence firm itself recognizes that Sora still has some very obvious flaws.

You may struggle to accurately simulate the physics of a complex scene and may not understand specific cases of cause and effect.

He gives as an example that a person can take a bite of a cookie, but then the cookie may not have the bite mark.

The model may also confuse spatial details of a cue, for example, confusing left and right, and may have problems with accurate descriptions of events that take place over time, such as following a specific camera path.

Before making the tool available to the public, OpenAI promises to take some precautions.

Among them is taking into account the instructions of the red teams.

Additionally, it is creating tools to help detect misleading content, with detectors that can tell when a video has been generated by Sora.

It has also developed powerful image classifiers that are used to review the frames of all generated videos and ensure that they comply with its usage policies before displaying them to the user.

Additionally, you will reuse the security methods you created for your products that use DALL-E 3. For example, the text classifier will check and reject text entry requests that violate your usage policies, such as those that request extreme violence, content sexual images, images that incite hatred, images of celebrities or the intellectual property of third parties.

“We will reach out to policy makers, educators and artists around the world to hear their concerns and identify positive use cases for this new technology.

Despite extensive research and testing, we cannot predict all the beneficial ways people will use our technology, nor all the ways they will abuse it.

That is why we believe that learning from real-world use is a fundamental component to creating and launching increasingly safer AI systems over time,” concludes OpenAI.

You can follow

EL PAÍS Tecnología

on

Facebook

and

X

or sign up here to receive our

weekly newsletter

.

Source: elparis

All tech articles on 2024-02-15

You may like

Trends 24h

Latest

© Communities 2019 - Privacy

The information on this site is from external sources that are not under our control.
The inclusion of any links does not necessarily imply a recommendation or endorse the views expressed within them.