Meet DALL-E, the Artificial Intelligence That Takes Everything at Your Command

[ad_1]

SAN FRANCISCO — At OpenAI, one of the world’s most ambitious artificial intelligence labs, researchers develop technologies that allow you to create digital images by simply describing what you want to see.

They both say hi to DALL-E 2008 animated movie “WALL-E” about an autonomous robot and surrealist painter Salvador Dalí.

open AI, Backed by a billion dollar fund from Microsoft, does not share the technology with the public yet. But recently, Alex Nichol, one of the researchers behind the system, demonstrated how the system works.

When he asked for an “avocado-shaped teapot,” by typing those words onto a largely blank computer screen, the system generated 10 different images of a dark green avocado teapot, some with seeds and some without. “DALL-E is good on avocado,” said Mr. Nichol.

When she wrote “Cats playing chess,” she placed two fluffy kittens on either side of a checkered game board, with 32 chess pieces lined up between them. As she called it “a teddy bear playing a trumpet underwater,” an image showed tiny air bubbles rising from the tip of the bear’s trumpet to the surface of the water.

DALL-E can also edit photos. A guitar appeared between his furry arms when Mr. Nichol wiped the teddy bear’s trumpet and asked for a guitar instead.

A team of seven researchers spent two years developing the technology that OpenAI plans to eventually offer as a tool for people like graphic artists, providing new shortcuts and new ideas when creating and editing digital images. Computer programmers already use the tool Copilot Based on OpenAI’s similar technologyto generate software code snippets.

But for many experts, DALL-E is worrisome. As this type of technology continues to evolve, they say it could help spread disinformation on the internet by fueling the kind of online campaigns that could help shake up the 2016 presidential election.

Like, “You can use it for good things, but you can certainly use it for all kinds of crazy, worrying applications, and that includes deep frauds.” misleading photos and videossaid Subbarao Kambhampati, a professor of computer science at Arizona State University.

Half a decade ago, the world’s leading artificial intelligence labs, identify objects in digital images and even creating images by yourself, including flowers, dogs, cars and faces. A few years later, they installed the following systems: could do the same with written languagesummarizing articles, answering questions, creating tweets and even writing blog posts.

Now researchers are combining these technologies to create new forms of AI. DALL-E is a remarkable step forward as it confuses both language and images, and in some cases grasps the relationship between the two.

“We can now use multiple, intersecting streams of information to create better and better technology,” said Oren Etzioni, president of the Allen Institute for Artificial Intelligence, an artificial intelligence lab in Seattle.

Technology is not perfect. Nichol didn’t quite grasp the idea when he asked DALL-E to “put the Eiffel Tower on the moon”. He placed the moon in the sky above the tower. When he asked for a “living room full of sand”, a scene arose that looked more like a construction site than a living room.

But when Mr Nichol changed his wishes a bit by adding or subtracting a few words here or there, he got what he wanted. When he asked for “a piano in a living room filled with sand,” the image looked more like a beach in the living room.

DALL-E, artificial intelligence researchers plexusIt is a mathematical system loosely modeled on the network of neurons in the brain. It’s the same technology that recognizes spoken commands on smartphones and identifies the presence of pedestrians as self-driving cars navigate city streets.

A neural network learns skills by analyzing large amounts of data. For example, he can learn to recognize an avocado by detecting patterns in thousands of avocado photos. DALL-E looks for patterns when analyzing millions of digital images and text captions that describe what each image represents. In this way, he learns to recognize the connections between images and words.

When someone defines an image for DALL-E, it creates a set of key features that image can contain. A feature might be the line on the edge of a trumpet. Another might be the curve at the top of a teddy bear’s ear.

Next, a second neural network, called the diffusion model, generates the image and generates the pixels needed to realize these features. The latest version of DALL-E, introduced Wednesday in a new research paper announcing the system, produces high-resolution images that, in most cases, look like photographs.

Although DALL-E often fails to understand what someone is describing and sometimes confuses the image it produces, OpenAI continues to improve the technology. Researchers can often improve the capabilities of a neural network by feeding larger amounts of data.

They can also build more powerful systems by applying the same concepts to new data types. The Allen Institute has recently created a system that can analyze audio as well as image and text. After analyzing millions of YouTube videos, including audio tracks and subtitles, Identify specific moments in TV shows or movieslike a barking dog or a closing door.

Experts believe that researchers will continue to develop such systems. As a result, these systems can help companies improve search engines, digital assistants, and other common technologies and automate new tasks for graphic artists, programmers, and other professionals.

But there are caveats to this potential. AI systems, in part because they can show bias against women and people of color. They learn their skills from enormous online repositories of text, images, and other data that show bias.. They can be used to create pornography, hate speech and other offensive material. And many experts believe that technology will eventually make this very easy. create disinformationPeople will have to be skeptical of almost everything they see online.

“We can fake text. We can put text in someone’s voice. We can create images and videos,” said Dr. etzioni “There is already disinformation online, but the concern is that this disinformation is taking to new levels.”

OpenAI is keeping a tight leash on the DALL-E. He would not allow outsiders to use the system on their own. It puts a watermark in the corner of every image it creates. The group will be small, though the lab plans to open the system to testers this week.

The system also includes filters that prevent users from creating images they find inappropriate. When “sheep-headed pig” was asked, he refused to produce an image. According to the lab, the combination of the words “pig” and “head” likely triggered OpenAI’s anti-bullying filters.

“It’s not a product,” said Mira Murati, OpenAI’s head of research. “The idea is to understand capabilities and limitations and give us the opportunity to improve mitigation.”

OpenAI can control the behavior of the system in some ways. But others around the world could soon create a similar technology that puts the same powers in the hands of nearly everyone. Based on a research paper describing an early version of DALL-E, Boris Dayma, an independent researcher in Houston, has already simpler version of technology.

“People need to know that the images they see may not be real,” he said.

[ad_2]

Source link

Leave a Reply

Your email address will not be published. Required fields are marked *