Draw what you imagine!
What’s
New: A picture worth a thousand words now takes just three or four words to
create, thanks to GauGAN2, the latest version of NVIDIA Research’s wildly
popular AI painting demo.
Key
Insight: The deep
learning model behind GauGan allows anyone to channel their imagination into
photorealistic masterpieces — and it’s easier than ever. Simply type a phrase
like “sunset at a beach” and AI generates the scene in real time. Add an
additional adjective like “sunset at a rocky beach,” or swap “sunset” to “afternoon” or “rainy
day” and the model, based on generative adversarial networks, instantly modifies the picture
With the press of
a button, users can generate a segmentation map, a high-level outline that
shows the location of objects in the scene. From there, they can switch to
drawing, tweaking the scene with rough sketches using labels like sky, tree,
rock and river, allowing the smart paintbrush to incorporate these doodles into
stunning images.
How it works: GauGAN2 combines segmentation
mapping, inpainting and text-to-image generation in a single model, making it a
powerful tool to create photorealistic art with a mix of words and drawings.
The demo is one of the first to combine multiple modalities —
text, semantic segmentation, sketch and style — within a single GAN framework.
This makes it faster and easier to turn an artist’s vision into a high-quality
AI-generated image.
Why
it matters: Compared to state-of-the-art models specifically
for text-to-image or segmentation map-to-image applications, the neural network
behind GauGAN2 produces a greater variety and higher quality of images.
NVIDIA Research has more than 200 scientists around the globe, focused on areas including AI, computer vision, self-driving cars, robotics and graphics. Learn more about their work. Check out yourself click here
Comments
Post a Comment