Text, image, audio and video... How powerful is Microsoft's cross-modal model CoDi?

Image source: Generated by Unbounded AI

The Microsoft Azure research team and the University of North Carolina researchers published a paper "Arbitrary Generation Through Composable Diffusion", introducing a new multimodal generation model - CoDi (Composable Diffusion).

CoDi is capable of generating any combination of output modalities from any combination of input modalities, such as language, image, video, or audio. Unlike existing generative AI systems, CoDi can generate multiple modalities in parallel, and its input is not limited to subsets of modalities such as text or images. CoDi is free to condition any combination of inputs and generate any set of modalities, even if they are not present in the training data.

CoDi introduces an unprecedented level of content generation by simultaneously processing and generating multimodal content such as text, images, audio, and video. Using diffusion models and composable techniques, CoDi can generate high-quality, diverse outputs from single or multiple inputs, transforming content creation, accessibility, and personalized learning.

CoDi is highly customizable and flexible, enabling robust joint modality generation quality that outperforms or rivals state-of-the-art single modality synthesis.

Recently, CoDi has made new progress and has been officially available on the Microsoft Azure platform. It can be used for free for 12 months.

How powerful is CoDi

CoDi emerged as part of Microsoft's ambitious i-Code project, a research initiative dedicated to advancing multimodal AI capabilities. CoDi's ability to seamlessly integrate information from various sources and generate consistent output is expected to revolutionize multiple areas of human-computer interaction.

One of the areas where CoDi could bring about change is assistive technology, enabling people with disabilities to interact with computers more effectively. By seamlessly generating content across text, images, video, and audio, CoDi can provide users with a more immersive and accessible computing experience.

Additionally, CoDi has the potential to reinvent custom learning tools by providing a comprehensive interactive learning environment. Students engage with multimodal content that seamlessly integrates information from a variety of sources, enhancing their understanding and engagement with the topic.

CoDi will also revolutionize content generation. The model is able to generate high-quality output across multiple modalities, which can simplify the content creation process and reduce the burden on creators. Whether generating engaging social media posts, crafting interactive multimedia presentations, or creating engaging storytelling experiences, CoDi's capabilities have the potential to reshape the content generation landscape.

To address the limitations of traditional unimodal AI models, CoDi provides a solution to the tedious and slow process of combining modality-specific generative models.

This novel model employs a unique composable generation strategy that bridges alignment during diffusion and facilitates simultaneous generation of interwoven modalities, such as time-aligned video and audio.

CoDi's model training process is also quite distinctive. It involves projecting input modalities such as image, video, audio, and language into a common semantic space. This allows for flexible handling of multimodal inputs, and through the cross-attention module and the environment encoder, it is able to simultaneously generate arbitrary combinations of output modalities.

(Above) CoDi's model architecture: CoDi uses a multi-stage training scheme capable of training only on a linear number of tasks but inferring on all combinations of input and output modalities.

丨Single or multiple inputs --> multiple outputs

CoDi models can take single or multiple cues (including video, image, text, or audio) to generate multiple aligned outputs, such as video with accompanying sound.

For example:

1. Text+Image+Audio——>Audio+Video

"A teddy bear on a skateboard, 4k, high resolution" + a picture of Times Square in New York + a rainy audio --> After CoDi generation, a piece of "A teddy bear skateboards in Times Square in the rain, Accompanied by the simultaneous sound of rain and street noise."

How is it generated?

CoDi can jointly generate any combination of video, image, audio, and text via composable diffusion. CoDi first receives audio tracks to generate text subtitles, then receives images for image+audio-audio, and then receives image+audio+text to combine their information to generate a new joint image+subtitle. Finally, CoDi can also receive image+audio+text and generate video+audio.

2 text+audio+image -->text+image

1. 3. Audio + Image --> Text + Image

1. 4. Text+Image ——>Text+Image

5. Text——>Video+Audio

6. Text——>Text+Audio+Image

丨Multiple inputs --> single output

1. Text+Audio——Image

2. Text + Image --> Image

3 Text+Audio -->Video

4 text + image --> video

5. There are also video + audio --> text, image + audio --> audio, text + image --> audio...etc

丨Single input——single output

1 Text --> Image

2 Audio --> Image

3 Image --> Video

4 Image --> Audio

5 Audio --> Text

6 Image --> Text

References:

View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Repost
  • Share
Comment
0/400
No comments
Trade Crypto Anywhere Anytime
qrCode
Scan to download Gate App
Community
English
  • 简体中文
  • English
  • Tiếng Việt
  • 繁體中文
  • Español
  • Русский
  • Français (Afrique)
  • Português (Portugal)
  • Bahasa Indonesia
  • 日本語
  • بالعربية
  • Українська
  • Português (Brasil)