Sora: OpenAI’s Groundbreaking Text-to-Image Diffusion Model

18 February 2024
openai sora

Sora: OpenAI’s Groundbreaking Text-to-Image Diffusion Model

OpenAI has unveiled Sora, a diffusion-based text-to-image model capable of generating 60-second videos. Compared to competitors like Runway, Pika, Stability AI, and Google, OpenAI’s model boasts high-resolution (Full HD) output,…

Microsoft DragNUWA: Video Generation via Object Trajectories

15 January 2024

Microsoft DragNUWA: Video Generation via Object Trajectories

Microsoft has released the DragNUWA weights – a cross-domain video generation model that offers more precise control over the resulting output compared to similar models. Control is achieved by simultaneously…

Pika 1.0: A Web Platform for Video Generation

7 January 2024
pika

Pika 1.0: A Web Platform for Video Generation

Pika Labs startup has launched Pika 1.0 – a free web platform for generating and editing videos using text-based queries. The service creates both realistic videos and 3D animation in…

VideoPoet: Google’s Language Model for Video Generation and Editing

23 December 2023
videopoet

VideoPoet: Google’s Language Model for Video Generation and Editing

Google has unveiled VideoPoet, a language model for multimodal video content processing capable of turning text and images into clips, styling pre-existing videos, and generating soundtrack for them without any…

Stable Video Diffusion: Stability AI’s Image-Based Video Generator

26 November 2023
Stable Video Diffusion

Stable Video Diffusion: Stability AI’s Image-Based Video Generator

Stability AI has announced the release of Stable Video Diffusion, a duo of models that generate up to 4-second videos from an input image. Both models are available publicly. Importantly,…

“Deepdub Go” Empowers Content Creators with AI for Video Dubbing

9 July 2023
ai for video dubbing - neural network based service

“Deepdub Go” Empowers Content Creators with AI for Video Dubbing

Israeli startup Deepdub has unveiled its groundbreaking service, Deepdub Go, which utilizes AI for dubbing to automatically dub videos in 65 languages. This innovative platform targets game development studios, advertising…

MAGVIT: Open Source Generative Video Transformer 10-in-1

29 June 2023
MAGVIT

MAGVIT: Open Source Generative Video Transformer 10-in-1

Researchers from Carnegie Mellon University, Google Research, and the University of Georgia have introduced MAGVIT (Masked Generative Video Transformer), an open-source video generation model. MAGVIT is a unified model that…

New Datasets for Object Tracking

8 November 2018

New Datasets for Object Tracking

Object tracking in the wild is far from being solved. Existing object trackers do quite a good job on the established datasets (e.g., VOT, OTB), but these datasets are relatively…

3D Hair Reconstruction Out of In-the-Wild Videos

22 October 2018
hair reconstruction from video

3D Hair Reconstruction Out of In-the-Wild Videos

3D hair reconstruction is a problem with numerous applications in different areas such as Virtual Reality, Augmented Reality, video games, medical software, etc. As a non-trivial problem, researchers have proposed…

Vid2Vid – Conditional GANs for Video-to-Video Synthesis

3 September 2018
vid2vid-video-to-video-synthesis-e1535641547242

Vid2Vid – Conditional GANs for Video-to-Video Synthesis

Researchers from NVIDIA and MIT’s Computer Science and Artificial Intelligence Lab have proposed a novel method for video-to-video synthesis, showing impressive results. The proposed method – Vid2Vid – can synthesize…

Everybody Dance Now: a New Approach to “Do As I Do” Motion Transfer

30 August 2018
everybody dance now

Everybody Dance Now: a New Approach to “Do As I Do” Motion Transfer

Not very good at dancing? Not a problem anymore! Now you can easily impress your friends with a stunning video, where you dance like a superstar. Researchers from UC Berkeley…

ReCoNet: Fast and Accurate Real-time Video Style Transfer

25 July 2018
video style transfer reconet

ReCoNet: Fast and Accurate Real-time Video Style Transfer

A real-time coherent video style transfer network – ReCoNet – is proposed by a group of researchers from the University of Hong Kong as a state-of-the-art approach to video style…