AI ML DL

AI ML DL

Low-Cost Deepfake Video Detection With H.264 Motion Vectors

Though it’s possible to detect deepfake video by studying movement in a deepfaked face, it takes a lot of resources to do it, and it’s not easy or cheap to use it to detect potential ‘live’ deepfakes in video streaming calls. However, the most popular video codec in the world, H.264, has already done a lot of the work necessary to discern deepfake content, and new research offers a way to cheaply use these ‘motion vectors’ to discern simulated faces.

Read More »
AI ML DL

Detecting When ‘Fake’ Images Are Actually Real

A new generation of fake image detectors hope to discern AI-generated photos from real ones. But how do they cope when real photos exhibit qualities normally found in fake photos? A new an seminal database of ‘impossible-but-true’ photos hopes to stir deeper interest in the nascent field of photo verification.

Read More »
AI ML DL

Improving Deepfaked Eye Contact via Sound Analysis

Bad rendering of eye direction in neural facial synthesis can make the difference between an image looking photoreal and looking like bad CGI. We are experts in eye contact, but machine learning systems only have small pixel spaces in which to attempt to recreate our survival-level understanding of eye gaze, and often make rudimentary mistakes, such as conflating head position with eye position. But a new paper is using sound as an extra factor to guess where people in a video are really looking, offering potential for improved gaze estimation in downstream applications.

Read More »
AI ML DL

Restoring Facial Expressions with CycleGAN

Researchers from Germany have developed a new method of restoring obstructed faces using only Generative Adversarial Networks (GANs), instead of needing to undertake expensive and time-consuming fine-training of existing models.

Read More »
AI ML DL

CGI-Style Object Control With Stable Diffusion

Though Stable Diffusion is an impressive generative system, it has difficulty performing the same operation twice, which makes it difficult to render the same subject consistently across frames – a major obstacle to the development of latent diffusion models aimed at producing coherent video. Now, a new system is offering a method that can output image concepts in a way that’s more consistent and repeatable, as CGI is.

Read More »
AI ML DL

A Dedicated Loss Function for Neural Face Training

Loss functions, which determine the ways that a machine learning network should develop during the course of training, are at the heart of generative AI and computer vision systems. However, the loss functions commonly used in facial neural synthesis are not specifically designed for faces, which can be a retarding and even confusing factor in framework development. Now, Disney Research and EFL Zurich are proposing an novel architecture-agnostic loss function designed specifically for facial workflows, and which is based on 1990s research into the way humans perceive 3D faces based on shading.

Read More »
AI ML DL

Improving Stable Diffusion With Better Captions

New work from Google Research rewrites the terrible alternative captions that are often found in images on the web, and which subsequently get used as pertinent information for generative models such as Stable Diffusion, so that they contain richer and more useful information. The authors found that fine-tuning Stable Diffusion on these better-annotated images wrought significant improvements in the generated images.

Read More »
AI ML DL

Restoring Archival Video with CLIP

New research from Italy leverages Transformers to offer a more intelligent and acerbic method of restoring damaged archival video – a goal that could help VFX practitioners that need to acquire better historical data of actors and other potential subjects for whom archive material is often not in the best state.

Read More »
AI ML DL

The Struggle for Salient Image-Cropping in Generative AI

Generative AI very often produces badly-cropped images, because the models are trained on non-square images which have been automatically (and, usually, quite randomly) cropped by the training process. Though it’s possible to devise methods for better cropping, many of them are compute-intensive, and would radically raise the cost for training hyperscale datasets in systems such as Stable Diffusion. But now, researchers from Japan propose a more economical yet intelligent way of performing salient crops on large amounts of data.

Read More »