Can Deepfake Tech Train Computer Vision AIs?

A recent survey has revealed that 99 percent of computer-vision engineers have had a machine-learning project completely canceled due to insufficient training data. Researchers have already been looking toward synthetic data to fill this breach. In fact, Gartner foresees that by 2024, more than half of data used for AI and analytics projects will be synthetically generated.

A research group from the Massachusetts Institute of Technology has explored whether, given that generative models are now capable of producing highly realistic images, they can replace data sets to train computer-vision AI models. In their paper presented at the 2022 International Conference on Learning Representations, the MIT researchers asked whether, “given access only to a trained generative model, and no access to the dataset that trained it, can we learn effective visual representations?”

As generative models, specifically generative adversarial networks (GANs), get better and better, researchers are indeed getting performance close to that of real data for learning representations, says Ali Jahanian, a research scientist at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), and one of the authors of the paper.

GANs are deep-learning models that use two neural networks working against each other—a generator that creates synthetic data, and a discriminator that distinguishes between real and synthetic data—to generate synthetic images that are almost indistinguishable from real ones. GANs are popular for generating images and videos, including deepfakes.

The research used these pretrained models for multiview representation learning. “Their method uses a contrastive objective that pushes the model to produce a presentation that’s similar for different views of the same object,” says Anna Rumshisky, a computer scientist at the University of Massachusetts, Lowell. Jahanian and his colleagues’ study is a valuable proof of concept showing that applying contrastive methods to synthetic data allows the model to learn well without having the data set, she adds.

On some specific tasks, GANs are able to outperform real data when used for a downstream transfer learning task. For example, Nvidia’s StyleGAN can disentangle color from the objects and can even rotate the objects. “Because it is not only learning the data,” he says, “but also the transformations.”

MIT researchers have demonstrated the use of a generative machine-learning model to create synthetic data, based on real data, that can be used to train another model for image classification. This image shows examples of the generative model’s transformation methods.MIT

Synthetic data has some advantages over traditional data sets, the researchers note. For instance, not all research teams have the resources to access high-volume, high-quality data, especially when it could involve sensitive information, such as personal data. In that case, pretrained generative models can be more efficient and accessible. Synthetic data is useful to simulate conditions that may not (yet) exist. These data sets can also be edited, and potentially at scale. This allows researchers to, for example, remove biases that exist in real data sets.

Developing methods to control what the model is generating is an active research area, says Rumshisky. Speaking from the point of view of natural language processing, which is her area of expertise, she adds, “There’s been quite a few efforts in the past three years to come up with models that allow you to control what’s getting generated…to push them all towards generating certain kinds of keywords, or certain kinds of attributes.” If you had a mechanism to control the attributes of the text that the model is generating, she explains, you could perhaps push it away from generating certain kinds of data, such as, personal details like names, social security numbers, and phone numbers, for instance.

The real test is whether the performance of ML models trained on synthetic data are on par with those trained on real data. That is still a work in progress at this stage. “I am using these models on data that, for instance, comes from self-driving vehicle data,” Jahanian says. “Eventually we want the AI agent to be able to go to the real world and…solve [real] tasks.” [READ MORE]