![]() ![]() The technology to make fake videos of real people is becoming increasingly available.Īdvances in generative AI will soon mean that fake but visually convincing content will proliferate online, leading to an even messier information ecosystem. In this new world, it will be a snap to generate a video of a CEO saying her company’s profits are down 20%, which could lead to billions in market-share loss, or to generate a video of a world leader threatening military action, which could trigger a geopolitical crisis, or to insert the likeness of anyone into a sexually explicit video. When this happens, and it is all but guaranteed to, it will become increasingly easier to distort reality. Although audio and video have not yet fully passed through the uncanny valley – images or models of people that are unsettling because they are close to but not quite realistic – they are likely to soon. My colleague Sophie Nightingale and I found that the average person is unable to reliably distinguish an image of a real person from an AI-generated person. By combining clever machine-learning algorithms with billions of pieces of human-generated content, these systems can do anything from create an eerily realistic image from a caption, synthesise a speech in President Joe Biden’s voice, replace one person’s likeness with another in a video, or write a coherent 800-word op-ed from a title prompt.Įven in these early days, generative AI is capable of creating highly realistic content. Generative AI, in the form of image generators like DALL-E, Midjourney and Stable Diffusion, and text generators like Bard, ChatGPT, Chinchilla and LLaMA, has exploded in the public sphere. They were created by a generative artificial intelligence system. These images looked like news photos, but they were fake. Shortly after rumours leaked of former President Donald Trump’s impending indictment, images purporting to show his arrest appeared online. ![]()
0 Comments
Leave a Reply. |