{"id":13645,"date":"2021-10-27T14:57:56","date_gmt":"2021-10-27T14:57:56","guid":{"rendered":"https:\/\/test.nahtnow.com\/?p=13645"},"modified":"2021-10-27T14:57:56","modified_gmt":"2021-10-27T14:57:56","slug":"fake-news","status":"publish","type":"post","link":"https:\/\/test.nahtnow.com\/en\/fake-news\/","title":{"rendered":"FAKE NEWS"},"content":{"rendered":"<h2 class=\"hP\" tabindex=\"-1\" data-thread-perm-id=\"thread-f:1713345866829273912|msg-f:1713345866829273912\" data-legacy-thread-id=\"17c706fcf9be1b38\">FAKE NEWS Is Going To Wreak Havoc On Society. We Are Not Prepared.<\/h2>\n<p>Last month during ESPN\u2019s hit documentary series <em>The Last Dance<\/em>, State Farm debuted <a title=\"https:\/\/www.youtube.com\/watch?v=FzOVqClci_s\" href=\"https:\/\/www.youtube.com\/watch?v=FzOVqClci_s\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"a TV commercial\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.youtube.com\/watch?v%3DFzOVqClci_s&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNHAC-cbUb3wKxq1zsT-Ii8YE8OXcA\">a TV commercial<\/a> that has become one of the <a title=\"https:\/\/www.adweek.com\/brand-marketing\/state-farm-and-kenny-mayne-brilliantly-faked-us-all-out-during-the-last-dance\/\" href=\"https:\/\/www.adweek.com\/brand-marketing\/state-farm-and-kenny-mayne-brilliantly-faked-us-all-out-during-the-last-dance\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"most widely discussed\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.adweek.com\/brand-marketing\/state-farm-and-kenny-mayne-brilliantly-faked-us-all-out-during-the-last-dance\/&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNFN6pBHcg4vhU8aO8NbA18oyHtCgA\">most widely discussed<\/a> ads in recent memory. It appeared to show footage from 1998 of an ESPN analyst making shockingly accurate predictions about the year 2020.<\/p>\n<p>As it turned out, the clip was not genuine: it was generated using cutting-edge AI. The commercial surprised, amused and delighted viewers.<\/p>\n<p>What viewers should have felt, though, was deep concern.<\/p>\n<p>The State Farm ad was a benign example of an important and dangerous new phenomenon in AI: deepfakes. Deepfake technology enables anyone with a computer and an Internet connection to create realistic-looking photos and videos of people saying and doing things that they did not actually say or do.<\/p>\n<div>\n<p aria-hidden=\"true\">A combination of the phrases \u201cdeep learning\u201d and \u201cfake\u201d, deepfakes <a title=\"https:\/\/knowyourmeme.com\/memes\/cultures\/deepfakes\" href=\"https:\/\/knowyourmeme.com\/memes\/cultures\/deepfakes\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"first emerged\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/knowyourmeme.com\/memes\/cultures\/deepfakes&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNFaDzBk_oSBWjXsXDY35iVOjOmIPA\">first emerged<\/a> on the Internet in late 2017, powered by an innovative new deep learning method known as generative adversarial networks (GANs).<\/p>\n<\/div>\n<p>Several deepfake videos have gone viral recently, giving millions around the world their first taste of this new technology: President Obama <a title=\"https:\/\/www.youtube.com\/watch?v=cQ54GDm1eL0&amp;feature=youtu.be\" href=\"https:\/\/www.youtube.com\/watch?v=cQ54GDm1eL0&amp;feature=youtu.be\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"using\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.youtube.com\/watch?v%3DcQ54GDm1eL0%26feature%3Dyoutu.be&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNG8TIbvKB3WwkJ52q68_uaI1dXJCA\">using<\/a> an expletive to describe President Trump, Mark Zuckerberg <a title=\"https:\/\/www.instagram.com\/p\/BypkGIvFfGZ\/?utm_source=ig_embed&amp;utm_campaign=embed_video_watch_again\" href=\"https:\/\/www.instagram.com\/p\/BypkGIvFfGZ\/?utm_source=ig_embed&amp;utm_campaign=embed_video_watch_again\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"admitting\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.instagram.com\/p\/BypkGIvFfGZ\/?utm_source%3Dig_embed%26utm_campaign%3Dembed_video_watch_again&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNE-6rkHeV1-O6Q34t-VYRvUYUCBpQ\">admitting<\/a> that Facebook&#8217;s true goal is to manipulate and exploit its users, Bill Hader <a title=\"https:\/\/www.youtube.com\/watch?v=kjI-JaRWG7s&amp;feature=youtu.be\" href=\"https:\/\/www.youtube.com\/watch?v=kjI-JaRWG7s&amp;feature=youtu.be\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"morphing into\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.youtube.com\/watch?v%3DkjI-JaRWG7s%26feature%3Dyoutu.be&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNGwZggYo30eU-Y2mdVuQmQUjjkx9Q\">morphing into<\/a> Al Pacino on a late-night talk show.<\/p>\n<div>\n<div id=\"m_-3288479769505317452gmail-recirc-unit\">\n<div>MORE FOR YOU<\/div>\n<div>\n<div>\n<h3>Why It\u2019s Important To Push Back On \u2018Plandemic\u2019\u2014And How To Do It<\/h3>\n<\/div>\n<div>\n<h3>These Are The Startups Applying AI To Transform Healthcare<\/h3>\n<\/div>\n<div>\n<h3>Trump Threatens To Issue Executive Order Preventing Biden From Being Elected President<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<p>The amount of deepfake content online is growing at a rapid rate. At the beginning of 2019 there were 7,964 deepfake videos online, according to <a title=\"https:\/\/deeptracelabs.com\/mapping-the-deepfake-landscape\/\" href=\"https:\/\/deeptracelabs.com\/mapping-the-deepfake-landscape\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"a report\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/deeptracelabs.com\/mapping-the-deepfake-landscape\/&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNHl6pzzrOXdmy8FrFOs33f7F44qDg\">a report<\/a> from startup Deeptrace; just nine months later, that figure had jumped to 14,678. It has no doubt continued to balloon since then.<\/p>\n<p>While impressive, today&#8217;s deepfake technology is still not quite to parity with authentic video footage\u2014by looking closely, it is typically possible to tell that a video is a deepfake. But the technology is improving at a breathtaking pace. Experts predict that deepfakes will be indistinguishable from real images before long.<\/p>\n<p>\u201cIn January 2019, deep fakes were buggy and flickery,\u201d <a title=\"https:\/\/www.ft.com\/content\/4bf4277c-f527-11e9-a79c-bc9acae3b654\" href=\"https:\/\/www.ft.com\/content\/4bf4277c-f527-11e9-a79c-bc9acae3b654\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"said\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.ft.com\/content\/4bf4277c-f527-11e9-a79c-bc9acae3b654&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNEn22bAxXPy-_5y3jRz9pwbObPbjg\">said<\/a> Hany Farid, a UC Berkeley professor and deepfake expert. \u201cNine months later, I\u2019ve never seen anything like how fast they\u2019re going. This is the tip of the iceberg.\u201d<\/p>\n<p>Today we stand at an inflection point. In the months and years ahead, deepfakes threaten to grow from an Internet oddity to a widely destructive political and social force. Society needs to act now to prepare itself.<\/p>\n<p>When Seeing Is Not Believing<\/p>\n<p>The first use case to which deepfake technology has been widely applied\u2014<a title=\"https:\/\/www.theatlantic.com\/technology\/archive\/2016\/06\/how-porn-leads-people-to-upgrade-their-tech\/486032\/\" href=\"https:\/\/www.theatlantic.com\/technology\/archive\/2016\/06\/how-porn-leads-people-to-upgrade-their-tech\/486032\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"as is often the case\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.theatlantic.com\/technology\/archive\/2016\/06\/how-porn-leads-people-to-upgrade-their-tech\/486032\/&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNGMt0cPPuVHrSyre98Fm23TYM1hLQ\">as is often the case<\/a> with new technologies\u2014is pornography. As of September 2019, 96% of deepfake videos online were pornographic, according to the Deeptrace report.<\/p>\n<p>A handful of websites dedicated specifically to deepfake pornography have emerged, collectively garnering hundreds of millions of views over the past two years. Deepfake pornography is almost always non-consensual, involving the <a title=\"https:\/\/www.bbc.com\/news\/technology-42912529\" href=\"https:\/\/www.bbc.com\/news\/technology-42912529\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"artificial synthesis\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.bbc.com\/news\/technology-42912529&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNFvcCISWS7YHz2ecOV2-5MneAfqtQ\">artificial synthesis<\/a> of explicit videos that feature famous celebrities or personal contacts.<\/p>\n<p>From these dark corners of the web, the use of deepfakes has begun to spread to the political sphere, where the potential for mayhem is even greater.<\/p>\n<p>It does not require much imagination to grasp the harm that could be done if entire populations can be shown fabricated videos that they believe are real. Imagine deepfake footage of a politician engaging in bribery or sexual assault right before an election; or of U.S. soldiers committing atrocities against civilians overseas; or of President Trump declaring the launch of nuclear weapons against North Korea. In a world where even some uncertainty exists as to whether such clips are authentic, the consequences could be catastrophic.<\/p>\n<p>Because of the technology\u2019s widespread accessibility, such footage could be created by anyone: state-sponsored actors, political groups, lone individuals.<\/p>\n<p>In a recent report, The Brookings Institution <a title=\"https:\/\/www.brookings.edu\/research\/is-seeing-still-believing-the-deepfake-challenge-to-truth-in-politics\/#cancel\" href=\"https:\/\/www.brookings.edu\/research\/is-seeing-still-believing-the-deepfake-challenge-to-truth-in-politics\/#cancel\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"grimly summed up\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.brookings.edu\/research\/is-seeing-still-believing-the-deepfake-challenge-to-truth-in-politics\/%23cancel&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNGO2svFMnFRYBlB0AltjSYh08Fxng\">grimly summed up<\/a> the range of political and social dangers that deepfakes pose: \u201cdistorting democratic discourse; manipulating elections; eroding trust in institutions; weakening journalism; exacerbating social divisions; undermining public safety; and inflicting hard-to-repair damage on the reputation of prominent individuals, including elected officials and candidates for office.\u201d<\/p>\n<p>Given the stakes, U.S. lawmakers have begun to pay attention.<\/p>\n<p>\u201cIn the old days, if you wanted to threaten the United States, you needed 10 aircraft carriers, and nuclear weapons, and long-range missiles,\u201d U.S. Senator Marco Rubio <a title=\"https:\/\/www.csoonline.com\/article\/3293002\/deepfake-videos-how-and-why-they-work.html\" href=\"https:\/\/www.csoonline.com\/article\/3293002\/deepfake-videos-how-and-why-they-work.html\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"said\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.csoonline.com\/article\/3293002\/deepfake-videos-how-and-why-they-work.html&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNGxG9uDoEdi65L9jg8dtFnvtzlbmQ\">said<\/a> recently. \u201cToday&#8230;.all you need is the ability to produce a very realistic fake video that could undermine our elections, that could throw our country into tremendous crisis internally and weaken us deeply.\u201d<\/p>\n<p>Technologists agree. <a title=\"https:\/\/www.cnn.com\/interactive\/2019\/01\/business\/pentagons-race-against-deepfakes\/\" href=\"https:\/\/www.cnn.com\/interactive\/2019\/01\/business\/pentagons-race-against-deepfakes\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"In the words\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.cnn.com\/interactive\/2019\/01\/business\/pentagons-race-against-deepfakes\/&amp;source=gmail&amp;ust=1635342840828000&amp;usg=AFQjCNGwRDhW5T-0AniOQqExVMaYDgbaew\">In the words<\/a> of Hani Farid, one of the world&#8217;s leading experts on deepfakes: \u201cIf we can&#8217;t believe the videos, the audios, the image, the information that is gleaned from around the world, that is a serious national security risk.\u201d<\/p>\n<p>This risk is no longer just hypothetical: there are early examples of deepfakes influencing politics in the real world. Experts warn that these incidents are canaries in a coal mine.<\/p>\n<p>Last month, a political group in Belgium released <a title=\"https:\/\/tube.rebellion.global\/videos\/watch\/2ad12b6b-bb53-473c-ad74-14eef02874b5?title=0&amp;warningTitle=0\" href=\"https:\/\/tube.rebellion.global\/videos\/watch\/2ad12b6b-bb53-473c-ad74-14eef02874b5?title=0&amp;warningTitle=0\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"a deepfake video\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/tube.rebellion.global\/videos\/watch\/2ad12b6b-bb53-473c-ad74-14eef02874b5?title%3D0%26warningTitle%3D0&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNGhP1G2tDKRfkFsQ-anciZl6WXjoA\">a deepfake video<\/a> of the Belgian prime minister giving a speech that linked the COVID-19 outbreak to environmental damage and called for drastic action on climate change. At least some viewers believed the speech was real.<\/p>\n<p>Even more insidiously, the mere possibility that a video <em>could<\/em> be a deepfake can stir confusion and facilitate political deception regardless of whether deepfake technology has actually been used. The most dramatic example of this comes from Gabon, a small country in central Africa.<\/p>\n<p>In late 2018, Gabon&#8217;s president Ali Bongo had not been seen in public for months. Rumors were swirling that he was no longer healthy enough for office or even that he had died. In an attempt to allay these concerns and reassert Bongo\u2019s leadership over the country, his administration announced that he would give a nationwide televised address on New Years Day.<\/p>\n<p>In the video address (which is worth <a title=\"https:\/\/www.facebook.com\/watch\/?v=324528215059254\" href=\"https:\/\/www.facebook.com\/watch\/?v=324528215059254\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"examining firsthand\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.facebook.com\/watch\/?v%3D324528215059254&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNHvJKoduwu1354td2qJoJp9ppBHMg\">examining firsthand<\/a> yourself), Bongo appears stiff and stilted, with unnatural speech and facial mannerisms. The video immediately inflamed suspicions that the government was concealing something from the public. Bongo\u2019s political opponents <a title=\"https:\/\/www.washingtonpost.com\/politics\/2020\/02\/13\/how-sick-president-suspect-video-helped-sparked-an-attempted-coup-gabon\/\" href=\"https:\/\/www.washingtonpost.com\/politics\/2020\/02\/13\/how-sick-president-suspect-video-helped-sparked-an-attempted-coup-gabon\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"declared\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.washingtonpost.com\/politics\/2020\/02\/13\/how-sick-president-suspect-video-helped-sparked-an-attempted-coup-gabon\/&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNHwCeIOrMeZ5tnXpsYtXHuA64JNvA\">declared<\/a> that the footage was a deepfake and that the president was incapacitated or dead. Rumors of a deepfake conspiracy <a title=\"https:\/\/twitter.com\/Archyppe\/status\/1080483842383904768\" href=\"https:\/\/twitter.com\/Archyppe\/status\/1080483842383904768\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"spread quickly\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/twitter.com\/Archyppe\/status\/1080483842383904768&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNHC4nwOZfPMuzyVAHq67Na8uDIB8Q\">spread quickly<\/a> on social media.<\/p>\n<p>The political situation in Gabon rapidly destabilized. Within a week, the military had launched a coup\u2014the first in the country since 1964\u2014citing the New Years video as proof that something was amiss with the president.<\/p>\n<p>To this day experts <a title=\"https:\/\/www.motherjones.com\/politics\/2019\/03\/deepfake-gabon-ali-bongo\/\" href=\"https:\/\/www.motherjones.com\/politics\/2019\/03\/deepfake-gabon-ali-bongo\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"cannot definitively say\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.motherjones.com\/politics\/2019\/03\/deepfake-gabon-ali-bongo\/&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNGpqTGMBsl56_MRSlIkEyhhV3ypUQ\">cannot definitively say<\/a> whether the New Years video was authentic, though most believe that it was. (The coup proved unsuccessful; Bongo has since appeared in public and remains in office today).<\/p>\n<p>But whether the video was real is almost beside the point. The larger lesson is that the emergence of deepfakes will make it increasingly difficult for the public to distinguish between what is real and what is fake, a situation that political actors will inevitably exploit\u2014with potentially devastating consequences.<\/p>\n<p>\u201cPeople are already using the fact that deepfakes exist to discredit genuine video evidence,\u201d <a title=\"https:\/\/www.bbc.com\/news\/business-51204954\" href=\"https:\/\/www.bbc.com\/news\/business-51204954\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"said\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.bbc.com\/news\/business-51204954&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNEfPRaV5iVUM-NpYDlxu1fUJJPUxg\">said<\/a> USC professor Hao Li. \u201cEven though there\u2019s footage of you doing or saying something, you can say it was a deepfake and it&#8217;s very hard to prove otherwise.\u201d<\/p>\n<p>In two recent incidents, politicians <a title=\"https:\/\/www.malaymail.com\/news\/malaysia\/2019\/06\/12\/is-the-political-aide-viral-sex-video-confession-real-or-a-deepfake\/1761422\" href=\"https:\/\/www.malaymail.com\/news\/malaysia\/2019\/06\/12\/is-the-political-aide-viral-sex-video-confession-real-or-a-deepfake\/1761422\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"in Malaysia\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.malaymail.com\/news\/malaysia\/2019\/06\/12\/is-the-political-aide-viral-sex-video-confession-real-or-a-deepfake\/1761422&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNEcwiytoSe-J5towgJ5zfJYrqLskQ\">in Malaysia<\/a> and <a title=\"https:\/\/www.bbc.com\/news\/business-51204954\" href=\"https:\/\/www.bbc.com\/news\/business-51204954\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"in Brazil\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.bbc.com\/news\/business-51204954&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNEfPRaV5iVUM-NpYDlxu1fUJJPUxg\">in Brazil<\/a> have sought to evade the consequences of compromising video footage by claiming that the videos were deepfakes. In both cases, no one has been able to definitively establish otherwise\u2014and public opinion has remained divided.<\/p>\n<p>Researcher Aviv Ovadya warns of what <a title=\"https:\/\/www.washingtonpost.com\/technology\/2019\/06\/12\/top-ai-researchers-race-detect-deepfake-videos-we-are-outgunned\/\" href=\"https:\/\/www.washingtonpost.com\/technology\/2019\/06\/12\/top-ai-researchers-race-detect-deepfake-videos-we-are-outgunned\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"she terms\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.washingtonpost.com\/technology\/2019\/06\/12\/top-ai-researchers-race-detect-deepfake-videos-we-are-outgunned\/&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNFJj92NyuXGN4QoMALZgNkL0Bc9rA\">she terms<\/a> \u201creality apathy\u201d: \u201cIt\u2019s too much effort to figure out what\u2019s real and what\u2019s not, so you\u2019re more willing to just go with whatever your previous affiliations are.\u201d<\/p>\n<p>In a world in which seeing is no longer believing, the ability for a large community to agree on what is true\u2014much less to engage in constructive dialogue about it\u2014suddenly seems precarious.<\/p>\n<p>A Game of Technological Cat-And-Mouse<\/p>\n<p>The core technology that makes deepfakes possible is a branch of deep learning known as generative adversarial networks (GANs). GANs <a title=\"https:\/\/arxiv.org\/abs\/1406.2661\" href=\"https:\/\/arxiv.org\/abs\/1406.2661\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"were invented\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/arxiv.org\/abs\/1406.2661&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNGb-DDjvrZWT7Z09vY9xncGnNU-Gg\">were invented<\/a> by Ian Goodfellow in 2014 during his PhD studies at the University of Montreal, one of the world&#8217;s top AI research institutes.<\/p>\n<p>In 2016, AI great Yann LeCun <a title=\"https:\/\/www.quora.com\/What-are-some-recent-and-potentially-upcoming-breakthroughs-in-deep-learning\" href=\"https:\/\/www.quora.com\/What-are-some-recent-and-potentially-upcoming-breakthroughs-in-deep-learning\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"called GANs\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.quora.com\/What-are-some-recent-and-potentially-upcoming-breakthroughs-in-deep-learning&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNHS2EyEGR0FPInXCmdehUuQ019BhA\">called GANs<\/a> \u201cthe most interesting idea in the last ten years in machine learning.\u201d<\/p>\n<p>Before the development of GANs, neural networks were adept at classifying existing content (for instance, understanding speech or recognizing faces) but not at creating new content. GANs gave neural networks the power not just to perceive, but to create.<\/p>\n<p>Goodfellow\u2019s conceptual breakthrough was to architect GANs using two separate neural networks\u2014one known as the \u201cgenerator\u201d, the other known as the \u201cdiscriminator\u201d\u2014and pit them against one another.<\/p>\n<p>Starting with a given dataset (say, a collection of photos of human faces), the generator begins generating new images that, in terms of pixels, are mathematically similar to the existing images. Meanwhile, the discriminator is fed photos without being told whether they are from the original dataset or from the generator&#8217;s output; its task is to identify which photos have been synthetically generated.<\/p>\n<p>As the two networks iteratively work against one another\u2014the generator trying to fool the discriminator, the discriminator trying to suss out the generator\u2019s creations\u2014they hone one another\u2019s capabilities. Eventually the discriminator\u2019s classification success rate falls to 50%, no better than random guessing, meaning that the synthetically generated photos have become indistinguishable from the originals.<\/p>\n<p>One reason deepfakes have proliferated is the machine learning community\u2019s open-source ethos: starting with Goodfellow\u2019s original paper, whenever a research advance in generative modeling occurs, the technology is generally made available for free for anyone in the world to download and make use of.<\/p>\n<p>Given that deepfakes are based on AI in the first place, some look to AI as a solution to harmful deepfake applications. For instance, researchers have built <a title=\"https:\/\/arxiv.org\/abs\/1912.11035\" href=\"https:\/\/arxiv.org\/abs\/1912.11035\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"sophisticated deepfake detection systems\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/arxiv.org\/abs\/1912.11035&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNFszGp9P8KHbDGuC93N_D0OJ6KazA\">sophisticated deepfake detection systems<\/a> that assess lighting, shadows, facial movements, and other features in order to flag images that are fabricated. Another <a title=\"https:\/\/arxiv.org\/pdf\/2003.01279.pdf\" href=\"https:\/\/arxiv.org\/pdf\/2003.01279.pdf\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"innovative defensive approach\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/arxiv.org\/pdf\/2003.01279.pdf&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNF8Jdg5xRYVpAc7cBKccQc3YMVxVg\">innovative defensive approach<\/a> is to add a filter to an image file that makes it impossible to use that image to generate a deepfake.<\/p>\n<p>A handful of startups have emerged that offer software to defend against deepfakes, including Truepic and Deeptrace.<\/p>\n<p>Yet such technological solutions are not likely to stem the spread of deepfakes over the long term. At best they will lead to an endless cat-and-mouse dynamic, similar to what exists in cybersecurity today, in which breakthroughs on the deepfake detection side spur further innovation in deepfake generation. The open-source nature of AI research makes this all the more likely.<\/p>\n<p>To give <a title=\"https:\/\/www.brookings.edu\/research\/fighting-deepfakes-when-detection-fails\/\" href=\"https:\/\/www.brookings.edu\/research\/fighting-deepfakes-when-detection-fails\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"one example\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.brookings.edu\/research\/fighting-deepfakes-when-detection-fails\/&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNFDFvMXs5ZbefQblwa5V0B53uKFcA\">one example<\/a>, in 2018 researchers at the University of Albany published analysis showing that blinking irregularities were often a telltale sign that a video was fake. It was a helpful breakthrough in the fight against deepfakes\u2014until, within months, new deepfake videos began to emerge that corrected for this blinking imperfection.<\/p>\n<p>\u201cWe are outgunned,\u201d said Farid. \u201cThe number of people working on the video-synthesis side, as opposed to the detector side, is 100 to 1.\u201d<\/p>\n<p>The Path Forward<\/p>\n<p>Looking beyond purely technological remedies, what legislative, political, and social steps can we take to defend against deepfakes\u2019 dangers?<\/p>\n<p>One tempting, simple solution is to pass laws that make it illegal to create or spread deepfakes. The state of California has experimented with this approach, <a title=\"https:\/\/www.theguardian.com\/us-news\/2019\/oct\/07\/california-makes-deepfake-videos-illegal-but-law-may-be-hard-to-enforce\" href=\"https:\/\/www.theguardian.com\/us-news\/2019\/oct\/07\/california-makes-deepfake-videos-illegal-but-law-may-be-hard-to-enforce\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"enacting a law last year\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.theguardian.com\/us-news\/2019\/oct\/07\/california-makes-deepfake-videos-illegal-but-law-may-be-hard-to-enforce&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNH4fZVWg96I0owTjRQtcUv2ViHg1w\">enacting a law last year<\/a> that makes it illegal to create or distribute deepfakes of politicians within 60 days of an election. But a blanket deepfake ban has both constitutional and practical challenges.<\/p>\n<p>The First Amendment of the U.S. Constitution enshrines the freedom of expression. Any law proscribing online content, particularly political content, risks running afoul of these constitutional protections.<\/p>\n<p>\u201cPolitical speech enjoys the highest level of protection under U.S. law,\u201d said law professor Jane Kirtley. \u201cThe desire to protect people from deceptive content in the run-up to an election is very strong and very understandable, but I am skeptical about whether they are going to be able to enforce [the California] law.\u201d<\/p>\n<p>Beyond constitutional concerns, deepfake bans will likely prove impracticable to enforce due to the anonymity and borderlessness of the Internet.<\/p>\n<p>Other existing legal frameworks that might be deployed to combat deepfakes include copyright, defamation and <a title=\"https:\/\/corporate.findlaw.com\/litigation-disputes\/right-of-publicity.html\" href=\"https:\/\/corporate.findlaw.com\/litigation-disputes\/right-of-publicity.html\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"the right of publicity\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/corporate.findlaw.com\/litigation-disputes\/right-of-publicity.html&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNHcffRnOs1yZTweaCwvQrupfpVd6g\">the right of publicity<\/a>. But given the broad applicability of the <a title=\"https:\/\/www.copyright.gov\/fair-use\/more-info.html\" href=\"https:\/\/www.copyright.gov\/fair-use\/more-info.html\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"fair use doctrine\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.copyright.gov\/fair-use\/more-info.html&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNH_WhAl_pk1_XTbEG5Sfy_7uu5_cA\">fair use doctrine<\/a>, the usefulness of these legal avenues may be limited.<\/p>\n<p>In the short term, the most effective solution may come from major tech platforms like Facebook, Google and Twitter voluntarily taking more rigorous action to limit the spread of harmful deepfakes.<\/p>\n<p>Relying on private companies to solve broad political and societal problems understandably makes many deeply uncomfortable. Yet as legal scholars Bobby Chesney and Danielle Citron <a title=\"https:\/\/www.brookings.edu\/research\/is-seeing-still-believing-the-deepfake-challenge-to-truth-in-politics\/#cancel\" href=\"https:\/\/www.brookings.edu\/research\/is-seeing-still-believing-the-deepfake-challenge-to-truth-in-politics\/#cancel\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"put it\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.brookings.edu\/research\/is-seeing-still-believing-the-deepfake-challenge-to-truth-in-politics\/%23cancel&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNFx4h7VV8KiOXEMP3Re13L6CVhm9w\">put it<\/a>, these tech platforms\u2019 terms-of-service agreements are \u201cthe single most important documents governing digital speech in today\u2019s world.\u201d As a result, these companies\u2019 content policies may be \u201cthe most salient response mechanism of all\u201d to deepfakes.<\/p>\n<p>A related legislative option is to amend the controversial <a title=\"https:\/\/www.law.cornell.edu\/uscode\/text\/47\/230\" href=\"https:\/\/www.law.cornell.edu\/uscode\/text\/47\/230\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"Section 230 of the Communications Decency Act\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.law.cornell.edu\/uscode\/text\/47\/230&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNFg8cg9Ada9X8QXFdySXLZ3ol2xdA\">Section 230 of the Communications Decency Act<\/a>. Written in the early days of the commercial Internet, Section 230 gives Internet companies almost complete civil immunity for any content posted on their platforms by third parties. Walking these protections back would make companies like Facebook legally responsible for limiting the spread of damaging content on their sites. But such an approach raises complex free speech and censorship concerns.<\/p>\n<p>In the end, no single solution will suffice. An essential first step is simply to increase public awareness of the possibilities and dangers of deepfakes. An informed citizenry is a crucial defense against widespread misinformation.<\/p>\n<p>The recent rise of fake news has led to fears that we are entering a \u201c<a title=\"https:\/\/en.wikipedia.org\/wiki\/Post-truth\" href=\"https:\/\/en.wikipedia.org\/wiki\/Post-truth\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"post-truth\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/en.wikipedia.org\/wiki\/Post-truth&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNEv3rP6Mc3GpeTQFE0Q84mSR2t6Gw\">post-truth<\/a>\u201d world. Deepfakes threaten to intensify and accelerate this trajectory. The next major chapter in this drama is likely just around the corner: the 2020 elections. The stakes could hardly be higher.<\/p>\n<p>\u201cThe man in front of the tank at Tiananmen Square moved the world,\u201d <a title=\"https:\/\/www.washingtonpost.com\/technology\/2019\/06\/12\/top-ai-researchers-race-detect-deepfake-videos-we-are-outgunned\/\" href=\"https:\/\/www.washingtonpost.com\/technology\/2019\/06\/12\/top-ai-researchers-race-detect-deepfake-videos-we-are-outgunned\/\" target=\"_blank\" rel=\"nofollow noopener noreferrer\" aria-label=\"said\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/www.washingtonpost.com\/technology\/2019\/06\/12\/top-ai-researchers-race-detect-deepfake-videos-we-are-outgunned\/&amp;source=gmail&amp;ust=1635342840829000&amp;usg=AFQjCNFJj92NyuXGN4QoMALZgNkL0Bc9rA\">said<\/a> NYU professor Nasir Memon. \u201cNixon on the phone cost him his presidency. Images of horror from concentration camps finally moved us into action. If the notion of not believing what you see is under attack, that is a huge problem. One has to restore truth in seeing again.\u201d<\/p>\n<ul>\n<li><\/li>\n<\/ul>\n","protected":false},"excerpt":{"rendered":"<p>FAKE NEWS Is Going To Wreak Havoc On Society. We Are Not Prepared. Last month during ESPN\u2019s hit documentary series The Last Dance, State Farm debuted a TV commercial that has become one of the most widely discussed ads in recent memory. It appeared to show footage from 1998 of an ESPN analyst making shockingly<a class=\"read-more-link\" href=\"https:\/\/test.nahtnow.com\/en\/fake-news\/\"> Read More&#8230;<\/a><\/p>\n","protected":false},"author":1,"featured_media":13651,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_links_to":"","_links_to_target":""},"categories":[39,52,40,5,51,1],"tags":[1171],"_links":{"self":[{"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/posts\/13645"}],"collection":[{"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/comments?post=13645"}],"version-history":[{"count":1,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/posts\/13645\/revisions"}],"predecessor-version":[{"id":13646,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/posts\/13645\/revisions\/13646"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/media\/13651"}],"wp:attachment":[{"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/media?parent=13645"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/categories?post=13645"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/test.nahtnow.com\/en\/wp-json\/wp\/v2\/tags?post=13645"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}