Welcome!

By registering with us, you'll be able to discuss, share and private message with other members of our community.

SignUp Now!
  • You Can Use Demo as Registered User :
    user:demo pass:dohdemo

“This is about love gone wrong,” says Kara Swisher

demo

Member
Joined
Dec 5, 2017
Messages
574
“This is about love gone wrong,” says Kara Swisher, looking back on a life spent studying the giants of Big Tech. “I saw the possibilities of tech being the saviour of humanity – or at the very least, really helping people, in terms of community and knowledge and education. And instead, you know …”

She pauses, and wearily exhales. “It’s like that old expression: ‘They promised us jetpacks, and this is what we got?’ Like, are you kidding me?” But, she adds: “The problem isn’t tech. It’s people.”


Swisher is essentially a business journalist, but her speciality is human beings and what they do with wealth and power. She has been scrutinising Silicon Valley for around three decades – writing ferociously and insightfully for the Washington Post, the Wall Street Journal and the New York Times, founding the tech news website Recode, and blazing a trail into podcasting. Her writing has always struck a delicate balance between insider knowledge and biting irreverence. Now, though, her iconoclastic side has won out, given free rein in an extremely readable memoir, Burn Book.

The title, she tells me, comes from the 2004 movie Mean Girls, in which the leading characters keep a shared diary full of slights and gossip about their classmates. As well as telling her own story, the book centres on pen-portraits of people – men, mostly – she has closely observed as their wealth and influence has ballooned: Bill Gates, Steve Jobs, Mark Zuckerberg, Jeff Bezos, Elon Musk, and Sergey Brin and Larry Page, the founders of Google.

A few of them emerge as rounded, deep-thinking people who at least understand the huge questions of power that swirl around them. Others, by contrast, are more cuttingly portrayed – as “fresh-faced wunderkinds I had mostly rooted for” who eventually made the author “feel like a parent whose progeny had turned into, well, assholes”.


n ,n,n,lnmlmlmlm , ,n ,n ,m k, k,n ,knmnn,n
n m m m
nkn,knk
kih
n mn nn bkn
m m mm n m mn n nm m m
 
سوف اقوم بجتربة اللغة العربية
 
In recent years, the use of artificial intelligence to create deepfake videos has increased dramatically. Deepfakes are synthetic media where a person's likeness is inserted into existing video or photo content using powerful AI algorithms. While some uses of this technology can be benign or even positive, one concerning application is the creation of nonconsensual deepfake videos of celebrities.


The implications of widespread celebrity deepfakes could be hugely detrimental both to the celebrities themselves as well as society at large. Unlike typical photo editing, deepfake videos can portray individuals doing or saying things they never actually did. The realism of these AI-generated videos makes them difficult to distinguish from authentic footage, even upon close examination.


Several instances of celebrity deepfake videos have already surfaced online, depicting famous actresses and singers in compromising sexual scenarios. These graphic deepfakes are nonconsensual violations of privacy and autonomy. Their creation and distribution essentially amount to revenge porn, made even more violating by the use of the target's likeness. The mental anguish and emotional distress this can cause celebrities is immense.


Beyond the personal trauma to individuals, celebrity deepfakes risk undermining public trust across many institutions. The authenticity of video and photo evidence has long been an important factor in journalism, law enforcement investigations, and court trials. Widespread adoption of deepfake technology threatens to erode faith in these institutions. If audiovisual records are doubted as potential deepfakes, then reporting, criminal proceedings, and testimonies may lose credibility.


Some argue that because celebrities are public figures, they open themselves up to satire and parody. However, nonconsensual deepfakes go far beyond parody. The ability to realistically portray individuals doing or saying anything fundamentally denies personal autonomy over one's public image and reputation. Additionally, the harms extend beyond just the targeted celebrities themselves. The normalization of graphic deepfakes can encourage further violence and exploitation toward women in society as well.


Laws around deepfake technology remain unclear and inadequate. Some argue that deepfakes qualifying as parody are protected free speech, while others contend they should be regulated as copyright infringement. These legal gray areas urgently need clarification in order to protect both public figures and society from dangers associated with deepfakes. Platforms hosting celebrity deepfakes against the will of those depicted should be obligated to remove the content promptly before it can spread virally across the internet.


The individuals creating and uploading nonconsensual celebrity deepfakes may feel they are harmless pranks, but the truth is these videos have lasting and traumatic impacts. As deepfake technology becomes more advanced and accessible, the threat of personalized disinformation and abuse will keep growing. Society must take steps to recognize the unique harm posed by celebrity and other involuntary deepfakes compared to traditional parody. With a balanced approach to regulation and ethics around synthetic media, perhaps we can mitigate dangers to both prominent individuals as well as the public trust.
 
Back
Top