How Aviator is Being Advertised Using Deepfake Technology

The rapid advancement of artificial intelligence (AI) and deepfake technology has opened up new avenues for marketing and advertising, but it has also led to a troubling trend: the unauthorized use of celebrity likenesses to promote gambling apps, including the increasingly popular Aviator game. Over the last six months, several reports have surfaced highlighting how deepfake technology is being exploited to create convincing but fraudulent advertisements featuring well-known celebrities endorsing Aviator and similar gambling platforms.

How Aviator is being advertised using deepfake technology

Understanding Deepfake Technology

Deepfake technology uses AI and machine learning algorithms to create highly realistic video, audio, or images that appear to be real but are, in fact, completely fabricated. This technology can superimpose a person’s face onto another body, mimic their voice, and even generate entirely new footage that looks authentic. While deepfakes have some positive applications in entertainment and media, their misuse in advertising—particularly in the gambling industry—has become a significant concern.

The Case of Aviator and Celebrity Deepfakes

Aviator, a crash-style game that has gained substantial popularity in online gambling circles, has become a target for unethical marketers who use deepfakes to promote the game. Deepfake advertisements featuring celebrities like MrBeast, Akshay Kumar, and Virat Kohli have been circulating on social media, deceiving audiences into believing that these figures endorse the Aviator game.

MrBeast and the gambling app

In October 2023, MrBeast, one of the most influential YouTubers with millions of followers, publicly condemned a deepfake video that falsely depicted him promoting a gambling app. The video was cleverly designed, using advanced AI techniques to mimic MrBeast’s appearance and voice, making it appear as though he was genuinely endorsing the app. The ad falsely claimed that users could win substantial amounts of money by playing the gambling games, enticing viewers to download the app and make bets​ (Vegas Slots Online).

MrBeast took to social media to warn his fans about the scam, urging them to be cautious and questioning whether social media platforms were prepared to handle the rising threat of AI deepfakes. His statement highlighted the broader issue of how easily such technology can be abused, particularly in the gambling sector, where trust and credibility are crucial for attracting new users.

Akshay Kumar’s Deepfake Dilemma

Similarly, Bollywood actor Akshay Kumar became another victim of deepfake misuse. In August 2023, a video began circulating on social media that appeared to show Kumar endorsing a gambling application linked to the Aviator game. The deepfake video was created by altering the audio and lip movements of a genuine video of Kumar promoting his movie, “OMG 2.” The manipulated video misled viewers into thinking that the actor was endorsing the Aviator game, a claim that Kumar and his team quickly refuted​ (FACTLY).

This incident underscores the sophistication of deepfake technology and the ease with which it can be used to manipulate content for nefarious purposes. Kumar’s team took legal action against the creators of the deepfake, but the damage to his reputation and the potential harm to unsuspecting viewers had already been done.

Moreover, we subjected the viral video to deep fake video detectors, and they identified it as potentially being a deep fake.

Virat Kohli and the Aviator Promotion

In another alarming instance, a deepfake video of Indian cricket star Virat Kohli promoting the Aviator game surfaced online. The video used AI technology to alter footage from a legitimate interview with Kohli, making it seem as though he was endorsing the gambling app. This deepfake was particularly convincing because it incorporated voice cloning technology to replicate Kohli’s voice, further lending an air of authenticity to the fraudulent advertisement​ (FACTLY).

Like the other celebrities targeted, Kohli had no knowledge of or involvement in the promotion of the Aviator game. The deepfake not only misled fans but also posed a significant risk to those who might have been tempted to download and use the app based on the false endorsement.

The Impact of Deepfake Advertising

The use of deepfakes to advertise gambling apps like Aviator has far-reaching implications. For consumers, these ads create a false sense of security and trust, leading them to engage with platforms they might otherwise avoid. The association with a trusted celebrity can be incredibly persuasive, especially in a context where gambling apps are often viewed with skepticism.

Moreover, these deepfake advertisements can lead to financial losses and even legal trouble for users who fall victim to the scams. The deceptive nature of the ads means that users might not realize they are being misled until it is too late, at which point they may have already lost money or compromised their personal information.

For the celebrities involved, deepfake ads represent a serious invasion of privacy and an attack on their personal brand. These individuals have built careers on their public personas, and the unauthorized use of their likeness in such a deceptive manner can damage their reputation and lead to a loss of trust among their fanbase.

Legal and Ethical Challenges

The rise of deepfake technology in advertising raises significant legal and ethical questions. Currently, there is a lack of comprehensive legislation that addresses the misuse of deepfake technology, particularly in the context of advertising. While some countries have begun to introduce laws aimed at curbing the spread of deepfakes, enforcement remains a challenge, especially when the perpetrators are located in different jurisdictions.

Furthermore, the rapid development of AI technology means that deepfakes are becoming increasingly difficult to detect. This creates a pressing need for digital platforms, such as social media and video-sharing sites, to develop more advanced tools for identifying and removing deepfake content before it can cause harm.

Conclusion

The use of deepfake technology to advertise gambling apps like Aviator is a troubling trend that highlights the darker side of AI advancements. While the technology itself is neutral, its application in creating deceptive advertisements is a clear misuse that poses risks to consumers, celebrities, and the integrity of the advertising industry.

As deepfakes become more prevalent, it is crucial for digital platforms, lawmakers, and the public to remain vigilant. Increased awareness, stronger legal frameworks, and more sophisticated detection tools are essential in combating the misuse of deepfake technology. In the meantime, consumers should exercise caution and skepticism when encountering online advertisements, particularly those involving celebrity endorsements.

For those interested in learning more about the implications of deepfake technology, reputable sources like Popular Science and Factly offer detailed insights and updates on the latest developments in this rapidly evolving field​ (FACTLY) (Popular Science).

About author
Ernest Shpilevoy
Expert / Author of the article
An expert in the gambling and betting industry. He used to be a professional soccer player, but ended his career after an injury. Worked for a major publication with a 20-year history in gambling news. His hobbies are fishing and betting on sports; he also occasionally gambles at casinos.
Share
Send this to a friend