Opinions expressed by Entrepreneur contributors are their very own.
In January 2024, a finance director at a UK agency transferred $25 million to fraudsters after a video name with what gave the impression to be the corporate’s CFO — a really subtle deepfake. This incident is way from remoted. The worldwide losses projected as a consequence of ecommerce fraud surged from US$44 billion in 2024 to $107 billion in 2029; a large 141% enhance, based on a report from Juniper Analysis.
Compounding the issue, a staggering 60% of shoppers now doubt the authenticity of on-line content material, citing considerations over AI-generated misinformation, deepfakes and content material overload, as revealed in Accenture’s Life Traits 2025 report survey.
For entrepreneurs and enterprise leaders, the threats are twofold: reputational injury from misleading artificial media and authorized legal responsibility as governments worldwide enact stringent AI disclosure legal guidelines.
Present options like watermarking and AI detection instruments are reactive by nature. Watermarking might be simply eliminated or solid, whereas AI detectors fail to determine manipulated content material by practically 30%, based on College of Pennsylvania researchers, by easy tweaks like added whitespace or typos. Authorized actions, in the meantime, usually come too late to mitigate injury.
Fortunately, an answer exists: blockchain + AI-powered digital twins.
Associated: Hackers Focused a $12 Billion Cybersecurity Firm With a Deepfake of Its CEO. Here is Why Small Particulars Made It Unsuccessful.
The rise of digital twins
Digital twins or avatars are a bridge between the bodily and digital world, serving to optimize methods and personalize our day-to-day experiences (journey, work, store, healthcare system and past). By embedding AI avatars with NFT passports, which act as tamper-proof digital certificates saved on the blockchain, these IDs create a verifiable report of an avatar’s origin and any subsequent modifications — making certain entrepreneurs and companies can confirm authenticity at supply slightly than scrambling to detect fakes after the very fact.
Shopper belief is inextricably linked to transparency. A 2024 Edelman report discovered that 62% of shoppers solely belief AI-generated content material if its provenance might be verified. Blockchain-based authentication addresses this demand head-on. There’s a rising development of integrating NFT-based digital passports throughout varied industries to boost product authenticity, traceability and buyer engagement.
The company world is already adopting this strategy. For example, Breitling, a Swiss luxurious watchmaker, partnered with Arianee to supply every watch with a blockchain-based digital passport. Since then, there have been over 200,000 NFTs created, with roughly 30% buyer adoption. Within the Artwork & Collectibles business, Arteïa launched its Arteïa Join, a safe digital passport for artworks anchored on the blockchain and securely linked to the piece through encrypted NFC tags, making certain authenticity and provenance.
Even the healthcare business is adopting related frameworks. The UK NHS’s Digital Workers Passport makes use of blockchain to authenticate medical professionals, whereas Mayo Clinic partnered with Secure Well being Methods to deploy AI-powered supplier IDs for telehealth — important steps to fight impersonation scams. By certifying authenticity on the level of creation, companies are thus fostering belief whereas mitigating the chance of reputational injury.
Associated: Blockchain, NFTs and the New Normal for Identification and Safety
Navigating the AI wave
For entrepreneurs, navigating the AI panorama requires a proactive strategy.
First, acknowledge that public-facing figures are prime targets for deepfake manipulation. Common monitoring for fraudulent content material is important. Second, anchor digital identities to verifiable applied sciences like blockchain. Third, staying forward of regulatory shifts is equally important. The EU AI Act, the world’s first complete AI legislation set to take impact this 12 months, imposes fines of as much as 7% of worldwide income for undisclosed artificial media.
Related measures are rising worldwide, from the US Deepfake Act to India’s draft deepfake penalties, signaling a world development towards stricter oversight. Regulators need proof, not guarantees. The blockchain’s immutable audit trails present precisely that. And, lastly, deal with digital id safety with the identical rigor as cybersecurity. Assign accountability inside your crew, conduct common audits of AI instruments, and think about partnerships with insurers specializing in deepfake legal responsibility.
The long run path is obvious: Deepfake fraud is not a hypothetical risk however a present-day legal responsibility. Whereas detection instruments play an vital function, blockchain-based authentication provides a proactive protection mechanism. Similar to SSL certificates safe ecommerce, NFT passports can do the identical for AI by securing id and authenticity.
For entrepreneurs and companies, the selection is obvious: to construct belief by way of verification now, or threat dropping all the pieces to an artificial rip-off tomorrow. Within the quickly advancing age of AI, authenticity will not simply be vital, it will likely be the muse of belief and aggressive benefit.
In January 2024, a finance director at a UK agency transferred $25 million to fraudsters after a video name with what gave the impression to be the corporate’s CFO — a really subtle deepfake. This incident is way from remoted. The worldwide losses projected as a consequence of ecommerce fraud surged from US$44 billion in 2024 to $107 billion in 2029; a large 141% enhance, based on a report from Juniper Analysis.
Compounding the issue, a staggering 60% of shoppers now doubt the authenticity of on-line content material, citing considerations over AI-generated misinformation, deepfakes and content material overload, as revealed in Accenture’s Life Traits 2025 report survey.
For entrepreneurs and enterprise leaders, the threats are twofold: reputational injury from misleading artificial media and authorized legal responsibility as governments worldwide enact stringent AI disclosure legal guidelines.
The remainder of this text is locked.
Be part of Entrepreneur+ at the moment for entry.