Michael Selig, chair of the US Commodity Futures Buying and selling Fee, mentioned blockchain might play a key position in verifying AI-generated content material, contending the expertise may help distinguish genuine media from artificial outputs as issues over misinformation develop.
Throughout an look on The Pomp Podcast on Thursday, Selig was requested by host Anthony Pompliano about the usage of AI-generated memes and pictures in markets, and whether or not intent issues or such content material must be restricted altogether. He informed Pompliano:
The non-public markets have options — blockchain expertise is a superb one. When you can timestamp issues and ensure there’s an identifier for every meme or AI generated posts, you’ll be able to confirm if it’s actual or generated by AI… Having these applied sciences right here within the US is vital.
He mentioned regulators are centered on sustaining US management in crypto, including that “you’ll be able to’t have AI with out blockchain.”

Supply: The Pomp Podcast
Relating to how regulators are approaching AI brokers, as autonomous buying and selling turns into extra prevalent in monetary markets and authorities are being pressed to tell apart between automated instruments and totally autonomous brokers, and the way the latter must be regulated, Selig responded:
I’m involved that we over-regulate and strangle a few of the expertise right here within the US… I’m taking a really a lot minimal efficient dose of regulation strategy, the place we’re… ensuring that we’re regulating the actors… and never the software program builders. The software program builders are those constructing the instruments, however they’re not really partaking within the monetary transactions.
Selig mentioned the CFTC is assessing how AI fashions are utilized in markets, emphasizing that enforcement ought to concentrate on members partaking in monetary exercise.
Associated: AI and stablecoins are successful regardless of 2026 crypto market stoop
Blockchain and proof-of-personhood instruments emerge for AI verification
A central problem amid the surge in synthetic intelligence use is distinguishing actual content material from artificial media. Selig’s feedback may very well be seen to mirror a broader push amongst policymakers and builders to make use of blockchain for content material verification and provenance.
One strategy is proof-of-personhood programs, which purpose to verify that an account belongs to an actual, distinctive human relatively than a bot. Essentially the most outstanding instance is Sam Altman’s World, whose World ID protocol permits customers to show their humanity with out revealing private knowledge. The system makes use of encrypted biometric iris scans saved on the person’s gadget, although it has drawn criticism over privateness dangers and potential coercion.
In March, World launched AgentKit, a toolkit that enables AI brokers to show they’re linked to a verified human whereas interacting with on-line companies. It integrates proof-of-personhood credentials with the x402 micropayments protocol developed by Coinbase and Cloudflare, enabling brokers to pay for entry whereas presenting cryptographic proof of human backing.
Ethereum co-founder Vitalik Buterin has proposed utilizing cryptography and blockchain to make on-line programs extra verifiable, together with via zero-knowledge proofs and onchain timestamps that might assist validate how content material is generated and distributed with out exposing delicate knowledge.
The proposals come as US policymakers weigh broader AI regulation. On March 20, the Trump administration launched a nationwide framework calling for a unified federal strategy, warning {that a} patchwork of state legal guidelines might hinder innovation and competitiveness.
Journal: Agent wastes 14 hours of scammers’ time, LLMs ‘poisoned’ by Iran: AI Eye

