Sunday, July 27, 2025

Faith vs Fakery

Why Islamic Ethics Can’t Fix the Deepfake Crisis

💥 A Brutally Honest Breakdown of “Combating Fake News… Insights from the Islamic Ethical Tradition”


🚨 Introduction: Noble Intentions, Flawed Execution

A 2019 paper titled “Combating Fake News, Misinformation, and Machine Learning Generated Fakes: Insights from the Islamic Ethical Tradition” attempts to apply classical Islamic teachings to modern tech problems like deepfakes and AI-generated lies.

Spoiler: It doesn’t work.

This is a critical breakdown of the paper’s claims — with no sugar-coating, no pinches pulled, and no blind reverence for tradition where logic fails.


⚠️ What the Paper Gets Right (Bare Minimum Credit)

  • Fake news and deepfakes are dangerous — Correct.

  • Islamic teachings condemn deception — Also true.

  • Moral responsibility matters — Sure, at the individual level.

But…


🧱 The Core Problem: Dragging a 7th-Century Toolset Into a 21st-Century Battlefield

The authors argue that the science of Hadith (how Muslims validated sayings of the Prophet) can be used to fight fake news and AI-generated fakes.

That’s like using a sundial to detect cybercrime.

The Hadith verification system:

  • Was oral, manual, and based on character trust, not hard evidence.

  • Is historically disputed, politicized, and far from airtight.

  • Is completely unsuited for automated bot networks, algorithmic propaganda, or deepfakes.


🧠 Misuse of Analogy: Hadith ≠ Fact-Checking

The paper claims we can rate digital sources the way scholars rated hadith narrators. But:

  • A narrator’s “piety” doesn’t equal data integrity.

  • Sectarian bias polluted much of hadith grading.

  • Deepfakes aren’t oral stories — they’re pixel-perfect digital forgeries.

This is a false analogy, plain and simple.


📉 What the Paper Completely Misses

Despite talking about AI, the authors ignore:

  • Explainability

  • Model transparency

  • Bias detection

  • Algorithmic auditing

  • Data provenance

  • AI risk classification

  • Human-in-the-loop systems

  • Regulatory frameworks like the EU AI Act

In short, they name-drop AI terms but never engage with AI as a real-world engineering problem.


🛡️ Islamic Ethics ≠ Technological Defense

Quoting Quran verses like "speak the truth" or "verify news" may guide personal behavior. But they won’t stop:

  • A GAN model creating fake porn

  • A chatbot spreading conspiracy theories

  • A state actor deploying AI for disinformation warfare

This is like bringing a moral compass to a drone strike.


🧨 The Dangerous Oversight: Romanticizing Hadith Science

They treat Hadith criticism as a model of rigorous truth-seeking.

Reality check:

  • Hadith collections are riddled with contradictions.

  • Scholars disagreed constantly on who was reliable.

  • Politics shaped what got preserved — not just truth.

Building an AI fact-checking system on that foundation?

That’s intellectual malpractice.


🔚 Final Verdict: Sermon Disguised as a Solution

AspectReality Check
Islamic ethicsGood for behavior, not for detection
Hadith scienceHistorically biased, methodologically outdated
AI technical engagementVirtually nonexistent
Practical recommendationsShallow and wishful
Value for AI policy debatesZero

🎯 Conclusion: This Isn’t the Blueprint We Need

If you want to stop deepfakes and AI-generated lies, you need:

  • Tech literacy

  • Evidence-based systems

  • Transparent algorithms

  • Cross-disciplinary cooperation

Not a rehash of ancient oral traditions wrapped in moral preaching. 

No comments:

Post a Comment

Islam on Trial It Collapses Under Both External and Internal Critique “You can’t critique Islam unless you believe in it.” That’s the fam...