BOOM

Trending Searches

    SUPPORT
    BOOM

    Trending News

      • Fact Check 
        • Fast Check
        • Politics
        • Business
        • Entertainment
        • Social
        • Sports
        • World
      • Law
      • Explainers
      • News 
        • All News
      • Decode 
        • Impact
        • Scamcheck
        • Life
        • Voices
      • Media Buddhi 
        • Digital Buddhi
        • Senior Citizens
        • Videos
      • Web Stories
      • BOOM Research
      • BOOM Labs
      • Deepfake Tracker
      • Videos 
        • Facts Neeti
      • Home-icon
        Home
      • About Us-icon
        About Us
      • Authors-icon
        Authors
      • Team-icon
        Team
      • Careers-icon
        Careers
      • Internship-icon
        Internship
      • Contact Us-icon
        Contact Us
      • Methodology-icon
        Methodology
      • Correction Policy-icon
        Correction Policy
      • Non-Partnership Policy-icon
        Non-Partnership Policy
      • Cookie Policy-icon
        Cookie Policy
      • Grievance Redressal-icon
        Grievance Redressal
      • Republishing Guidelines-icon
        Republishing Guidelines
      • Fact Check-icon
        Fact Check
        Fast Check
        Politics
        Business
        Entertainment
        Social
        Sports
        World
      • Law-icon
        Law
      • Explainers-icon
        Explainers
      • News-icon
        News
        All News
      • Decode-icon
        Decode
        Impact
        Scamcheck
        Life
        Voices
      • Media Buddhi-icon
        Media Buddhi
        Digital Buddhi
        Senior Citizens
        Videos
      • Web Stories-icon
        Web Stories
      • BOOM Research-icon
        BOOM Research
      • BOOM Labs-icon
        BOOM Labs
      • Deepfake Tracker-icon
        Deepfake Tracker
      • Videos-icon
        Videos
        Facts Neeti
      Trending Tags
      TRENDING
      • #Operation Sindoor
      • #Pahalgam Terror Attack
      • #Narendra Modi
      • #Rahul Gandhi
      • #Waqf Amendment Bill
      • #Arvind Kejriwal
      • #Deepfake
      • #Artificial Intelligence
      • Home
      • Fact Check
      • Sheikh Hasina's Confession On...
      Fact Check

      Sheikh Hasina's Confession On Detention Centres In Bangladesh Is A Deepfake

      BOOM found that claims about Sheikh Hasina admitting to 'aynaghor' creation are false, with evidence pointing to AI manipulation of old footage.

      By -  Archis Chowdhury
      Published -  18 Feb 2025 3:30 PM IST
    • Boomlive
      Listen to this Article
      Sheikh Hasinas Confession On Detention Centres In Bangladesh Is A Deepfake
      CLAIMSheikh Hasina has admitted to having create detention centers called 'aynaghor' in Bangladesh.
      FACT CHECKVideo of Sheikh Hasina admitting to creating 'aynaghor' is a deepfake, made by manipulating a 2019 interview she gave to BBC News Bangla.

      A video is being shared widely on Facebook, purportedly showing ousted Bangladeshi Prime Minister Sheikh Hasina admitting to have created "aynaghor" or "house of mirrors" - a colloquial term for clandestine detention centers.

      BOOM found this claim to be false; our investigation provides strong evidence of the viral video being edited using artificial intelligence, by manipulating a 5-year-old interview of Hasina. Our partners at the Deepfakes Analysis Unit also confirmed to us through expert consultation that the viral video is AI-manipulated, created by overlaying a voice clone, and adding a lip-sync to match her lip movements.

      Hasina's regime was marred with allegations of forced disappearances and incarcerations of dissidents at clandestine detention centers, where they were allegedly tortured. These detention centers were colloquially called "aynaghor" in Bangla, which translates to 'house of mirrors' in English.

      Also Read:TISS And JNU Release Flawed, Biased Reports Ahead Of Key Polls

      In the viral video, Sheikh Hasina can be heard saying, "Tarek Siddiqui gave the idea for the 'Aynaghor' (House of Mirror). The 'Aynaghor' was planned experimentally as there was a lot of publicity and criticism due to the crossfire. A three-foot by three-foot house was built and opposition party leaders and journalists were first brought in." She can be heard making other comments on the "aynaghor" in the video.

      The video was widely shared with a Bangla caption, translates to English as, "Hasina's confession—how House of Mirror is built, how (people) were kept in House of Mirror day after day through torture and fear."


      Click here to see the above post, and here to view an archive.

      Verdict: Deepfake Made Using Old Interview Footage

      BOOM analysed the video and found several jump cuts throughout, indicating presence of manipulation. We ran a few keyframes of the video through a reverse image search on Google, which led us to a 2019 interview of Hasina by BBC News Bangla.



      We then compared keyframes from the YouTube video of the BBC interview with those in the viral video and observed that Hasina's clothing and the background were the same - leading us to conclude that the BBC interview was the original clip of the interview.

      In the original 22-minute BBC interview, Hasina does not mention 'aynaghor' or the 'house of mirrors'.

      Additionally, the search on the video keyframe led us to a video uploaded on YouTube on February 13, 2025. This video carried a label of 'Altered or Synthetic Media', added by YouTube. YouTube adds this disclaimer/notice on content created or manipulted using Artificial Intelligence


      This indicated to us that the viral video has been altered, and that it has likely been manipulated through artificial intelligence.

      BOOM also consulted our partners at the Deepfakes Analysis Unit, who analysed the video using detection tools and spoke to experts at Contrails AI. Both concluded that the audio and video were manipulated using AI.

      Also Read:It Took Nearly Five Hours For Govt To Admit Stampede At New Delhi Station

      The audio analysis predicted one per cent chance of it being real, showing a very high likelihood of it being a voice clone.

      The video analysis predicted 31 per cent chance of the video being real, showing a moderately high likelihood of it being altered using AI. The experts at Contrails AI also pointed out to DAU that "there are clear visible manipulation artifacts near the lips are also shaking unnaturally when the audio is silent."

      Tags

      Sheikh HasinaBangladeshDeepfakeArtificial Intelligence
      Read Full Article
      Claim :   Video shows Sheikh Hasina confession to having made 'aynaghor' or detention centers.
      Claimed By :  Facebook users
      Fact Check :  False
      Next Story
      Our website is made possible by displaying online advertisements to our visitors.
      Please consider supporting us by disabling your ad blocker. Please reload after ad blocker is disabled.
      X

      Subscribe to BOOM Newsletters

      👉 No spam, no paywall — but verified insights.

      Please enter a Email Address
      Subscribe for free!

      Stay Ahead of Misinformation!

      Please enter a Email Address
      Subscribe Now🛡️ 100% Privacy Protected | No Spam, Just Facts
      By subscribing, you agree with the Terms & conditions and Privacy Policy connected to the offer

      Thank you for subscribing!

      You’re now part of the BOOM community.

      Or, Subscribe to receive latest news via email
      Subscribed Successfully...
      Copy HTMLHTML is copied!
      There's no data to copy!