Designing Responsible AI Technologies to Protect Information Integrity


The rise of social media and massive information sharing online have led to a dramatic increase in the spread of both inadvertent misinformation and strategic disinformation (e.g., foreign influence operations seeking to undermine democratic nations). Additional challenges arise in helping decision-makers navigate conflicting information (e.g., information coming from different sources or evolving during a crisis, such as a national disaster or pandemic). To meet this challenge, our mission is to design, build, and test innovative AI technologies to support journalists, professional fact-checkers, and information analysts. Our use-inspired research to protect information integrity world-wide drives our broader work to develop responsible AI technologies that are both fair (in protecting different stakeholders who may bear disproportionately impacts) and explainable (so that stakeholders can best capitalize upon AI speed and scalability alongside their own knowledge, experience, and human ingenuity).

Team Members



April 17, 2023
“They’re Coming to Take over Our Country”: Researching Global Circuits of Racist Misinformation
“The others are coming. They are coming to get us, take over our country, colonize us, and replace us. They’re an existential threat.” These types of racist logics are a regular trope around the world. Some governments incorporate this type of messaging into disinformation campaigns, which have ripple effects of unintentional misinformation on social media.
March 2, 2023
Understanding the Ethical Future of AI
Dr. Matt Lease, professor in the School of Information at the University of Texas at Austin, provides a better understanding of AI and Chat GPT.
Nov. 11, 2022
Op-Ed: Social Media Platforms’ Struggles with Misinformation and Racism: Challenges and Paths Forward
From “fake news” screenshots to conspiratorial claims, the lead up to the 2022 Midterm elections has shown that misinformation remains a problem in public discourse. This is especially harmful for minority groups and underrepresented populations, as they tend to be the target of misinformation-motivated vitriol.