• Fact-Checking Found to Influence Recommender Algorithms

    By Tom Fleischman

    Researchers have shown that urging individuals to actively participate in the news they consume can reduce the spread of these kinds of falsehoods. “We don’t have to think of ourselves as captive to tech platforms and algorithms,” said a researcher.

  • Fighting Fake News: Using Machine Learning, Blockchain to Counter Misinformation

    By Anthony Borrelli

    False information can lead to harmful consequences. How can content creators focus their efforts on areas where the misinformation is likely to do the most public harm? Research offers possible solutions through a proposed machine learning framework, as well as expanded use of blockchain technology.

  • Using AI to Protect Against AI Image Manipulation

    By Rachel Gordon

    As we enter a new era where technologies powered by artificial intelligence can craft and manipulate images with a precision that blurs the line between reality and fabrication, the specter of misuse looms large. “PhotoGuard,” developed by MIT CSAIL researchers, prevents unauthorized image manipulation, safeguarding authenticity in the era of advanced generative models.

  • New Cipher System Protects Computers Against Spy Programs

    Researchers have achieved a breakthrough in computer security with the development of a new and highly efficient cipher for cache randomization. The innovative cipher addresses the threat of cache side-channel attacks, offering enhanced security and exceptional performance.

  • De-Risking Authoritarian AI

    By Simeon Gilding

    You may not be interested in artificial intelligence, but it is interested in you. AI-enabled systems make many invisible decisions affecting our health, safety and wealth. They shape what we see, think, feel and choose, they calculate our access to financial benefits as well as our transgressions. In a technology-enabled world, opportunities for remote, large-scale foreign interference, espionage and sabotage —via internet and software updates—exist at a ‘scale and reach that is unprecedented’.

  • Regulate National Security AI Like Covert Action

    Congress is trying to roll up its sleeves and get to work on artificial intelligence (AI) regulation. Ashley Deeks writes that only a few of these proposed provisions, however, implicate national security-related AI, and none create any kind of framework regulation for such tools. She proposes crafting a law similar to the War Powers Act to govern U.S. intelligence and military agencies use of AI tools.

  • Bringing Resilience to Small-Town Hydropower

    Using newly developed technologies, researchers demonstrated how hydropower with advanced controls and use of a mobile microgrid, can enable small communities to maintain critical services during emergencies.

  • U.S. Voluntary AI Code of Conduct and Implications for Military Use

    By Akshat Upadhyay

    Seven technology companies including Microsoft, OpenAI, Anthropic and Meta, with major artificial intelligence (AI) products made voluntary commitments regarding the regulation of AI. These are non-binding, unenforceable and voluntary, but they may form the basis for a future Executive Order on AI, which will become critical given the increasing military use of AI.

  • Geoscientists Aim to Improve Human Security Through Planet-Scale POI Modeling

    Geoinformatics engineering researchers developed MapSpace, a publicly available, scalable land-use modeling framework. By providing data characteristics broader and deeper than satellite imagery alone, MapSpace can generate population analytics invaluable for urban planning and disaster response.

  • Sandia Helps Develop Digital Tool to Track Cloud Hackers

    By Michael Ellis Langley

    Sandia programmers are helping the federal Cybersecurity and Infrastructure Security Agency (CISA) through an innovative program that enlists Microsoft cloud users everywhere to track down hackers and cyberterrorists.

  • Closer Look at “Father of Atomic Bomb”

    By Samantha Laine Perfas

    Robert Oppenheimer is often referred to as the “father of the atomic bomb.” But he also had his federal security clearance revoked during the McCarthy era, a disputed decision that was only posthumously reversed last year. Harvard historian unwinds the complexities of J. Robert Oppenheimer as scientist, legend.

  • U.S., Artificial Intelligence Companies Work to Mitigate Risks

    By Nike Ching

    Can artificial intelligence wipe out humanity? A senior U.S. official said the United States government is working with leading AI companies and at least 20 countries to set up guardrails to mitigate potential risks.

  • Roles and Implications of AI in the Russian-Ukrainian Conflict

    By Sam Bendett

    Artificial Intelligence (AI) is emerging as a significant asset in the ongoing Russian-Ukrainian conflict. Specifically, it has become a key data analysis tool that helps operators and warfighters make sense of the growing volume and amount of information generated by numerous systems, weapons and soldiers in the field.

  • Six Ways AI Can Make Political Campaigns More Deceptive Than Ever

    By David E. Clementson

    Political campaign ads and donor solicitations have long been deceptive. These days, the internet has gone wild with deceptive political ads. Ads often pose as polls and have misleading clickbait headlines. Campaigns are now rapidly embracing artificial intelligence for composing and producing ads and donor solicitations, even as there are growing fears that AI will make politics more deceptive than ever.

  • Protecting Health Security from Potential Threats Arising from Advances in Biotechnology

    The Johns Hopkins Center for Health Security said it was applauding congressional leaders in health security for their introduction of two pieces of legislation to strengthen security around emerging threats from advances in biotechnology.