The U.S. Department of Defense has officially designated Anthropic as a supply chain risk, marking a significant escalation in government scrutiny of major AI companies. The decision comes as OpenAI faces mounting criticism for signing a contract with the military, sparking a growing "cancel ChatGPT" movement among users and advocacy groups concerned about AI deployment in defense applications. The contrasting regulatory treatment of the two leading AI firms highlights deepening tensions between commercial AI development and national security considerations. The DOD's classification of Anthropic suggests concerns about the company's position in critical infrastructure and defense technology supply chains. Meanwhile, OpenAI's military partnership has galvanized opposition from civil liberties advocates and AI safety proponents who argue that deploying large language models in military contexts raises ethical and technical concerns. The dual developments underscore a pivotal moment for AI governance, as policymakers and the private sector grapple with competing priorities of innovation, security, and accountability.