The AI Forum promotes thoughtful discussion of the legal and societal challenges posed by AI … read more About Us
Responsible AI in the News
California Governor Gavin Newsom has signed a landmark AI safety law requiring advanced AI companies to publicly disclose their safety measures and safeguard whistleblowers. This move positions California as a leader in AI regulation, sparking debates over innovation versus consumer protection, and prompting scrutiny of state legislative authority within a rapidly evolving technology sector.
Eve, a San Francisco-based legal AI startup, has reached a $1 billion valuation after raising $103 million in new funding. The firm specializes in AI-driven tools for plaintiffs’ law firms, reflecting accelerated demand for automation in legal practice as more firms turn to generative AI to streamline case work and document analysis.
The FTC has launched a sweeping inquiry into seven major tech firms, including Meta, OpenAI, and Alphabet, demanding details on how their AI chatbots protect children and teens. The investigation seeks to uncover companies’ practices for monitoring safety, disclosing risks, and complying with privacy laws amid rising concern about AI-powered companion bots’ impacts on youth.
Anthropic just became the first major AI company to endorse California’s groundbreaking SB 53 bill, which would require unprecedented transparency, reporting, and safety standards for cutting-edge AI developers. If passed, the legislation promises to reshape how AI risks are managed—not just in California, but potentially nationwide. The final vote is imminent.
Anthropic has reached a historic $1.5 billion settlement in a landmark copyright dispute over its use of pirated and lawfully acquired books to train AI models. The deal sets a precedent for the industry, highlighting the need for robust data governance, licensing, and legal compliance for AI developers and enterprise users navigating evolving copyright rules.
Governments are taking unprecedented stakes in major tech companies, as seen by the U.S. acquiring 10% of Intel. This marks a bold shift from regulation to state ownership of vital technologies. The “Intel precedent” signals new legal and compliance hurdles for multinationals as competition heats up globally over AI chips, sovereignty, and market control
The US, like many countries, faces a complicated challenge: how to regulate the rise of AI and AGI in the midst of fierce geopolitical competition. Software-focused controls face the decades-old dilemma of unpredictable outcomes, and America’s hardware advantages are quickly fading away. Policymakers must balance innovation, security, and ethical governance in an era of unprecedented technological upheaval.
Two federal court rulings in New Jersey and Mississippi were abruptly withdrawn after major factual and legal errors, likely triggered by unvetted generative AI research, surfaced within hours. These high-profile incidents highlight the growing risks of AI “hallucinations” in judicial work and underscore urgent calls for transparent, enforceable rules governing courtroom use of AI tools
Elon Musk's xAI has filed an explosive antitrust lawsuit against tech giants Apple and OpenAI, accusing them of conspiring to create an AI monopoly that stifles competition. The billionaire claims their partnership illegally locks up markets while harming his Grok chatbot's prospects in the intensifying artificial intelligence arms race.
What’s on the Podcast?
Join former Washington State Governor Jay Inslee on the Responsible AI podcast as he delves into technology regulation, clean energy, and the challenges of misinformation. We discover his insights on fostering innovation and what he thinks is more dangerous for politics than AI. Tune in for a thought-provoking conversation on shaping a sustainable future.
In this thought-provoking episode of the Responsible AI podcast, we explore the philosophical implications of AI and consciousness with Blaise Agüera y Arcas, Google's CTO of Technology and Society. Delve into the profound questions surrounding AI's role in understanding intelligence, the nature of consciousness, and the ethical considerations of treating AI as moral entities.
Dive into the world of creativity and technology with Russell Ginns, a prolific author and inventor. Known for his work with Sesame Street and NASA, Russell shares his journey from traditional media to embracing AI's transformative power. "You can't beat them, so you got to join them," he advises. Find out how he went from fear to excitement in this episode of the Responsible AI podcast.
Finding the Intersection of Law and Society
Can we truly trust AI if we don't understand it? The "black box paradox" in AI poses a significant challenge to transparency, making it impossible to fully grasp decision-making processes. Samy Yacoubi delves into the problem of demanding AI transparency while acknowledging the inherent opacity of complex algorithms and explores whether there are alternative approaches.
You can’t put the genie back in the bottle. As students embrace AI, teachers must help them understand things to watch for, including algorithmic bias, data hallucination, and gaps in the AI’s source material. And educators must ensure students' privacy rights are respected and address possible abuses as policies evolve. In this final excerpt from his paper, Andrew Rosston addresses safe AI usage for underage students.
Personalization is a major benefit of AI and can be used to students’ benefit with the development of tutoring programs that are tailored to each learner’s gaps and needs. Recent studies show that a combination approach of human tutors and personalized AI learning tracks are preferred by students, but more study is needed. Andrew Rosston digs into the challenges of AI in tutoring.
Responsible AI Shorts
Hear more conversations about responsible AI on our YouTube channel or explore our podcast.