πŸ“– tl;dr Responsible AI

      • πŸ“„ Case Study Aggregators
      • πŸ“„ Generative AI and Labor: Power, Hype, and Value at Work
      • πŸ“„ Red Teaming Methods in AI Security
      • πŸ“„ Responsible AI Glossary
      • πŸ“„ Azure AI Foundry Risk and Safety
      • πŸ“„ The AI Risk Repository: A Comprehensive Meta-Review, Database, and Taxonomy of Risks From Artificial Intelligence
        • πŸ“„ Agent-SafetyBench - Evaluating the Safety of LLM Agents
        • πŸ“„ BBQ: A Hand-Built Bias Benchmark for Question Answering
        • πŸ“„ CrowS-Pairs - A Challenge Dataset for Measuring Social Biases in Masked Language Models
        • πŸ“„ HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal
        • πŸ“„ AART AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications
        • πŸ“„ Building Safe GenAI Applications - An End-to-End Overview of Red Teaming for Large Language Models
        • πŸ“„ On Verbalized Confidence Scores for LLMs
        • πŸ“„ STAR: SocioTechnical Approach to Red Teaming Language Models
        • πŸ“„ Azure AI Foundry Agent Evaluate SDK
        • πŸ“„ LLM Comparator
        • πŸ“„ Microsoft RAI Impact Assessment Guide Summary
        • πŸ“„ SafeArena: Evaluating the Safety of Autonomous Web Agents
        • πŸ“„ WildGuard: Open One-stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs
      • πŸ“„ Evaluating the Social Impact of Generative AI Systems
      • πŸ“„ NIST AI 600-1: Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile
      • πŸ“„ Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!
      • πŸ“„ Know Thy Judge - On the Robustness Meta-Evaluation of LLM Safety Judges
      • πŸ“„ Red Teaming Language Models to Reduce Harms - Methods, Scaling Behaviors, and Lessons Learned
      • πŸ“„ Red-Teaming in the Public Interest
      • πŸ“„ Safety Alignment Should Be Made More Than Just a Few Tokens Deep
      • πŸ“„ Sociotechnical Safety Evaluation of Generative AI Systems
      • πŸ“„ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers
      • πŸ“„ Walk the Talk? Measuring the Faithfulness of Large Language Model Explanations
      • πŸ“„ LICENSE
      • πŸ“„ README
    Home

    ❯

    Case Studies

    ❯

    Case Study Aggregators

    Case Study Aggregators

    1 min read

    This page includes a curated list of aggregated case studies of AI in the real world.

    OrganizationLink
    Google CloudReal-world gen AI use cases from the world’s leading organizations
    DeloitteGen-AI Use Cases
    AmazonTop generative AI use cases

    Recent

    • Sociotechnical Safety Evaluation of Generative AI Systems

    • The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers

    • Walk the Talk? Measuring the Faithfulness of Large Language Model Explanations

    • Summaries of RAI concepts, research, and frameworks

    • AART AI-Assisted Red-Teaming with Diverse Data Generation for New LLM-powered Applications

    Created with Quartz v4.4.0 Β© 2025

    • GitHub
    • Bluesky