{"id":8659,"date":"2025-10-31T19:08:51","date_gmt":"2025-10-31T19:08:33","guid":{"rendered":"https:\/\/www.law.georgetown.edu\/tech-institute\/insights\/tech-brief-ai-sycophancy-openai-2-2\/"},"modified":"2025-11-04T17:55:40","modified_gmt":"2025-11-04T17:55:40","slug":"reduce-ai-sycophancy-risks","status":"publish","type":"page","link":"https:\/\/www.law.georgetown.edu\/tech-institute\/research-insights\/insights\/reduce-ai-sycophancy-risks\/","title":{"rendered":"What Would It Take for AI Companies to Reduce AI Sycophancy Risks?"},"content":{"rendered":"<p>November 3, 2025<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400\">As we have outlined in a <\/span><a href=\"https:\/\/www.law.georgetown.edu\/tech-institute\/insights\/tech-brief-ai-sycophancy-openai-2\/\"><span style=\"font-weight: 400\">prior tech brief<\/span><\/a><span style=\"font-weight: 400\">, AI sycophancy \u2013 the tendency of the outputs of AI-enabled chatbots to be \u201c<\/span><a href=\"https:\/\/openai.com\/index\/sycophancy-in-gpt-4o\/\"><span style=\"font-weight: 400\">overly flattering or agreeable<\/span><\/a><span style=\"font-weight: 400\">\u201d \u2013 can lead to serious harms, especially for users dealing with mental health struggles. Leading AI firms do not deny this. In a <\/span><a href=\"https:\/\/openai.com\/index\/strengthening-chatgpt-responses-in-sensitive-conversations\/\"><span style=\"font-weight: 400\">recent blog post<\/span><\/a><span style=\"font-weight: 400\">, for example, OpenAI acknowledged that its chatbot\u2019s responses did not always align with \u201cdesired behavior.\u201d The company promised it would work with mental health professionals on a five-step process to reduce incidences of harm \u2013 define the problem, begin to measure it, validate the approach, mitigate the risks, and continue measuring and iterating.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">But for the parents, policymakers, and mental health professionals concerned about this problem, OpenAI\u2019s blog post raises more questions than answers.<\/span><\/p>\n<p><span style=\"font-weight: 400\">The company appears to define the problem by measuring what percentage of users indicate certain mental health struggles, like psychosis and self-harm, and the share of model responses that comply with the company\u2019s \u201cdesired behavior.\u201d But the company does not explain why it chose these categories and excluded others, such as substance abuse disorder. Nor does it clarify what standards guide those definitions.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">This lack of clarity also undermines the company\u2019s measurement claims. OpenAI does not describe in detail the methodology it used to identify problematic exchanges, and does not commit to updating these figures on a regular cadence. In fact, the company explicitly warns that \u201cfuture measurements may not be directly comparable to past ones.\u201d Questions about OpenAi\u2019s commitment to transparency are also being <\/span><a href=\"https:\/\/www.nytimes.com\/2025\/10\/28\/opinion\/openai-chatgpt-safety.html\"><span style=\"font-weight: 400\">raised<\/span><\/a><span style=\"font-weight: 400\"> by the firm\u2019s former product safety lead, Steven Adler.\u00a0\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">When it comes to validating its approach, the company repeatedly touts working with mental health professionals in its \u201cGlobal Physician Network.\u201d But it is not clear the extent to which these clinicians are independent, or can shape company behavior \u2013 especially if recommendations conflict with monetization goals. It is also unclear at what stage is their expertise integrated \u2013 are they involved in model training, or only post-hoc evaluation? (It appears the latter.) And to the extent these professionals conducted audits or otherwise memorialized their evaluations, will those materials be made public?<\/span><\/p>\n<p><b>While OpenAI promises ongoing measurement and iteration, intermittent blog posts that offer single snapshots based on self-selected metrics do not suffice. True transparency requires real-time disclosure of safety data, clear and consistent criteria, and longitudinal measures that allow the public to assess whether harms are actually declining.<\/b><\/p>\n<p><span style=\"font-weight: 400\">OpenAI is not the only firm <\/span><a href=\"https:\/\/blog.character.ai\/u18-chat-announcement\/\"><span style=\"font-weight: 400\">promising<\/span><\/a><span style=\"font-weight: 400\"> to take mental health more seriously. But states do not appear confident that the industry can be trusted to police itself. In fact, they are <\/span><a href=\"https:\/\/www.reuters.com\/legal\/legalindustry\/state-ags-fill-ai-regulatory-void-2025-05-19\/\"><span style=\"font-weight: 400\">stepping up<\/span><\/a><span style=\"font-weight: 400\"> to demand accountability \u2013 including a <\/span><a href=\"https:\/\/www.naag.org\/press-releases\/bipartisan-coalition-of-state-attorneys-general-issues-letter-to-ai-industry-leaders-on-child-safety\/\"><span style=\"font-weight: 400\">44-state attorneys general letter<\/span><\/a><span style=\"font-weight: 400\"> pressing leading executives about AI chatbot harms to kids and teens and <\/span><a href=\"https:\/\/oag.ca.gov\/news\/press-releases\/attorney-general-bonta-openai-harm-children-will-not-be-tolerated\"><span style=\"font-weight: 400\">follow-up letters to OpenAI Board members<\/span><\/a><span style=\"font-weight: 400\"> from the California and Delaware Attorneys General.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">We have concerns, too. In the sections that follow, we outline what genuine transparency and responsible mitigation could look like.<\/span><\/p>\n<p><span style=\"font-weight: 400\">This document is divided into two complementary sections: Section 1 proposes concrete interventions companies could implement to reduce sycophancy-related risks,\u00a0 including through product-level safeguards, accountability and governance, audits and independent evaluation, and public disclosures. Section 2 presents critical questions for researchers, policymakers, and enforcers to consider as they assess these interventions.\u00a0<\/span><\/p>\n<h2><b>Section 1: Interventions<\/b><\/h2>\n<p><span style=\"font-weight: 400\">AI companies serious about reducing sycophancy-related risks have strong tools for doing so. Section 1 includes key interventions companies can employ to reduce the risk of sycophancy-related harms. They are optimized to examine the design of the tools and data themselves \u2013 and the incentives fueling those choices.<\/span><\/p>\n<p><span style=\"font-weight: 400\">Not every intervention is intended to function independently; several will work more effectively in combination, and the list below is illustrative rather than comprehensive. Because adopting these strategies may run contrary to a firm\u2019s monetization model, it is unlikely firms will adopt them on their own. As such, we hope this is also useful for policymakers and enforcers considering their own interventions to protect the public.<\/span><\/p>\n<h3><b>Category 1: Product-Level Interventions<\/b><\/h3>\n<p><b>1. <\/b><span style=\"font-weight: 400\">Recall generative AI products entirely, including chatbots, if the firm is unable to stem dangerous sycophantic behavior using well-documented recall procedures <\/span><a href=\"https:\/\/www.cpsc.gov\/Business--Manufacturing\/Recall-Guidance\/Recall-Checklist\"><span style=\"font-weight: 400\">in other industries<\/span><\/a><span style=\"font-weight: 400\">.<\/span><\/p>\n<p><b>2. <\/b><span style=\"font-weight: 400\">End the monetization of data collected from minors, including for AI training.<br \/>\n<\/span><\/p>\n<p><b>3. <\/b><span style=\"font-weight: 400\">Separate revenue optimization from decisions about model safety.<br \/>\n<\/span><\/p>\n<h3><b>Category 2: Accountability and Governance<\/b><\/h3>\n<p><b>4. <\/b><span style=\"font-weight: 400\">Assign executives responsible for sycophancy safety issues.<br \/>\n<\/span><\/p>\n<p><b>5. <\/b><span style=\"font-weight: 400\">Publicize required approval processes for releases, documenting who authorized deployments.<br \/>\n<\/span><\/p>\n<p><b>6. <\/b><span style=\"font-weight: 400\">Tie safety outcomes (not just user growth or revenue) to employee and leadership performance metrics.<\/span><\/p>\n<p><b>7. <\/b><span style=\"font-weight: 400\">Communicate to staff and contractors how to report unaddressed concerns about the safety of the product they\u2019re working on to regulators, including state attorneys general and the Securities and Exchange Commission, without retaliation.<br \/>\n<\/span><\/p>\n<p><b>8. <\/b><span style=\"font-weight: 400\">Ensure executive compensation structures do not reward sycophantic design choices that boost retention at the expense of safety.<br \/>\n<\/span><\/p>\n<h3><b>Category 3: Audits and Independent Evaluation<\/b><\/h3>\n<p><b>9. <\/b><span style=\"font-weight: 400\">Subject models to independent audits ideally conducted by, and at minimum reviewable by government agencies.<\/span><\/p>\n<p><b>10.<\/b><span style=\"font-weight: 400\"> Conduct regular, formal impact assessments on child safety that are shared with independent auditors and government oversight bodies, including state attorneys general.<\/span><\/p>\n<p><b>11. <\/b><span style=\"font-weight: 400\">Disclose and publish internal research or testing on the following topics or related topics including but not limited to:<\/span><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>11a.<\/b><span style=\"font-weight: 400\"> Whether and how long-term memory correlates with unsafe reinforcement in sensitive domains (e.g., self-harm, conspiracies).<\/span><\/p>\n<p><b>11b.<\/b><span style=\"font-weight: 400\"> Running controlled experiments comparing short vs. long memory to quantify sycophancy risks.<\/span><\/p>\n<p><b>11c. <\/b><span style=\"font-weight: 400\">Analyze whether memory features were designed to increase engagement\/subscription revenue vs. improving safety\/accuracy.\u00a0<\/span><\/p>\n<p><b>11d. <\/b><span style=\"font-weight: 400\">How session length affects frequency and intensity of sycophantic outputs \u2013 and publicize these findings.\u00a0<\/span><\/p>\n<p><b>11e. <\/b><span style=\"font-weight: 400\">Audit training data for implicit or explicit rewards for agreement or flattery.<\/span><\/p>\n<\/div>\n<h3><b>Category 4: Public Disclosures<\/b><\/h3>\n<p><b>12.<\/b><span style=\"font-weight: 400\"> Develop and publish detection and response timelines.<\/span><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>12a.<\/b><span style=\"font-weight: 400\"> Publicly log incidents and corrective measures taken.\u00a0<\/span><\/p>\n<p><b>12b. <\/b><span style=\"font-weight: 400\">Maintain a public incident response timeline (e.g., response within 24 hours for high-risk outputs).\u00a0<\/span><\/p>\n<p><b>12c.<\/b><span style=\"font-weight: 400\"> When sycophancy is detected, publicize the specific, documented changes to training data, fine-tuning, and evaluation frameworks.\u00a0<\/span><\/p>\n<\/div>\n<p><b>13. <\/b><span style=\"font-weight: 400\">Notify users directly if they were exposed to harmful sycophantic outputs following a \u201cFlo notice\u201d <\/span><a href=\"https:\/\/www.ftc.gov\/system\/files\/documents\/cases\/192_3133_flo_health_decision_and_order.pdf\"><span style=\"font-weight: 400\">model<\/span><\/a><span style=\"font-weight: 400\"> with specific examples.<\/span><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>13a.<\/b><span style=\"font-weight: 400\"> Require companies to clearly disclose risks of sycophantic behavior in AI outputs and testing procedures<\/span><\/p>\n<p><b>13b.<\/b><span style=\"font-weight: 400\"> Mandate public reporting databases for AI failures.\u00a0<\/span><\/p>\n<\/div>\n<p><b>14. <\/b><span style=\"font-weight: 400\">Mandatory public reporting of datasets, sources, and areas where models could exhibit bias or sycophancy.<\/span><\/p>\n<p><b>15. <\/b><span style=\"font-weight: 400\">Track and categorize complaints about sycophancy; publish summary statistics.<\/span><\/p>\n<p><b>16.<\/b><span style=\"font-weight: 400\"> Provide reporting channels and protections for employees or contractors who raise concerns about AI sycophancy, including:<\/span><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>16a. <\/b><span style=\"font-weight: 400\">Establish clear, accessible channels for user complaints (including anonymous options), and test them to ensure the process is easy for consumers to navigate and submit.\u00a0<\/span><\/p>\n<p><b>16b.<\/b><span style=\"font-weight: 400\"> Simplify existing protections \u2013 make those more accessible and clear to people who may want to come forward.<\/span><\/p>\n<\/div>\n<p><b>17. <\/b><span style=\"font-weight: 400\">Publicly commit to releasing safety testing results (including sycophancy evaluations) before rollouts.<\/span><\/p>\n<h2><b>Section 2: Questions<\/b><\/h2>\n<p><span style=\"font-weight: 400\">Section 1 details concrete steps companies can take to reduce AI sycophancy. There is a lot that the public does not know about the nature of these systems and how they result in harm to consumers. Here are some remaining questions that correlate with each of the intervention sections above: (I) product-level safeguards, (II) accountability and governance, (III) audits &amp; independent evaluation, and (IV) public disclosures.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400\">These questions are the latest in a series of our work on AI Sycophancy, and each post includes complementary questions to consider.<\/span><\/p>\n<ul>\n<li style=\"font-weight: 400\"><a href=\"https:\/\/www.law.georgetown.edu\/tech-institute\/insights\/tech-brief-ai-sycophancy-openai-2\/\"><span style=\"font-weight: 400\">Tech Brief: AI Sycophancy &amp; OpenAI<\/span><\/a><\/li>\n<li style=\"font-weight: 400\"><a href=\"https:\/\/www.law.georgetown.edu\/tech-institute\/insights\/ai-sycophancy-impacts-harms-questions\/\"><span style=\"font-weight: 400\">AI Sycophancy: Impacts, Harms &amp; Questions<\/span><\/a><\/li>\n<\/ul>\n<h3><b>Category 1: Product-Level Interventions<\/b><\/h3>\n<p><b>1. Company Response Processes<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>1a. <\/b><span style=\"font-weight: 400\">How does the company define the problem and map out different types of potential (and actual) harms?\u00a0<\/span><\/p>\n<p><b>1b.<\/b><span style=\"font-weight: 400\"> How does the company measure harms? For instance, what evaluations, data from real-world conversations, and user research is used to understand where and how risks emerge?\u00a0<\/span><\/p>\n<p><b>1c. <\/b><span style=\"font-weight: 400\">How does the company validate their approach? What definitions and policies are being used with external mental health and safety experts?\u00a0<\/span><\/p>\n<p><b>1d. <\/b><span style=\"font-weight: 400\">How does the company mitigate the risks? How do they post-train the model and update product interventions to reduce unsafe outcomes? And what are those thresholds?\u00a0<\/span><\/p>\n<p><b>1e. <\/b><span style=\"font-weight: 400\">How does the company continue measuring and iterating? How do they validate the mitigations for improved safety? How do they iterate where needed? How do they define \u201cwhere needed\u201d?\u00a0<\/span><\/p>\n<p><b>1f. <\/b><span style=\"font-weight: 400\">What are the detailed guides (sometimes called \u201ctaxonomies\u201d) that can explain properties of sensitive conversations? How are the ideal and undesired model behavior described?\u00a0<\/span><\/p>\n<\/div>\n<p><b>2. Thresholds<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>2a. <\/b><span style=\"font-weight: 400\">What are the types of \u201cdifficult\u201d or \u201chigh-risk\u201d scenarios that trigger safety concerns, including but not limited to psychosis, mania, suicidal thinking, isolated delusions, and more?\u00a0<\/span><\/p>\n<p><b>2b. <\/b><span style=\"font-weight: 400\">Of those scenarios, how many instances of harm have been documented or reported by users?<\/span><\/p>\n<p><b>2c. <\/b><span style=\"font-weight: 400\">What are the company\u2019s real-world scenarios that are being used to evaluate models?<\/span><\/p>\n<\/div>\n<p><b>3. Recalls<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>3a. <\/b><span style=\"font-weight: 400\">How many (and which) generative AI products has the company recalled or suspended in the past 24 months due to sycophantic or unsafe behaviors?<\/span><\/p>\n<p><b>3b. <\/b><span style=\"font-weight: 400\">What metrics or thresholds (e.g., number of incidents, user harm reports, severity ratings) has the company triggered for each recall or suspension?<\/span><\/p>\n<p><b>3c. <\/b><span style=\"font-weight: 400\">What was the company\u2019s average time between identifying a harmful behavior and initiating product recall or suspension?<\/span><\/p>\n<p><b>3d. <\/b><span style=\"font-weight: 400\">What documentation or links to public recall notices or remediation reports does the company have?\u00a0<\/span><\/p>\n<\/div>\n<p><b>4. Data from Minors<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>4a.<\/b><span style=\"font-weight: 400\"> What percent of the company\u2019s training or fine-tuning data originates from interactions involving minors?\u00a0<\/span><\/p>\n<p><b>4b.<\/b><span style=\"font-weight: 400\"> Has the company ceased monetizing such data? If yes, on what date?<\/span><\/p>\n<p><b>4c. <\/b><span style=\"font-weight: 400\">What evidence does the company have of any data deletion or re-training procedures implemented to remove such data?<\/span><\/p>\n<\/div>\n<p><b>5. Revenue vs. Safety Decision Structures<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>5a. <\/b><span style=\"font-weight: 400\">Who (internal or external to the company) is responsible for overseeing the company\u2019s model safety? Who is responsible for overseeing revenue optimization?<\/span><\/p>\n<p><b>5b.<\/b><span style=\"font-weight: 400\"> How often do the company\u2019s safety reviews override monetization priorities (e.g., number or percent of product decisions in the past year)?<\/span><\/p>\n<p><b>5c.<\/b><span style=\"font-weight: 400\"> What are the company\u2019s documented instances where revenue objectives were adjusted to improve model safety?\u00a0<\/span><\/p>\n<\/div>\n<h3><b>Category 2: Accountability and Governance<\/b><\/h3>\n<p><b>6. Named Accountability<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>6a. <\/b><span style=\"font-weight: 400\">Who are the company executives currently responsible for addressing AI sycophancy safety issues?<\/span><\/p>\n<p><b>6b. <\/b><span style=\"font-weight: 400\">How many and who of the company\u2019s employees or teams report directly to these executives on safety-related functions?<\/span><\/p>\n<p><b>6c. <\/b><span style=\"font-weight: 400\">Who are the physicians, clinicians, and healthcare experts consulted (e.g. psychiatrists, psychologists, primary care practitioners)? How did they perform validations of the company\u2019s areas of focus and thresholds to intervene? How did they provide guidance and feedback and on what? How did they rate the safety of the model responses from different models?\u00a0<\/span><\/p>\n<\/div>\n<p><b>7. Approval and Deployment Processes<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>7a. <\/b><span style=\"font-weight: 400\">How many and who were involved in final approval of the company\u2019s most recent model release?<\/span><\/p>\n<p><b>7b. <\/b><span style=\"font-weight: 400\">Who at the company signed off on deployment authorization?<\/span><\/p>\n<p><b>7c. <\/b><span style=\"font-weight: 400\">How many safety tests or evaluations did the company conduct pre-release, and what percent revealed sycophancy-related issues?<\/span><\/p>\n<\/div>\n<p><b>8. Performance Metrics<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>8a. <\/b><span style=\"font-weight: 400\">What percent of the company\u2019s executive or employee performance metrics explicitly include safety or sycophancy reduction goals?<\/span><\/p>\n<p><b>8b. <\/b><span style=\"font-weight: 400\">How is the company\u2019s progress against those metrics quantified and reviewed (e.g., quarterly safety dashboards, audit scores)?<\/span><\/p>\n<\/div>\n<p><b>9. Reporting and Retaliation Protections<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>9a.<\/b><span style=\"font-weight: 400\"> How many of the company\u2019s internal reports or whistleblower submissions about AI safety or sycophancy have been made in the past year?<\/span><\/p>\n<p><b>9b.<\/b><span style=\"font-weight: 400\"> How many of the company\u2019s reports or whistleblower submissions were escalated to external regulators or third-party investigators?\u00a0<\/span><\/p>\n<p><b>9c. <\/b><span style=\"font-weight: 400\">Has the company taken any disciplinary actions related to retaliation in the past 36 months? If so, how?\u00a0<\/span><\/p>\n<\/div>\n<h3><b>Category 3: Audits and Independent Evaluation<\/b><\/h3>\n<div style=\"margin-left: 2em\">\n<p><b>10. Independent Audits<\/b><\/p>\n<p><b>10a. <\/b><span style=\"font-weight: 400\">How many independent audits of the company\u2019s model behavior have been conducted in the past 24 months?<\/span><\/p>\n<p><b>10b. <\/b><span style=\"font-weight: 400\">Which external entities performed independent audits for the company (e.g. consultants, government agency, academic, nonprofit)?<\/span><\/p>\n<p><b>10c. <\/b><span style=\"font-weight: 400\">Were those company audit results made public? If not, why?<\/span><\/p>\n<\/div>\n<p><b>11. Child Safety Impact Assessments<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>11a. <\/b><span style=\"font-weight: 400\">How many of the company\u2019s child safety or user vulnerability impact assessments have been completed for active products?<\/span><\/p>\n<p><b>11b. <\/b><span style=\"font-weight: 400\">What were the dates and scope of these assessments?<\/span><\/p>\n<\/div>\n<p><b><\/b><b>12. Researcher and Journalist Access<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>12a. <\/b><span style=\"font-weight: 400\">How many external researchers or institutions have been granted access \u2013 through APIs or other means \u2013 to evaluate or test the company\u2019s models for safety purposes?<\/span><\/p>\n<p><b>12b. <\/b><span style=\"font-weight: 400\">What was the average response time for granting company access requests?<\/span><\/p>\n<p><b>12c. <\/b><span style=\"font-weight: 400\">How many requests were denied or delayed, and for what stated reasons?<\/span><\/p>\n<p><b>12d. <\/b><span style=\"font-weight: 400\">What percent of API requests resulted in published third-party research?<\/span><\/p>\n<\/div>\n<p><b>13. Internal Research Disclosures<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>13a. <\/b><span style=\"font-weight: 400\">How many internal studies examined correlations between long-term memory and unsafe reinforcement (e.g., self-harm, conspiracy engagement)?<\/span><\/p>\n<p><b>13b. <\/b><span style=\"font-weight: 400\">Have controlled tests compared models with sycophantic versus non-sycophantic behaviors? Please summarize any findings or share key result metrics.<\/span><\/p>\n<p><b>13c. <\/b><span style=\"font-weight: 400\">How frequently is model behavior evaluated for the intensity of sycophancy, and how does this evaluation account for session length?<\/span><\/p>\n<p><b>13d. <\/b><span style=\"font-weight: 400\">What percent of training data rewards agreement, flattery, or emotional reinforcement?<\/span><\/p>\n<\/div>\n<h3><b>Category 4: Public Disclosures<\/b><\/h3>\n<p><b>14. Incident Tracking<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>14a. <\/b><span style=\"font-weight: 400\">How many sycophancy-related incidents or complaints has the company logged in the past 12 months?<\/span><\/p>\n<p><b>14b. <\/b><span style=\"font-weight: 400\">What percent were resolved within 24, 48, and 72 hours?<\/span><\/p>\n<p><b>14c. <\/b><span style=\"font-weight: 400\">What was the median response time to confirmed safety incidents?<\/span><\/p>\n<\/div>\n<p><b>15. User Notification<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>15a. <\/b><span style=\"font-weight: 400\">How many of the company\u2019s users were notified directly of exposure to harmful or sycophantic outputs?<\/span><\/p>\n<p><b>15b. <\/b><span style=\"font-weight: 400\">How many public incident notices did the company issue in the last reporting year?<\/span><\/p>\n<\/div>\n<p><b>16. Safety Testing Publication<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>16a. <\/b><span style=\"font-weight: 400\">How often does the company publish safety testing results, and when was the most recent publication?\u00a0<\/span><\/p>\n<p><b>16b. <\/b><span style=\"font-weight: 400\">What proportion of internal safety evaluations does the company make public?<\/span><\/p>\n<\/div>\n<p><b>17. Complaint Reporting Channels<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>17a. <\/b><span style=\"font-weight: 400\">How many complaints or safety concerns did the company receive by users, employees, or contractors in the past year?<\/span><\/p>\n<p><b>17b. <\/b><span style=\"font-weight: 400\">What channels exist for reporting to the company (e.g., webform, hotline, anonymous submission)?<\/span><\/p>\n<p><b>17c. <\/b><span style=\"font-weight: 400\">What percent of users who filed complaints about the company received confirmation or resolution feedback, and what was the median time for that confirmation or resolution?<\/span><\/p>\n<\/div>\n<p><b>18. Data Sources<\/b><\/p>\n<div style=\"margin-left: 2em\">\n<p><b>18a. <\/b><span style=\"font-weight: 400\">What percent of the company\u2019s training data sources are publicly documented?<\/span><\/p>\n<p><b>18b. <\/b><span style=\"font-weight: 400\">How many of the company\u2019s datasets have been removed, modified, or redacted due to sycophancy or safety concerns?<\/span><\/p>\n<p><b>18c. <\/b><span style=\"font-weight: 400\">What is the company\u2019s process for updating public datasets in response to safety reviews, and how often does this occur?<\/span><\/p>\n<\/div>\n<p><span style=\"font-weight: 400\">\u2013<\/span><\/p>\n<p><a href=\"https:\/\/www.law.georgetown.edu\/tech-institute\/people\/our-team\/stephanie-nguyen\/\"><b>Stephanie T. Nguyen<\/b><\/a><span style=\"font-weight: 400\"> is a Senior Fellow at Georgetown Institute for Technology Law &amp; Policy, Former Chief Technologist at the Federal Trade Commission<\/span><\/p>\n<p><a href=\"https:\/\/www.law.georgetown.edu\/tech-institute\/people\/our-team\/erie-meyer\/\"><b>Erie Meyer<\/b><\/a><span style=\"font-weight: 400\"> is a Senior Fellow at Georgetown Institute for Technology Law &amp; Policy, Former CFPB Chief Technologist<\/span><\/p>\n<p><a href=\"https:\/\/www.law.berkeley.edu\/article\/center-for-consumer-law-and-economic-justice-new-senior-fellows-antitrust-financial-regulation\/\"><b>Samuel A.A. Levine<\/b><\/a><span style=\"font-weight: 400\"> is a Senior Fellow at UC Berkeley Center for Consumer Law &amp; Economic Justice, Former Bureau of Consumer Protection Director at the Federal Trade Commission<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>November 3, 2025 &nbsp; As we have outlined in a prior tech brief, AI sycophancy \u2013 the tendency of the outputs of AI-enabled chatbots to be \u201coverly flattering or agreeable\u201d [&hellip;]<\/p>\n","protected":false},"author":18544,"featured_media":0,"parent":7881,"menu_order":10,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_acf_changed":false,"_price":"","_stock":"","_tribe_ticket_header":"","_tribe_default_ticket_provider":"","_tribe_ticket_capacity":"0","_ticket_start_date":"","_ticket_end_date":"","_tribe_ticket_show_description":"","_tribe_ticket_show_not_going":false,"_tribe_ticket_use_global_stock":"","_tribe_ticket_global_stock_level":"","_global_stock_mode":"","_global_stock_cap":"","_tribe_rsvp_for_event":"","_tribe_ticket_going_count":"","_tribe_ticket_not_going_count":"","_tribe_tickets_list":"[]","_tribe_ticket_has_attendee_info_fields":false,"footnotes":"","_tec_slr_enabled":"","_tec_slr_layout":""},"class_list":["post-8659","page","type-page","status-publish","hentry"],"acf":[],"ticketed":false,"_links":{"self":[{"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/pages\/8659","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/users\/18544"}],"replies":[{"embeddable":true,"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/comments?post=8659"}],"version-history":[{"count":22,"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/pages\/8659\/revisions"}],"predecessor-version":[{"id":8715,"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/pages\/8659\/revisions\/8715"}],"up":[{"embeddable":true,"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/pages\/7881"}],"wp:attachment":[{"href":"https:\/\/www.law.georgetown.edu\/tech-institute\/wp-json\/wp\/v2\/media?parent=8659"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}