OpenAI introduces new ‘Trusted Contact’ safeguard for cases of possible self-harm
OpenAI introduces new ‘Trusted Contact’ safeguard for cases of possible self-harm
OpenAI 推出全新的“受信任联系人”安全保障功能,应对潜在的自残风险
On Thursday OpenAI announced a new feature called Trusted Contact, designed to alert a trusted third party if mentions of self-harm are expressed within a conversation. The feature allows an adult ChatGPT user to designate another person as a trusted contact within their account, such as a friend or family member. In cases where a conversation may turn to self-harm, OpenAI will now encourage the user to reach out to that contact. It also sends an automated alert to the contact, encouraging them to check in with the user.
周四,OpenAI 宣布推出一项名为“受信任联系人”(Trusted Contact)的新功能,旨在当对话中出现自残相关言论时,向受信任的第三方发出警报。该功能允许 ChatGPT 的成年用户在账户中指定另一人(如朋友或家人)作为受信任联系人。当对话可能涉及自残倾向时,OpenAI 现在会鼓励用户联系该受信任对象,并向该联系人发送自动提醒,鼓励他们主动关心用户。
OpenAI has faced a wave of lawsuits from the families of people who have committed suicide after talking with its chatbot. In a number of cases, the families say ChatGPT encouraged their loved one to kill themselves — or even helped them plan it out.
OpenAI 此前遭遇了一波来自自杀者家属的诉讼,这些死者生前曾与该公司的聊天机器人进行过交流。在多起案件中,家属称 ChatGPT 鼓励了他们的亲人自杀,甚至协助他们策划了自杀过程。
OpenAI currently uses a combination of automation and human review to handle potentially harmful incidents. Certain conversational triggers alert the company’s system to suicidal ideations, which then relay the information to a human safety team. The company claims that every time it receives this kind of notification, the incident is reviewed by a human. “We strive to review these safety notifications in under one hour,” the company says.
OpenAI 目前采用自动化与人工审核相结合的方式来处理潜在的有害事件。特定的对话触发词会向公司系统发出自杀意念预警,随后信息会被转交给人工安全团队。该公司声称,每当收到此类通知,都会由人工进行审核。公司表示:“我们力争在一小时内完成对这些安全通知的审核。”
If OpenAI’s internal team decides that the situation represents a serious safety risk, ChatGPT proceeds to send the trusted contact an alert — either by email, text message, or an in-app notification. The alert is designed to be brief and to encourage the contact to check in with the person in question. It does not include detailed information about what was being discussed, as a means of protecting the user’s privacy, the company says.
如果 OpenAI 的内部团队判定情况存在严重安全风险,ChatGPT 将会向受信任联系人发送警报——通过电子邮件、短信或应用内通知。该警报设计简洁,旨在鼓励联系人主动关心当事人。公司表示,为了保护用户隐私,警报中不会包含对话内容的详细信息。
The Trusted Contact feature follows the safeguards the company introduced last September that gave parents the power to have some oversight of their teens’ accounts, including receiving safety notifications designed to alert the parent if OpenAI’s system believes their child is facing a “serious safety risk.” For some time now, ChatGPT has also included automated alerts to seek professional health services, should a conversation trend toward the topic of self-harm.
“受信任联系人”功能是继去年 9 月公司推出的安全保障措施之后的又一举措。此前的措施赋予了家长对青少年账户的一定监管权,包括在 OpenAI 系统认为孩子面临“严重安全风险”时,向家长发送安全通知。此外,ChatGPT 此前已包含自动提醒功能,若对话涉及自残话题,系统会建议用户寻求专业医疗服务。
Crucially, Trust Contact is optional and, even if the protection is activated on a particular account, any user can have multiple ChatGPT accounts. OpenAI’s parental controls are also optional, presenting a similar limitation.
关键在于,“受信任联系人”功能是可选的。即便某个账户开启了此保护功能,用户仍可拥有多个 ChatGPT 账户。OpenAI 的家长控制功能同样是可选的,也存在类似的局限性。
“Trusted Contact is part of OpenAI’s broader effort to build AI systems that help people during difficult moments,” the company wrote in the announcement post. “We will continue to work with clinicians, researchers, and policymakers to improve how AI systems respond when people may be experiencing distress.”
“‘受信任联系人’是 OpenAI 更广泛努力的一部分,旨在构建能够在人们困难时刻提供帮助的 AI 系统,”该公司在公告中写道。“我们将继续与临床医生、研究人员和政策制定者合作,改进 AI 系统在人们可能处于困境时的响应方式。”