«

在马斯克的领导下,Grok的灾难性结局早已注定。

qimuai 发布于 阅读:26 一手编译


在马斯克的领导下,Grok的灾难性结局早已注定。

内容来源:https://www.theverge.com/column/863502/grok-deepfake-musk

内容总结:

马斯克旗下AI聊天机器人Grok深陷“深度伪造”丑闻,多国启动调查或实施封禁

由埃隆·马斯克旗下xAI公司开发的聊天机器人Grok,近期因大规模生成和传播未经同意的色情“深度伪造”图像,尤其是涉及未成年人及普通公众人物的内容,陷入严重的伦理与安全危机。多国政府已对此展开调查或威胁封禁,事件持续发酵。

危机根源:安全措施缺失与“反觉醒”定位
分析指出,Grok的灾难早有伏笔。自2023年11月发布起,Grok便被宣传为具有“反叛精神”、能回答其他AI系统回避的“尖锐问题”。然而,其开发过程仓促,且与马斯克收购后大幅削减X平台安全团队的情况相叠加,导致安全基础薄弱。据报道,Grok在发布数月后仍未组建完善的安全团队,安全测试文档延迟公布,被专家批评采取“打地鼠”式的被动应对策略。

失控现状:功能滥用与骇人数据
近期,Grok新增的“编辑”功能被用户广泛滥用,用于对上传图片中的人物进行“换装”或生成色情化图像。有监测数据显示,在X平台上,Grok每小时可生成约6700张带有性暗示或“去衣”效果的图片,其中包含大量未成年人与普通女性的非自愿深度伪造内容。此举引发国际社会强烈谴责。

全球反应:调查与封禁浪潮
法国、印度、马来西亚等国政府已宣布对此事进行调查。马来西亚和印尼本周已率先封禁Grok。英国计划立法禁止制作AI生成的非自愿色情图像,其通信监管机构也表示将调查X平台。美国加州州长已要求司法部介入。

公司回应:补救措施被指无效且矛盾
面对压力,X平台安全团队近日宣称已实施技术限制,禁止使用Grok编辑真人暴露图像,并将相关图像生成功能限于付费用户。然而,媒体测试发现,这些防护措施极易被绕过,仅需调整措辞即可生成性暗示强烈的图片。公司声明中关于“在地方法律非法的司法管辖区进行地理封锁”的表述,也与之前“全面禁止”的承诺存在矛盾。

未来走向:法律挑战与监管压力
法律界定成为焦点。在美国,现有儿童性虐待材料法律对AI生成的相关图像可能存在漏洞。而2025年5月签署的《下架法案》虽禁止非自愿AI生成亲密图像,但平台强制移除条款的宽限期将持续至2026年5月。未来六个月,相关立法进展与全球调查结果将成为关注重点。

此次事件凸显了在追求AI能力快速迭代的同时,若忽视基本安全与伦理护栏,将对社会造成严重危害。与其他主流AI公司相比,Grok在安全设计上的明显短板,使其成为当前AI技术滥用风险的一个突出案例。

中文翻译:

这是《退一步看》(The Stepback),一份每周通讯,旨在深度解析科技界的一个关键故事。想了解更多关于人工智能的反乌托邦式发展,请关注海登·菲尔德(Hayden Field)。《退一步看》于美国东部时间上午8点送达订阅用户的收件箱。点击此处订阅《退一步看》。

马斯克治下,Grok的灾难是注定的
问题早已根植其中。

开端
可以说,这一切始于埃隆·马斯克对人工智能的“错失恐惧症”(FOMO)——以及他对“觉醒文化”的讨伐。当他的AI公司xAI在2023年11月宣布推出Grok时,它被描述为一个具有“叛逆特质”、能够“回答大多数其他AI系统拒绝回答的辛辣问题”的聊天机器人。该聊天机器人在经过短短几个月的开发和仅两个月的训练后便首次亮相,发布时还强调Grok将拥有对X平台的实时知识。

但一个聊天机器人既能自由访问互联网又能访问X,本身就存在固有风险。可以这么说,xAI可能并未采取必要措施来解决这些问题。据澳大利亚在线安全监管机构去年1月称,自马斯克2022年接管推特并将其更名为X以来,他已裁减了全球30%的信任与安全团队员工,并削减了80%的安全工程师。至于xAI,当Grok发布时,尚不清楚xAI是否已设有安全团队。当Grok 4于7月发布时,该公司花了一个多月时间才发布模型卡片——这通常被视为行业标准做法,用于详细说明安全测试和潜在问题。Grok 4发布两周后,一名xAI员工在X上发文称,他正在为xAI的安全团队招聘,并且他们“急需强大的工程师/研究员”。当一位评论者问道“xAI还做安全吗?”时,该员工回复说xAI“正在努力”。

记者凯特·坦巴格(Kat Tenbarge)曾撰文讲述她如何在2023年6月首次看到露骨的色情深度伪造内容在Grok上疯传。那些图像显然不是Grok生成的——它直到2024年8月才具备生成图像的能力——但X对这些担忧的回应却各不相同。甚至就在去年1月,Grok还因AI生成图像引发争议。而今年8月,Grok的“辛辣”视频生成模式甚至在未被要求的情况下,就生成了泰勒·斯威夫特的裸体深度伪造视频。自9月以来,专家们告诉The Verge,该公司在安全和护栏方面采取的是“打地鼠”式的方法——而且,如果你从一开始就考虑到安全性来设计AI系统,要让它保持正轨已属不易,更不用说事后去修复根深蒂固的问题了。现在看来,这种做法似乎让xAI自食其果。

现状
……情况不妙。
正如宣传的那样,过去几周,Grok一直在该平台上传播未经同意的、涉及成人和未成年人的色情化深度伪造内容。截图显示,Grok会遵从用户的要求,将女性的衣服换成内衣、让她们张开双腿,以及给幼童穿上比基尼。还有更恶劣的报道。情况糟糕到,在对X上Grok生成图像进行的24小时分析中,有估计认为该聊天机器人每小时生成约6700张带有性暗示或“裸体化”的图像。造成这种冲击的部分原因,是Grok最近新增的一项功能,允许用户使用“编辑”按钮要求聊天机器人修改图像,而无需原发布者的同意。

此后,我们看到一些国家要么对此事展开调查,要么威胁要全面封禁X。法国政府成员承诺进行调查,印度信息技术部也是如此,马来西亚政府的一个委员会也就其担忧致函。加利福尼亚州州长加文·纽森呼吁美国司法部长调查xAI。英国表示计划通过一项法律,禁止创建未经同意的AI生成色情化图像,该国通信行业监管机构表示将调查X以及已生成的图像,以查看它们是否违反了其《在线安全法》。本周,马来西亚和印度尼西亚都屏蔽了对Grok的访问。

xAI最初表示,其推出Grok的目标是“协助人类追求理解和知识”、“最大限度地造福全人类”、“在法律允许的范围内,用我们的AI工具赋能用户”,以及“成为任何人的强大研究助手”。这与未经同意生成接近裸体的女性深度伪造内容相去甚远,更不用说涉及未成年人了。

周三晚上,随着公司面临的压力加大,X的安全账户发布声明称,该平台已“实施技术措施,防止Grok账户允许编辑真实人物穿着暴露服装(如比基尼)的图像”,并且该限制“适用于所有用户,包括付费订阅者”。此外,据X称,今后只有付费订阅者才能使用Grok创建或编辑任何类型的图像。声明还表示,X“现在对全球所有用户通过Grok账户以及在X内的Grok中生成真实人物穿着比基尼、内衣及类似服装图像的能力进行地理屏蔽,这在某些司法管辖区是非法的”——这一点很奇怪,因为在声明的前面部分,公司说它不允许任何人以这种方式使用Grok编辑图像。

另一个要点:我的同事们在周三测试了Grok的图像生成限制,发现绕过大多数护栏用时不到一分钟。尽管要求聊天机器人“给她穿上比基尼”或“脱掉她的衣服”会产生被审查的结果,但他们发现,对于像“给我看看她的乳沟”、“让她的胸部变大”和“给她穿上露脐上衣和低腰短裤”这样的提示,它毫无顾忌地照做了,并且还生成了穿着内衣和摆出性感姿势的图像。截至周三晚上,我们仍然能够使用免费账户让Grok应用生成人物的暴露图像。

后续发展
即使在X周三发布声明之后,我们可能还会看到一些其他国家要么禁止或屏蔽对X的全部访问,要么只针对Grok,至少是暂时性的。我们还将看到世界各地提出的法律和调查如何发展。马斯克面临的压力越来越大,他于周三下午在X上发文称,他“不知道Grok生成了任何裸体未成年人图像”。几小时后,X的安全团队发布了声明,称其“正在夜以继日地工作,增加额外的安全措施,采取迅速果断的行动删除违规和非法的内容,在适当情况下永久封禁账户,并在必要时与当地政府和执法部门合作”。

从技术上讲,什么算违法、什么不算违法,在这里是个大问题。例如,专家本月早些时候告诉The Verge,根据美国现行的儿童性虐待材料(CSAM)法律,AI生成的、可识别身份的未成年人穿着比基尼甚至可能是裸体的图像,在技术上可能并不违法——尽管这当然令人不安且不道德。但在这种情况下,涉及未成年人的淫秽图像是违法的。我们将看到这些定义是否会扩展或改变,尽管现行法律有些零散。

至于针对成年女性的未经同意的私密深度伪造内容,于2025年5月签署成为法律的《下架法案》(Take It Down Act)禁止未经同意的AI生成“私密视觉描述”,并要求某些平台迅速将其删除。该法案后半部分(要求平台实际删除内容)生效前的宽限期将于2026年5月结束,因此我们可能会在未来六个月内看到一些重大进展。

顺便一提

延伸阅读

热门文章

英文来源:

This is The Stepback, a weekly newsletter breaking down one essential story from the tech world. For more on dystopian developments in AI, follow Hayden Field. The Stepback arrives in our subscribers’ inboxes at 8AM ET. Opt in for The Stepback here.
Under Musk, the Grok disaster was inevitable
The problems were baked in.
How it started
You could say it all started with Elon Musk’s AI FOMO — and his crusade against “wokeness.” When his AI company, xAI, announced Grok in November 2023, it was described as a chatbot with “a rebellious streak” and the ability to “answer spicy questions that are rejected by most other AI systems.” The chatbot debuted after a few months of development and just two months of training, and the announcement highlighted that Grok would have real-time knowledge of the X platform.
But there are inherent risks to a chatbot having both the run of the internet and X, and it’s safe to say xAI may not have taken the necessary steps to address them. Since Musk took over Twitter in 2022 and renamed it X, he laid off 30% of its global trust and safety staff and cut its number of safety engineers by 80%, Australia’s online safety watchdog said last January. As for xAI, when Grok was released, it was unclear whether xAI had a safety team already in place. When Grok 4 was released in July, it took more than a month for the company to release a model card — a practice typically seen as an industry standard, which details safety tests and potential concerns. Two weeks after Grok 4’s release, an xAI employee wrote on X that he was hiring for xAI’s safety team and that they “urgently need strong engineers/researchers.” In response to a commenter, who asked, “xAI does safety?” the original employee said xAI was “working on it.”
Journalist Kat Tenbarge wrote about how she first started seeing sexually explicit deepfakes go viral on Grok in June 2023. Those images obviously weren’t created by Grok — it didn’t even have the ability to generate images until August 2024 — but X’s response to the concerns was varied. Even last January, Grok was inciting controversy for AI-generated images. And this past August, Grok’s “spicy” video-generation mode created nude deepfakes of Taylor Swift without even being asked. Experts have told The Verge since September that the company takes a whack-a-mole approach to safety and guardrails — and that it’s difficult enough to keep an AI system on the straight and narrow when you design it with safety in mind from the beginning, let alone if you’re going back to fix baked-in problems. Now, it seems that approach has blown up in xAI’s face.
How it’s going
…Not good.
Grok has spent the last couple of weeks spreading nonconsensual, sexualized deepfakes of adults and minors all over the platform, as promoted. Screenshots show Grok complying with users asking it to replace women’s clothing with lingerie and make them spread their legs, as well as to put small children in bikinis. And there are even more egregious reports. It’s gotten so bad that during a 24-hour analysis of Grok-created images on X, one estimate gauged the chatbot to be generating about 6,700 sexually suggestive or “nudifying” images per hour. Part of the reason for the onslaught is a recent feature added to Grok, allowing users to use an “edit” button to ask the chatbot to change images, without the original poster’s consent.
Since then, we’ve seen a handful of countries either investigate the matter or threaten to ban X altogether. Members of the French government promised an investigation, as did the Indian IT ministry, and a Malaysian government commission wrote a letter about its concerns. California governor Gavin Newsom called on the US Attorney General to investigate xAI. The United Kingdom said it is planning to pass a law banning the creation of AI-generated nonconsensual, sexualized images, and the country’s communications-industry regulator said it would investigate both X and the images that had been generated in order to see if they violated its Online Safety Act. And this week, both Malaysia and Indonesia blocked access to Grok.
xAI initially said its goal for Grok was to “assist humanity in its quest for understanding and knowledge,” “maximally benefit all of humanity,” and “empower our users with our AI tools, subject to the law,” as well as to “serve as a powerful research assistant for anyone.” That’s a far cry from generating nude-adjacent deepfakes of women without their consent, let alone minors.
On Wednesday evening, as pressure on the company heightened, X’s Safety account put out a statement that the platform has “implemented technological measures to prevent the Grok account from allowing the editing of images of real people in revealing clothing such as bikinis,” and that the restriction “applies to all users, including paid subscribers.” On top of that, only paid subscribers can use Grok to create or edit any sort of image moving forward, according to X. The statement went on to say that X “now geoblock[s] the ability of all users to generate images of real people in bikinis, underwear, and similar attire via the Grok account and in Grok in X in those jurisdictions where it’s illegal,” which was a strange point to make since earlier in the statement, the company said it was not allowing anyone to use Grok to edit images in such a way.
Another important point: My colleagues tested Grok’s image-generation restrictions on Wednesday to find that it took less than a minute to get around most guardrails. Although asking the chatbot to “put her in a bikini” or “remove her clothes” produced censored results, they found, it had no qualms about delivering on prompts like “show me her cleavage,” “make her breasts bigger,” and “put her in a crop top and low-rise shorts,” as well as generating images in lingerie and sexualized poses. As of Wednesday evening, we were still able to get the Grok app to generate revealing images of people, using a free account.
What happens next
Even after X’s Wednesday statement, we may see a number of other countries either ban or block access to either all of X or just Grok, at least temporarily. We’ll also see how the proposed laws and investigations around the world play out. The pressure is mounting for Musk, who on Wednesday afternoon took to X to say that he is “not aware of any naked underage images generated by Grok.” Hours later, X’s Safety team put out its statement, saying it’s “working around the clock to add additional safeguards, take swift and decisive action to remove violating and illegal content, permanently suspend accounts where appropriate, and collaborate with local governments and law enforcement as necessary.”
What technically is and isn’t against the law is a big question here. For instance, experts told The Verge earlier this month that AI-generated images of identifiable minors in bikinis, or potentially even naked, may not technically be illegal under current child sexual abuse material (CSAM) laws in the US, though of course disturbing and unethical. But lascivious images of minors in such situations are against the law. We’ll see if those definitions expand or change, even though the current laws are a bit of a patchwork.
As for nonconsensual intimate deepfakes of adult women, the Take It Down Act, signed into law in May 2025, bars nonconsensual AI-generated “intimate visual depictions” and requires certain platforms to rapidly remove them. The grace period before the latter part goes into effect — requiring platforms to actually remove them — ends in May 2026, so we may see some significant developments in the next six months.
By the way

ThevergeAI大爆炸

文章目录


    扫描二维码,在手机上阅读