<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/">
    <channel>
        <title>Rocket Lu</title>
        <link>http://rocketlu.cn/</link>
        <description>AIGC+ART+DATAVIS</description>
        <lastBuildDate>Fri, 06 Mar 2026 06:18:00 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>https://github.com/jpmonette/feed</generator>
        <language>zh-CN</language>
        <copyright>All rights reserved 2026, Rocketlu</copyright>
        <item>
            <title><![CDATA[Kyutai STT 一种专为实时应用优化的语音转文字模型]]></title>
            <link>http://rocketlu.cn/article/Kyutai STT</link>
            <guid>http://rocketlu.cn/article/Kyutai STT</guid>
            <pubDate>Fri, 04 Jul 2025 00:00:00 GMT</pubDate>
            <description><![CDATA[一种专为实时应用优化的语音转文字模型]]></description>
            <content:encoded><![CDATA[<div id="notion-article" class="mx-auto overflow-hidden "><main class="notion light-mode notion-page notion-block-22637cbb8d10800ba23ffc420977d6e5"><div class="notion-viewport"></div><div class="notion-collection-page-properties"></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d108008ae45d6446adffd82" data-id="22637cbb8d108008ae45d6446adffd82"><span><div id="22637cbb8d108008ae45d6446adffd82" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d108008ae45d6446adffd82" title="Kyutai STT"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Kyutai STT</span></span></h4><div class="notion-text notion-block-22637cbb8d108056a629cb570b1ded00">一种专为实时应用优化的语音转文字（Speech-to-Text）模型。</div><div class="notion-text notion-block-22637cbb8d1080f2bda7d9fc41f9e26b">👉 你可以在 <a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://unmute.sh/">unmute.sh</a> 上试用</div><div class="notion-text notion-block-22637cbb8d1080e8b3b0e0cb27f54a19">👉 查看代码：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://github.com/">GitHub</a></div><div class="notion-blank notion-block-22637cbb8d10801d8f87e6aa74316f97"> </div><div class="notion-text notion-block-22637cbb8d108059bb27de2352c1467e">Kyutai STT 是一种流式语音转文字模型架构，在延迟和准确性之间取得了出色的平衡，非常适合交互式应用。它支持批处理（batching），因此只需一块 GPU 就能同时处理数百个对话。</div><div class="notion-blank notion-block-22637cbb8d1080aea816d4b06bdde909"> </div><div class="notion-text notion-block-22637cbb8d1080e5b623e7f18176932f">我们发布了两个模型：</div><ul class="notion-list notion-list-disc notion-block-22637cbb8d1080f1900bc9f2a7236929"><li><b>kyutai/stt-1b-en_fr</b>：低延迟模型，支持英文和法文，内置语义语音活动检测（VAD）。</li></ul><ul class="notion-list notion-list-disc notion-block-22637cbb8d1080759ce1da0edf642e0a"><li><b>kyutai/stt-2.6b-en</b>：更大的英文专用模型，追求极致准确性。</li></ul><div class="notion-blank notion-block-22637cbb8d108039a307c40f5bae0154"> </div><hr class="notion-hr notion-block-22637cbb8d1080c595dcec159bd44780"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d1080a1b51dcaae150e3468" data-id="22637cbb8d1080a1b51dcaae150e3468"><span><div id="22637cbb8d1080a1b51dcaae150e3468" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d1080a1b51dcaae150e3468" title="1.实时且准确"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">1.实时且准确</span></span></h4><div class="notion-text notion-block-22637cbb8d108083b6c9ca0819db6c2a">字错误率（WER）图表</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d1080a98595d2024ba9df08"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A2c45b8d1-16d0-4c0b-a938-d06370fd72cc%3Aimage.png?table=block&amp;id=22637cbb-8d10-80a9-8595-d2024ba9df08&amp;t=22637cbb-8d10-80a9-8595-d2024ba9df08" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-22637cbb8d1080aab040f699ac0dd125">字错误率越低越好。</div><div class="notion-blank notion-block-22637cbb8d10805fa74aed738927ccd7"> </div><div class="notion-text notion-block-22637cbb8d108022a80fe73a63f50882">Kyutai STT 是一种“流式”模型，意味着它会一边接收音频一边实时转录，而不是等到整段音频输入完成后再开始处理。因此非常适用于实时应用，比如 <a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://unmute.sh/">Unmute</a>。</div><div class="notion-text notion-block-22637cbb8d1080248bfbd88bd6ce6918">它能输出格式规范、带有标点的转录结果，还支持逐词时间戳。</div><div class="notion-text notion-block-22637cbb8d1080048f00f3a2d98709f4">在准确率方面，它的表现与目前最先进的非流式模型相当，后者通常需要整段音频数据。</div><hr class="notion-hr notion-block-22637cbb8d10805fb4e5e513b5743871"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d1080049f70d94c1851ad03" data-id="22637cbb8d1080049f70d94c1851ad03"><span><div id="22637cbb8d1080049f70d94c1851ad03" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d1080049f70d94c1851ad03" title="2.语义语音活动检测（Semantic VAD）"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">2.语义语音活动检测（Semantic VAD）</span></span></h4><div class="notion-text notion-block-22637cbb8d108030a5dfe8437ad461ee">对于像 Unmute 这样需要语音对话的应用，我们需要判断用户是否已经说完话，以便系统可以开始回应。</div><div class="notion-text notion-block-22637cbb8d10806d8043de8e1fb6788a">常见的方法是使用一个单独的语音活动检测模型，判断用户是否正在说话，然后在检测到用户停止说话后等待一段固定时间。</div><div class="notion-text notion-block-22637cbb8d108083993afa34cd9250f8">但这种方法有缺陷——人们说话时经常会暂停，固定等待时间很难适配所有情况，容易误判。</div><div class="notion-blank notion-block-22637cbb8d1080039b4ae99268caf2e9"> </div><div class="notion-text notion-block-22637cbb8d10804daf84e575f506f323">Kyutai STT 的解决方案是：不仅预测文本，还预测用户是否已经说完。系统会根据说话内容和语调，智能调整等待时间。</div><div class="notion-text notion-block-22637cbb8d1080719975cf61c05a1a56">你可以在上面的演示中体验这一功能，注意提示“End of speech detected”。</div><div class="notion-text notion-block-22637cbb8d1080c2af7ccdc7440d5c31">目前，语义 VAD 只在 Rust 版本的服务器中提供，其他实现尚未支持。</div><hr class="notion-hr notion-block-22637cbb8d1080b39a46ddf785b1dc56"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d10805c9118db2be9a3c120" data-id="22637cbb8d10805c9118db2be9a3c120"><span><div id="22637cbb8d10805c9118db2be9a3c120" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d10805c9118db2be9a3c120" title="3.超低延迟"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">3.超低延迟</span></span></h4><ul class="notion-list notion-list-disc notion-block-22637cbb8d1080fd8904decd69b96132"><li><b>kyutai/stt-1b-en_fr</b> 模型的延迟为 500 毫秒，即说出一个词后，大约 0.5 秒内就能转录出来。</li></ul><ul class="notion-list notion-list-disc notion-block-22637cbb8d1080819e74e8487a6eb698"><li><b>kyutai/stt-2.6b-en</b> 的延迟为 2.5 秒，换取更高的准确率。</li></ul><div class="notion-text notion-block-22637cbb8d1080b8aee7e06280a8944b">在 Unmute 中，我们使用一种叫做 “flush trick” 的技术进一步降低响应延迟：</div><div class="notion-text notion-block-22637cbb8d1080abaf9ce6ef8a5a4813">当语音活动检测器判断用户说完后，虽然还要等 500ms（模型的延迟），但我们会让 STT 服务器尽快处理已有音频。</div><div class="notion-text notion-block-22637cbb8d1080188c05e5ba3f19419a">服务器的处理速度约为 4 倍实时速率，因此只需 125ms（500ms ÷ 4）即可处理完现有音频。通过这种方式，我们“加速了时间”，只需等 125ms 就能确保转录完成。</div><hr class="notion-hr notion-block-22637cbb8d1080f4a710d55221e1deee"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d1080089031fa2221d4fdd4" data-id="22637cbb8d1080089031fa2221d4fdd4"><span><div id="22637cbb8d1080089031fa2221d4fdd4" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d1080089031fa2221d4fdd4" title="4.高并发能力"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">4.高并发能力</span></span></h4><div class="notion-text notion-block-22637cbb8d10803bb145c2d5d3909ae0">Kyutai STT 非常适合生产环境：</div><div class="notion-text notion-block-22637cbb8d108042be53f836787ab912">在一块 H100 GPU 上，它可以<b>同时处理 400 条实时音频流</b>。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d1080b2917dd1c8c90a8bd9"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A762746bc-9efa-4ac5-b1ca-b76166a299e9%3Aimage.png?table=block&amp;id=22637cbb-8d10-80b2-917d-d1c8c90a8bd9&amp;t=22637cbb-8d10-80b2-917d-d1c8c90a8bd9" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-22637cbb8d108060a483ea5c08ec1997">这得益于我们独创的 <b>延迟流建模架构（delayed streams modeling）</b>，让模型本身就能高效地批量处理数据，无需额外代码支持流式处理。</div><hr class="notion-hr notion-block-22637cbb8d10809ebd08dbb31003dc02"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d108070a688d6682606e87a" data-id="22637cbb8d108070a688d6682606e87a"><span><div id="22637cbb8d108070a688d6682606e87a" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d108070a688d6682606e87a" title="5.单通道语音转文字"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">5.单通道语音转文字</span></span></h4><div class="notion-text notion-block-22637cbb8d108068a1eff9c75b3b6494">相比之下，将 OpenAI 的 Whisper 模型变为流式（Whisper-Streaming）则需要单独的研究项目。这种方法是反复处理最后几秒音频，并拼接结果。</div><div class="notion-text notion-block-22637cbb8d10801e964fe241566c1dba">虽然技术上很强大，但 Whisper-Streaming 不支持批处理，因此吞吐量远低于 Kyutai STT。如果你希望延迟更低，它还需要更频繁地重新处理音频，进一步降低效率。</div><hr class="notion-hr notion-block-22637cbb8d108040bf4ef56af07ad2d3"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d1080ffbe55c2414cc22e29" data-id="22637cbb8d1080ffbe55c2414cc22e29"><span><div id="22637cbb8d1080ffbe55c2414cc22e29" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d1080ffbe55c2414cc22e29" title="6.多种实现方式"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">6.多种实现方式</span></span></h4><div class="notion-text notion-block-22637cbb8d1080e580f3fef055ae1aa0">根据你的需求，我们提供不同的实现版本，详细说明见<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://github.com/kyutai-labs/delayed-streams-modeling/"> GitHub</a>：</div><ul class="notion-list notion-list-disc notion-block-22637cbb8d108093a93ffd581f0f3394"><li><b>PyTorch 版</b>：适合研究和实验。如果你想在 Python 中调用模型，可选此版本。</li></ul><ul class="notion-list notion-list-disc notion-block-22637cbb8d10801e8d85d497a1f71131"><li><b>Rust 版</b>：适合生产环境部署。Unmute 就是使用这个版本。</li><ul class="notion-list notion-list-disc notion-block-22637cbb8d10801e8d85d497a1f71131"><li>我们的 Rust 服务支持通过 websocket 进行流式访问。</li><li>在 L40S GPU 上，可以以 3 倍实时速率服务 64 个并发连接。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-22637cbb8d10806d874ccd3e0fcbaaf6"><li><b>MLX 版</b>：适用于在 iPhone 和 Mac 上进行设备端推理。</li><ul class="notion-list notion-list-disc notion-block-22637cbb8d10806d874ccd3e0fcbaaf6"><li>MLX 是 Apple 的机器学习框架，支持 Apple Silicon 上的硬件加速。</li></ul></ul><hr class="notion-hr notion-block-22637cbb8d1080588ba7e55bfc9da251"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d108049a00ae5120e73e4a2" data-id="22637cbb8d108049a00ae5120e73e4a2"><span><div id="22637cbb8d108049a00ae5120e73e4a2" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d108049a00ae5120e73e4a2" title="7.延迟流建模（Delayed Streams Modeling）"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">7.延迟流建模（Delayed Streams Modeling）</span></span></h4><div class="notion-text notion-block-22637cbb8d1080a990c9e6df21593042">Kyutai STT 的核心创新，是我们在 Kyutai 首创的一项技术，称为“延迟流建模”，最初在 <a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://moshi.chat/">Moshi</a> 项目中提出。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d1080d180a7d41a1a0efc2d"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Abf2dd8e6-a1ac-4bd3-baa2-31d141618348%3Aimage.png?table=block&amp;id=22637cbb-8d10-80d1-80a7-d41a1a0efc2d&amp;t=22637cbb-8d10-80d1-80a7-d41a1a0efc2d" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-22637cbb8d1080ecbdffdf010009fd1d">传统的语音转文字方法，是把完整音频输入模型，然后逐步生成文本（如 Whisper 采用的编码器-解码器结构）。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d1080e59e8efe771a0169ab"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A1d721286-8b1e-456a-9833-55768ed45dda%3Aimage.png?table=block&amp;id=22637cbb-8d10-80e5-9e8e-fe771a0169ab&amp;t=22637cbb-8d10-80e5-9e8e-fe771a0169ab" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-22637cbb8d1080cbb66ff8423e431bda">而 Kyutai STT 则将音频与文本建模为“时间对齐”的两个流：音频流和文字流是并列的，而不是线性先后关系。我们会延迟文本流几个时间帧，让模型有“前瞻”能力。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d1080e9a489c1f3271f8a1f"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Aa80fe87c-3a6f-4390-88b8-32fe7c1010ac%3Aimage.png?table=block&amp;id=22637cbb-8d10-80e9-a489-c1f3271f8a1f&amp;t=22637cbb-8d10-80e9-a489-c1f3271f8a1f" alt="notion image" loading="lazy" decoding="async"/></div></figure><blockquote class="notion-quote notion-block-22637cbb8d1080db8530d502af7509d1"><div><b>训练时：</b>模型学会同时建模音频和文字两个流。</div><div class="notion-text notion-block-22637cbb8d108062af53eebb6abb2d9d"><b>推理时：</b>我们实时输入音频，模型根据音频预测文本。</div></blockquote><div class="notion-text notion-block-22637cbb8d1080f0ae0efbe1f49e9569">这种方式还有个好处是<b>对称性</b>：我们只要将延迟从文本流切换到音频流，再把文字作为输入固定，就可以变成一个文字转语音模型。我们只需在模型中加一点技巧，让它预测空白 token 来对齐时间轴。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d10808ab8b3f09582946742"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A0f09fbc5-5d9d-4170-991a-3ec10c9420ad%3Aimage.png?table=block&amp;id=22637cbb-8d10-808a-b8b3-f09582946742&amp;t=22637cbb-8d10-808a-b8b3-f09582946742" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-22637cbb8d1080eb8fb6fb2faace1ddc">我们稍后将开源文字转语音（TTS）模型，并发布论文介绍这两种模型的细节。</div><hr class="notion-hr notion-block-22637cbb8d1080f88cfccdfbc887bb57"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d1080b5a9e9f005decb9517" data-id="22637cbb8d1080b5a9e9f005decb9517"><span><div id="22637cbb8d1080b5a9e9f005decb9517" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d1080b5a9e9f005decb9517" title="了解更多："><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">了解更多：</span></span></h4><ul class="notion-list notion-list-disc notion-block-22637cbb8d10802b92f7d9139c8af312"><li><a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://kyutai.org/next/tts">文字转语音（Text-to-Speech）</a></li></ul><ul class="notion-list notion-list-disc notion-block-22637cbb8d10808b84bac9dbb8fe7f9b"><li><a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://unmute.sh/">Unmute 应用</a></li></ul><hr class="notion-hr notion-block-22637cbb8d10807ebea6ec599953c645"/><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-22637cbb8d1080feba63e7c57737c845" data-id="22637cbb8d1080feba63e7c57737c845"><span><div id="22637cbb8d1080feba63e7c57737c845" class="notion-header-anchor"></div><a class="notion-hash-link" href="#22637cbb8d1080feba63e7c57737c845" title="致谢"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">致谢</span></span></h4><div class="notion-text notion-block-22637cbb8d10803b9bb1cd4f12fc04af">Kyutai STT、TTS 和 Unmute 项目由以下成员创建：</div><div class="notion-text notion-block-22637cbb8d10808abc2efecf9ac07de3">Alexandre Défossez、Edouard Grave、Eugene Kharitonov、Laurent Mazare、Gabriel de Marmiesse、Emmanuel Orsini、Patrick Perez、Václav Volhejn 和 Neil Zeghidour，以及 Kyutai 团队的其他支持者。</div><hr class="notion-hr notion-block-22637cbb8d1080c78d9ffded6c31d2f1"/><div class="notion-blank notion-block-22637cbb8d1081fb93a8d2bc4ab61360"> </div><div class="notion-callout notion-gray_background_co notion-block-22637cbb8d10813ebdf7ee267cadcd27"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="💡">💡</span></div><div class="notion-callout-text"><b>对这个话题感兴趣的小伙伴，欢迎加我一起探索交流~  </b></div></div><div class="notion-row notion-block-22637cbb8d108128b98afd4c6bc05340"><div class="notion-column notion-block-22637cbb8d10815c8434ec20a3622d27" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-22637cbb8d1081a39b92f7d9c385cb46"> </div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-22637cbb8d10812e9787f8d9ee1fdfbf"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2F2666d03a-fb22-44f0-83a0-96826c4e3d2d%2F154b5cdb-df44-4c08-9984-33dbebdc5057%2Flittlerocketlu.jpg?table=block&amp;id=22637cbb-8d10-812e-9787-f8d9ee1fdfbf&amp;t=22637cbb-8d10-812e-9787-f8d9ee1fdfbf&amp;width=540.9921875&amp;cache=v2" alt="notion image" loading="lazy" decoding="async"/></div></figure></div><div class="notion-spacer"></div><div class="notion-column notion-block-22637cbb8d1081269919ed988e526a04" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-22637cbb8d108171aa25f1efb3cb67bc"> </div></div><div class="notion-spacer"></div></div></main></div>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[AIDaily 079/100 苹果与阿里巴巴合作、Perplexity发布Sonar模型、OpenAI自研芯片、苹果AI台灯、马斯克收购OpenAI被拒、Anthropic融资、巴黎AI峰会…]]></title>
            <link>http://rocketlu.cn/article/Daily79</link>
            <guid>http://rocketlu.cn/article/Daily79</guid>
            <pubDate>Wed, 12 Feb 2025 00:00:00 GMT</pubDate>
            <description><![CDATA[每天更新关于AI有趣有用的信息]]></description>
            <content:encoded><![CDATA[<div id="notion-article" class="mx-auto overflow-hidden "><main class="notion light-mode notion-page notion-block-19737cbb8d1080179a93c0238858609d"><div class="notion-viewport"></div><div class="notion-collection-page-properties"></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19737cbb8d108165893fdee920c6e9a7" data-id="19737cbb8d108165893fdee920c6e9a7"><span><div id="19737cbb8d108165893fdee920c6e9a7" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19737cbb8d108165893fdee920c6e9a7" title="AIDaily 079/100"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><span class="notion-orange">AIDaily 079/100</span></span></span></h4><div class="notion-callout notion-orange_background_co notion-block-19737cbb8d10812a966ee2054d8c19ca"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="🖼️">🖼️</span></div><div class="notion-callout-text">每天都能接收到无数条与AI、科技、艺术、经济相关的信息。
但是感觉自己就像那只掰玉米的熊，掰了一路，最后出来发现只剩下手里的两根玉米🌽。
今年希望能够以Newsletter的形式，给自己掰下来的玉米们找个背篓。

<span class="notion-orange">人们会被自己热爱的事物改变，而没有人因为给予而贫穷。</span></div></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19737cbb8d10816b8afffd79a99bf40c" data-id="19737cbb8d10816b8afffd79a99bf40c"><span><div id="19737cbb8d10816b8afffd79a99bf40c" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19737cbb8d10816b8afffd79a99bf40c" title="Vol.078"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>Vol.078</b></span></span></h4><div class="notion-text notion-block-19737cbb8d1081ecbf5ccd9d49806758"><span class="notion-gray">by Rocket</span></div><hr class="notion-hr notion-block-19737cbb8d10819baddad4c83f71491c"/><div class="notion-text notion-block-19737cbb8d1081639686f6f2be604e9f"><b>探索·AI产品</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1080449d59f300561538fc" data-id="19837cbb8d1080449d59f300561538fc"><span><div id="19837cbb8d1080449d59f300561538fc" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1080449d59f300561538fc" title="🍎 苹果与阿里巴巴合作，将Apple Intelligence引入中国市场"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">🍎 苹果与阿里巴巴合作，将Apple Intelligence引入中国市场</span></span></h4><div class="notion-text notion-block-19837cbb8d108065adb3db7e1e53c743">苹果计划将Apple Intelligence引入中国市场，此前曾与百度、字节跳动等公司探讨合作，但最终选择了阿里巴巴。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d10801c899aee315a030b68"><li><b>选择原因</b>：阿里巴巴拥有海量的用户购物和支付数据，能够为苹果提供更个性化和本地化的AI服务。此外，阿里巴巴的AI基础设施被认为更适合苹果的需求。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108088bfdfee4941512e44"><li><b>合作进展</b>：苹果与阿里巴巴已共同开发了一系列AI功能，并提交给中国网信办审批。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080cfa9e3ebedd8c646f2"><li><b>市场影响</b>：苹果在中国市场的iPhone销量因缺乏AI功能而下滑，此次合作有望提升其市场竞争力。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080d9a15de1c7777b54f8"><li><b>未来展望</b>：苹果计划在3月25日的中国开发者大会上正式推出这些AI功能，但具体发布时间仍需等待监管审批。</li></ul><div class="notion-blank notion-block-19837cbb8d1080bc80d2dd5db071db0e"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d108083b3c1dfdd141c030c" data-id="19837cbb8d108083b3c1dfdd141c030c"><span><div id="19837cbb8d108083b3c1dfdd141c030c" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d108083b3c1dfdd141c030c" title="💨 Perplexity发布新一代Sonar模型，性能大幅提升"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>💨 Perplexity发布新一代Sonar模型，性能大幅提升</b></span></span></h4><div class="notion-text notion-block-19837cbb8d10804287e0d451d4ce6b59">Perplexity推出了基于Llama 3.3 70B训练的新版Sonar搜索模型，运行在Cerebras的推理基础设施上，Token解码速度达到每秒1200个，是Gemini 2.0 Flash的8.5倍。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d10808e9325ed597a890bef"><li><b>性能优势</b>：Sonar在用户满意度、事实性和可读性方面表现卓越，超越了GPT-4o mini、Claude 3.5 Haiku等模型，甚至与Claude 3.5 Sonnet相当。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080c4af8ee51c227c588e"><li><b>用户体验</b>：Sonar能够提供更快速、准确和可读性强的回答，支持Markdown格式，优化了用户搜索体验。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10801ba7d9df16a07805d1"><li><b>市场布局</b>：新版Sonar已向Perplexity Pro用户开放，并将以API形式提供服务，未来有望进一步扩大市场份额。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080d48cd8c824901f5f21"><li><b>行业影响</b>：Sonar的推出标志着AI搜索领域的新突破，可能推动传统搜索提供商重新思考其硬件策略。</li></ul><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19737cbb8d1081c4a4b3ec5c0eebd7ac" data-id="19737cbb8d1081c4a4b3ec5c0eebd7ac"><span><div id="19737cbb8d1081c4a4b3ec5c0eebd7ac" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19737cbb8d1081c4a4b3ec5c0eebd7ac" title="🚀 OpenAI今年将完成第一代自研AI芯片设计，并计划与台积电合作制造"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">🚀 OpenAI今年将完成第一代自研AI芯片设计，并计划与台积电合作制造</span></span></h4><div class="notion-text notion-block-19837cbb8d1080b6a1acc34ed855e408">OpenAI的第一代自研AI芯片设计即将完成，目标是优化AI模型的计算效率。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d10806ab4eaf9a88f267895"><li><b>制造合作</b>：OpenAI计划与台积电合作，利用其先进的3nm工艺进行芯片制造。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080d492e9cebc6ba92825"><li><b>战略意义</b>：此举旨在减少对英伟达等现有芯片供应商的依赖，提升OpenAI在硬件领域的自主性。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080faa244fbf5d181e163"><li><b>未来展望</b>：自研芯片有望为OpenAI的AI模型提供更强大的计算支持，推动AI技术的进一步发展。</li></ul><div class="notion-blank notion-block-19737cbb8d1080ebad5dc36a69d32bc1"> </div><div class="notion-text notion-block-19737cbb8d1081b99638ea99df656ff4"><b>探索·AI硬件</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1080b09cddd1cec848798e" data-id="19837cbb8d1080b09cddd1cec848798e"><span><div id="19837cbb8d1080b09cddd1cec848798e" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1080b09cddd1cec848798e" title="💡 苹果工程师打造“皮克斯风”AI台灯，或为智能家居设备铺路"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>💡 苹果工程师打造“皮克斯风”AI台灯，或为智能家居设备铺路</b></span></span></h4><div class="notion-text notion-block-19837cbb8d1080d49500f38698762a76">苹果机器学习研究团队开发了一款AI台灯机器人，灵感来源于皮克斯的经典角色Luxo Jr.（小台灯）。该台灯不仅具备照明功能，还能通过语音、手势和摄像头与用户交互。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080cdac1ae2faadc43531"><li><b>情感表达</b>：台灯能够通过流畅、精细的动作传达情感，如委屈、点头哈腰等，增强了人机互动的自然感。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10807f898ff6a5d5478a39"><li><b>实用性与娱乐性</b>：除了情感表达，台灯还能完成实用任务，如提醒喝水、显示天气、辅助学习等。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080e69601cf4d98a19ab6"><li><b>技术框架</b>：苹果开发了名为“ELEGNT”的框架，让非人形机器人兼具表现力和功能性，提升用户的好感度。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108099a563e4635d39245f"><li><b>未来展望</b>：这款台灯可能是苹果未来智能家居设备的雏形，有传言称苹果的智能家居设备可能配备机器人手臂。</li></ul><div class="notion-blank notion-block-19737cbb8d1080568772edf7bc065ce2"> </div><div class="notion-text notion-block-19737cbb8d10805a982af5a55a296c28"><b>探索·AI投融资</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1080dd97c2dcb1ed24b384" data-id="19837cbb8d1080dd97c2dcb1ed24b384"><span><div id="19837cbb8d1080dd97c2dcb1ed24b384" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1080dd97c2dcb1ed24b384" title="📝 马斯克出价974亿美元收购OpenAI，遭奥特曼拒绝"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>📝 马斯克出价974亿美元收购OpenAI，遭奥特曼拒绝</b></span></span></h4><div class="notion-text notion-block-19837cbb8d1080d29897cd09b1c5346e">埃隆·马斯克（Elon Musk）领导的投资财团提出以974亿美元收购OpenAI的非营利组织。马斯克与OpenAI CEO山姆·奥特曼（Sam Altman）均为OpenAI的联合创始人，但马斯克于2018年离开。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080db8797f0baf4e3415d"><li><b>收购动机</b>：马斯克指责OpenAI背离了其非营利使命，隐藏源代码，并试图阻止其完全商业化。此次收购旨在将OpenAI重新定位为开源、以安全为导向的非营利研究机构。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080a8bc55ea521777206d"><li><b>奥特曼回应</b>：奥特曼在马斯克的X平台上拒绝了该提议，表示“不，谢谢，但如果你愿意，我们可以以97.4亿美元的价格收购Twitter”。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108002acaed10397841415"><li><b>董事会态度</b>：OpenAI的董事会尚未正式收到该收购提案，但奥特曼和董事会主席布雷特·泰勒（Bret Taylor）均表示公司“不待出售”，并认为马斯克的举动是对公司使命的干扰。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10806f809fc3f267d16c03"><li><b>潜在影响</b>：此次收购提议可能会对OpenAI正在进行的400亿美元融资以及5000亿美元的Stargate项目产生影响。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10803dbac8d86ed9825239"><li><b>法律与财务问题</b>：由于OpenAI的非营利性质，任何资产转移都需要确保公平市场价值，并且需要经过严格的法律程序。</li></ul><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19737cbb8d10806ebeb4c74fbbd54599" data-id="19737cbb8d10806ebeb4c74fbbd54599"><span><div id="19737cbb8d10806ebeb4c74fbbd54599" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19737cbb8d10806ebeb4c74fbbd54599" title="💰 Anthropic新一轮融资超20亿美元"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">💰 Anthropic新一轮融资超20亿美元</span></span></h4><div class="notion-text notion-block-19837cbb8d108022b64fdb4950cb9f6c">Anthropic预计在下一轮融资中筹集超过20亿美元，吸引风险投资公司General Catalyst和MGX等争夺投资份额。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d10807d8ef9c73659754e98"><li><b>Brookfield投资</b>：加拿大投资公司Brookfield计划在法国投资207亿美元建设新的AI数据中心。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080d7aa16d38157acf8bc"><li><b>法国AI计划</b>：这是法国政府吸引AI初创企业的更大计划的一部分，总投资额达1125亿美元，旨在推动AI产业发展。</li></ul><div class="notion-blank notion-block-19837cbb8d1080538461fc9436025f4a"> </div><div class="notion-text notion-block-19737cbb8d1080a293f2fbbb2b69b807"><b>探索·AI大事件</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1080948a37dcf102134991" data-id="19837cbb8d1080948a37dcf102134991"><span><div id="19837cbb8d1080948a37dcf102134991" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1080948a37dcf102134991" title="🌐 巴黎AI行动峰会落幕，美英拒绝签署多边宣言，欧洲宣布巨额投资计划"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">🌐 巴黎AI行动峰会落幕，美英拒绝签署多边宣言，欧洲宣布巨额投资计划</span></span></h4><div class="notion-text notion-block-19837cbb8d108073b5ecf9fdd1e6d67a">2月11日，巴黎AI行动峰会结束，包括中国、法国、德国、印度在内的60个国家签署了《巴黎人工智能宣言》，旨在确保AI技术是“安全、可靠和值得信任的”，并承诺以“开放”“包容”和“道德”的方式开发AI。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d108054a059c80881664a30"><li><b>美英立场</b>：美国和英国拒绝签署该宣言，理由是国家安全问题和对治理的分歧。美国副总统JD Vance警告欧洲不要过度监管AI，称这可能会扼杀行业发展。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1080849694d8ee91e60251"><li><b>欧洲计划</b>：欧盟委员会主席冯·德莱恩宣布了一项2000亿欧元的AI投资计划，旨在将欧洲定位为美国AI开发的开源替代品。加拿大投资公司Brookfield也计划在法国投资207亿美元建设AI数据中心。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108059a0f8ede4e37f5bc9"><li><b>行业反应</b>：Anthropic CEO Dario Amodei称峰会是“错失的机会”，强调需要加快AI治理进程，以应对技术快速发展的挑战。</li></ul><div class="notion-blank notion-block-19837cbb8d10801d8aa3ec034a5dbfbc"> </div><div class="notion-text notion-block-19737cbb8d10816ab8a4f1727a74df9c"><b>【🚀 精选内容】</b></div><div class="notion-row notion-block-19737cbb8d1081c28821e7df5757506e"><div class="notion-column notion-block-19737cbb8d1081fc8c69ed298e537a4d" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"><a class="notion-page-link 570c7c80-1df6-4ac1-afd8-2e20674ec6f7" href="/570c7c801df64ac1afd82e20674ec6f7"><span class="notion-page-title"><div class="notion-page-icon-inline notion-page-icon-image"><svg class="notion-page-title-icon notion-page-icon" alt="我用OpenGlass做了一个AI眼镜【上篇】" viewBox="0 0 30 30" width="16"><path d="M16,1H4v28h22V11L16,1z M16,3.828L23.172,11H16V3.828z M24,27H6V3h8v10h10V27z M8,17h14v-2H8V17z M8,21h14v-2H8V21z M8,25h14v-2H8V25z"></path></svg></div><span class="notion-page-title-text"><b>我用OpenGlass做了一个AI眼镜【上篇】</b></span></span></a></div><div class="notion-spacer"></div><div class="notion-column notion-block-19737cbb8d1081e8ab51cdd8031a4ea5" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"><a class="notion-page-link cfc49a0e-341c-4376-b680-0dc1f8f197a3" href="/cfc49a0e341c4376b6800dc1f8f197a3"><span class="notion-page-title"><div class="notion-page-icon-inline notion-page-icon-image"><svg class="notion-page-title-icon notion-page-icon" alt="Flux 完整的 Lora 设置和数据集指南 - 两周学习的事后分析[译文]" viewBox="0 0 30 30" width="16"><path d="M16,1H4v28h22V11L16,1z M16,3.828L23.172,11H16V3.828z M24,27H6V3h8v10h10V27z M8,17h14v-2H8V17z M8,21h14v-2H8V21z M8,25h14v-2H8V25z"></path></svg></div><span class="notion-page-title-text"><b>Flux 完整的 Lora 设置和数据集指南 - 两周学习的事后分析[译文]</b></span></span></a></div><div class="notion-spacer"></div><div class="notion-column notion-block-19737cbb8d1081d3b9ede7ab579f079f" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"><a class="notion-page-link 470debdd-454d-4443-8d5a-2d2ab4672ed2" href="/470debdd454d44438d5a2d2ab4672ed2"><span class="notion-page-title"><div class="notion-page-icon-inline notion-page-icon-image"><svg class="notion-page-title-icon notion-page-icon" alt="Midjourney上线人物一致功能 我拥有了一个IP宇宙" viewBox="0 0 30 30" width="16"><path d="M16,1H4v28h22V11L16,1z M16,3.828L23.172,11H16V3.828z M24,27H6V3h8v10h10V27z M8,17h14v-2H8V17z M8,21h14v-2H8V21z M8,25h14v-2H8V25z"></path></svg></div><span class="notion-page-title-text">Midjourney上线人物一致功能 我拥有了一个IP宇宙</span></span></a></div><div class="notion-spacer"></div></div><div class="notion-text notion-block-19737cbb8d1081d19e0bf2ca539b4702"><b>❤ 如果对你有帮助，欢迎分享或者Buy Me A Coffee ❤</b></div><figure class="notion-asset-wrapper notion-asset-wrapper-embed notion-block-19737cbb8d10813c8ce8d963748a2aba"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:690px"><iframe class="notion-asset-object-fit" src="https://buymeacoffee.com/rocketlu?spaceId=2666d03a-fb22-44f0-83a0-96826c4e3d2d" title="iframe embed" frameBorder="0" allowfullscreen="" loading="lazy" scrolling="auto"></iframe></div></figure><div class="notion-blank notion-block-19737cbb8d10819a8dfbc4bb9a2ffc69"> </div><div class="notion-callout notion-gray_background_co notion-block-19737cbb8d10813ebadaee0cac78b248"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="💡">💡</span></div><div class="notion-callout-text"><b>对这个话题感兴趣的小伙伴，欢迎加我一起探索交流~</b></div></div><div class="notion-row notion-block-19737cbb8d108180b2c7d54cfd318aaa"><div class="notion-column notion-block-19737cbb8d1081e79f7cdd0c90736e80" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19737cbb8d1081caa7f8f8f5da8a3006"> </div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19737cbb8d108164afe4e189457c3066"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2F2666d03a-fb22-44f0-83a0-96826c4e3d2d%2F154b5cdb-df44-4c08-9984-33dbebdc5057%2Flittlerocketlu.jpg?table=block&amp;id=19737cbb-8d10-8164-afe4-e189457c3066&amp;t=19737cbb-8d10-8164-afe4-e189457c3066&amp;width=540.9921875&amp;cache=v2" alt="notion image" loading="lazy" decoding="async"/></div></figure></div><div class="notion-spacer"></div><div class="notion-column notion-block-19737cbb8d1081a4b230f3285c8b1949" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19737cbb8d1081e29107c642fd085e61"> </div></div><div class="notion-spacer"></div></div><div class="notion-blank notion-block-19737cbb8d108152910be49999aef07b"> </div><div class="notion-blank notion-block-19737cbb8d1081a78a4cff519eb2be88"> </div><div class="notion-blank notion-block-19737cbb8d10816b8238de32d38781a3"> </div><div class="notion-blank notion-block-19737cbb8d10813ba34fc7109bb6e02a"> </div><div class="notion-blank notion-block-19737cbb8d10814799fcef72424ede66"> </div></main></div>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[AIDaily 078/100 OpenAI超级碗首秀，字节发布Goku广告模型、艺术家抵制佳士得AI拍卖、Anthropic经济指数发布…]]></title>
            <link>http://rocketlu.cn/article/Daily78</link>
            <guid>http://rocketlu.cn/article/Daily78</guid>
            <pubDate>Tue, 11 Feb 2025 00:00:00 GMT</pubDate>
            <description><![CDATA[每天更新关于AI有趣有用的信息]]></description>
            <content:encoded><![CDATA[<div id="notion-article" class="mx-auto overflow-hidden "><main class="notion light-mode notion-page notion-block-19837cbb8d1080309fd7f11e044e3875"><div class="notion-viewport"></div><div class="notion-collection-page-properties"></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d108100b066fe7aaade78d0" data-id="19837cbb8d108100b066fe7aaade78d0"><span><div id="19837cbb8d108100b066fe7aaade78d0" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d108100b066fe7aaade78d0" title="AIDaily 078/100"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><span class="notion-orange">AIDaily 078/100</span></span></span></h4><div class="notion-callout notion-orange_background_co notion-block-19837cbb8d1081429c71d142798d76e4"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="🖼️">🖼️</span></div><div class="notion-callout-text">每天都能接收到无数条与AI、科技、艺术、经济相关的信息。
但是感觉自己就像那只掰玉米的熊，掰了一路，最后出来发现只剩下手里的两根玉米🌽。
今年希望能够以Newsletter的形式，给自己掰下来的玉米们找个背篓。

<span class="notion-orange">人们会被自己热爱的事物改变，而没有人因为给予而贫穷。</span></div></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081b0b82fc89fec55b259" data-id="19837cbb8d1081b0b82fc89fec55b259"><span><div id="19837cbb8d1081b0b82fc89fec55b259" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081b0b82fc89fec55b259" title="Vol.078"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>Vol.078</b></span></span></h4><div class="notion-text notion-block-19837cbb8d10814f8289dffb5fbe9c2c"><span class="notion-gray">by Rocket</span></div><hr class="notion-hr notion-block-19837cbb8d1081008ecbf9ce0d6e28ef"/><div class="notion-text notion-block-19837cbb8d1081a8adbec8d4f682fad2"><b>探索·AI产品</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081068d63cb462c03c202" data-id="19837cbb8d1081068d63cb462c03c202"><span><div id="19837cbb8d1081068d63cb462c03c202" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081068d63cb462c03c202" title="飞书多维表格接入DeepSeek R1模型，功能强大且实用"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">飞书多维表格接入DeepSeek R1模型，功能强大且实用</span></span></h4><div class="notion-text notion-block-19837cbb8d1081378e35e2d03115fa04">飞书多维表格已正式接入DeepSeek R1模型，这一集成极大地提升了其在数据处理与分析上的智能化水平。用户现在可以在多维表格的AI字段捷径中直接调用DeepSeek R1模型，实现以下功能：</div><ol start="1" class="notion-list notion-list-numbered notion-block-19837cbb8d10815e869be97bbb851c23" style="list-style-type:decimal"><li><b>论文解读与扩展</b>：用户可以上传PDF文件，DeepSeek R1能够自动分析论文的关键信息，包括优点和不足，并生成社交媒体的发布文案。</li></ol><ol start="2" class="notion-list notion-list-numbered notion-block-19837cbb8d1081a58484fc95c22eed1f" style="list-style-type:decimal"><li><b>虚拟数字人视频生成</b>：支持直接生成虚拟数字人视频，适用于广告视频生成。</li></ol><ol start="3" class="notion-list notion-list-numbered notion-block-19837cbb8d10814aac59d4621833763b" style="list-style-type:decimal"><li><b>产品展示视频生成</b>：可以从产品图片生成人物互动视频，保持产品样式。</li></ol><ol start="4" class="notion-list notion-list-numbered notion-block-19837cbb8d108155990ce52813b98a8d" style="list-style-type:decimal"><li><b>批量处理任务</b>：用户可以在多维表格中一次性输入多个提示词，实现批量处理任务，无需逐个调用API。</li></ol><div class="notion-blank notion-block-19837cbb8d1081c899c2ebdfb175cae0"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d108159ac88d9c2ebcc63f4" data-id="19837cbb8d108159ac88d9c2ebcc63f4"><span><div id="19837cbb8d108159ac88d9c2ebcc63f4" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d108159ac88d9c2ebcc63f4" title="Google DeepMind推出AlphaGeometry2：在IMO几何问题上超越金牌得主"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Google DeepMind推出AlphaGeometry2：在IMO几何问题上超越金牌得主</span></span></h4><div class="notion-text notion-block-19837cbb8d1081eba978de2648dde50c">Google DeepMind的AlphaGeometry2在解决国际数学奥林匹克竞赛（IMO）几何问题方面取得了重大突破，其表现超过了平均金牌得主。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d10818eab8cc9c5251b3416"><li><b>性能提升</b>：AlphaGeometry2解决了过去25年IMO中84%的几何问题，显著高于其前身AlphaGeometry1的54%解题率。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10812cbe06d8f7435ac99d"><li><b>技术细节</b>：该系统结合了谷歌Gemini系列的语言模型和符号引擎，通过神经符号混合方法解决复杂的几何问题。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10815fb76acd50c6e5363c"><li><b>训练数据</b>：为了解决几何问题，DeepMind为AlphaGeometry2生成了超过3亿个不同复杂度的合成定理和证明。</li></ul><div class="notion-blank notion-block-19837cbb8d1081068cd8e9910d6e689c"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d10817faef9f7c3d649201d" data-id="19837cbb8d10817faef9f7c3d649201d"><span><div id="19837cbb8d10817faef9f7c3d649201d" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d10817faef9f7c3d649201d" title="Perplexity AI推出“百万美元问题”抽奖活动"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Perplexity AI推出“百万美元问题”抽奖活动</span></span></h4><div class="notion-text notion-block-19837cbb8d1081c08017ead3e6948ce1">Perplexity AI宣布将在2025年2月9日超级碗期间举办“百万美元问题”抽奖活动。参与者需要在比赛期间通过Perplexity应用程序向AI提问，以获得抽奖资格。具体参与条件包括：</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081a78016df453970a0d2"><li>必须是18岁及以上的美国居民。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108183968bddeb891f2646"><li>需要下载Perplexity移动应用，并在比赛期间（太平洋时间下午3点至晚上7点半）至少向AI提问5个问题。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081dbaf56fb0f2673bce2"><li>每人最多可获得5次抽奖机会。活动结束后，将在比赛结束后约1小时内随机抽取获奖者。</li></ul><div class="notion-blank notion-block-19837cbb8d1081a99a11fd749648f12f"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081eea425e0c3fee4bf47" data-id="19837cbb8d1081eea425e0c3fee4bf47"><span><div id="19837cbb8d1081eea425e0c3fee4bf47" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081eea425e0c3fee4bf47" title="🏈 OpenAI超级碗首秀：用动画点讲述人类发明史"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">🏈 OpenAI超级碗首秀：用动画点讲述人类发明史</span></span></h4><div class="notion-text notion-block-19837cbb8d1081d1b84fceebe1ee1d26">OpenAI在2025年超级碗上发布了首条电视广告，这是其首次大规模的付费广告活动，该广告时长为60秒，投放成本约为1400万美元。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d10811c83c4d6cbd3ca316a"><li><b>创意呈现</b>：广告通过独特的点画风格动画，展示了人类技术的进化，从早期的火和轮子到现代的DNA测序和太空探索，最终聚焦于ChatGPT处理日常任务的能力，创意源自ChatGPT的光标点。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10815aa0cbee59ae76e355"><li><b>聚焦应用</b>：广告未过度宣传AI能力，而是聚焦于其实际应用，避免引发争议。</li></ul><div class="notion-blank notion-block-19837cbb8d10815eae2fe0b7af45bfdb"> </div><div class="notion-blank notion-block-19837cbb8d108110819bf1164a9633aa"> </div><div class="notion-text notion-block-19837cbb8d108180938dc7a2a1eb0ca6"><b>探索·AI开源</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d10813295dcc463a2bdb477" data-id="19837cbb8d10813295dcc463a2bdb477"><span><div id="19837cbb8d10813295dcc463a2bdb477" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d10813295dcc463a2bdb477" title="字节跳动发布Goku和Goku+视频生成模型"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">字节跳动发布Goku和Goku+视频生成模型</span></span></h4><div class="notion-text notion-block-19837cbb8d1081ff9c06f68bbe2a58ac">字节跳动与香港大学联合发布了名为Goku的视频生成模型，以及其广告视频生成的扩展版本Goku+。这些模型能够根据文本提示生成高质量的视频内容，支持多种生成任务，包括文本到视频、图像到视频、文本到图像等。</div><figure class="notion-asset-wrapper notion-asset-wrapper-video notion-block-19837cbb8d108121aaa1c2ab6bc87730"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:400px;max-width:100%;flex-direction:column;min-height:100px"><video playsinline="" controls="" preload="metadata" style="object-fit:contain" src="https://saiyan-world.github.io/goku/assets/videos/moviegen_bench/moviegen_bench_0001.mp4?spaceId=2666d03a-fb22-44f0-83a0-96826c4e3d2d" title="video"></video></div></figure><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081479bb2def2cee1feb3" data-id="19837cbb8d1081479bb2def2cee1feb3"><span><div id="19837cbb8d1081479bb2def2cee1feb3" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081479bb2def2cee1feb3" title="Goku模型特点"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Goku模型特点</span></span></h4><ul class="notion-list notion-list-disc notion-block-19837cbb8d108109ab2edc394e4fcd73"><li><b>高质量视频生成</b>：Goku能够生成包括动画、自然风光、动物行为等多种场景的视频，效果生动真实。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10818b9de7e542b6127a98"><li><b>虚拟数字人视频</b>：支持直接生成虚拟数字人视频，Goku+将文本转换为超现实的人类视频，具有稳定的手部动作和丰富的面部及身体动作。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10810fb743e587ee0c9a06"><li><b>产品展示视频</b>：可以从产品图片生成人物互动视频，保持产品样式，适用于广告视频生成。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108137a0c7e9c1c3552213"><li><b>广告视频生成</b>：Goku+专注于广告场景，能够根据文本描述生成高质量的广告视频，支持人物与产品的自然互动，显著降低制作成本。</li></ul><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081a4930ac5de39d179f0" data-id="19837cbb8d1081a4930ac5de39d179f0"><span><div id="19837cbb8d1081a4930ac5de39d179f0" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081a4930ac5de39d179f0" title="技术细节"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">技术细节</span></span></h4><ul class="notion-list notion-list-disc notion-block-19837cbb8d108175a181d4a5eb1a4520"><li><b>数据集</b>：Goku团队构建了包含3600万视频和1.6亿图像的数据集，通过多种技术严格筛选数据质量。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081a2881afc630d0c205f"><li><b>架构</b>：基于Rectified Flow Transformer框架，支持多模态训练，能够处理复杂的时空依赖关系。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081ae8cafe91f2463de31"><li><b>性能</b>：在多个视频生成基准测试中表现优异，特别是在VBench测试中，Goku-T2V模型以84.85分获得第二名。</li></ul><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d10810ea729f2a44fb5572f" data-id="19837cbb8d10810ea729f2a44fb5572f"><span><div id="19837cbb8d10810ea729f2a44fb5572f" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d10810ea729f2a44fb5572f" title="应用场景"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">应用场景</span></span></h4><ul class="notion-list notion-list-disc notion-block-19837cbb8d108105977fe56108755441"><li><b>广告视频制作</b>：Goku+能够根据文本描述生成高质量的广告视频，支持从文本直接生成视频、从产品图片生成人物互动视频，以及生成产品展示视频。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081339202fe9391e655cd"><li><b>虚拟数字人视频生成</b>：Goku+可以将文本转换为超现实的人类视频，生成超过20秒的视频，具有稳定的手部动作和丰富的面部及身体动作。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10813e810aea12b6d560ae"><li><b>内容创作</b>：Goku能生成多种场景的视频，为艺术创作者提供丰富的灵感和创作素材。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10818da862f6927cb5a9a1"><li><b>教育与培训</b>：可用于制作教育视频和培训课程，提高教育培训的效果和趣味性。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10819fb7a2d44b6a990c55"><li><b>娱乐产业</b>：在电影、电视剧、动画等娱乐产业中，Goku可用于内容制作和特效生成。</li></ul><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081e19b60ce6340e3be1f" data-id="19837cbb8d1081e19b60ce6340e3be1f"><span><div id="19837cbb8d1081e19b60ce6340e3be1f" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081e19b60ce6340e3be1f" title="项目资源"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">项目资源</span></span></h4><ul class="notion-list notion-list-disc notion-block-19837cbb8d108190aa96d0895ed71751"><li><b>项目官网</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://saiyan-world.github.io/goku/">https://saiyan-world.github.io/goku/</a></li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081c5a42dca7f8baf2b44"><li><b>Github仓库</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://github.com/Saiyan-World/goku">https://github.com/Saiyan-World/goku</a></li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108113bf4cf2291a9c3c61"><li><b>HuggingFace模型库</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://huggingface.co/datasets/saiyan-world/Goku">https://huggingface.co/datasets/saiyan-world/Goku</a></li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081339abad701e98e2b6e"><li><b>arXiv技术论文</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://arxiv.org/pdf/2502.04896">https://arxiv.org/pdf/2502.04896</a></li></ul><div class="notion-blank notion-block-19837cbb8d1081dfa770c53d046bfb83"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d108111a581e2d71b6f407e" data-id="19837cbb8d108111a581e2d71b6f407e"><span><div id="19837cbb8d108111a581e2d71b6f407e" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d108111a581e2d71b6f407e" title="cursor-tools增强Cursor Agent的能力"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">cursor-tools增强Cursor Agent的能力</span></span></h4><div class="notion-text notion-block-19837cbb8d108140afc7c2221d7f5e57"><b>cursor-tools</b>是一个强大的扩展工具集，旨在显著增强Cursor Agent的能力。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19837cbb8d1081999808c806eaac6646"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Af224d8f1-70ee-4344-a843-4ffdf76ddaa2%3Aimage.png?table=block&amp;id=19837cbb-8d10-8199-9808-c806eaac6646&amp;t=19837cbb-8d10-8199-9808-c806eaac6646" alt="notion image" loading="lazy" decoding="async"/></div></figure><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081a7b31ed85b014ddf16"><li><b>网页搜索功能</b>：通过Perplexity AI，cursor-tools为Cursor提供了网页搜索功能，帮助开发者快速获取最新信息。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10815d91efcce9545fb8db"><li><b>大规模代码库分析</b>：支持使用Gemini 2.0进行大规模代码库分析，提升代码理解和生成的效率。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108116ba7cc08913c10eed"><li><b>浏览器自动化</b>：cursor-tools支持浏览器自动化操作，包括打开网页、执行操作、观察交互元素和提取数据等。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108189be67f003b356c8c2"><li><b>GitHub Issues和Pull Requests支持</b>：cursor-tools允许AI编码助手直接从命令行访问和处理GitHub Issues和Pull Requests。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10811aae8df54cce6b429e"><li><b>Github仓库</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://github.com/eastlondoner/cursor-tools">https://github.com/eastlondoner/cursor-tools</a></li></ul><div class="notion-blank notion-block-19837cbb8d1081b69457c94e02961bf9"> </div><div class="notion-text notion-block-19837cbb8d1081d99492cc5fb6a49ff0"><b>探索·AI版权</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081299405e2f4507d8c48" data-id="19837cbb8d1081299405e2f4507d8c48"><span><div id="19837cbb8d1081299405e2f4507d8c48" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081299405e2f4507d8c48" title="🎨 超过2000名艺术家联名抵制佳士得AI艺术拍卖"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>🎨 超过2000名艺术家联名抵制佳士得AI艺术拍卖</b></span></span></h4><div class="notion-text notion-block-19837cbb8d108171909df63f524ae3cf">佳士得即将举办的“增强情报”AI艺术拍卖会引发了广泛争议：</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081f98426d2b4492bc89f"><li><b>版权争议</b>：艺术家指责AI模型在训练中使用了受版权保护的作品，未经艺术家授权。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081d49929edca02488cf9"><li><b>公开信</b>：超过2800名艺术家签署公开信，要求佳士得取消此次拍卖。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081a986c5fa555d0a6c8a"><li><b>拍卖背景</b>：此次拍卖会计划于2月20日开始，展出现场作画机器人等作品。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108155832cc118c51f577f"><li><b>佳士得回应</b>：拍卖行表示，参与拍卖的艺术家作品使用AI是为了增强创作，且数据训练基于艺术家自身的输入。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081d786dafdb7bd31b38c"><li><b>艺术家态度</b>：部分艺术家对AI在艺术创作中的应用持开放态度，但许多传统艺术家担心其权益受损。</li></ul><div class="notion-text notion-block-19837cbb8d10819083ecd40728f7dab7">此次事件凸显了AI技术在艺术领域应用中的版权和伦理问题，引发了艺术界对AI创作与传统艺术创作之间界限的深刻思考。</div><div class="notion-blank notion-block-19837cbb8d1081adbf8cff35a0b1858e"> </div><div class="notion-text notion-block-19837cbb8d108178a55ac601d46a6809"><b>探索·新研究</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d10816599f9e41e6140cda6" data-id="19837cbb8d10816599f9e41e6140cda6"><span><div id="19837cbb8d10816599f9e41e6140cda6" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d10816599f9e41e6140cda6" title="🌐 OpenAI高管在东京大学探讨AI未来"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">🌐 OpenAI高管在东京大学探讨AI未来</span></span></h4><div class="notion-text notion-block-19837cbb8d108121abc0f05a2d653030">OpenAI首席执行官Sam Altman和首席产品官Kevin Weil在东京大学全球教育中心发表演讲，回答了关于AI未来、GPT-5等新模型以及Stargate项目等问题。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d108115b4ddca233d2e86e2"><li><b>Stargate项目</b>：Altman表示，价值5000亿美元的Stargate项目将使下一代模型能够开发出首批“科学知识的新位”。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081e7b712f65802bd5a66"><li><b>模型能力</b>：OpenAI的内部模型目前是全球第50大程序员，预计到今年年底可能达到第1名。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081039444ce387ea014ae"><li><b>开源方向</b>：Altman认为OpenAI将朝着开源方向发展，因为社会愿意接受开放模型带来的权衡。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081d0bcb2da847bd38432"><li><b>AI能力提升</b>：Altman将AI能力的提升比作“试图超越计算器”，认为AI将在每个通用领域超越人类。</li></ul><div class="notion-blank notion-block-19837cbb8d10819dbcdbdd31bb8b555a"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d10817894c4f8d3cee33e32" data-id="19837cbb8d10817894c4f8d3cee33e32"><span><div id="19837cbb8d10817894c4f8d3cee33e32" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d10817894c4f8d3cee33e32" title="📈 Anthropic发布“Anthropic经济指数”：AI使用模式深度分析"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">📈 Anthropic发布“Anthropic经济指数”：AI使用模式深度分析</span></span></h4><div class="notion-text notion-block-19837cbb8d1081d2b5befa8e181b30de">Anthropic通过分析Claude上的数百万条匿名聊天记录，发布了“Anthropic经济指数”，揭示了AI在不同职业中的使用模式：</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081419f04f4a9929289ef"><li><b>职业分布</b>：在22个职业类别中，“计算机与数学”占比最高（37.2%），而“办公室与行政支持”在劳动力市场中占比最高（12.2%）。渔业、林业在两个维度中的占比都最低（0.3%与0.1%）。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10812e83f0f0c715ab86e3"><li><b>薪资与AI使用</b>：AI使用集中在中等至中高收入群体，低收入和高收入职业的AI使用率较低。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d108167a7d0d93a733b14e2"><li><b>增强与自动化</b>：AI使用更偏向“增强”（57%），即与人类协同完成任务，而非“自动化”（43%），即AI直接执行任务。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10816f849ac20b6b509469"><li><b>使用深度</b>：仅4%的职业在超过75%的任务中使用AI，但36%的职业在至少25%的任务中使用AI，说明中等程度的使用更为普遍。</li></ul><div class="notion-text notion-block-19837cbb8d10812c8d4df7dfdf04cd19">Anthropic还开源了数据集，供研究人员进一步分析。</div><div class="notion-blank notion-block-19837cbb8d10817cb24cc8527a41627d"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d1081e1bef5cf7a4a726928" data-id="19837cbb8d1081e1bef5cf7a4a726928"><span><div id="19837cbb8d1081e1bef5cf7a4a726928" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d1081e1bef5cf7a4a726928" title="AI技术的快速进步：五年内基准测试的显著提升"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">AI技术的快速进步：五年内基准测试的显著提升</span></span></h4><div class="notion-text notion-block-19837cbb8d108154b183d4b3a927aaca">Jason Wei发布的图表展示了过去五年AI技术在各种基准测试中的快速进步，直观地反映了AI能力的显著提升。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19837cbb8d1081ed84c6c966fe8a96a1"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A155f7db6-a8c1-4fa6-84ba-3faea323119d%3Aimage.png?table=block&amp;id=19837cbb-8d10-81ed-84c6-c966fe8a96a1&amp;t=19837cbb-8d10-81ed-84c6-c966fe8a96a1" alt="notion image" loading="lazy" decoding="async"/></div></figure><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081e9a933fd4572d86029"><li><b>基准测试（Benchmark）</b>：</li><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081e9a933fd4572d86029"><li><b>定义</b>：基准测试是评估AI能力的一种方法，通过让AI回答问题或完成任务并打分，类似于学校的考试。</li><li><b>类型</b>：</li><ul class="notion-list notion-list-disc notion-block-19837cbb8d10812b95cce8b2f276171c"><li><b>常识问答</b>：如TriviaQA，测试AI对琐事或常识性问题的回答能力。</li><li><b>多科目知识</b>：如MMLU，测试AI在不同科目上的知识水平。</li><li><b>基础数学</b>：如GSM8K，测试AI解决小学或基础数学问题的能力。</li><li><b>高级理工科考试</b>：如MATH、AIME、GPQA等，测试AI在更专业、难度更高的数学和理工科问题上的表现。</li><li><b>“人类的最终考试”</b>：假设性的“终极测试”，代表AI面临的极具挑战性的重要考试。</li></ul></ul></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081eb99b7ed4e7efe82ef"><li><b>测试结果</b>：图表中的每条彩色线代表一种特定的AI测试，展示了AI在这些测试中的表现随时间的快速提升。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d10811c997bfd20f0436a70"><li><b>技术进步</b>：AI技术在短短几年内不断更新升级，准确率不断提高，甚至超过人类平均水平或接近满分。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081b1a3b1fa2ceb667110"><li><b>行业影响</b>：</li><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081b1a3b1fa2ceb667110"><li><b>学习能力</b>：AI的学习和推理能力正在以惊人的速度提升。</li><li><b>复杂任务处理</b>：AI在处理复杂任务方面的能力也在迅速增强。</li><li><b>未来展望</b>：AI的快速进步预示着未来可能在更多领域超越人类，带来深远的变革。</li></ul></ul><div class="notion-text notion-block-19837cbb8d1081f6bf82deceeef12090">这张图表不仅展示了AI技术的快速进步，也提醒我们AI的发展速度可能远超我们的预期。未来，AI可能会在更多领域实现突破，带来更多的可能性和挑战。</div><div class="notion-blank notion-block-19837cbb8d10811b8f4dd5df3735ce9f"> </div><div class="notion-text notion-block-19837cbb8d10811a91caf103cd2c8486"><b>探索·AI人员变动</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19837cbb8d108176885acf32bd301b60" data-id="19837cbb8d108176885acf32bd301b60"><span><div id="19837cbb8d108176885acf32bd301b60" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19837cbb8d108176885acf32bd301b60" title="Ilya Sutskever的Safe Superintelligence寻求200亿美元估值融资"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Ilya Sutskever的Safe Superintelligence寻求200亿美元估值融资</span></span></h4><div class="notion-text notion-block-19837cbb8d1081e3ae84d1bc17b22c0c">Safe Superintelligence（SSI）由OpenAI前首席科学家Ilya Sutskever于2024年6月创立，目标是开发具有“超级智能”的AI模型，并防止有害输出。</div><ul class="notion-list notion-list-disc notion-block-19837cbb8d108134af32de76d19a6d7e"><li><b>融资情况</b>：SSI正在与投资者洽谈，计划以至少200亿美元的估值进行融资。这将是该公司继2024年9月以50亿美元估值完成10亿美元融资后的又一轮重大融资。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081348d19ea441f458ce7"><li><b>技术方向</b>：SSI专注于开发安全的超级智能AI模型，其技术路线与现有AI开发方法不同，旨在探索新的研究方向。</li></ul><ul class="notion-list notion-list-disc notion-block-19837cbb8d1081dda170d56fef11674d"><li><b>团队构成</b>：SSI的创始团队还包括前OpenAI研究员Daniel Levy和前Y Combinator合伙人Daniel Gross，后者曾领导苹果公司的AI开发工作。</li></ul><div class="notion-blank notion-block-19837cbb8d108101a4ddc88fe41d433d"> </div><div class="notion-blank notion-block-19837cbb8d1081f998d3f873e2edbee1"> </div><div class="notion-text notion-block-19837cbb8d1081188411e3c332791a6d"><b>【🚀 精选内容】</b></div><div class="notion-row notion-block-19837cbb8d1081f0ba54f8e395b68a01"><div class="notion-column notion-block-19837cbb8d108199ac23ccb52861b2ea" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"></div><div class="notion-spacer"></div><div class="notion-column notion-block-19837cbb8d10817183a1f676131caba3" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"></div><div class="notion-spacer"></div><div class="notion-column notion-block-19837cbb8d10818cae12e4ec34f0416c" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"></div><div class="notion-spacer"></div></div><div class="notion-text notion-block-19837cbb8d1081fd8eaceb88f427ebd5"><b>❤ 如果对你有帮助，欢迎分享或者Buy Me A Coffee ❤</b></div><figure class="notion-asset-wrapper notion-asset-wrapper-embed notion-block-19837cbb8d10819287d7decff6a5959d"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:690px"><iframe class="notion-asset-object-fit" src="https://buymeacoffee.com/rocketlu?spaceId=2666d03a-fb22-44f0-83a0-96826c4e3d2d" title="iframe embed" frameBorder="0" allowfullscreen="" loading="lazy" scrolling="auto"></iframe></div></figure><div class="notion-blank notion-block-19837cbb8d108199be60fd6d8e3e6706"> </div><div class="notion-callout notion-gray_background_co notion-block-19837cbb8d1081fe86ead4c1820e911a"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="💡">💡</span></div><div class="notion-callout-text"><b>对这个话题感兴趣的小伙伴，欢迎加我一起探索交流~</b></div></div><div class="notion-row notion-block-19837cbb8d1081839293fb61ddccb7a5"><div class="notion-column notion-block-19837cbb8d1081798c00d6ca1c2ccb5b" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19837cbb8d10812e8b68fc768d16ed71"> </div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19837cbb8d10818d9056c8e67fa155fc"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2F2666d03a-fb22-44f0-83a0-96826c4e3d2d%2F154b5cdb-df44-4c08-9984-33dbebdc5057%2Flittlerocketlu.jpg?table=block&amp;id=19837cbb-8d10-818d-9056-c8e67fa155fc&amp;t=19837cbb-8d10-818d-9056-c8e67fa155fc&amp;width=540.9921875&amp;cache=v2" alt="notion image" loading="lazy" decoding="async"/></div></figure></div><div class="notion-spacer"></div><div class="notion-column notion-block-19837cbb8d108106b443e535a0ef2488" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19837cbb8d1081728a12fb83cc852d60"> </div></div><div class="notion-spacer"></div></div><div class="notion-blank notion-block-19837cbb8d1081ddb3ccd1906d79f53b"> </div><div class="notion-blank notion-block-19837cbb8d1081518ea6e990df758db8"> </div><div class="notion-blank notion-block-19837cbb8d1081f0b648fe36db127c4b"> </div><div class="notion-blank notion-block-19837cbb8d10815286dfde4207ff7bf8"> </div><div class="notion-blank notion-block-19837cbb8d10814391f7fb0642cd8e66"> </div></main></div>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[AIDaily 077/100 ElevenLabs Studio开放使用、OpenAI Canvas开放共享、Pika Labs Pikadditions发布、OpenAI超级碗广告、Mistral Le Chat移动应用发布…]]></title>
            <link>http://rocketlu.cn/article/Daily77</link>
            <guid>http://rocketlu.cn/article/Daily77</guid>
            <pubDate>Mon, 10 Feb 2025 00:00:00 GMT</pubDate>
            <description><![CDATA[每天更新关于AI有趣有用的信息]]></description>
            <content:encoded><![CDATA[<div id="notion-article" class="mx-auto overflow-hidden "><main class="notion light-mode notion-page notion-block-18437cbb8d1080559ce0c83cb0076a5e"><div class="notion-viewport"></div><div class="notion-collection-page-properties"></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-18437cbb8d10819182b3fc5229d4af36" data-id="18437cbb8d10819182b3fc5229d4af36"><span><div id="18437cbb8d10819182b3fc5229d4af36" class="notion-header-anchor"></div><a class="notion-hash-link" href="#18437cbb8d10819182b3fc5229d4af36" title="AIDaily 077/100"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><span class="notion-orange">AIDaily 077/100</span></span></span></h4><div class="notion-callout notion-orange_background_co notion-block-18437cbb8d1081bdb0d6f2b71aa18302"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="🖼️">🖼️</span></div><div class="notion-callout-text">每天都能接收到无数条与AI、科技、艺术、经济相关的信息。
但是感觉自己就像那只掰玉米的熊，掰了一路，最后出来发现只剩下手里的两根玉米🌽。
今年希望能够以Newsletter的形式，给自己掰下来的玉米们找个背篓。

<span class="notion-orange">人们会被自己热爱的事物改变，而没有人因为给予而贫穷。</span></div></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-18437cbb8d1081b090b4e84d1016f144" data-id="18437cbb8d1081b090b4e84d1016f144"><span><div id="18437cbb8d1081b090b4e84d1016f144" class="notion-header-anchor"></div><a class="notion-hash-link" href="#18437cbb8d1081b090b4e84d1016f144" title="Vol.077"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>Vol.077</b></span></span></h4><div class="notion-text notion-block-18437cbb8d108187a730d456d21c96da"><span class="notion-gray">by Rocket</span></div><hr class="notion-hr notion-block-18437cbb8d10815e9a38c57131f148d3"/><div class="notion-text notion-block-18437cbb8d1081db9b7fcbdfc5bbab9d"><b>探索·AI产品</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d108071b9bded8b92b989d1" data-id="19637cbb8d108071b9bded8b92b989d1"><span><div id="19637cbb8d108071b9bded8b92b989d1" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d108071b9bded8b92b989d1" title="ElevenLabs Studio工具开放使用 可制作长播客"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">ElevenLabs Studio工具开放使用 可制作长播客</span></span></h4><div class="notion-text notion-block-19637cbb8d108006a170c56de9fdb206">ElevenLabs宣布其Studio工具现已向所有用户开放。Studio是一个长格式文本到音频的编辑器，专为创作者和讲故事的人设计，能够将有声读物、旁白、文章甚至播客等带入生活。用户可以利用其多角色、精细的声音定制、节奏控制等功能，将文字内容转化为生动的音频作品。</div><div class="notion-text notion-block-19637cbb8d1080dda1abc15be6ee0811">此次更新包括以下亮点：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080dbbb81ead39bbd35a4"><li><b>节奏控制</b>：用户现在可以通过菜单栏添加持续时间为0.1秒至3秒的停顿。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080b2bea1fdbb439fb0f9"><li><b>自动角色声音分配</b>：当导入包含多个角色的有声读物或剧本时，系统会自动为每个角色分配不同的声音。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108044b147c2cd716cbe66"><li><b>GenFM功能</b>：用户可以轻松创建播客风格的讨论，通过上传文档或从URL导入内容，选择主持人与嘉宾之间的对话形式或仅由主持人播报的形式。</li></ul><div class="notion-text notion-block-19637cbb8d10806a803ff024bb9996a6">此外，ElevenLabs Studio还支持多种文件格式，包括EPUB、TXT、PDF、HTML等，用户可以直接从URL初始化项目。该工具提供了一个全面的编辑器，用户可以添加多个章节，为不同部分分配独特的声音，重新生成并下载选定的短语，并锁定已完成的部分以便更好地组织内容。</div><div class="notion-text notion-block-19637cbb8d10801c807deae51586dd6c">ElevenLabs Studio的开放，为创作者和企业提供了更强大的工具，用于制作高质量的音频内容，无论是有声读物、播客还是其他长格式音频项目。</div><div class="notion-blank notion-block-19637cbb8d108018b2a2e42724a2ed02"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d10806e8508f1790914e1dc" data-id="19637cbb8d10806e8508f1790914e1dc"><span><div id="19637cbb8d10806e8508f1790914e1dc" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10806e8508f1790914e1dc" title="OpenAI Canvas及o1模型更新"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">OpenAI Canvas及o1模型更新</span></span></h4><div class="notion-text notion-block-19637cbb8d10806fb27fdb17b89c800d">OpenAI宣布用户现在可以与合作者共享Canvas工作区，方便团队协作和项目共享。<div class="notion-text-children"><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080528118cfd544020daa"><li><b>o1模型集成</b>：Canvas现在集成了o1模型，支持更复杂的任务处理，如代码生成和文档编辑。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10803184a8dfa915aa50ee"><li><b>HTML与React支持</b>：Canvas新增了对HTML和React代码的渲染支持，提升开发效率。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080e0bd82c8e2e340fd02"><li><b>o3-mini模型更新</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080e0bd82c8e2e340fd02"><li><b>思维链更新</b>：OpenAI更新了o3-mini模型的思维链展示方式，使其更透明、更详细，帮助用户理解模型的推理过程。</li></ul></ul></div></div><div class="notion-text notion-block-19637cbb8d1080e89ebeca5c9f55e98c">此次更新面向所有用户，包括免费和付费用户，付费用户还将体验到更高级的o3-mini-high版本。</div><div class="notion-blank notion-block-19637cbb8d1080f6b9d3fdf652374d5c"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d1080d88557e077ca9dddad" data-id="19637cbb8d1080d88557e077ca9dddad"><span><div id="19637cbb8d1080d88557e077ca9dddad" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080d88557e077ca9dddad" title="Pika Labs推出Pikadditions：视频编辑新突破"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Pika Labs推出Pikadditions：视频编辑新突破</span></span></h4><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080afa124d4942bd8c532"><li><b>功能介绍</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080afa124d4942bd8c532"><li><b>无缝集成</b>：Pikadditions允许用户将任何人物或物体无缝融入现有视频中，同时保留原始视频和声音。</li><li><b>操作简单</b>：用户只需上传一张图片和一个视频，并输入简单的提示（Prompt），AI将自动完成合成。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080298f3af51d3f020793"><li><b>技术特点</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080298f3af51d3f020793"><li><b>现实与想象融合</b>：通过先进的AI技术，Pikadditions在1080p高清分辨率下，实现虚拟元素与现实画面的完美融合。</li><li><b>智能场景整合</b>：系统会根据视频中的动作和场景上下文，自动调整添加元素的大小、位置和动作，确保自然和引人入胜的整合。</li><li><b>专业级增强引擎</b>：利用最新的2.1模型技术，Pikadditions提供工作室级别的效果，确保每个添加元素自然地融入视频。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10804495eeef15f54f7508"><li><b>使用方法</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d10804495eeef15f54f7508"><li><b>上传视频和图片</b>：用户需上传一个至少五秒长的视频和一张图片。</li><li><b>输入提示</b>：描述如何将添加的元素融入视频，考虑当前动作和场景上下文。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10808dadd1edb1af9f7b3b"><li><b>免费体验</b>：新用户注册可获得15次免费体验机会。</li></ul><div class="notion-blank notion-block-19637cbb8d1080d4a503deba4c5b887d"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d10800d9589df5ef0a9561c" data-id="19637cbb8d10800d9589df5ef0a9561c"><span><div id="19637cbb8d10800d9589df5ef0a9561c" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10800d9589df5ef0a9561c" title="OpenAI将在超级碗期间首次投放广告"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">OpenAI将在超级碗期间首次投放广告</span></span></h4><div class="notion-text notion-block-19637cbb8d108077813aec0ed174d09d">OpenAI将在2025年超级碗期间播出其首个电视广告，这是其首次大规模的付费广告活动。</div><div class="notion-text notion-block-19637cbb8d108074be76cd062f481b12">该广告时长为60秒，投放成本约为1400万美元，预计观看人数将达到1.3亿。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a8b390e6371792ec6f"><li><b>广告创意</b>：广告通过独特的点画风格动画，展示了人类技术的进化，从早期的火和轮子到现代的DNA测序和太空探索，最终聚焦于ChatGPT处理日常任务的能力。</li></ul><div class="notion-blank notion-block-19637cbb8d1080159788f844ea6d0906"> </div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d1080d78846f21b62241586" data-id="19637cbb8d1080d78846f21b62241586"><span><div id="19637cbb8d1080d78846f21b62241586" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080d78846f21b62241586" title="Mistral推出Le Chat移动应用及平台更新"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Mistral推出Le Chat移动应用及平台更新</span></span></h4><div class="notion-text notion-block-19637cbb8d10804e9fd3d2c676d261ac">法国AI创业公司Mistral AI发布了iOS和Android版的Le Chat移动应用，这是其聊天机器人Le Chat的手机版，具备网页搜索、多模态输入、图像生成和代码解释器等功能。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080a18129ca960a11b6be"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Add33f999-e01e-41a7-baea-2a295ac027e4%3Aimage.png?table=block&amp;id=19637cbb-8d10-80a1-8129-ca960a11b6be&amp;t=19637cbb-8d10-80a1-8129-ca960a11b6be" alt="notion image" loading="lazy" decoding="async"/></div></figure><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080d389c7de6886a4c96a"><li><b>核心功能</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080d389c7de6886a4c96a"><li><b>网页搜索</b>：结合网页搜索、新闻、社交媒体平台及其他来源的新近资讯，确保回应的正确及时效性。</li><li><b>多模态输入</b>：支持视觉和文件理解，基于视觉和光学字符识别（OCR）模型，可准确读取PDF、试算表、log文件与复杂图片文件。</li><li><b>代码解释器</b>：内置Code interpreter功能，用户可在沙箱环境中运行程序、进行科学分析、执行虚拟化及模拟。</li><li><b>图像生成</b>：基于Black Forest Labs Flux Ultra模型，生成逼真的高品质图片。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10808e9956c313546c86c3"><li><b>速度优势</b>：Le Chat引入了“Flash Answers”功能，处理速度超过ChatGPT和Claude等竞争对手的10倍，1秒可输出1,000字。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080f99e39e4923511253c"><li><b>定价方案</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080f99e39e4923511253c"><li><b>免费版</b>：提供最新模型、网页搜索、图像生成、文件上传、打开URL及Canvas功能。</li><li><b>Pro版</b>：每月14.99美元，提供Mistral最高性能模型、无限消息、更高速率限制，并可选择不与Mistral共享数据。</li><li><b>Team版</b>：每月每用户24.99美元，面向企业用户。</li><li><b>企业版</b>：提供本地部署、VPC或SaaS部署选项，支持自定义模型和工具集成，目前处于私密预览阶段。</li><div class="notion-blank notion-block-19637cbb8d108038905bce7c2ee18019"> </div></ul></ul><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d108099bdebdc19d96f35a1" data-id="19637cbb8d108099bdebdc19d96f35a1"><span><div id="19637cbb8d108099bdebdc19d96f35a1" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d108099bdebdc19d96f35a1" title="GitHub Copilot全面整合智能体技术"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">GitHub Copilot全面整合智能体技术</span></span></h4><div class="notion-text notion-block-19637cbb8d10809ba593d3144a4b4cfd">GitHub Copilot引入了新的代理模式，能够自主迭代代码，自动识别并修复运行时错误，同时推断并执行额外任务，显著提升开发效率。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080c38dc3e1d679557771"><li><b>视觉功能（Vision）</b>：新增“视觉”功能，支持用户上传截图、照片或图表等非文本文件，并据此生成界面、代码及替代文本。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080cbb50bd0c789cbe23f"><li><b>Copilot Edits功能</b>：该功能现全面可用，支持在VS Code中通过自然语言指令进行多文件编辑，提供代码更改建议。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108082862bf1091fc1bfa9"><li><b>下一步编辑建议（Next Step Editing Suggestions）</b>：分析开发者最近的操作，预测其下一步需求，并提供智能编辑建议。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080da9f10e64168fff8ac"><li><b>Project Padawan</b>：预告了一个名为Project Padawan的自主编码代理，预计今年推出，能够处理GitHub问题并生成经过全面测试的拉取请求。</li></ul><div class="notion-blank notion-block-19637cbb8d1080a88aaeea7aff6236e0"> </div><div class="notion-text notion-block-19337cbb8d1080e1be01d8d701c29edb"><b>探索·AI数字人</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d10804d809cf818eefcc38b" data-id="19637cbb8d10804d809cf818eefcc38b"><span><div id="19637cbb8d10804d809cf818eefcc38b" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10804d809cf818eefcc38b" title="AI生成的贾斯汀·比伯歌曲引发关注"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">AI生成的贾斯汀·比伯歌曲引发关注</span></span></h4><div class="notion-text notion-block-19637cbb8d10802da9faf933ac0ac7fd">2024年4月，一首据称由贾斯汀·比伯演唱的AI生成歌曲开始在社交媒体平台上广泛传播，尤其是在TikTok和YouTube上。假AI生成的贾斯汀·比伯（Justin Bieber）的歌曲在YouTube上有近200万次观看。</div><div class="notion-blank notion-block-19637cbb8d10808ca182db690d84bdc1"> </div><div class="notion-text notion-block-19637cbb8d10808ebd8ccaa9cb664904"><b>探索·新研究</b></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d1080829f25ce96b8441b7d" data-id="19637cbb8d1080829f25ce96b8441b7d"><span><div id="19637cbb8d1080829f25ce96b8441b7d" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080829f25ce96b8441b7d" title="Neuralink揭示其三名患者如何“在日常生活中使用心灵感应”"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">Neuralink揭示其三名患者如何“在日常生活中使用心灵感应”</span></span></h4><div class="notion-text notion-block-19637cbb8d10804bb0a4f9868b855aa8">Neuralink的实验性脑接口技术取得了显著进展，其三名患者通过植入的脑机接口设备实现了“心灵感应”功能。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080c99bb9d1900f151cb3"><li><b>患者体验</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080c99bb9d1900f151cb3"><li><b>首位患者Noland Arbaugh</b>：通过植入设备，Noland能够仅凭意念控制电脑，观看视频、阅读、下棋和玩游戏，甚至在飞机上使用该技术。</li><li><b>第二位患者</b>：成功植入了400个电极，设备运行良好，患者能够通过意念控制外部设备。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080228aa5cb1a0cec53ea"><li><b>技术细节</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080228aa5cb1a0cec53ea"><li><b>植入设备</b>：Neuralink的设备通过植入大脑的电极捕捉神经信号，并将其转化为计算机可识别的指令。</li><li><b>产品Telepathy</b>：该产品旨在帮助神经元受损的人恢复身体功能，使他们能够通过意念控制设备。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108042b01df9ea9e3c451c"><li><b>未来展望</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d108042b01df9ea9e3c451c"><li><b>更多植入手术</b>：Neuralink计划在2025年完成至少8次植入手术，进一步验证技术的安全性和实用性。</li><li><b>终极目标</b>：马斯克表示，Neuralink的终极目标是让人类与AI建立共生关系，降低AI带来的风险。</li></ul></ul><div class="notion-blank notion-block-19337cbb8d1080dbb8cce12e1ea4e4bf"> </div><div class="notion-blank notion-block-19337cbb8d108008b18ddf57b500ef8c"> </div><div class="notion-text notion-block-18437cbb8d108167b99ef3b0ee976f5d"><b>【🚀 精选内容】</b></div><div class="notion-row notion-block-18437cbb8d108167a0d2e76a217cc105"><div class="notion-column notion-block-18437cbb8d1081fa8d9cfaf73edf9459" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"><a class="notion-page-link 570c7c80-1df6-4ac1-afd8-2e20674ec6f7" href="/570c7c801df64ac1afd82e20674ec6f7"><span class="notion-page-title"><div class="notion-page-icon-inline notion-page-icon-image"><svg class="notion-page-title-icon notion-page-icon" alt="我用OpenGlass做了一个AI眼镜【上篇】" viewBox="0 0 30 30" width="16"><path d="M16,1H4v28h22V11L16,1z M16,3.828L23.172,11H16V3.828z M24,27H6V3h8v10h10V27z M8,17h14v-2H8V17z M8,21h14v-2H8V21z M8,25h14v-2H8V25z"></path></svg></div><span class="notion-page-title-text"><b>我用OpenGlass做了一个AI眼镜【上篇】</b></span></span></a></div><div class="notion-spacer"></div><div class="notion-column notion-block-18437cbb8d10818380a7fd5f2f4cbb45" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"><a class="notion-page-link cfc49a0e-341c-4376-b680-0dc1f8f197a3" href="/cfc49a0e341c4376b6800dc1f8f197a3"><span class="notion-page-title"><div class="notion-page-icon-inline notion-page-icon-image"><svg class="notion-page-title-icon notion-page-icon" alt="Flux 完整的 Lora 设置和数据集指南 - 两周学习的事后分析[译文]" viewBox="0 0 30 30" width="16"><path d="M16,1H4v28h22V11L16,1z M16,3.828L23.172,11H16V3.828z M24,27H6V3h8v10h10V27z M8,17h14v-2H8V17z M8,21h14v-2H8V21z M8,25h14v-2H8V25z"></path></svg></div><span class="notion-page-title-text"><b>Flux 完整的 Lora 设置和数据集指南 - 两周学习的事后分析[译文]</b></span></span></a></div><div class="notion-spacer"></div><div class="notion-column notion-block-18437cbb8d108121a954f5d282e6d923" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"><a class="notion-page-link 470debdd-454d-4443-8d5a-2d2ab4672ed2" href="/470debdd454d44438d5a2d2ab4672ed2"><span class="notion-page-title"><div class="notion-page-icon-inline notion-page-icon-image"><svg class="notion-page-title-icon notion-page-icon" alt="Midjourney上线人物一致功能 我拥有了一个IP宇宙" viewBox="0 0 30 30" width="16"><path d="M16,1H4v28h22V11L16,1z M16,3.828L23.172,11H16V3.828z M24,27H6V3h8v10h10V27z M8,17h14v-2H8V17z M8,21h14v-2H8V21z M8,25h14v-2H8V25z"></path></svg></div><span class="notion-page-title-text">Midjourney上线人物一致功能 我拥有了一个IP宇宙</span></span></a></div><div class="notion-spacer"></div></div><div class="notion-text notion-block-18437cbb8d1081b3921ed49305f421ff"><b>❤ 如果对你有帮助，欢迎分享或者Buy Me A Coffee ❤</b></div><figure class="notion-asset-wrapper notion-asset-wrapper-embed notion-block-18437cbb8d10812585f8e946aa7439ea"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:690px"><iframe class="notion-asset-object-fit" src="https://buymeacoffee.com/rocketlu?spaceId=2666d03a-fb22-44f0-83a0-96826c4e3d2d" title="iframe embed" frameBorder="0" allowfullscreen="" loading="lazy" scrolling="auto"></iframe></div></figure><div class="notion-blank notion-block-18437cbb8d108114b4e0ea974f9a2f77"> </div><div class="notion-callout notion-gray_background_co notion-block-18437cbb8d10814e84b2c328ab6ee006"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="💡">💡</span></div><div class="notion-callout-text"><b>对这个话题感兴趣的小伙伴，欢迎加我一起探索交流~</b></div></div><div class="notion-row notion-block-18437cbb8d1081c0ab4ac2247ce2a2b8"><div class="notion-column notion-block-18437cbb8d1081a08bcef8fd66fde5cc" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-18437cbb8d1081ffaa58e4ac8f554217"> </div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-18437cbb8d10817b8a99d139a9b20621"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2F2666d03a-fb22-44f0-83a0-96826c4e3d2d%2F154b5cdb-df44-4c08-9984-33dbebdc5057%2Flittlerocketlu.jpg?table=block&amp;id=18437cbb-8d10-817b-8a99-d139a9b20621&amp;t=18437cbb-8d10-817b-8a99-d139a9b20621&amp;width=540.9921875&amp;cache=v2" alt="notion image" loading="lazy" decoding="async"/></div></figure></div><div class="notion-spacer"></div><div class="notion-column notion-block-18437cbb8d1081fc83ffe50723987b11" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-18437cbb8d108140907bd61eb460ad80"> </div></div><div class="notion-spacer"></div></div><div class="notion-blank notion-block-18437cbb8d108142a5a2ef1e1b41bb3f"> </div><div class="notion-blank notion-block-18437cbb8d1081a6b419eb4eb483d3a1"> </div><div class="notion-blank notion-block-18437cbb8d1081359369e05d2227f298"> </div><div class="notion-blank notion-block-18437cbb8d10816a9494e1bab5dac314"> </div><div class="notion-blank notion-block-18437cbb8d108185ba41e7cb801621ac"> </div></main></div>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[AIDaily 076/100 DeepSeek VL2-Small发布、阿里巴巴Qwen2.5-Max超越DeepSeek、Meta推出Apollo和Motivo、Google Gemini 2.0系列发布…]]></title>
            <link>http://rocketlu.cn/article/Daily76</link>
            <guid>http://rocketlu.cn/article/Daily76</guid>
            <pubDate>Fri, 07 Feb 2025 00:00:00 GMT</pubDate>
            <description><![CDATA[每天更新关于AI有趣有用的信息]]></description>
            <content:encoded><![CDATA[<div id="notion-article" class="mx-auto overflow-hidden "><main class="notion light-mode notion-page notion-block-19637cbb8d1080d3bf2be867cadb06db"><div class="notion-viewport"></div><div class="notion-collection-page-properties"></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d108159a572e67f3fbb7432" data-id="19637cbb8d108159a572e67f3fbb7432"><span><div id="19637cbb8d108159a572e67f3fbb7432" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d108159a572e67f3fbb7432" title="AIDaily 076/100"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><span class="notion-orange">AIDaily 076/100</span></span></span></h4><div class="notion-callout notion-orange_background_co notion-block-19637cbb8d108108a061db356beb8df6"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="🖼️">🖼️</span></div><div class="notion-callout-text">每天都能接收到无数条与AI、科技、艺术、经济相关的信息。
但是感觉自己就像那只掰玉米的熊，掰了一路，最后出来发现只剩下手里的两根玉米🌽。
今年希望能够以Newsletter的形式，给自己掰下来的玉米们找个背篓。

<span class="notion-orange">人们会被自己热爱的事物改变，而没有人因为给予而贫穷。</span></div></div><h4 class="notion-h notion-h3 notion-h-indent-0 notion-block-19637cbb8d10814a95fbdfee9fadc8f3" data-id="19637cbb8d10814a95fbdfee9fadc8f3"><span><div id="19637cbb8d10814a95fbdfee9fadc8f3" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10814a95fbdfee9fadc8f3" title="Vol.076"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>Vol.076</b></span></span></h4><div class="notion-text notion-block-19637cbb8d108150972ae70b5a6a81aa"><span class="notion-gray">by Rocket</span></div><hr class="notion-hr notion-block-19637cbb8d108140888ef15aea49a3ae"/><div class="notion-text notion-block-19637cbb8d1081c295f1edad5aff8a10"><b>探索·AI产品</b></div><div class="notion-text notion-block-19637cbb8d108179ae06ec09cdff7016"><b>Perplexity推出o3-mini：支持Web搜索的AI推理模型</b></div><div class="notion-text notion-block-19637cbb8d1081568b9cebe28712eaa4">Perplexity的o3-mini模型现在支持Web搜索功能，用户可以利用其推理能力获取最新答案，并提供相关网页链接。免费用户每天最多可以执行五次搜索。Pro用户每天可获得多达500次搜索。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d108151a9f0f386fe04d8ae"><li><b>推理能力</b>：o3-mini在数学、编程和科学等STEM领域表现优异，支持复杂任务的分解与多步决策。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108118b9b2d5bf9b19e196"><li><b>性能提升</b>：与o1-mini相比，o3-mini的响应速度提升了24%，答案准确性也有所提高。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10817fb511f3c142cd807c"><li><b>成本效益</b>：o3-mini的定价为每百万输入tokens 1.10美元，每百万输出tokens 4.40美元，相比o1-mini降低了63%。</li></ul><div class="notion-blank notion-block-19637cbb8d10814a9c81e07397778fa1"> </div><div class="notion-text notion-block-19637cbb8d1081238e78d8eca4cc4890"><b>DeepSeek R1引发的全球关注与争议</b></div><div class="notion-text notion-block-19637cbb8d1081d7b623f129a74f0a32">DeepSeek的R1模型发布后，美国科技股出现了大幅波动，纳斯达克综合指数下跌约3%，英伟达的股价下跌超过15%。这一现象表明市场对DeepSeek的低成本、高性能模型的出现感到不安，担心其可能对现有AI巨头构成威胁。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d108159acf1f3f6a0abbffe"><li><b>技术细节与优势</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d108159acf1f3f6a0abbffe"><li>DeepSeek-R1采用了纯强化学习的方法，实现了接近OpenAI o1的性能，但训练成本仅为557万美元，研发周期不到两个月。</li><li>DeepSeek-R1的开发过程包含了两个强化学习阶段和两个监督微调阶段，这些阶段相辅相成，为模型的推理和非推理能力打下了基础。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081a0ad7ffa5fd29d1655"><li><b>创新与竞争</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081a0ad7ffa5fd29d1655"><li><b>创新激励</b>：DeepSeek的R1模型可能激励美国AI公司加快创新速度，Y Combinator的Garry Tan表示，更便宜、更易获取的训练方法将加速AI应用的需求。</li><li><b>开源合作的力量</b>：Meta的首席AI科学家Yann LeCun强调，DeepSeek的成功得益于开源研究和合作，这表明开源模型在AI发展中的重要性。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081eebdceff50bf09b141"><li><b>企业AI采用率</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081eebdceff50bf09b141"><li>数百家公司，尤其是与政府相关的企业，已屏蔽了DeepSeek。Armis表示大约70%的客户已请求阻止访问DeepSeek，Netskope威胁实验室总监Ray Canzanese透露，52%的Netskope客户完全阻止了对DeepSeek网站的访问。</li><li>微软、亚马逊、英伟达等科技巨头已将DeepSeek-R1集成到其产品和服务中。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081f2beadf08c43da53ef"><li><b>监管与市场考量</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081f2beadf08c43da53ef"><li>美国政府正在调查DeepSeek是否通过新加坡的第三方购买英伟达（NVIDIA）的先进半导体，以规避美国对中国用于AI任务的芯片销售限制。</li><li>美国国会众议院首席行政事务官已向国会办公室发出通知，警告不要使用DeepSeek的服务。</li><li>美国海军、国会、国防部以及NASA等机构已相继禁止在政府官方设备上使用DeepSeek。</li><li>美国政府还在考虑发布新法案，下载DeepSeek将被定为犯罪，最高判处20年监禁。</li><blockquote class="notion-quote notion-block-19637cbb8d1081b5b02ef36c7f6f12d5"><div>法案链接：https://www.hawley.senate.gov/wp-content/uploads/2025/01/Hawley-Decoupling-Americas-Artificial-Intelligence-Capabilities-from-China-Act.pdf</div></blockquote></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10815c97c6c2df7e74b4e9"><li><b>国际社会的反应</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d10815c97c6c2df7e74b4e9"><li>意大利、爱尔兰等多国政府已开展封锁行动或计划对DeepSeek进行审查。</li><li>比利时政府隐私监管机构确认收到有关DeepSeek的投诉，但未就是否已对DeepSeek启动调查作出评论。</li></ul></ul><div class="notion-text notion-block-19637cbb8d10816c9c10e7cda31e51ba">DeepSeek R1的发布不仅在技术上取得了突破，也在AI行业内部掀起了关于成本效益、开源合作以及全球竞争格局的广泛讨论。</div><div class="notion-blank notion-block-19637cbb8d108129a403d84ef3112502"> </div><div class="notion-text notion-block-19637cbb8d10810dbedcf7f04d3d0eb4"><b>探索·AI大模型</b></div><div class="notion-text notion-block-19637cbb8d10817f8717da5910314b06"><b>DeepSeek VL2-Small：强大的视觉语言模型</b></div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081a6870cd0b4b12a9e5c"><li><b>功能特点</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081a6870cd0b4b12a9e5c"><li><b>视觉理解</b>：DeepSeek-VL2-Small特别擅长分析文档和图像，能够处理视觉问答、光学字符识别、文档/表格/图表理解以及视觉定位等任务。</li><li><b>多图像对话</b>：可以分析多张图像之间的关联和差异，并整合内容进行简单推理，例如根据几张图像编写创意故事。</li><li><b>视觉定位</b>：能够识别图像中的物体位置，即使在不同场景下也能准确识别。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081d8999ad26f1c11f004"><li><b>技术优势</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081d8999ad26f1c11f004"><li><b>动态高分辨率视觉编码</b>：采用动态平铺视觉编码策略，有效处理不同纵横比的高分辨率图像。</li><li><b>多头潜在注意力机制</b>：优化语言模型架构，减少计算开销，提高推理速度。</li><li><b>混合专家（MoE）架构</b>：在任务执行期间仅激活必要的参数子集，提高可扩展性和效率。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108124a079ea9aa1ff8609"><li><b>模型变体</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d108124a079ea9aa1ff8609"><li><b>DeepSeek-VL2-Tiny</b>：1.0B激活参数，适合轻量级部署。</li><li><b>DeepSeek-VL2-Small</b>：2.8B激活参数，适合中等计算需求。</li><li><b>DeepSeek-VL2</b>：4.5B激活参数，适合资源密集型任务。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10815aa3dadcca2ef1ca45"><li><b>开源与社区支持</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d10815aa3dadcca2ef1ca45"><li><b>Hugging Face</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://huggingface.co/spaces/deepseek-ai/deepseek-vl2-small?utm_source=superhuman&amp;utm_medium=referral&amp;utm_campaign=gemini-2-0-pro-is-here">https://huggingface.co/spaces/deepseek-ai/deepseek-vl2-small</a></li><li><b>GitHub仓库</b>：<a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://github.com/deepseek-ai/DeepSeek-VL2">https://github.com/deepseek-ai/DeepSeek-VL2</a></li></ul></ul><div class="notion-text notion-block-19637cbb8d1081fcb3a1f6ae52fe4b8e">DeepSeek-VL2-Small的发布为视觉语言模型领域带来了新的突破，其强大的功能和高效的架构使其成为开发者和研究人员的有力工具。</div><div class="notion-blank notion-block-19637cbb8d1081b581f3fa46edffe598"> </div><div class="notion-text notion-block-19637cbb8d1081a8b5f9c9968e36674f"><b>阿里巴巴Qwen2.5-Max超越DeepSeek V3和Meta Llama 3.1</b></div><div class="notion-text notion-block-19637cbb8d108111b8ced5e53842e701">阿里巴巴的Qwen2.5-Max在多个基准测试中取得了优异成绩，超越了DeepSeek V3和Meta的Llama 3.1。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d108176aa27f25416824518"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A47d5224f-29f1-471d-8443-12d5d846f0d0%3Aimage.png?table=block&amp;id=19637cbb-8d10-8176-aa27-f25416824518&amp;t=19637cbb-8d10-8176-aa27-f25416824518" alt="notion image" loading="lazy" decoding="async"/></div></figure><ul class="notion-list notion-list-disc notion-block-19637cbb8d108173ad6fc7fda731340c"><li><b>测评结果</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d108173ad6fc7fda731340c"><li>在数学和编程方面排名第一。</li><li>在处理复杂任务的硬提示（hard prompts）方面排名第二。</li><li>在总体排名中跻身第7，领先于DeepSeek V3、O1-Mini和Claude-3.5-Sonnet等顶级专有大语言模型。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10818181c1f16c4bb73373"><li><b>技术架构</b>：Qwen2.5-Max采用了超大规模的专家混合（MoE）模型架构，预训练数据量超过20万亿个token，运用监督微调（SFT）和人类反馈强化学习（RLHF）技术进行优化。</li></ul><div class="notion-blank notion-block-19637cbb8d108135ae20d552943f3494"> </div><div class="notion-text notion-block-19637cbb8d10812c8fb9e6715a659608"><b>Meta推出Apollo：专注于视频理解的全新Video-LMM系列</b></div><div class="notion-text notion-block-19637cbb8d108113a579e0dd9c51b03f">尽管多模态模型（LMMs）在文本和图像任务上取得了显著进展，但基于视频的模型仍面临挑战。视频结合了空间和时间维度，需要更多计算资源，且现有方法难以捕捉运动和时间模式。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d10818ebbbede6d32eaea42"><li><b>Apollo模型</b>：Meta AI和斯坦福大学的研究人员开发了Apollo，这是一系列专注于视频的LMMs，旨在突破视频理解的界限。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081818e0cd545367bc97f"><li><b>主要创新</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081818e0cd545367bc97f"><li><b>每秒帧采样（fps）</b>：与均匀帧采样相比，fps采样保持一致的时间流，使Apollo能够更好地理解视频中的运动。</li><li><b>双视觉编码器</b>：结合SigLIP和InternVideo2，实现视频数据的平衡表示。</li><li><b>标记重采样</b>：通过感知器重采样器减少视频标记，处理长视频时无需过多计算开销。</li><li><b>优化训练</b>：采用三阶段训练流程，确保稳定有效的学习。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081279f72e696a890cdec"><li><b>性能优势</b>：Apollo在多个基准测试中取得了优异成绩，超越了DeepSeek V3和Meta的Llama 3.1。</li></ul><div class="notion-blank notion-block-19637cbb8d1081b48453da9101be0b00"> </div><div class="notion-text notion-block-19637cbb8d10817b9a65fd445565d5ed"><b>Meta Motivo：连接人类动作与人形机器人的新算法</b></div><div class="notion-text notion-block-19637cbb8d10810d8711fc8105bd8429">Meta Motivo基于一种名为Forward-Backward Representations with Conditional Policy Regularization（FB-CPR）的特殊算法，通过分析大量未标记的动作数据来学习人类的运动。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d10816689c1df932e01e3c1"><li><b>主要特点</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d10816689c1df932e01e3c1"><li><b>适应性</b>：能够适应不同条件，如重力变化或风等干扰，仍能执行类似人类的动作。</li><li><b>无监督学习</b>：不依赖于每个任务的具体指令，减少了为不同用途编程这些系统的工作量。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081e08a0cd3c5890a7821"><li><b>应用场景</b>：可用于创建逼真的视频游戏角色动画、辅助物理康复或控制人形机器人执行精确动作。</li></ul><div class="notion-blank notion-block-19637cbb8d1081af901bfacab27a2f8b"> </div><div class="notion-text notion-block-19637cbb8d10811dbe4fc8df013f26a5"><b>Google Gemini 2.0系列发布，推动AI技术普及化</b></div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081ce8fe0dc35a691e7ad"><li><b>Gemini 2.0 Flash Thinking Experimental</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081ce8fe0dc35a691e7ad"><li>Google发布了Gemini 2.0 Flash Thinking Experimental模型，该模型能够将复杂任务分解为一系列步骤，并展示其推理过程。这使得用户可以更清楚地了解模型为何以某种方式响应，其假设是什么，以及推理的逻辑路径。</li><li><b>性能优势</b>：该模型在LMArena的基准测试中被评为“世界上最好的模型”，并且对所有用户免费开放。</li><li><b>应用场景</b>：适用于需要深入研究、构建自己的AI代理以及部署编码项目的人群。</li><li><b>对比 DeepSeek V3 （粗算）</b>，成本低 6 倍；输出速度快 60 倍；上下文长 16 倍；原生全模态支持；Google 亲生，算力管够</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081beb6b4c38258948892"><li><b>Gemini 2.0 Pro Experimental</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081beb6b4c38258948892"><li><b>功能特点</b>：专为复杂任务设计，提供更好的事实性和更强的编程和数学提示性能。</li><li><b>适用人群</b>：目前仅对开发人员和Gemini高级用户开放。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081f49e0de5ff699cb6ca"><li><b>Gemini 2.0 Flash-Lite</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081f49e0de5ff699cb6ca"><li><b>功能特点</b>：性能优于其前身Gemini 1.5 Flash，但保持了相同的速度和成本。它是一个多模态模型，能够处理图像和文本输入，并生成文本输出。</li><li><b>性能优势</b>：在大部分基准测试中优于Gemini 1.5 Flash，能够为4万帧图片生成一行的文本图说，在AI Studio中花费不到1美元。</li></ul></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081379bb9d83b848f5c0f"><li><b>行业影响</b>：</li><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081379bb9d83b848f5c0f"><li><b>AI技术普及化</b>：Google的这一系列发布标志着AI技术向更广泛的用户群体开放，让任何人进行深入研究、建立自己的代理并免费部署编码项目。</li><li><b>市场竞争</b>：Google计划在2025年投入约750亿美元用于AI项目，显示出其在AI领域的持续投入和竞争决心。</li></ul></ul><div class="notion-text notion-block-19637cbb8d1081b59200cf3557ec7f65"><div class="notion-text-children"><div class="notion-text notion-block-19637cbb8d10810cb900de5b3f62e0f5">Google预告将在未来几个月内逐步增加新的模态能力。随着AI技术的快速发展，预计OpenAI或Anthropic等竞争对手将很快推出更先进的模型。</div></div></div><div class="notion-blank notion-block-19637cbb8d10819085dcea18a3086e9a"> </div><div class="notion-text notion-block-19637cbb8d10816e89c8e423582c0f68"><b>探索·AI数字人</b></div><div class="notion-text notion-block-19637cbb8d1081d8b166f2b363a106a9"><b>ByteDance推出OmniHuman-1：从单张照片生成逼真人像视频</b></div><div class="notion-text notion-block-19637cbb8d108125959be58fa81d2c86">ByteDance的OmniHuman-1模型能够通过单张照片生成逼真的全身人像视频，支持自然的说话、唱歌和动作。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d10816ba21aec580303f6d3"><li><b>多模态输入</b>：该模型支持多种输入方式，包括音频驱动的动画（生成同步的唇部动作和手势）、视频驱动的动画（复制参考视频中的动作），以及多模态融合（结合音频和视频信号，精确控制身体各部位）。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081319fb5ea942e465283"><li><b>技术架构</b>：OmniHuman-1采用扩散变换器（Diffusion Transformer）架构，结合“全条件训练策略”（Omni-Conditions Training），能够处理多种输入条件，生成自然流畅的动作。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081038d52c4e5208c415c"><li><b>性能优势</b>：OmniHuman-1在多个基准测试中表现优异，包括唇部同步精度、手势表现力和手部关键点置信度等指标，均优于其他领先模型。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108180a070e4c47b6b55c7"><li><b>应用场景</b>：该模型适用于虚拟形象、数字故事创作、游戏开发和AI辅助电影制作等领域，能够根据不同的身体比例和宽高比生成视频。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108126abd1ed6a77699ea3"><li><b>数据利用</b>：OmniHuman-1利用了大量数据进行训练，包括18,700小时的人类动作数据，这使得模型能够生成更真实、更自然的动作。</li></ul><div class="notion-text notion-block-19637cbb8d1081bfa7ddce6a697c083f">OmniHuman-1的发布标志着AI视频生成技术的重大进步，其生成的视频在视觉上几乎与真实视频无法区分。</div><div class="notion-blank notion-block-19637cbb8d108161850bf6389eea5482"> </div><div class="notion-text notion-block-19637cbb8d1081b0813cd98774c83c1a"><b>探索·新投融资</b></div><div class="notion-text notion-block-19637cbb8d10816a883ed51ea55ec03d"><b>ElevenLabs完成2.5亿美元C轮融资</b></div><div class="notion-text notion-block-19637cbb8d1081a9a982e7dc6566b0df">AI音频技术公司ElevenLabs完成了2.5亿美元的C轮融资，公司估值达到30亿至33亿美元。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081619cdec058346a9453"><li><b>投资方</b>：此轮融资由ICONIQ Growth领投，其他参与方包括a16z、NEA、World Innovation Lab等。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081d29699e1929c1a2046"><li><b>技术特点</b>：ElevenLabs专注于AI语音克隆和配音技术，其产品包括对话AI、语音设计、音效模型、多语言配音等。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081f3909ef5d1f0bc526b"><li><b>应用场景</b>：ElevenLabs的技术被广泛应用于内容创作、客户支持、游戏、教育和无障碍等领域。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108154b5c9d6d79d0589cb"><li><b>未来计划</b>：融资将用于继续开发音频工具和业务拓展</li></ul><div class="notion-blank notion-block-19637cbb8d1081dcaeeacee324ac6d2c"> </div><div class="notion-text notion-block-19637cbb8d1081b3942af40f59777f74"><b>Cursor成为最快达到1亿美元年收入的AI编程助手</b></div><div class="notion-text notion-block-19637cbb8d1081caa223d308d118b403">根据Sacra的估计，Cursor在2024年底达到了1亿美元的年收入，比2023年的100万美元增长了9900%。这一增长速度使其成为历史上从100万美元到1亿美元年收入增长最快的SaaS公司。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081f0b9a7f5118d87449e"><li><b>用户基础</b>：Cursor通过大约36万名主要为个人开发者的客户群体达到了这一里程碑，这些客户每月支付20-40美元，平均合同价值为276美元。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081afb0dadc68a80f21f5"><li><b>市场表现</b>：Cursor的AI代码编辑器在开发者中迅速被采用，拥有超过4万名付费客户，包括OpenAI、Midjourney、Perplexity和Shopify等知名科技公司的工程师。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10811eb7eee66679e6285b"><li><b>产品特点</b>：Cursor基于VS Code构建，并扩展了AI功能。它通过智能代码补全、AI助手回答编程问题以及一系列工具来简化编码过程。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081b9b4c4e9de184310a4"><li><b>公司背景</b>：Cursor由Anysphere开发，该公司由几位麻省理工学院的学生于2022年创立。Anysphere在2024年12月完成了1.05亿美元的B轮融资，投后估值达到26亿美元。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10813b98e8d1d2bf15b8f0"><li><b>行业影响</b>：Cursor的快速增长和市场表现表明，AI编程助手正在成为软件开发中不可或缺的工具。许多风险投资家预测，由于这些工具带来的效率提升，未来初创公司可能需要更少的软件开发人员。</li></ul><div class="notion-text notion-block-19637cbb8d1081fba6edfa307bb35e43">Cursor的快速崛起不仅展示了AI技术在提高编程效率方面的潜力，也反映了市场对AI辅助开发工具的强烈需求。</div><div class="notion-blank notion-block-19637cbb8d10812eb7a1c4dbbbff54a8"> </div><div class="notion-text notion-block-19637cbb8d108154b2b2f787b4ee809f"><b>探索·新合作</b></div><div class="notion-text notion-block-19637cbb8d1081ffa33cf45f04970672"><b>OpenAI与SoftBank成立新公司SB OpenAI Japan</b></div><div class="notion-text notion-block-19637cbb8d108106acd9ffcf97784866">OpenAI与SoftBank集团宣布成立一家名为SB OpenAI Japan的合资企业，旨在为日本企业提供定制化的人工智能服务。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081ec8935f53c3e863ebe"><li><b>投资规模</b>：SoftBank集团将每年支付高达30亿美元，以将OpenAI的最新模型整合到其企业产品中。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10817c9885f3ddd33ad87b"><li><b>技术目标</b>：合资企业将专注于开发和推广名为“Cristal intelligence”的先进企业级AI解决方案，该方案能够安全地整合企业的系统和数据。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108123bbb7fc6bfde433e5"><li><b>市场定位</b>：SB OpenAI Japan将独家向日本的主要公司销售Cristal intelligence，帮助企业实现任务自动化和业务转型。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108187a2fdc9efa326e924"><li><b>合作意义</b>：此次合作不仅将推动SoftBank集团的运营变革，还将为日本乃至全球的企业工作方式带来革命性变化。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10819d991ecdead9e7e322"><li><b>技术基础</b>：Arm公司将提供计算平台，支持从云端到边缘的AI代理的性能、效率和可扩展性。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108185b12eeba66f5e31a9"><li><b>未来展望</b>：通过在日本的合作经验，双方计划创建一个可在全球范围内复制的AI驱动转型模型。</li></ul><div class="notion-text notion-block-19637cbb8d1081baad3fcebb3d5883fe">这一合作标志着OpenAI和SoftBank集团在AI领域迈出了重要一步，旨在通过定制化的AI解决方案提升企业的运营效率和创新能力。</div><div class="notion-blank notion-block-19637cbb8d108112a646de58030efad4"> </div><div class="notion-text notion-block-19637cbb8d10813981efe2c9406696e9"><b>Retro Biosciences与OpenAI合作开发GPT-4b Micro</b></div><div class="notion-text notion-block-19637cbb8d1081a0aea4faaf207dec73">Retro Biosciences与OpenAI合作，开发了一种名为GPT-4b Micro的AI模型，旨在通过重新设计蛋白质来延长人类寿命。</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d108113887ee66bebcea636"><li><b>技术特点</b>：GPT-4b Micro专注于Yamanaka因子，能够将成人皮肤细胞重新编程为多能干细胞，效率比传统方法提高了50倍以上。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1081c0a0cae459349b95a2"><li><b>应用场景</b>：该技术有望在器官再生和细胞替代疗法中取得突破，可能将人类寿命延长多达十年。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10819f8e99f64af3c52149"><li><b>项目进展</b>：目前没有明确的完成时间表，但OpenAI和Retro Biosciences承诺将与科学界分享研究成果。</li></ul><div class="notion-blank notion-block-19637cbb8d10815fa250c0966afe21d7"> </div><div class="notion-blank notion-block-19637cbb8d1081239f8ffdb403e532d8"> </div><div class="notion-text notion-block-19637cbb8d10815ebf71f8f0111d9a4d"><b>【🚀 精选内容】</b></div><div class="notion-row notion-block-19637cbb8d1081e28825cfd0f898d1d9"><div class="notion-column notion-block-19637cbb8d1081439108c2cbf459d386" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"></div><div class="notion-spacer"></div><div class="notion-column notion-block-19637cbb8d108190b501fe6f35921592" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"></div><div class="notion-spacer"></div><div class="notion-column notion-block-19637cbb8d10811a9e3bcd58430131a0" style="width:calc((100% - (2 * min(32px, 4vw))) * 0.3333333333333333)"></div><div class="notion-spacer"></div></div><div class="notion-text notion-block-19637cbb8d108128a5c8ca9650d54601"><b>❤ 如果对你有帮助，欢迎分享或者Buy Me A Coffee ❤</b></div><figure class="notion-asset-wrapper notion-asset-wrapper-embed notion-block-19637cbb8d1081b494daefdb93a26571"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:690px"><iframe class="notion-asset-object-fit" src="https://buymeacoffee.com/rocketlu?spaceId=2666d03a-fb22-44f0-83a0-96826c4e3d2d" title="iframe embed" frameBorder="0" allowfullscreen="" loading="lazy" scrolling="auto"></iframe></div></figure><div class="notion-blank notion-block-19637cbb8d1081fd9a27f3f2936d03d9"> </div><div class="notion-callout notion-gray_background_co notion-block-19637cbb8d1081a1ad75c1bc0549d28a"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="💡">💡</span></div><div class="notion-callout-text"><b>对这个话题感兴趣的小伙伴，欢迎加我一起探索交流~</b></div></div><div class="notion-row notion-block-19637cbb8d1081a19035c17d08ac78a2"><div class="notion-column notion-block-19637cbb8d108185af56c81e00ae8136" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19637cbb8d1081c2a4daceafdfc692d1"> </div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1081b0bcc0df0a626d0a75"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2F2666d03a-fb22-44f0-83a0-96826c4e3d2d%2F154b5cdb-df44-4c08-9984-33dbebdc5057%2Flittlerocketlu.jpg?table=block&amp;id=19637cbb-8d10-81b0-bcc0-df0a626d0a75&amp;t=19637cbb-8d10-81b0-bcc0-df0a626d0a75&amp;width=540.9921875&amp;cache=v2" alt="notion image" loading="lazy" decoding="async"/></div></figure></div><div class="notion-spacer"></div><div class="notion-column notion-block-19637cbb8d108145afe3f884e88cb7b0" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19637cbb8d10810da378d5def41d5e8d"> </div></div><div class="notion-spacer"></div></div><div class="notion-blank notion-block-19637cbb8d1081e8a4dad1a867f33950"> </div><div class="notion-blank notion-block-19637cbb8d1081848e2dc5df0dfd1931"> </div><div class="notion-blank notion-block-19637cbb8d1081999805ddc662658f67"> </div><div class="notion-blank notion-block-19637cbb8d10817bac5eda17e0089bcc"> </div><div class="notion-blank notion-block-19637cbb8d1081abb964df7526cc58d9"> </div></main></div>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[DeepSeek-R1 详解：如何打造推世界顶级推理大模型]]></title>
            <link>http://rocketlu.cn/article/DeepSeekR1</link>
            <guid>http://rocketlu.cn/article/DeepSeekR1</guid>
            <pubDate>Mon, 10 Feb 2025 00:00:00 GMT</pubDate>
            <description><![CDATA[如何打造推世界顶级推理大模型]]></description>
            <content:encoded><![CDATA[<div id="notion-article" class="mx-auto overflow-hidden "><main class="notion light-mode notion-page notion-block-19637cbb8d1080dca26eee7932bd0622"><div class="notion-viewport"></div><div class="notion-collection-page-properties"></div><h3 class="notion-h notion-h2 notion-h-indent-0 notion-block-19637cbb8d10809c8285c2d21d72c3a3" data-id="19637cbb8d10809c8285c2d21d72c3a3"><span><div id="19637cbb8d10809c8285c2d21d72c3a3" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10809c8285c2d21d72c3a3" title="引言"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>引言</b></span></span></h3><div class="notion-text notion-block-19637cbb8d1080ddb2b3e3144d410fbf">DeepSeek-R1 是 AI 发展的又一重要里程碑，尤其对机器学习（ML）和研究社区而言，它的发布具有以下几个关键意义：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a4bbf6fe01b601ff3a"><li>它是一个开源权重模型，并提供了更小、更精简的版本。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10805fa7c2ff9d73a3cec0"><li>它分享并探讨了一种训练方法，可用于复现类似 OpenAI O1 这样的推理型大模型。</li></ul><div class="notion-text notion-block-19637cbb8d1080c4a45dc5cfbd9ed023">本文将带您深入了解 DeepSeek-R1 的构建过程。</div><h3 class="notion-h notion-h2 notion-h-indent-0 notion-block-19637cbb8d1080a6a5eef22455a020de" data-id="19637cbb8d1080a6a5eef22455a020de"><span><div id="19637cbb8d1080a6a5eef22455a020de" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080a6a5eef22455a020de" title="核心内容"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title">核心内容</span></span></h3><ol start="1" class="notion-list notion-list-numbered notion-block-19637cbb8d108083b1b0d6e6461940b8" style="list-style-type:decimal"><li><b>回顾：大语言模型（LLM）是如何训练的</b></li></ol><ol start="2" class="notion-list notion-list-numbered notion-block-19637cbb8d10800a885cefde2ff123e1" style="list-style-type:decimal"><li><b>DeepSeek-R1 训练方法</b></li><ol class="notion-list notion-list-numbered notion-block-19637cbb8d10800a885cefde2ff123e1" style="list-style-type:lower-alpha"><ul class="notion-list notion-list-disc notion-block-19637cbb8d10806eacb1f4afabeefc08"><li><b>长链推理 SFT 数据</b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108062acdfd0b70681ff16"><li><b>一个高质量的推理模型（但在非推理任务上表现较差）</b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10803b954acca361ad98ea"><li><b>通过大规模强化学习（RL）打造推理模型</b></li><ul class="notion-list notion-list-disc notion-block-19637cbb8d10803b954acca361ad98ea"><li><b>R1-Zero：大规模推理导向的强化学习</b></li><li><b>用中间推理模型生成 SFT 推理数据</b></li><li><b>通用 RL 训练阶段</b></li></ul></ul></ol></ol><ol start="3" class="notion-list notion-list-numbered notion-block-19637cbb8d1080ec96a3dfb562ec4e3f" style="list-style-type:decimal"><li><b>模型架构</b></li></ol><ol start="4" class="notion-list notion-list-numbered notion-block-19637cbb8d10801d8ac7e7857fbfc8e9" style="list-style-type:decimal"><li><b>总结</b></li></ol><div class="notion-blank notion-block-19637cbb8d10802790d6df54ecbe4dea"> </div><h3 class="notion-h notion-h2 notion-h-indent-0 notion-block-19637cbb8d10800aa71bc2c48498cecf" data-id="19637cbb8d10800aa71bc2c48498cecf"><span><div id="19637cbb8d10800aa71bc2c48498cecf" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10800aa71bc2c48498cecf" title="回顾：大语言模型的训练方式"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>回顾：大语言模型的训练方式</b></span></span></h3><div class="notion-text notion-block-19637cbb8d108006bd8cd03fc800df6b">DeepSeek-R1 的生成方式与大多数 LLM 类似，每次预测一个 token（单词或字符片段）。但它在数学和推理任务上表现突出，因为它能够生成“思维 token”，即用于解释推理过程的额外内容，从而更深入地理解问题。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080f886d9d8507287e174"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A68c3f6c1-4a03-4ef1-b4c3-b11d0ba5d713%3Aimage.png?table=block&amp;id=19637cbb-8d10-80f8-86d9-d8507287e174&amp;t=19637cbb-8d10-80f8-86d9-d8507287e174" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d10801d87cfdfada7c4deff">下图（摘自《Hands-On Large Language Models》第 12 章）展示了训练高质量 LLM 的三个关键步骤：</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d10805f9026eb36b46c25db"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A679d8256-258c-4c18-a49b-013a77a2f367%3Aimage.png?table=block&amp;id=19637cbb-8d10-805f-9026-eb36b46c25db&amp;t=19637cbb-8d10-805f-9026-eb36b46c25db" alt="notion image" loading="lazy" decoding="async"/></div></figure><ol start="1" class="notion-list notion-list-numbered notion-block-19637cbb8d10803693cdda9ced0d9a21" style="list-style-type:decimal"><li><b>语言建模（Language Modeling）</b></li><ol class="notion-list notion-list-numbered notion-block-19637cbb8d10803693cdda9ced0d9a21" style="list-style-type:lower-alpha"><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080d8b9c4ffc4456c9f0d"><li>该步骤使用海量互联网数据训练模型，使其能够预测下一个单词。这一阶段的产物被称为<b>基础模型（Base Model）</b>。</li></ul></ol></ol><ol start="2" class="notion-list notion-list-numbered notion-block-19637cbb8d10807fb602e6f298b3f119" style="list-style-type:decimal"><li><b>监督微调（Supervised Fine-Tuning, SFT）</b></li><ol class="notion-list notion-list-numbered notion-block-19637cbb8d10807fb602e6f298b3f119" style="list-style-type:lower-alpha"><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a9af57ec351cf425bb"><li>通过<b>人工标注数据</b>进一步训练模型，使其更擅长回答问题和执行指令，从而得到<b>指令微调模型（Instruction-Tuned Model）</b>，即 SFT 模型。</li></ul></ol></ol><ol start="3" class="notion-list notion-list-numbered notion-block-19637cbb8d1080a7a015c724c88e5ca8" style="list-style-type:decimal"><li><b>偏好调整（Preference Tuning）</b></li><ol class="notion-list notion-list-numbered notion-block-19637cbb8d1080a7a015c724c88e5ca8" style="list-style-type:lower-alpha"><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080e4a548cdeca4965b18"><li>进一步优化模型，使其行为更符合人类偏好，最终形成用户在各种 AI 应用和交互界面中体验到的 LLM。</li></ul></ol></ol><hr class="notion-hr notion-block-19637cbb8d10800fbf72dc985bc57a0a"/><h3 class="notion-h notion-h2 notion-h-indent-0 notion-block-19637cbb8d108032921aef59c6bedf6d" data-id="19637cbb8d108032921aef59c6bedf6d"><span><div id="19637cbb8d108032921aef59c6bedf6d" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d108032921aef59c6bedf6d" title="DeepSeek-R1 训练方法"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>DeepSeek-R1 训练方法</b></span></span></h3><div class="notion-text notion-block-19637cbb8d1080d9b346c9bf0a82859d">DeepSeek-R1 遵循上述通用方法，但其细节有所不同。它的基础模型源自 DeepSeek-V3，但并未使用最终版本，而是基于最初的基础模型，然后经历 SFT 和偏好调整阶段。</div><div class="notion-text notion-block-19637cbb8d1080bc8383fa6a5c535f3d">在DeepSeek-R1 训练过程中有三个值得关注的关键点。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080f09687f1a910b74df9"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A3cdb935c-37fe-4e33-a94f-9ac1308d690d%3Aimage.png?table=block&amp;id=19637cbb-8d10-80f0-9687-f1a910b74df9&amp;t=19637cbb-8d10-80f0-9687-f1a910b74df9" alt="notion image" loading="lazy" decoding="async"/></div></figure><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d1080deb641ece10d4c479a" data-id="19637cbb8d1080deb641ece10d4c479a"><span><div id="19637cbb8d1080deb641ece10d4c479a" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080deb641ece10d4c479a" title="1. 长链推理 SFT 数据"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>1. 长链推理 SFT 数据</b></span></span></h4><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d108059ab69e2e6e2b1d047"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A6ff7694b-0f38-42bd-9af4-fdb7e0e1f259%3Aimage.png?table=block&amp;id=19637cbb-8d10-8059-ab69-e2e6e2b1d047&amp;t=19637cbb-8d10-8059-ab69-e2e6e2b1d047" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d108020a187eebc0d79cd3f">DeepSeek-R1 训练了 <b>60 万条</b> 复杂的“思维链”推理示例（Chain-of-Thought, CoT），这些数据非常稀缺，并且人工标注成本极高。因此，如何创建这些数据成为一个关键问题。</div><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d10802fadb6cae7e6cfadf4" data-id="19637cbb8d10802fadb6cae7e6cfadf4"><span><div id="19637cbb8d10802fadb6cae7e6cfadf4" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10802fadb6cae7e6cfadf4" title="2. 一个高质量的推理模型（但在非推理任务上表现较差）"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>2. 一个高质量的推理模型（但在非推理任务上表现较差）</b></span></span></h4><div class="notion-text notion-block-19637cbb8d10807da180e7a65f0e6109">DeepSeek-R1 训练数据部分来自一个“前置推理模型”（Interim Reasoning Model），它是 DeepSeek-R1 的一个兄弟版本，专门用于推理任务。</div><div class="notion-blank notion-block-19637cbb8d10804aafc1c59ec17d912b"> </div><div class="notion-text notion-block-19637cbb8d108038ad6bd246885023bb">这个模型的灵感来自 <b>R1-Zero</b>（将在后面讨论）。虽然这个模型不适合作为通用 LLM，但它在极少人工标注数据的情况下，通过大规模强化学习，实现了卓越的推理能力。</div><div class="notion-blank notion-block-19637cbb8d1080048cace32a0a3fc53d"> </div><div class="notion-text notion-block-19637cbb8d1080cfbec8ff04c57d6ac2">然后，我们可以用这个推理专家模型的输出，训练一个更加通用的 LLM，使其既能处理推理任务，也能胜任非推理任务，以满足用户需求。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080c59bbae62f18c3fa52"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A2c9b7421-fc26-45a1-b696-4e2d0736f3cd%3Aimage.png?table=block&amp;id=19637cbb-8d10-80c5-9bba-e62f18c3fa52&amp;t=19637cbb-8d10-80c5-9bba-e62f18c3fa52" alt="notion image" loading="lazy" decoding="async"/></div></figure><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d10808ebb33fac8bb811a75" data-id="19637cbb8d10808ebb33fac8bb811a75"><span><div id="19637cbb8d10808ebb33fac8bb811a75" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10808ebb33fac8bb811a75" title="3. 通过大规模强化学习（RL）打造推理模型"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>3. 通过大规模强化学习（RL）打造推理模型</b></span></span></h4><div class="notion-text notion-block-19637cbb8d10807a824ffe2f1315ae96">这一过程包括两个关键步骤：</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080e1856af9de2b5c7c8c"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A4201c00b-c615-4992-b611-e48406b3fa0d%3Aimage.png?table=block&amp;id=19637cbb-8d10-80e1-856a-f9de2b5c7c8c&amp;t=19637cbb-8d10-80e1-856a-f9de2b5c7c8c" alt="notion image" loading="lazy" decoding="async"/></div></figure><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d1080d38211fd1fcbbad06d" data-id="19637cbb8d1080d38211fd1fcbbad06d"><span><div id="19637cbb8d1080d38211fd1fcbbad06d" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080d38211fd1fcbbad06d" title="3.1 R1-Zero：大规模推理导向的强化学习"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>3.1 R1-Zero：大规模推理导向的强化学习</b></span></span></h4><div class="notion-text notion-block-19637cbb8d10806b8ed7e75616448375">R1-Zero 是一个特殊的推理模型，它的独特之处在于：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d108068a173cc32a1bea450"><li><b>它没有使用 SFT 训练数据</b>，而是直接从预训练的基础模型出发，通过强化学习（RL）训练。</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a48ce6d99dfdf295b0"><li>它的推理能力足以媲美 OpenAI O1 级别的模型。</li></ul><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d10804ea472da43670ed79d"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Aca684072-c4f3-49a7-b7ac-65402835e000%3Aimage.png?table=block&amp;id=19637cbb-8d10-804e-a472-da43670ed79d&amp;t=19637cbb-8d10-804e-a472-da43670ed79d" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d1080e3851ee56fe25542fb">这表明，现代基础模型已经达到了一定的质量和能力门槛（例如，DeepSeek-R1 的基础模型训练使用了 <b>14.8 万亿个高质量 token</b>）。此外，相比于通用聊天任务或写作请求，推理问题可以<b>自动验证</b>，从而减少对人工标注数据的依赖。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080358a55d45031cea723"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A6eccdf6d-8af7-4383-b1af-ad8fda8e08b0%3Aimage.png?table=block&amp;id=19637cbb-8d10-8035-8a55-d45031cea723&amp;t=19637cbb-8d10-8035-8a55-d45031cea723" alt="notion image" loading="lazy" decoding="async"/></div></figure><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d10800babaaf49d1e9ebb8c" data-id="19637cbb8d10800babaaf49d1e9ebb8c"><span><div id="19637cbb8d10800babaaf49d1e9ebb8c" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10800babaaf49d1e9ebb8c" title="示例：自动验证推理任务"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>示例：自动验证推理任务</b></span></span></h4><div class="notion-text notion-block-19637cbb8d1080e49adcf46fbf09bd8b">假设训练任务是：</div><blockquote class="notion-quote notion-block-19637cbb8d1080e08c5cf55d73cca9ee"><div>编写一个 Python 代码，输入一个数字列表，返回排序后的结果，并在开头添加数字 42。</div></blockquote><div class="notion-text notion-block-19637cbb8d1080b6abb6cf6f72fbaaf8">这种问题可以用多种方法自动验证，例如：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d10808198f7f50478b17276"><li><b>语法检查</b>：代码是否符合 Python 语法？</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080d7b5bcefdc87d4fe61"><li><b>可执行性检查</b>：代码是否能正确运行？</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d10808ead61e88e41a0e1b4"><li><b>单元测试</b>：自动化测试是否能验证代码行为？</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080c7bf29d4c7ebab5c4b"><li><b>性能优化</b>：是否能生成运行速度更快的代码？</li></ul><div class="notion-blank notion-block-19637cbb8d10809a9e22f24a389b56fa"> </div><div class="notion-text notion-block-19637cbb8d108090a721d209a7fa4d59">我们可以在训练步骤中向模型提出这样的问题，并生成多种可能的解决方案。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080e5adcec54ecb9d5fa1"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Aab74320d-e7d5-4f93-a543-224708be3092%3Aimage.png?table=block&amp;id=19637cbb-8d10-80e5-adce-c54ecb9d5fa1&amp;t=19637cbb-8d10-80e5-adce-c54ecb9d5fa1" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d1080fab3fbd345a7c7afbd">我们可以自动检查（无需人工干预），发现第一个完成甚至不是代码。第二个是代码，但不是 python 代码。第三个是可能的解决方案，但无法通过单元测试，而第四个是正确的解决方案。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080a48032cac5bb8ae0ed"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Af01d4ff2-69f3-419d-ac56-3ef3bd04dc47%3Aimage.png?table=block&amp;id=19637cbb-8d10-80a4-8032-cac5bb8ae0ed&amp;t=19637cbb-8d10-80a4-8032-cac5bb8ae0ed" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d10801f82e9d52471aa5dda">这些自动化检查信号可以用于强化学习，使模型在不断优化中提高推理能力。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d10802ea5c1c73d9ae28c4d"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Aa7964afa-53bb-4651-b3aa-8a44e1f532e1%3Aimage.png?table=block&amp;id=19637cbb-8d10-802e-a5c1-c73d9ae28c4d&amp;t=19637cbb-8d10-802e-a5c1-c73d9ae28c4d" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d10806892a5f27d47a9bb01">正如本文图 2 所示，这些奖励信号和模型更新是模型在 RL 训练过程中不断改进任务的方式。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080c1bab3d8958a87711b"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A796b3d58-5c58-41e9-ba1e-beed9a335d35%3Aimage.png?table=block&amp;id=19637cbb-8d10-80c1-bab3-d8958a87711b&amp;t=19637cbb-8d10-80c1-bab3-d8958a87711b" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d10808b960afc57314976ef">与此功能的提高相对应的是生成响应的长度，即模型生成更多的思考标记来处理问题。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d10804c9c07cc791922b5be"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Ab01f726a-3f83-45d4-92f2-a07bafd3f6bf%3Aimage.png?table=block&amp;id=19637cbb-8d10-804c-9c07-cc791922b5be&amp;t=19637cbb-8d10-804c-9c07-cc791922b5be" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d1080019c6edc045008752c">这一过程非常有用，但 R1-Zero 模型尽管在这些推理问题上得分很高，却面临着其他问题，使其可用性低于预期。</div><blockquote class="notion-quote notion-block-19637cbb8d1080bd8520f35f89475161"><div>尽管DeepSeek-R1-Zero表现出强大的推理能力，并能自主开发出意想不到的强大推理行为，但它也面临着一些问题。例如，DeepSeek-R1-Zero面临着可读性差和语言混杂等难题。</div></blockquote><div class="notion-text notion-block-19637cbb8d10807a9e2ae70033a70a74">R1 的目的是成为一个更实用的模型。因此，它并不完全依赖于 RL 过程，而是在本节前面提到的两个地方使用：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d10803db2f0cd3ec4fe92cd"><li>1 创建一个临时推理模型，以生成 SFT 数据点</li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108071b571f564b75a4652"><li>2 训练 R1 模型以改善推理和非推理问题（使用其他类型的验证器）</li></ul><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d10807097a2f0ac46f5602d"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A2baa7e1b-39b1-41dd-9264-a19d782ec13c%3Aimage.png?table=block&amp;id=19637cbb-8d10-8070-97a2-f0ac46f5602d&amp;t=19637cbb-8d10-8070-97a2-f0ac46f5602d" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-blank notion-block-19637cbb8d1080e592e6c381a647b229"> </div><hr class="notion-hr notion-block-19637cbb8d1080fda17efe17896243b3"/><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d1080af97d1d5f6a3c97a9c" data-id="19637cbb8d1080af97d1d5f6a3c97a9c"><span><div id="19637cbb8d1080af97d1d5f6a3c97a9c" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080af97d1d5f6a3c97a9c" title="3.2 用中间推理模型生成 SFT 数据"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>3.2 用中间推理模型生成 SFT 数据</b></span></span></h4><div class="notion-text notion-block-19637cbb8d1080e992fbdd2c4a1f42f9">为了使推理模型更加稳定，DeepSeek-R1 先进行少量 SFT 训练（大约 5000 条数据），然后利用该模型合成 <b>60 万条</b> 训练数据。</div><blockquote class="notion-quote notion-block-19637cbb8d108013a19aed785f799181"><div><b>2.3.1. 冷启动</b>
与DeepSeek-R1-Zero不同，为了防止从基础模型开始的RL训练出现早期不稳定的冷启动阶段，我们为DeepSeek-R1构建并收集了少量长CoT数据，以微调作为初始RL行为者的模型。

为了收集这些数据，我们探索了几种方法：以长 CoT 为例，使用少量提示；直接提示模型生成带有反思和验证的详细答案；以可读格式收集 DeepSeek-R1- Zero 输出；通过人工注释者的后期处理完善结果。</div></blockquote><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080c7b4e5c04b0806f225"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Af249a30a-2ceb-4463-ac4d-396fc3f7f952%3Aimage.png?table=block&amp;id=19637cbb-8d10-80c7-b4e5-c04b0806f225&amp;t=19637cbb-8d10-80c7-b4e5-c04b0806f225" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d1080868cfff097c6d8959a">这一步被称为 <b>冷启动（Cold Start）</b>。初始数据来源包括：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d108059b184fc839d84b7ec"><li><b>使用少样本提示（few-shot prompting）生成推理示例</b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080b880c5edef9177cb30"><li><b>让模型生成详细答案并进行自我验证</b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a391f7eeb6d04b70f3"><li><b>提取 DeepSeek-R1-Zero 生成的内容，并进行人工后处理</b></li></ul><div class="notion-text notion-block-19637cbb8d10804fbe44f5a41bc32286">但是，等等，既然我们掌握了这些数据，为什么还要依赖 RL 流程呢？</div><div class="notion-text notion-block-19637cbb8d10802b9843d8777b979a36">这是因为数据的规模。尽管 5000 条 SFT 数据可手动标注，但 60 万条数据的规模则需要模型自动生成。
这个临时模型弥补了这一差距，并能综合生成极其宝贵的数据。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080a1a4c3d8218d36ebd4"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A3253240a-b4a0-4fe7-b8bd-5310fb972bae%3Aimage.png?table=block&amp;id=19637cbb-8d10-80a1-a4c3-d8218d36ebd4&amp;t=19637cbb-8d10-80a1-a4c3-d8218d36ebd4" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d10809da42ef5cf80e47df5">如果你对 “监督微调”（Supervised Fine-Tuning，SFT）的概念还不太了解，那么它就是以提示和正确完成的形式向模型提供训练示例的过程。</div><div class="notion-text notion-block-19637cbb8d1080118595dd3e211d13f5">这张图展示了几个 SFT 训练示例：</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d108033a8b0fa8fe7791302"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A9fb64f84-ca6f-443b-a25e-0087229afbb2%3Aimage.png?table=block&amp;id=19637cbb-8d10-8033-a8b0-fa8fe7791302&amp;t=19637cbb-8d10-8033-a8b0-fa8fe7791302" alt="notion image" loading="lazy" decoding="async"/></div></figure><hr class="notion-hr notion-block-19637cbb8d10807a908cfe064b9e9fba"/><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d1080fab021dee3fd3ae3f4" data-id="19637cbb8d1080fab021dee3fd3ae3f4"><span><div id="19637cbb8d1080fab021dee3fd3ae3f4" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d1080fab021dee3fd3ae3f4" title="3.3 通用 RL 训练阶段"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>3.3 通用 RL 训练阶段</b></span></span></h4><div class="notion-text notion-block-19637cbb8d10809bb7e1f96301506644">最终，DeepSeek-R1 进入通用强化学习训练，使其不仅能处理推理任务，还能胜任其他非推理任务。例如：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d108083b531e6fbda186ffc"><li><b>帮助性（Helpfulness）奖励模型</b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d108042b4e4efd111267389"><li><b>安全性（Safety）奖励模型</b></li></ul><div class="notion-text notion-block-19637cbb8d108050918bce68b1190fc1">这种方法类似于 Llama 模型的 RL 训练策略，使 DeepSeek-R1 既能推理，也能胜任更广泛的任务。</div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d108001bac9dde2facba2be"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Abd84ea0b-7a6f-4b47-944a-0216f2084198%3Aimage.png?table=block&amp;id=19637cbb-8d10-8001-bac9-dde2facba2be&amp;t=19637cbb-8d10-8001-bac9-dde2facba2be" alt="notion image" loading="lazy" decoding="async"/></div></figure><hr class="notion-hr notion-block-19637cbb8d108097a896e0245762470a"/><h3 class="notion-h notion-h2 notion-h-indent-0 notion-block-19637cbb8d10805d8168fc1945203f36" data-id="19637cbb8d10805d8168fc1945203f36"><span><div id="19637cbb8d10805d8168fc1945203f36" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10805d8168fc1945203f36" title="DeepSeek-R1 的模型架构"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>DeepSeek-R1 的模型架构</b></span></span></h3><div class="notion-text notion-block-19637cbb8d108013963df202c6864fae">DeepSeek-R1 的底层结构依然是 <b>Transformer 解码器（Decoder-Only Transformer）</b>，这一点与 GPT-2、GPT-3 以及其他 LLM 相似，由 <b>61 层 Transformer 解码块</b> 组成。其中：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d10804f8ce8dbae6b2d7deb"><li><b>前三层是密集层（Dense Layers），即传统的 Transformer 层。</b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080338707dcb4597d7382"><li><b>其余部分采用专家混合（MoE）层</b>，这意味着并非所有参数在每次计算时都被激活，而是仅启用特定的专家网络，以提高计算效率。</li></ul><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080c09e42c8e9a529be27"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3A17db5f9c-1bcb-4531-88c2-2100af694909%3Aimage.png?table=block&amp;id=19637cbb-8d10-80c0-9e42-c8e9a529be27&amp;t=19637cbb-8d10-80c0-9e42-c8e9a529be27" alt="notion image" loading="lazy" decoding="async"/></div></figure><h4 class="notion-h notion-h3 notion-h-indent-1 notion-block-19637cbb8d10806faed7eeae23e1fd38" data-id="19637cbb8d10806faed7eeae23e1fd38"><span><div id="19637cbb8d10806faed7eeae23e1fd38" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d10806faed7eeae23e1fd38" title="专家混合（MoE）结构的优势"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>专家混合（MoE）结构的优势</b></span></span></h4><div class="notion-text notion-block-19637cbb8d1080df9541c6789594d6a2">这种 MoE 结构允许模型在不同任务上动态调度计算资源，提高效率。</div><ol start="1" class="notion-list notion-list-numbered notion-block-19637cbb8d1080279f64f770911a1707" style="list-style-type:decimal"><li><b>计算资源优化</b>：</li><ol class="notion-list notion-list-numbered notion-block-19637cbb8d1080279f64f770911a1707" style="list-style-type:lower-alpha"><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080d8903ae7e208513c4d"><li>传统 Transformer 每一层都会激活所有参数，而 MoE 结构<b>只激活部分专家层</b>，从而降低计算开销，提高推理效率。</li></ul></ol></ol><ol start="2" class="notion-list notion-list-numbered notion-block-19637cbb8d1080138fcbd074c35ee289" style="list-style-type:decimal"><li><b>模型能力增强</b>：</li><ol class="notion-list notion-list-numbered notion-block-19637cbb8d1080138fcbd074c35ee289" style="list-style-type:lower-alpha"><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a0ae39de81d5d90dc4"><li>通过 MoE，模型可以在不同任务上分配不同的专家，使其在<b>推理、对话、代码生成等任务上表现更优</b>。</li></ul></ol></ol><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1080ae91d1ca0b8b712d5e"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/attachment%3Abe319a5b-4271-4ae2-addd-825f13cc1ea6%3Aimage.png?table=block&amp;id=19637cbb-8d10-80ae-91d1-ca0b8b712d5e&amp;t=19637cbb-8d10-80ae-91d1-ca0b8b712d5e" alt="notion image" loading="lazy" decoding="async"/></div></figure><div class="notion-text notion-block-19637cbb8d10801ea529ddcaaaf79821">更多技术细节可参考：</div><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080dd8fb2eaf838742748"><li><b>DeepSeek-V3 技术报告 </b><b><a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://arxiv.org/pdf/2412.19437v1">[查看报告]</a></b></li></ul><ul class="notion-list notion-list-disc notion-block-19637cbb8d1080a1a3b9cc31bfe4616a"><li><b>DeepSeekMoE：专家混合语言模型的终极优化 [</b><b><a target="_blank" rel="noopener noreferrer" class="notion-link" href="https://arxiv.org/pdf/2401.06066">查看报告</a></b><b>]</b></li></ul><hr class="notion-hr notion-block-19637cbb8d10804199f9d325ce7fa703"/><h3 class="notion-h notion-h2 notion-h-indent-0 notion-block-19637cbb8d108037941beb7bebf833da" data-id="19637cbb8d108037941beb7bebf833da"><span><div id="19637cbb8d108037941beb7bebf833da" class="notion-header-anchor"></div><a class="notion-hash-link" href="#19637cbb8d108037941beb7bebf833da" title="总结"><svg viewBox="0 0 16 16" width="16" height="16"><path fill-rule="evenodd" d="M7.775 3.275a.75.75 0 001.06 1.06l1.25-1.25a2 2 0 112.83 2.83l-2.5 2.5a2 2 0 01-2.83 0 .75.75 0 00-1.06 1.06 3.5 3.5 0 004.95 0l2.5-2.5a3.5 3.5 0 00-4.95-4.95l-1.25 1.25zm-4.69 9.64a2 2 0 010-2.83l2.5-2.5a2 2 0 012.83 0 .75.75 0 001.06-1.06 3.5 3.5 0 00-4.95 0l-2.5 2.5a3.5 3.5 0 004.95 4.95l1.25-1.25a.75.75 0 00-1.06-1.06l-1.25 1.25a2 2 0 01-2.83 0z"></path></svg></a><span class="notion-h-title"><b>总结</b></span></span></h3><div class="notion-text notion-block-19637cbb8d10804caa7cf1591972d65d">通过本篇文章，您应该对 DeepSeek-R1 的核心概念有所了解。</div><div class="notion-text notion-block-19637cbb8d108073af2cf2bbb4c3faca">DeepSeek-R1 代表了推理型大模型训练的新范式，它通过强化学习、大规模推理数据合成以及专家混合架构，实现了卓越的推理能力和通用性。</div><div class="notion-blank notion-block-19637cbb8d10815b86cafd61a5d7c644"> </div><div class="notion-callout notion-gray_background_co notion-block-19637cbb8d10818499a5c6b5999518b1"><div class="notion-page-icon-inline notion-page-icon-span"><span class="notion-page-icon" role="img" aria-label="💡">💡</span></div><div class="notion-callout-text"><b>对这个话题感兴趣的小伙伴，欢迎加我一起探索交流~  </b></div></div><div class="notion-row notion-block-19637cbb8d1081c89347d0b7c8559fdd"><div class="notion-column notion-block-19637cbb8d1081eaabeaf1e76c7f52a1" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19637cbb8d1081e7bb2ecb6a084634ab"> </div><figure class="notion-asset-wrapper notion-asset-wrapper-image notion-block-19637cbb8d1081da88b5ed99e86dfdaa"><div style="position:relative;display:flex;justify-content:center;align-self:center;width:100%;max-width:100%;flex-direction:column;height:100%"><img style="object-fit:cover" src="https://www.notion.so/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2F2666d03a-fb22-44f0-83a0-96826c4e3d2d%2F154b5cdb-df44-4c08-9984-33dbebdc5057%2Flittlerocketlu.jpg?table=block&amp;id=19637cbb-8d10-81da-88b5-ed99e86dfdaa&amp;t=19637cbb-8d10-81da-88b5-ed99e86dfdaa&amp;width=540.9921875&amp;cache=v2" alt="notion image" loading="lazy" decoding="async"/></div></figure></div><div class="notion-spacer"></div><div class="notion-column notion-block-19637cbb8d1081f7b246f13169954715" style="width:calc((100% - (1 * min(32px, 4vw))) * 1)"><div class="notion-blank notion-block-19637cbb8d108152a832cb866858e151"> </div></div><div class="notion-spacer"></div></div><div class="notion-blank notion-block-19637cbb8d1081e0b0e7fbe12984e990"> </div></main></div>]]></content:encoded>
        </item>
    </channel>
</rss>