{"id":362,"date":"2026-04-28T07:29:14","date_gmt":"2026-04-27T23:29:14","guid":{"rendered":"http:\/\/www.faiyi.com\/?p=362"},"modified":"2026-04-28T07:29:14","modified_gmt":"2026-04-27T23:29:14","slug":"ai%e5%8a%a8%e6%80%81%e6%af%8f%e6%97%a5%e7%ae%80%e6%8a%a5-2026-04-28","status":"publish","type":"post","link":"http:\/\/www.faiyi.com\/?p=362","title":{"rendered":"AI\u52a8\u6001\u6bcf\u65e5\u7b80\u62a5 2026-04-28"},"content":{"rendered":"<p>\u65e5\u671f\uff1a2026-04-28<\/p>\n<p>\u672c\u671f\u805a\u7126\uff1a\u91cd\u70b9\u5173\u6ce8\u6a21\u578b\u53d1\u5e03\u4e0e release notes\u3001\u5b98\u65b9 engineering blog\u3001AI coding \/ agent \/ SRE\u3001\u8bc4\u6d4b\u699c\u5355\u53d8\u5316\u3001\u5f00\u53d1\u8005\u5b9e\u8df5\u535a\u5ba2\u3001\u6846\u67b6\u751f\u6001\u3001\u5f00\u6e90\u6a21\u578b\u4e0e\u771f\u5b9e\u7528\u6237\u89c6\u89d2\uff1b\u5f53 HN\u3001Reddit\u3001Hugging Face \u7b49\u793e\u533a\u6e90\u53ef\u8bbf\u95ee\u65f6\u4f18\u5148\u7eb3\u5165\u3002<\/p>\n<hr \/>\n<ol>\n<li>\n<p><strong>Artificial Analysis \u6700\u65b0\u6a21\u578b\u6392\u540d\u89c2\u5bdf<\/strong>\uff08Artificial Analysis\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>Artificial Analysis \u6700\u65b0\u6a21\u578b\u6392\u540d\u663e\u793a\uff0cGPT-5.5 (xhigh) \u4ee5 60 \u5206\u4f4d\u5c45\u667a\u80fd\u6307\u6570\u699c\u9996\uff0cGPT-5.5 (high) \u4ee5 59 \u5206\u7d27\u968f\u5176\u540e\uff0cClaude Opus 4.7 (Max Effort) \u4e0e Gemini 3.1 Pro Preview \u5e76\u5217\u7b2c\u4e09\uff0857 \u5206\uff09\u3002\u901f\u5ea6\u65b9\u9762\uff0cMercury 2 \u4ee5 687 tokens\/\u79d2\u9886\u5148\uff0cGranite 3.3 8B \u8fbe 333 tokens\/\u79d2\u3002\u5ef6\u8fdf\u6700\u4f4e\u7684\u662f Ministral 3 3B\uff080.45 \u79d2\uff09\u3002\u8be5\u5e73\u53f0\u7684 Intelligence Index v4.0 \u6db5\u76d6 GDPval-AA\u3001Terminal-Bench Hard\u3001Humanity&#039;s Last Exam\u3001GPQA Diamond \u7b49 10 \u9879\u8bc4\u6d4b\uff0c\u4e3a\u5f00\u53d1\u8005\u63d0\u4f9b\u6a21\u578b\u9009\u578b\u53c2\u8003\u3002<\/p>\n<p><strong>English Summary:<\/strong> Artificial Analysis&#039; latest rankings show GPT-5.5 (xhigh) leading the Intelligence Index at 60 points, followed by GPT-5.5 (high) at 59, with Claude Opus 4.7 (Max Effort) and Gemini 3.1 Pro Preview tied at 57. For speed, Mercury 2 leads at 687 tokens\/s, while Ministral 3 3B has the lowest latency at 0.45s. The Intelligence Index v4.0 covers 10 benchmarks including GDPval-AA, Terminal-Bench Hard, and Humanity&#039;s Last Exam.<\/p>\n<p><a href=\"https:\/\/artificialanalysis.ai\/models\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Introducing Claude Opus 4.7<\/strong>\uff08Anthropic News\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>Anthropic \u53d1\u5e03 Claude Opus 4.7\uff0c\u5728\u591a\u6b65\u9aa4\u4efb\u52a1\u6548\u7387\u4e0a\u521b\u4e0b\u5185\u90e8\u7814\u7a76\u667a\u80fd\u4f53\u57fa\u51c6\u65b0\u9ad8\uff0c\u516d\u9879\u6a21\u5757\u603b\u5206\u8fbe 0.715\uff0c\u957f\u4e0a\u4e0b\u6587\u8868\u73b0\u6700\u4e3a\u7a33\u5b9a\u3002\u5728 General Finance \u6a21\u5757\u5f97\u5206\u4ece Opus 4.6 \u7684 0.767 \u63d0\u5347\u81f3 0.813\uff0c\u6570\u636e\u62ab\u9732\u4e0e\u7eaa\u5f8b\u6027\u8868\u73b0\u6700\u4f73\u3002Quantium \u8bc4\u4f30\u663e\u793a\u5176\u5728\u63a8\u7406\u6df1\u5ea6\u3001\u7ed3\u6784\u5316\u95ee\u9898\u6846\u67b6\u548c\u590d\u6742\u6280\u672f\u5de5\u4f5c\u65b9\u9762\u8fdb\u6b65\u663e\u8457\uff1bDatabricks \u7684 OfficeQA Pro \u6d4b\u8bd5\u8868\u660e\u6587\u6863\u63a8\u7406\u9519\u8bef\u51cf\u5c11 21%\uff1bRamp \u53cd\u9988\u79f0\u5176\u5728\u4ee3\u7406\u56e2\u961f\u5de5\u4f5c\u6d41\u4e2d\u89d2\u8272\u5fe0\u5b9e\u5ea6\u3001\u6307\u4ee4\u9075\u5faa\u548c\u8de8\u5de5\u5177\u8c03\u8bd5\u80fd\u529b\u66f4\u5f3a\uff0c\u6240\u9700\u9010\u6b65\u6307\u5bfc\u5927\u5e45\u51cf\u5c11\u3002<\/p>\n<p><strong>English Summary:<\/strong> Anthropic released Claude Opus 4.7, achieving a new high on internal research-agent benchmarks with 0.715 across six modules and the most consistent long-context performance. General Finance scores improved from 0.767 to 0.813. Quantium noted gains in reasoning depth and structured problem-framing; Databricks reported 21% fewer document reasoning errors on OfficeQA Pro; Ramp highlighted stronger role fidelity and reduced need for step-by-step guidance in agent-team workflows.<\/p>\n<p><a href=\"https:\/\/www.anthropic.com\/news\/claude-opus-4-7\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Featured An update on recent Claude Code quality reports<\/strong>\uff08Anthropic Engineering\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>Anthropic \u5de5\u7a0b\u56e2\u961f\u53d1\u5e03 Claude Code \u8d28\u91cf\u62a5\u544a\uff0c\u590d\u76d8\u4e86 4 \u6708 16 \u65e5 Opus 4.7 \u53d1\u5e03\u540e\u51fa\u73b0\u7684\u95ee\u9898\u3002\u56e2\u961f\u4e3a\u964d\u4f4e\u6a21\u578b\u5197\u957f\u5ea6\uff0c\u5728\u7cfb\u7edf\u63d0\u793a\u4e2d\u52a0\u5165\u4e86\u5b57\u6570\u9650\u5236\uff08\u5de5\u5177\u8c03\u7528\u95f4 \u226425 \u8bcd\uff0c\u6700\u7ec8\u56de\u590d \u2264100 \u8bcd\uff09\uff0c\u5bfc\u81f4\u667a\u80fd\u8868\u73b0\u610f\u5916\u4e0b\u964d\u3002\u53e6\u4e00\u7f13\u5b58\u4f18\u5316\u8bef\u5220\u4e86\u5386\u53f2\u63a8\u7406\u5185\u5bb9\uff0c\u5f71\u54cd\u4ee3\u7801\u5ba1\u67e5\u8d28\u91cf\u3002\u95ee\u9898\u5df2\u5728 4 \u6708 10 \u65e5\u7684 v2.1.101 \u7248\u672c\u4e2d\u4fee\u590d\u3002\u56e2\u961f\u73b0\u5df2\u5c06 Opus 4.7 \u9ed8\u8ba4 effort \u6a21\u5f0f\u6062\u590d\u4e3a xhigh\uff0c\u5e76\u8ba1\u5212\u4e3a\u4ee3\u7801\u5ba1\u67e5\u589e\u52a0\u989d\u5916\u4ed3\u5e93\u4e0a\u4e0b\u6587\u652f\u6301\uff0c\u4ee5\u9632\u6b62\u7c7b\u4f3c\u95ee\u9898\u518d\u6b21\u53d1\u751f\u3002<\/p>\n<p><strong>English Summary:<\/strong> Anthropic Engineering published a postmortem on Claude Code quality issues following the April 16 Opus 4.7 launch. A system prompt change adding word limits (\u226425 words between tool calls, \u2264100 for final responses) unexpectedly reduced intelligence, while a caching optimization dropped prior reasoning from context, degrading code review quality. Both issues were fixed in v2.1.101 on April 10. The team reverted Opus 4.7 defaults to xhigh effort and plans to add multi-repository context for code reviews.<\/p>\n<p><a href=\"https:\/\/www.anthropic.com\/engineering\/april-23-postmortem\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Scaling Managed Agents: Decoupling the brain from the hands<\/strong>\uff08Anthropic Engineering\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>Anthropic \u5de5\u7a0b\u535a\u5ba2\u53d1\u6587\u9610\u8ff0 Managed Agents \u67b6\u6784\u7406\u5ff5\uff0c\u63d0\u51fa\u5c06&quot;\u5927\u8111&quot;\uff08Claude \u667a\u80fd\uff09\u4e0e&quot;\u53cc\u624b&quot;\uff08\u5177\u4f53\u5de5\u5177\u6267\u884c\uff09\u89e3\u8026\u3002\u8be5\u5143\u6846\u67b6\u4e0d\u9884\u8bbe\u7279\u5b9a\u5de5\u5177\u94fe\uff0c\u800c\u662f\u63d0\u4f9b\u901a\u7528\u63a5\u53e3\u652f\u6301\u591a\u79cd\u4ee3\u7406\u6846\u67b6\uff08\u5982 Claude Code \u6216\u9886\u57df\u4e13\u7528\u4ee3\u7406\uff09\u3002\u6838\u5fc3\u521b\u65b0\u5728\u4e8e\u4f1a\u8bdd\u65e5\u5fd7\u4f5c\u4e3a\u6301\u4e45\u5316\u4e0a\u4e0b\u6587\u5b58\u50a8\uff0c\u901a\u8fc7 getEvents() \u63a5\u53e3\u5141\u8bb8\u6a21\u578b\u6309\u4f4d\u7f6e\u5207\u7247\u68c0\u7d22\u4e8b\u4ef6\u6d41\uff0c\u5b9e\u73b0\u7075\u6d3b\u7684\u72b6\u6001\u6062\u590d\u3001\u56de\u6eaf\u548c\u4e0a\u4e0b\u6587\u91cd\u8bfb\u3002\u8fd9\u79cd\u8bbe\u8ba1\u8ba9\u4ee3\u7406\u80fd\u968f\u6a21\u578b\u80fd\u529b\u63d0\u5347\u800c\u8fdb\u5316\uff0c\u65e0\u9700\u9891\u7e41\u91cd\u5199\u5de5\u5177\u94fe\u5047\u8bbe\u3002<\/p>\n<p><strong>English Summary:<\/strong> Anthropic Engineering introduced Managed Agents, an architecture decoupling the &quot;brain&quot; (Claude&#039;s intelligence) from the &quot;hands&quot; (tool execution). This meta-harness provides general interfaces accommodating various agent frameworks without encoding brittle assumptions. The key innovation is durable session logs serving as persistent context storage outside Claude&#039;s window, with a getEvents() interface enabling flexible event stream slicing for state recovery, rewinding, and context re-reading as models improve.<\/p>\n<p><a href=\"https:\/\/www.anthropic.com\/engineering\/managed-agents\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Physical AI that Moves the World \u2014 Qasar Younis &amp; Peter Ludwig, Applied Intuition<\/strong>\uff08Latent Space\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>Latent Space \u64ad\u5ba2\u8bbf\u8c08 Applied Intuition CEO Qasar Younis \u4e0e CTO Peter Ludwig\uff0c\u63a2\u8ba8\u7269\u7406 AI \u5728\u91c7\u77ff\u3001\u65e0\u4eba\u673a\u3001\u5361\u8f66\u3001\u6218\u8230\u7b49\u6781\u7aef\u73af\u5883\u4e2d\u7684\u5e94\u7528\u3002\u8be5\u516c\u53f8\u4ece YC \u65f6\u671f\u7684\u81ea\u52a8\u9a7e\u9a76\u4eff\u771f\u5de5\u5177\u8d77\u6b65\uff0c\u53d1\u5c55\u4e3a\u4f30\u503c 150 \u4ebf\u7f8e\u5143\u7684\u7269\u7406 AI \u5e73\u53f0\uff0c\u81f4\u529b\u4e8e\u6210\u4e3a&quot;\u673a\u5668\u754c\u7684 Android&quot;\u3002\u8bbf\u8c08\u6db5\u76d6\uff1a\u8f66\u8f86\u8f6f\u4ef6\u6808\u788e\u7247\u5316\u7684\u6574\u5408\u613f\u666f\u3001Cursor \u4e0e Claude Code \u7b49 AI \u7f16\u7a0b\u5de5\u5177\u5728\u5d4c\u5165\u5f0f\u548c\u5b89\u5168\u5173\u952e\u8f6f\u4ef6\u4e2d\u7684\u91c7\u7528\u3001\u7aef\u5230\u7aef\u81ea\u52a8\u9a7e\u9a76\u5bf9\u4eff\u771f\u9a8c\u8bc1\u7684\u65b0\u8981\u6c42\uff0c\u4ee5\u53ca\u795e\u7ecf\u4eff\u771f\u9700\u8db3\u591f\u5feb\u548c\u4fbf\u5b9c\u4ee5\u652f\u6491\u5f3a\u5316\u5b66\u4e60\u8bad\u7ec3\u7b49\u8bdd\u9898\u3002<\/p>\n<p><strong>English Summary:<\/strong> Latent Space interviewed Applied Intuition CEO Qasar Younis and CTO Peter Ludwig on physical AI powering mining rigs, drones, trucks, and warships. The $15B company evolved from YC-era autonomy tooling toward becoming &quot;Android for machines,&quot; consolidating fragmented vehicle software stacks. Topics include AI coding tools (Cursor, Claude Code) adoption in embedded systems, how end-to-end autonomy changes simulation requirements, and why neural simulation must be fast and cheap enough to make RL practical for physical AI.<\/p>\n<p><a href=\"https:\/\/www.latent.space\/p\/appliedintuition\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Automate repetitive tasks with Amazon Quick Flows<\/strong>\uff08AWS ML Blog\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>AWS \u53d1\u5e03 Amazon Quick Flows\uff0c\u4e00\u6b3e\u9762\u5411\u975e\u6280\u672f\u7528\u6237\u7684 AI \u5de5\u4f5c\u6d41\u81ea\u52a8\u5316\u5de5\u5177\u3002\u7528\u6237\u53ef\u901a\u8fc7\u81ea\u7136\u8bed\u8a00\u63cf\u8ff0\u4efb\u52a1\u9700\u6c42\uff0c\u7cfb\u7edf\u81ea\u52a8\u751f\u6210\u5305\u542b\u6570\u636e\u6536\u96c6\u3001AI \u5206\u6790\u3001\u5916\u90e8\u7cfb\u7edf\u96c6\u6210\u7684\u5b8c\u6574\u5de5\u4f5c\u6d41\u3002\u6587\u7ae0\u4ee5\u8d22\u52a1\u5206\u6790\u5de5\u5177\u548c\u5458\u5de5\u5165\u804c\u81ea\u52a8\u5316\u4e3a\u4f8b\uff0c\u6f14\u793a\u4e86\u4ece\u6587\u672c\u8f93\u5165\u3001\u7f51\u7edc\u641c\u7d22\u3001\u6570\u636e\u6d1e\u5bdf\u5230\u90ae\u4ef6\/Slack \u901a\u77e5\u7684\u5168\u6d41\u7a0b\u6784\u5efa\u3002Quick Flows \u63d0\u4f9b\u53ef\u89c6\u5316\u7f16\u8f91\u5668\uff0c\u652f\u6301\u6761\u4ef6\u903b\u8f91\u3001\u5faa\u73af\u3001\u53d8\u91cf\u4f20\u9012\u7b49\u9ad8\u7ea7\u529f\u80fd\uff0c\u5e76\u53ef\u8fde\u63a5 SharePoint\u3001S3\u3001HR\/IT \u7cfb\u7edf\u7b49\u5916\u90e8\u670d\u52a1\uff0c\u65e0\u9700\u7f16\u5199\u4ee3\u7801\u5373\u53ef\u5b9e\u73b0\u590d\u6742\u7684\u4e1a\u52a1\u81ea\u52a8\u5316\u3002<\/p>\n<p><strong>English Summary:<\/strong> AWS introduces Amazon Quick Flows, an AI-powered workflow automation tool for non-technical users. It allows users to describe tasks in natural language, automatically generating complete workflows including data collection, AI analysis, and external system integrations. The post demonstrates building a financial analysis tool and employee onboarding automation, showcasing features like visual editing, conditional logic, loops, variable passing, and connections to external services like SharePoint, S3, and HR\/IT systems\u2014all without requiring code.<\/p>\n<p><a href=\"https:\/\/aws.amazon.com\/blogs\/machine-learning\/automate-repetitive-tasks-with-amazon-quick-flows\/\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>OpenAI ends Microsoft legal peril over its $50B Amazon deal<\/strong>\uff08TechCrunch AI\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>OpenAI \u4e0e\u5fae\u8f6f\u8fbe\u6210\u4fee\u8ba2\u534f\u8bae\uff0c\u89e3\u51b3\u4e86\u6b64\u524d\u56e0 OpenAI \u4e0e\u4e9a\u9a6c\u900a\u9ad8\u8fbe 500 \u4ebf\u7f8e\u5143\u5408\u4f5c\u800c\u5f15\u53d1\u7684\u6cd5\u5f8b\u98ce\u9669\u3002\u6839\u636e\u65b0\u6761\u6b3e\uff0c\u5fae\u8f6f\u5bf9 OpenAI \u77e5\u8bc6\u4ea7\u6743\u7684\u72ec\u5bb6\u6388\u6743\u8f6c\u4e3a\u975e\u72ec\u5bb6\uff0c\u6709\u6548\u671f\u81f3 2032 \u5e74\uff1bOpenAI \u4ea7\u54c1\u5c06\u4f18\u5148\u5728 Azure \u53d1\u5e03\uff0c\u4f46\u53ef\u540c\u65f6\u901a\u8fc7\u4efb\u4f55\u4e91\u670d\u52a1\u5546\u5411\u5ba2\u6237\u63d0\u4f9b\u3002\u5fae\u8f6f\u505c\u6b62\u5411 OpenAI \u652f\u4ed8\u6536\u5165\u5206\u6210\uff0c\u4f46\u5c06\u7ee7\u7eed\u83b7\u5f97 OpenAI \u7684\u6536\u5165\u5206\u6210\u81f3 2030 \u5e74\uff08\u8bbe\u6709\u4e0a\u9650\uff09\u3002\u5fae\u8f6f\u4ecd\u6301\u6709 OpenAI \u7ea6 27% \u80a1\u4efd\uff0c\u53cc\u65b9\u7ee7\u7eed\u4fdd\u6301&quot;\u4e3b\u8981\u4e91\u5408\u4f5c\u4f19\u4f34&quot;\u5173\u7cfb\u3002\u6b64\u4e3e\u6d88\u9664\u4e86\u5fae\u8f6f\u5c31 AWS \u72ec\u5bb6\u4ee3\u7406 Frontier \u5de5\u5177\u63d0\u8d77\u8bc9\u8bbc\u7684\u53ef\u80fd\u6027\uff0c\u4f7f\u4f01\u4e1a\u5ba2\u6237\u80fd\u591f\u5728\u591a\u4e91\u73af\u5883\u4e2d\u81ea\u7531\u9009\u62e9\u3002<\/p>\n<p><strong>English Summary:<\/strong> OpenAI and Microsoft reached an amended agreement resolving legal risks from OpenAI&#039;s $50 billion deal with Amazon. Under new terms, Microsoft&#039;s exclusive license to OpenAI IP becomes non-exclusive through 2032; OpenAI products will launch first on Azure but can now be served across any cloud provider. Microsoft stops paying revenue share to OpenAI but continues receiving revenue share from OpenAI through 2030 (subject to a cap).<\/p>\n<p><a href=\"https:\/\/techcrunch.com\/2026\/04\/27\/openai-ends-microsoft-legal-peril-over-its-50b-amazon-deal\/\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>QCon San Francisco 2026: 12 Tracks Announced<\/strong>\uff08InfoQ AI\/ML\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>QCon San Francisco 2026 \u516c\u5e03 12 \u4e2a\u6280\u672f\u4e13\u9898\uff0c\u4f1a\u8bae\u5b9a\u4e8e 11 \u6708 16-20 \u65e5\u4e3e\u884c\u3002\u5176\u4e2d 4 \u4e2a\u4e13\u9898\u805a\u7126 AI \u751f\u4ea7\u5316\u5b9e\u8df5\uff1a\u667a\u80fd\u4f53\u67b6\u6784\u8bbe\u8ba1\u3001AI \u7cfb\u7edf\u5de5\u7a0b\u3001\u8bc4\u4f30\u4e0e\u5b89\u5168\u62a4\u680f\u3001\u6570\u636e\u5e73\u53f0\u91cd\u6784\uff1b\u5176\u4f59 8 \u4e2a\u6db5\u76d6\u5206\u5e03\u5f0f\u7cfb\u7edf\u3001\u67b6\u6784\u5256\u6790\u3001\u5f39\u6027\u5de5\u7a0b\u3001\u5e73\u53f0\u5de5\u7a0b\u3001\u5f00\u53d1\u8005\u4f53\u9a8c\u3001\u73b0\u4ee3 API \u8bbe\u8ba1\u3001Staff+ \u5de5\u7a0b\u5e08\u6280\u80fd\u4ee5\u53ca\u975e\u5de5\u7a0b\u5e08\u89d2\u8272\u7684\u4ee3\u7801\u80fd\u529b\u3002\u5927\u4f1a\u5f3a\u8c03\u5b9e\u6218\u6848\u4f8b\u800c\u975e\u4ea7\u54c1\u8def\u7ebf\u56fe\uff0c\u7531\u8d44\u6df1\u4ece\u4e1a\u8005\u7ec4\u6210\u7684\u59d4\u5458\u4f1a\u7b5b\u9009\u8bae\u9898\uff0c\u65e8\u5728\u5e2e\u52a9\u9ad8\u7ea7\u5de5\u7a0b\u5e08\u5e94\u5bf9\u81ea\u4e3b\u667a\u80fd\u4f53\u6545\u969c\u6a21\u5f0f\u3001\u6d41\u91cf\u5cf0\u503c\u4e0b\u7684 P99 \u5ef6\u8fdf\u4fdd\u969c\u3001API \u67b6\u6784\u6f14\u8fdb\u7b49\u73b0\u5b9e\u6311\u6218\u3002<\/p>\n<p><strong>English Summary:<\/strong> QCon San Francisco 2026 announced 12 tracks for its November 16-20 conference. Four tracks focus on production AI: Architecting for Agents, Engineering AI Systems, Guardrails &amp; Safety Nets (Evals), and Data Platforms Reimagined. The remaining eight cover Distributed Systems, Architecture Teardowns, Resilience Engineering, Platform Engineering, Developer Experience, Modern API Design, Staff+ Engineering Skills, and Code Beyond Engineers.<\/p>\n<p><a href=\"https:\/\/www.infoq.com\/news\/2026\/04\/qconsf-2026-tracks-announced\/?utm_campaign=infoq_content&#038;utm_source=infoq&#038;utm_medium=feed&#038;utm_term=AI%2C+ML+%26+Data+Engineering\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>OpenAI available at FedRAMP Moderate<\/strong>\uff08OpenAI News\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>OpenAI \u5ba3\u5e03 ChatGPT Enterprise \u548c API Platform \u83b7\u5f97 FedRAMP 20x Moderate \u6388\u6743\uff0c\u6210\u4e3a\u7f8e\u56fd\u8054\u90a6\u673a\u6784\u53ef\u5408\u89c4\u4f7f\u7528\u7684 AI \u670d\u52a1\u3002FedRAMP 20x \u662f GSA \u4e8e 2025 \u5e74 3 \u6708\u63a8\u51fa\u7684\u5feb\u901f\u6388\u6743\u8def\u5f84\uff0c\u91c7\u7528\u4e91\u539f\u751f\u5b89\u5168\u8bc1\u636e\u3001\u5173\u952e\u5b89\u5168\u6307\u6807\uff08KSI\uff09\u548c\u81ea\u52a8\u5316\u9a8c\u8bc1\u3002\u83b7\u5f97\u6388\u6743\u540e\uff0c\u8054\u90a6\u673a\u6784\u53ef\u5728\u5185\u90e8\u8fd0\u8425\u548c\u4efb\u52a1\u652f\u6301\u573a\u666f\u4e2d\u4f7f\u7528 GPT-5.5 \u7b49\u524d\u6cbf\u6a21\u578b\uff0c\u672a\u6765\u8fd8\u5c06\u901a\u8fc7 FedRAMP \u73af\u5883\u8bbf\u95ee Codex Cloud\u3002OpenAI \u5df2\u5728 FedRAMP Marketplace \u4e0a\u67b6\uff0c\u653f\u5e9c\u673a\u6784\u53ef\u901a\u8fc7 Carahsoft \u7b49\u6388\u6743\u7ecf\u9500\u5546\u91c7\u8d2d\u3002<\/p>\n<p><strong>English Summary:<\/strong> OpenAI announced FedRAMP 20x Moderate authorization for ChatGPT Enterprise and API Platform, making frontier AI available to U.S. federal agencies. FedRAMP 20x, launched by GSA in March 2025, uses cloud-native security evidence, Key Security Indicators, and automated validation for faster authorization. Federal agencies can now use GPT-5.5 and soon access Codex Cloud through FedRAMP environments for internal and mission-support use cases. OpenAI is listed on the FedRAMP Marketplace and available through authorized resellers like Carahsoft.<\/p>\n<p><a href=\"https:\/\/openai.com\/index\/openai-available-at-fedramp-moderate\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>The next phase of the Microsoft OpenAI partnership<\/strong>\uff08OpenAI News\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>OpenAI \u4e0e\u5fae\u8f6f\u5ba3\u5e03\u4fee\u8ba2\u5408\u4f5c\u534f\u8bae\uff0c\u4e3a\u53cc\u65b9\u957f\u671f\u5408\u4f5c\u63d0\u4f9b\u66f4\u5927\u786e\u5b9a\u6027\u3002\u6838\u5fc3\u6761\u6b3e\u5305\u62ec\uff1a\u5fae\u8f6f\u4fdd\u6301 OpenAI \u4e3b\u8981\u4e91\u5408\u4f5c\u4f19\u4f34\u5730\u4f4d\uff0cOpenAI \u4ea7\u54c1\u4f18\u5148\u5728 Azure \u53d1\u5e03\uff0c\u4f46\u53ef\u8de8\u4efb\u610f\u4e91\u670d\u52a1\u5546\u4ea4\u4ed8\u5ba2\u6237\uff1b\u5fae\u8f6f\u5bf9 OpenAI \u77e5\u8bc6\u4ea7\u6743\u7684\u6388\u6743\u4ece\u72ec\u5bb6\u8f6c\u4e3a\u975e\u72ec\u5bb6\uff0c\u6709\u6548\u671f\u81f3 2032 \u5e74\uff1b\u5fae\u8f6f\u505c\u6b62\u5411 OpenAI \u652f\u4ed8\u6536\u5165\u5206\u6210\uff0c\u4f46\u7ee7\u7eed\u83b7\u5f97 OpenAI \u7684\u6536\u5165\u5206\u6210\u81f3 2030 \u5e74\uff08\u8bbe\u6709\u603b\u989d\u4e0a\u9650\uff09\uff1b\u5fae\u8f6f\u4f5c\u4e3a\u5927\u80a1\u4e1c\u7ee7\u7eed\u53c2\u4e0e OpenAI \u589e\u957f\u3002\u53cc\u65b9\u5c06\u7ee7\u7eed\u5728\u6570\u636e\u4e2d\u5fc3\u6269\u5bb9\u3001\u4e0b\u4e00\u4ee3\u82af\u7247\u3001\u7f51\u7edc\u5b89\u5168\u7b49 ambitious \u9879\u76ee\u4e0a\u5408\u4f5c\u3002<\/p>\n<p><strong>English Summary:<\/strong> OpenAI and Microsoft announced an amended partnership agreement providing long-term clarity. Key terms: Microsoft remains OpenAI&#039;s primary cloud partner with products shipping first on Azure, but OpenAI can now serve customers across any cloud provider; Microsoft&#039;s license to OpenAI IP becomes non-exclusive through 2032; Microsoft stops paying revenue share to OpenAI but continues receiving revenue share from OpenAI through 2030 (subject to a cap); Microsoft remains a major shareholder.<\/p>\n<p><a href=\"https:\/\/openai.com\/index\/next-phase-of-microsoft-partnership\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>[AINews] DeepSeek V4 Pro (1.6T-A49B) and Flash (284B-A13B), Base and Instruct \u2014 runnable on Huawei Ascend chips<\/strong>\uff08Latent Space\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>DeepSeek \u6b63\u5f0f\u53d1\u5e03 V4 \u7cfb\u5217\u6a21\u578b\uff0c\u5305\u62ec V4 Pro\uff081.6T \u603b\u53c2\u6570 \/ 49B \u6fc0\u6d3b\uff09\u548c V4 Flash\uff08284B \u603b\u53c2\u6570 \/ 13B \u6fc0\u6d3b\uff09\uff0c\u91c7\u7528 MIT \u5f00\u6e90\u534f\u8bae\u3002\u8fd9\u662f\u81ea 2024 \u5e74 12 \u6708 V3 \u548c 2025 \u5e74 1 \u6708 R1 \u4ee5\u6765\u7684\u9996\u6b21\u91cd\u5927\u7248\u672c\u66f4\u65b0\u3002V4 \u7cfb\u5217\u652f\u6301 100 \u4e07 token \u8d85\u957f\u4e0a\u4e0b\u6587\uff0c\u901a\u8fc7 CSA\uff08\u538b\u7f29\u7a00\u758f\u6ce8\u610f\u529b\uff09\u548c HCA\uff08\u91cd\u5ea6\u538b\u7f29\u6ce8\u610f\u529b\uff09\u6280\u672f\uff0cKV \u7f13\u5b58\u76f8\u6bd4 V3.2 \u51cf\u5c11\u7ea6 10 \u500d\u3002\u6a21\u578b\u91c7\u7528 FP4\/FP8 \u6df7\u5408\u7cbe\u5ea6\uff0c\u8bad\u7ec3\u6570\u636e\u91cf\u8fbe 32T tokens\u3002\u72ec\u7acb\u8bc4\u6d4b\u663e\u793a V4 Pro \u5728\u5f00\u653e\u6743\u91cd\u6a21\u578b\u4e2d\u6392\u540d\u7b2c\u4e8c\uff0c\u4ec5\u6b21\u4e8e Kimi K2.6\uff0c\u5728 Agentic \u4efb\u52a1\u4e0a\u8868\u73b0\u9886\u5148\u3002\u503c\u5f97\u6ce8\u610f\u7684\u662f\uff0cDeepSeek \u540c\u65f6\u9002\u914d\u534e\u4e3a\u6607\u817e\u82af\u7247\uff0c\u8ba1\u5212\u4e0b\u534a\u5e74\u90e8\u7f72 Ascend 950 \u8d85\u7ea7\u8282\u70b9\u4ee5\u8fdb\u4e00\u6b65\u964d\u4f4e\u4ef7\u683c\u3002<\/p>\n<p><strong>English Summary:<\/strong> DeepSeek released the V4 family including V4 Pro (1.6T total \/ 49B active params) and V4 Flash (284B total \/ 13B active), under MIT license. This marks the first major release since V3 (Dec 2024) and R1 (Jan 2025). V4 supports 1M token context via CSA and HCA attention mechanisms, achieving ~10x KV cache reduction vs V3.2. Trained on 32T tokens with FP4\/FP8 mixed precision, V4 Pro ranks #2 among open-weight models per independent benchmarks, behind only Kimi K2.6, with leading agentic performance. DeepSeek also announced Huawei Ascend chip compatibility, with plans to deploy Ascend 950 supernodes in H2 to reduce pricing.<\/p>\n<p><a href=\"https:\/\/www.latent.space\/p\/ainews-deepseek-v4-pro-16t-a49b-and\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Reading today&#039;s open-closed performance gap<\/strong>\uff08Interconnects\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>\u6587\u7ae0\u63a2\u8ba8\u4e86\u5f00\u653e\u6a21\u578b\u4e0e\u95ed\u6e90\u6a21\u578b\u4e4b\u95f4\u7684\u6027\u80fd\u5dee\u8ddd\u8bc4\u4f30\u95ee\u9898\uff0c\u6307\u51fa\u5355\u4e00\u7efc\u5408\u8bc4\u5206\u63a9\u76d6\u4e86\u80fd\u529b\u8986\u76d6\u7684\u590d\u6742\u6027\u3002\u4ee5 Artificial Analysis Intelligence Index \u4e3a\u4f8b\uff0c\u4f5c\u8005\u5206\u6790\u4e86\u57fa\u51c6\u6d4b\u8bd5\u5982\u4f55\u968f\u65f6\u95f4\u6f14\u53d8\u3001\u4e0e\u5b9e\u9645\u4f7f\u7528\u573a\u666f\u7684\u76f8\u5173\u6027\u53d8\u5316\uff0c\u4ee5\u53ca\u4e0d\u540c\u8bad\u7ec3\u8303\u5f0f\u5bf9\u8bc4\u5206\u7684\u5f71\u54cd\u3002\u5f53\u524d\u884c\u4e1a\u7126\u70b9\u5df2\u4ece\u7b80\u5355\u5bf9\u8bdd\u548c\u6570\u5b66\u8f6c\u5411\u590d\u6742\u4ee3\u7801\u4e0e Agentic \u4efb\u52a1\uff0c\u800c\u95ed\u6e90\u5b9e\u9a8c\u5ba4\u6b63\u6295\u5165\u5de8\u8d44\u638c\u63e1\u8fd9\u4e9b\u9886\u57df\uff0c\u5e76\u5f00\u59cb\u5411\u4f1a\u8ba1\u3001\u6cd5\u5f8b\u3001\u533b\u7597\u7b49\u4e13\u4e1a\u9886\u57df\u6269\u5c55\u3002\u6587\u7ae0\u6307\u51fa\uff0c\u5f00\u653e\u6a21\u578b\u5b9e\u9a8c\u5ba4\uff08\u5c24\u5176\u662f\u4e2d\u56fd\u5b9e\u9a8c\u5ba4\uff09\u5728\u8ffd\u8d76\u8fc7\u7a0b\u4e2d\u9762\u4e34\u6570\u636e\u83b7\u53d6\u6210\u672c\u548c\u73af\u5883\u6784\u5efa\u7684\u6311\u6218\uff0c\u4f46 RLVR\uff08\u53ef\u9a8c\u8bc1\u5956\u52b1\u5f3a\u5316\u5b66\u4e60\uff09\u8bad\u7ec3\u65b9\u6cd5\u7684\u666e\u53ca\u4f7f\u5b83\u4eec\u4ecd\u80fd\u4fdd\u6301\u7ade\u4e89\u529b\u3002\u4f5c\u8005\u8ba4\u4e3a\uff0c\u968f\u7740\u4efb\u52a1\u96be\u5ea6\u589e\u52a0\u548c\u6570\u636e\u79c1\u6709\u5316\u8d8b\u52bf\uff0c\u5f00\u653e\u6a21\u578b\u80fd\u5426\u6301\u7eed\u8ffd\u8d76\u4ecd\u5b58\u4e0d\u786e\u5b9a\u6027\u3002<\/p>\n<p><strong>English Summary:<\/strong> The article examines how the open-closed model performance gap is often oversimplified into a single number, masking nuanced capability coverage dynamics. Using the Artificial Analysis Intelligence Index as an example, the author discusses how benchmarks evolve, their correlation with real-world usage, and how training paradigms shift scores. Industry focus has moved from chat and math to complex coding and agentic tasks, with closed labs investing heavily while expanding into specialized domains like law and healthcare. Open labs (especially Chinese ones) face challenges with data acquisition costs and environment building, but RLVR training methods help them remain competitive. The author questions whether open models can sustain this catch-up as tasks grow harder and data becomes more proprietary.<\/p>\n<p><a href=\"https:\/\/www.interconnects.ai\/p\/reading-todays-open-closed-performance\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Building an emoji list generator with the GitHub Copilot CLI<\/strong>\uff08GitHub AI\/ML\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>GitHub \u56e2\u961f\u5728 Rubber Duck Thursday \u76f4\u64ad\u6d3b\u52a8\u4e2d\u5c55\u793a\u4e86\u5982\u4f55\u4f7f\u7528 GitHub Copilot CLI \u6784\u5efa\u4e00\u4e2a\u8868\u60c5\u7b26\u53f7\u5217\u8868\u751f\u6210\u5668\u3002\u8be5\u9879\u76ee\u5141\u8bb8\u7528\u6237\u5728\u7ec8\u7aef\u4e2d\u7c98\u8d34\u6216\u8f93\u5165\u5217\u8868\uff0c\u901a\u8fc7 AI \u81ea\u52a8\u4e3a\u6bcf\u6761\u9879\u76ee\u5339\u914d\u76f8\u5173\u8868\u60c5\u7b26\u53f7\uff0c\u5e76\u5c06\u7ed3\u679c\u590d\u5236\u5230\u526a\u8d34\u677f\u3002\u5f00\u53d1\u8fc7\u7a0b\u4e2d\u4f7f\u7528\u4e86 @opentui\/core \u6784\u5efa\u7ec8\u7aef UI\u3001@github\/copilot-sdk \u63d0\u4f9b AI \u80fd\u529b\u3001clipboardy \u5904\u7406\u526a\u8d34\u677f\u529f\u80fd\u3002\u56e2\u961f\u91c7\u7528\u4e86 Plan \u6a21\u5f0f\u4e0e Claude Sonnet 4.6 \u8fdb\u884c\u89c4\u5212\uff0c\u7136\u540e\u4f7f\u7528 Claude Opus 4.7 \u5b9e\u73b0\u529f\u80fd\u3002\u6587\u7ae0\u8fd8\u4ecb\u7ecd\u4e86\u591a\u6a21\u578b\u5de5\u4f5c\u6d41\u3001Autopilot \u6a21\u5f0f\u3001allow-all \u5de5\u5177\u6807\u5fd7\u4ee5\u53ca GitHub MCP \u670d\u52a1\u5668\u7b49 Copilot CLI \u7279\u6027\u7684\u5b9e\u9645\u5e94\u7528\u3002<\/p>\n<p><strong>English Summary:<\/strong> GitHub&#039;s Rubber Duck Thursday livestream demonstrated building an emoji list generator using GitHub Copilot CLI. The tool lets users paste or type bullet points in the terminal, automatically matches relevant emojis to each item via AI, and copies the result to clipboard. The build used @opentui\/core for terminal UI, @github\/copilot-sdk for AI capabilities, and clipboardy for clipboard access. The team employed Plan mode with Claude Sonnet 4.6 for planning, then Claude Opus 4.7 for implementation. The article showcases practical applications of Copilot CLI features including multi-model workflows, Autopilot mode, the allow-all tools flag, and the GitHub MCP server.<\/p>\n<p><a href=\"https:\/\/github.blog\/ai-and-ml\/github-copilot\/building-an-emoji-list-generator-with-the-github-copilot-cli\/\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Build a personal organization command center with GitHub Copilot CLI<\/strong>\uff08GitHub AI\/ML\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>GitHub \u5de5\u7a0b\u5e08 Brittany Ellich \u5206\u4eab\u4e86\u5982\u4f55\u4f7f\u7528 GitHub Copilot CLI \u6784\u5efa\u4e2a\u4eba\u7ec4\u7ec7\u6307\u6325\u4e2d\u5fc3\uff0c\u4ee5\u89e3\u51b3\u6570\u5b57\u4fe1\u606f\u5206\u6563\u5728\u591a\u4e2a\u5e94\u7528\u4e2d\u7684\u95ee\u9898\u3002\u8be5\u9879\u76ee\u662f\u4e00\u4e2a Electron \u5e94\u7528\uff0c\u5c06\u5206\u6563\u5728\u4e0d\u540c\u5e73\u53f0\u7684\u4efb\u52a1\u3001\u65e5\u7a0b\u548c\u4fe1\u606f\u7edf\u4e00\u5230\u4e00\u4e2a\u96c6\u4e2d\u7684\u53ef\u89c6\u5316\u754c\u9762\u4e2d\u3002\u5f00\u53d1\u91c7\u7528&quot;\u5148\u89c4\u5212\u540e\u5b9e\u73b0&quot;\u7684\u5de5\u4f5c\u6d41\u7a0b\uff1a\u4f7f\u7528 Copilot \u8fdb\u884c\u9700\u6c42\u8bbf\u8c08\u548c\u89c4\u5212\u751f\u6210 plan.md\uff0c\u7136\u540e\u7531 Copilot \u6839\u636e\u89c4\u5212\u5b9e\u73b0\u529f\u80fd\u3002Brittany \u540c\u65f6\u4f7f\u7528 VS Code \u7684 Agent \u6a21\u5f0f\u8fdb\u884c\u540c\u6b65\u5f00\u53d1\uff0c\u4ee5\u53ca Copilot Cloud Agent \u5904\u7406\u5f02\u6b65\u4efb\u52a1\u3002\u5979\u5f3a\u8c03\uff0c\u5c3d\u7ba1\u8fd9\u662f\u5979\u7684\u7b2c\u4e00\u4e2a Electron \u5e94\u7528\uff0c\u4f46\u501f\u52a9 AI \u8f85\u52a9\u5f00\u53d1\uff0c\u4ece\u60f3\u6cd5\u5230\u53ef\u7528\u5de5\u5177\u4ec5\u7528\u4e86\u4e0d\u5230\u4e00\u5929\u65f6\u95f4\uff0c\u540c\u65f6\u5979\u4e5f\u624b\u52a8\u7b80\u5316\u4e86\u4ee3\u7801\u5e93\u4ee5\u63d0\u9ad8\u53ef\u7ef4\u62a4\u6027\u3002<\/p>\n<p><strong>English Summary:<\/strong> GitHub engineer Brittany Ellich shared how she built a personal organization command center using GitHub Copilot CLI to solve digital fragmentation across multiple apps. The Electron app unifies tasks, schedules, and information from various platforms into one centralized visual interface. The development followed a &quot;plan-then-implement&quot; workflow: using Copilot to interview her requirements and generate a plan.md, then having Copilot implement based on the plan. Brittany used VS Code Agent mode for synchronous development alongside Copilot Cloud Agent for asynchronous tasks. She noted that despite being her first Electron app, AI-assisted development took her from idea to working tool in under a day, though she manually simplified the codebase for maintainability.<\/p>\n<p><a href=\"https:\/\/github.blog\/ai-and-ml\/github-copilot\/build-a-personal-organization-command-center-with-github-copilot-cli\/\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<li>\n<p><strong>Ollama is now powered by MLX on Apple Silicon in preview<\/strong>\uff08Ollama Blog\uff09<\/p>\n<p><strong>\u4e2d\u6587\u6458\u8981\uff1a<\/strong>Ollama \u53d1\u5e03\u57fa\u4e8e Apple MLX \u6846\u67b6\u7684\u9884\u89c8\u7248\u672c\uff0c\u4e3a Apple Silicon \u8bbe\u5907\u5e26\u6765\u663e\u8457\u6027\u80fd\u63d0\u5347\u3002\u65b0\u7248\u672c\u5229\u7528 Apple \u7684\u7edf\u4e00\u5185\u5b58\u67b6\u6784\uff0c\u5728 M5\u3001M5 Pro \u548c M5 Max \u82af\u7247\u4e0a\u501f\u52a9 GPU Neural Accelerators \u52a0\u901f\u9996 token \u751f\u6210\u65f6\u95f4\u548c\u89e3\u7801\u901f\u5ea6\u3002\u6d4b\u8bd5\u663e\u793a\uff0c\u4f7f\u7528 Qwen3.5-35B-A3B \u6a21\u578b\u65f6\uff0cprefill \u6027\u80fd\u53ef\u8fbe 1851 tokens\/s\uff0cdecode \u6027\u80fd\u8fbe 134 tokens\/s\u3002Ollama \u65b0\u589e\u5bf9 NVIDIA NVFP4 \u683c\u5f0f\u7684\u652f\u6301\uff0c\u5728\u964d\u4f4e\u5185\u5b58\u5e26\u5bbd\u548c\u5b58\u50a8\u9700\u6c42\u7684\u540c\u65f6\u4fdd\u6301\u6a21\u578b\u7cbe\u5ea6\uff0c\u4f7f\u7528\u6237\u83b7\u5f97\u4e0e\u751f\u4ea7\u73af\u5883\u4e00\u81f4\u7684\u7ed3\u679c\u3002\u6b64\u5916\uff0c\u7f13\u5b58\u7cfb\u7edf\u5347\u7ea7\u5305\u62ec\u8de8\u4f1a\u8bdd\u590d\u7528\u7f13\u5b58\u3001\u667a\u80fd\u68c0\u67e5\u70b9\u5b58\u50a8\u548c\u66f4\u667a\u80fd\u7684\u6dd8\u6c70\u7b56\u7565\uff0c\u663e\u8457\u63d0\u5347\u4e86\u7f16\u7801\u548c Agentic \u4efb\u52a1\u7684\u54cd\u5e94\u901f\u5ea6\u3002<\/p>\n<p><strong>English Summary:<\/strong> Ollama released a preview version powered by Apple&#039;s MLX framework, delivering significant performance improvements on Apple Silicon devices. The new version leverages Apple&#039;s unified memory architecture and GPU Neural Accelerators on M5, M5 Pro, and M5 Max chips to accelerate time-to-first-token and decode speeds. Testing with Qwen3.5-35B-A3B showed prefill performance reaching 1851 tokens\/s and decode at 134 tokens\/s. Ollama now supports NVIDIA&#039;s NVFP4 format, maintaining model accuracy while reducing memory bandwidth and storage requirements for production parity. Cache system upgrades include cross-session cache reuse, intelligent checkpoint storage, and smarter eviction policies, significantly improving responsiveness for coding and agentic tasks.<\/p>\n<p><a href=\"https:\/\/ollama.com\/blog\/mlx\" target=\"_blank\" rel=\"noopener noreferrer\">\u539f\u6587\u94fe\u63a5<\/a><\/p>\n<\/li>\n<\/ol>\n","protected":false},"excerpt":{"rendered":"<p>\u65e5\u671f\uff1a2026-04-28 \u672c\u671f\u805a\u7126\uff1a\u91cd\u70b9\u5173\u6ce8\u6a21\u578b\u53d1\u5e03\u4e0e release notes\u3001\u5b98\u65b9 engineeri [&hellip;]<\/p>\n","protected":false},"author":0,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[7],"tags":[],"class_list":["post-362","post","type-post","status-publish","format-standard","hentry","category-ai-daily"],"_links":{"self":[{"href":"http:\/\/www.faiyi.com\/index.php?rest_route=\/wp\/v2\/posts\/362","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/www.faiyi.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.faiyi.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"http:\/\/www.faiyi.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=362"}],"version-history":[{"count":0,"href":"http:\/\/www.faiyi.com\/index.php?rest_route=\/wp\/v2\/posts\/362\/revisions"}],"wp:attachment":[{"href":"http:\/\/www.faiyi.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=362"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.faiyi.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=362"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.faiyi.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=362"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}