Skip to main content
Featured Newsletter

BestBlogs Issue #74: Deep Thinking

Hey there! Welcome to BestBlogs.dev Issue #75.

This week's theme is deep thinking , and it carries two meanings. First, AI is learning to truly think—DeepSeek V3.2 pioneered the fusion of thinking mode with tool calling, GPT-5.1 made reasoning models the default, and models are no longer just reacting quickly but starting to gather information, think it through, then respond, much like humans do.

Second, it's about us. As AI gets better at doing the work, humans need to return to our most essential capabilities: understanding how the world works, judging what's true, and making critical decisions. One article from Tencent Research Institute left a strong impression on me—the truly scary part isn't failing to keep up with change, it's charging ahead with outdated thinking. Evidence-first reasoning, logical thinking, embracing uncertainty, staying open to being proven wrong—these elements of modern thinking are our true operating system in the AI era.

Here are 10 highlights worth your attention this week:

  • 🤖 DeepSeek V3.2 officially launched , deeply integrating thinking mode with tool calling. The standard version balances reasoning depth with response speed, while V3.2-Speciale focuses on extreme reasoning—winning gold medals in both IMO and IOI. A significant step forward for open-source models in agent capabilities.
  • 🧠 OpenAI's podcast reveals GPT-5.1's core evolution: reasoning models are now the default configuration. The model shifts from intuitive responses to System 2-style chain-of-thought, significantly improving instruction-following even in simple interactions. An interesting reframe: model personality is now defined as a UX combination of memory, context window, and response style—not anthropomorphized traits.
  • 📚 Tencent's engineering team published an in-depth piece tracing the journey from Scaling Laws to CoT, then to internalization mechanisms like PPO, DPO, and GRPO. If you want a systematic understanding of how LLMs learn to think deeply, this is a rare technical roadmap.
  • ✨ The soul document used in Claude 4.5 Opus training has been revealed. Used during supervised learning, it shapes the model's core values and self-awareness through narrative and ethical guidance—even including defenses against prompt injection. A rare and fascinating lens into alignment.
  • 📁 Google's open-source Agent Development Kit proposes an important idea: context should be treated as a first-class system citizen with its own architecture and lifecycle . Separation of storage and representation, explicit transformations, default scoping—this context engineering methodology is invaluable for building long-running multi-agent systems.
  • 🛠️ Want to quickly master agent architectures? Datawhale compiled 17 mainstream implementations (including ReAct, PEV, blackboard systems) with end-to-end Jupyter notebooks. From concept to code in one package.
  • 🎬 Runway Gen-4.5 launched as instant SOTA, pushing video generation's physical realism to new heights—weight, dust, lighting all feel right. Community verdict: game-changing.
  • 🏢 LinkedIn's CPO reveals a paradigm shift in product development: from functional silos to AI-empowered full-stack builders . Facing predictions that 70% of skills will be disrupted by 2030, LinkedIn is replacing APM with APB and restructuring talent development. Not just tool upgrades—a radical experiment in human-AI collaboration culture.
  • 🌏 Joe Tsai's speech at HKU analyzed China's unique AI advantages: affordable energy, infrastructure edge, systems-level optimization talent, and open-source ecosystem. His key insight: AI competition's endgame isn't about model parameter counts—it's about actual adoption rates and data sovereignty.
  • 💡 Finally, a must-read on cognitive transformation from Tencent Research Institute. The author argues that current anxiety isn't about AI itself—it's because we're trying to understand new technology with pre-modern thinking that relies on authority and craves absolute certainty. As knowledge depreciates, humans should hand over the doing to AI while holding tight to the thinking. Only by building modern thinking—grounded in evidence, logic, and acceptance of uncertainty—can we find our irreplaceable role in human-AI collaboration.

Hope this issue sparks some new ideas. Stay curious, and see you next week!

DeepSeek
mp.weixin.qq.com
12-01
1719 words · 7 min
94
DeepSeek V3.2 Official Release: Enhanced Agent Capabilities, Featuring Integrated Reasoning

DeepSeek has officially released the V3.2 model series, marking a significant breakthrough in open-source Agent capabilities. The standard V3.2 balances reasoning depth with response speed and introduces a novel integration of "Thinking Mode" with tool usage, greatly enhancing generalization in complex tasks. Simultaneously, DeepSeek-V3.2-Speciale targets extreme reasoning, achieving gold medals in IMO and IOI, with performance rivaling Gemini-3.0-Pro. Both models are open-source, providing a powerful foundation for developers to build the next generation of AI applications that combine high intelligence with execution.

OpenAI
youtube.com
12-02
9850 words · 40 min
92
Shaping Model Behavior in GPT-5.1— the OpenAI Podcast Ep. 11

This episode of the OpenAI Podcast dives deep into the core evolution of GPT-5.1, marking the official shift of Reasoning Models as the default for all users. OpenAI’s Post-Training Research Lead, Christina Kim, and Product Manager, Laurentia Romaniuk, reveal how models are moving from "intuitive reactions" to System 2-like "Chain of Thought" patterns, significantly improving instruction following even in simple interactions. The conversation redefines "model personality" as a UX combination of Memory, Context Window, and response styles, rather than mere anthropomorphism. For readers focused on AI product evolution, this episode details the balance between maximizing user freedom (Steerability) and safety boundaries, explaining how "safe completions" are replacing hard refusals.

Claude 4.5 Opus’ Soul Document

This article uncovers a unique "Soul Document" used during the training of Claude 4.5 Opus. Confirmed by Anthropic, this internal guide is utilized during Supervised Learning to shape the model's core values, self-awareness, and safety protocols. The document outlines Anthropic's philosophy on balancing AI capability with safety and explicitly instills defenses against Prompt Injection. It offers a rare and fascinating glimpse into how SOTA LLMs are "aligned" through narrative and ethical grounding rather than just raw code.

量子位
qbitai.com
12-02
997 words · 4 min
92
Runway Gen-4.5 Sweeping Release: Accurately Renders Weight, Dust, and Lighting Effects; Netizens Say: Disruptive

Runway Gen-4.5 launches as the new SOTA video model, significantly enhancing physical realism and visual details like complex camera movements and mirror reflections. It strengthens the understanding of sequential instructions while maintaining Gen-4's speed. Rolling out at no extra cost, it is widely recognized by the community as a "disruptor" in AI video generation.

Architecting efficient context-aware multi-agent framework for production

This article offers a deep dive into the critical Context Management bottleneck faced when scaling AI agents to production-grade systems. It explicitly argues that simply increasing the context window size is not a sustainable solution and proposes "Context Engineering"—treating context as a first-class system with its own architecture and lifecycle. Google's open-source Agent Development Kit (ADK) framework is built on this thesis, providing a tiered model (Working Context, Session, Memory, and Artifacts) and a pipeline-based processing mechanism governed by three core principles: separation of storage and presentation, explicit transformation, and scoping by default. This represents a rigorous system engineering approach essential for building reliable, efficient, and long-horizon multi-agent systems, and is a must-read for all serious AI agent developers.

Datawhale
mp.weixin.qq.com
11-28
13064 words · 53 min
93
A Comprehensive Guide to Building AI Agents: 17 Architectures with Detailed Implementations

An essential practical handbook for AI developers. It details 17 mainstream Agent architectures (including ReAct, PEV, and Blackboard Systems) with accompanying end-to-end Jupyter Notebook implementations. Covering foundational patterns, multi-agent collaboration, and advanced memory/safety mechanisms, it directly addresses the pain point of moving from concept to code. Ideal for engineers looking to quickly master the ability to build complex, robust AI systems.

LangChain Blog
blog.langchain.com
12-03
1657 words · 7 min
92
Evaluating Deep Agents: Our Learnings

Drawing from experience building four Deep Agent applications, LangChain shares five essential patterns for evaluating complex, stateful AI agents. The article argues that traditional LLM evaluation methods are insufficient for dynamic agents. Instead, developers should adopt bespoke, code-based test logic for each datapoint. Key strategies include using single-step evaluations for decision validation, full-turn runs for end-state analysis, and maintaining clean, reproducible environments with mocked APIs.

腾讯技术工程
mp.weixin.qq.com
11-28
9820 words · 40 min
92
Must-Read Series of 2025: How AI Redefines Research? A Comprehensive Overview of Deep Research

This comprehensive 10,000-word article systematically dissects the core AI trend of 2025: Deep Research. It details the evolutionary path of AI Agents from passive retrieval (RAG) to active exploration, analyzing the universal architecture comprising Planning, Question Developing, Web Exploration, and Report Generation modules. Beyond comparing mainstream systems like OpenAI and Qwen, it critically identifies the limitations of relying solely on public web data. Using Tencent's Dola as a case study, the article demonstrates how fusing "unstructured public data" with "structured private data" can effectively mitigate hallucinations and enhance the reliability of business decision-making.

AI Engineer
youtube.com
12-02
4875 words · 20 min
92
Building Cursor Composer – Lee Robinson, Cursor

Lee Robinson details the engineering behind Cursor Composer, an AI agent built to reconcile the trade-off between speed and intelligence in software engineering. By leveraging Reinforcement Learning (RL) and custom kernels, the team achieved 4x token generation efficiency while maintaining frontier-level intelligence. The talk covers infrastructure challenges, parallel tool execution, and how semantic search enhances agentic capabilities.

Lenny's Podcast
youtube.com
12-04
23538 words · 95 min
92
Why LinkedIn is turning PMs into AI-powered "full stack builders” | Tomer Cohen (LinkedIn CPO)

LinkedIn CPO Tomer Cohen reveals a major paradigm shift in product development: moving from specialized functional silos to an AI-empowered "Full Stack Builder" model. Facing the prediction that 70% of current job skills will be disrupted by 2030, LinkedIn is re-architecting its platform, building deeply customized internal AI Agents, and overhauling its talent pipeline by replacing the traditional APM program. This episode offers a practical blueprint for large organizations navigating the deep waters of AI transformation, showcasing how to integrate platform, tools, and culture for a radical efficiency leap.

Founder Park
mp.weixin.qq.com
12-02
7859 words · 32 min
92
Superhuman's Methodology: Achieving $35M ARR with an AI Email Tool and PMF

This article deconstructs the "PMF Engine" framework developed by Superhuman founder Rahul Vohra, transforming the abstract goal of Product-Market Fit into actionable metrics. The core strategy revolves around Sean Ellis's 40% threshold (users who would be "very disappointed" without the product) and identifying "High-Expectation Customers". It details a tactical 50/50 roadmap approach: dedicating half the resources to doubling down on core differentiators like speed, and the other half to addressing blockers. An essential read for SaaS founders seeking a systematic path to validate market demand.

Founder Park
mp.weixin.qq.com
12-04
4317 words · 18 min
92
AI Voice Input Product Valued at $700 Million: Voice Input Focuses on Intent-Based Dictation, Not Just Transcription

Wispr Flow argues that the future of AI voice input lies in "dictation" (understanding intent via context) rather than simple "transcription." With an 89% zero-edit rate, it seeks to eliminate the cognitive load of keyboard input. The founder envisions a true AI assistant as an intelligent layer with global memory, not an isolated tool, paving the way for a "post-keyboard era" in communication.

阿真Irene
mp.weixin.qq.com
12-03
6071 words · 25 min
92
Cartoon Anime English Vocabulary Cards: New Prompt Templates!

A ready-to-use AI solution for creating English vocabulary flashcards, featuring high-quality prompt templates in four styles like watercolor and 3D clay. This workflow leverages LLMs to intelligently generate scene descriptions, supports custom IP characters, and is perfectly adapted for Nano Banana Pro and Lovart to generate complete educational infographics containing "scenes + words + vocabulary lists" in one click.

Product School
youtube.com
12-01
5778 words · 24 min
92
Classic Product Rules to Rethink in AI Era | fmr. Zalando Head of Product

Traditional PM rules often fail in AI. This video reveals 10 essential shifts: moving from deterministic outputs to probabilistic outcomes, from customer feedback to data feedback, and from UX-first to "Trust as Usability." It is a concise playbook for PMs transitioning from releasing features to training systems, emphasizing the critical need for dynamic KPIs and cross-disciplinary orchestration amidst uncertainty.

Web3天空之城
mp.weixin.qq.com
11-29
17382 words · 70 min
93
Full Version: Joe Tsai's In-Depth Interpretation at HKU in November - China's Unique AI Advantages and Technological Drivers for the Next Decade | Illustrated + Full Text of 17,000 Words, with Video

Joe Tsai analyzes China's unique AI edge: cheap energy, infrastructure advantages, talent in system-level optimization, and an open-source ecosystem. Identifying high-tech manufacturing and tech self-reliance as the decade's economic engines, he argues the AI endgame isn't about model parameters, but Adoption Rate and data sovereignty.

量子位
qbitai.com
12-03
20981 words · 84 min
92
Mark Chen on OpenAI's Strategy, Talent, and Future AI

In this in-depth interview, OpenAI Lead Researcher Mark Chen reveals the internal strategy of the company leading the charge toward AGI. Chen details how OpenAI retains its core team amidst a fierce talent war with Meta (featuring the "soup" anecdote) and emphasizes that OpenAI remains a "pure AI research company" at heart. Key insights include their doubled-down focus on Pre-training—countering claims that Scaling Laws are dead—to outperform Gemini 3, and the vision of AI for Science to accelerate discoveries.

Jina AI
mp.weixin.qq.com
12-02
11404 words · 46 min
92
Jina AI Startup Lessons: Understanding the Scaling Law for AI Teams

Jina AI founder Han Xiao reviews three strategic pivots and survival philosophies prior to the acquisition by Elastic. He points out that in the rapid AI wave, extreme focus and execution are the only moats. The article explores the commercial limits of "small models," the paradox of team scaling vs. output efficiency, and the reality that "productivity gains ≠ value capture," serving as a must-read field guide for AI founders.

121. Interview with Jie Tan from DeepMind: Robotics, Cross-Embodiment, World Models, Gemini Robotics 1.5, and Google

Google DeepMind's Tan Jie decodes the frontier of robotics. Key highlights: Gemini Robotics 1.5 achieves complex task decomposition via "Thinking Chains" and utilizes Motion Transfer to allow different robots to share learning experiences, fundamentally addressing data scarcity. Tan argues that Sim-to-Real and Synthetic Data are crucial for the future, predicting a "GPT moment" for robotics within 2-3 years, though widespread home adoption remains 5-10 years away.

腾讯研究院
mp.weixin.qq.com
11-28
9868 words · 40 min
92
The AI Era: The Peril of Outdated Thinking

This is a profound piece on cognitive restructuring in the AI era. Professor Ma Zhaoyuan argues that current societal anxiety stems not from AI technology itself, but from our persistence in using "pre-modern thinking"—which relies on authority and seeks absolute certainty—to understand new paradigms. The article sharply critiques the over-hyping of AGI by capital markets, using Turing machine theoretical limits to demonstrate AI's boundaries. The author asserts that as knowledge depreciates, humans should delegate "labor" to AI while retaining the dominance of "thinking." Only by establishing "modern thinking" based on evidence, logic, and the acceptance of uncertainty can we find an irreplaceable role in human-machine collaboration.

    BestBlogs Issue #74: Deep Thinking | BestBlogs.dev