New OpenAI Research: AI 'Chain-of-Thought' Can Be Monitored But Requires Deliberate Maintenance
AI SafetyLarge ModelsAI Interpretability
OpenAI research reveals that the 'chain-of-thought' reasoning generated by current leading large models (e.g., the GPT-5 series) can still be effectively read and monitored by AI systems to detect anomalous behaviors (such as reward hacking, capability concealment). However, as model capabilities improve, intentionally hiding the reasoning process during training will significantly increase monitoring difficulty. The study notes that longer reasoning chains are easier to monitor, and higher reasoning intensity in smaller models can enhance monitorability, albeit at a computational cost. OpenAI emphasizes that future safe deployment of large models requires proactive maintenance of reasoning interpretability; otherwise, AI behavior will become difficult to track and explain.
Waymo Robotaxi Service Halted in San Francisco Due to Major Blackout
Autonomous DrivingAI ImplementationSmart Mobility
Waymo's autonomous taxi service in San Francisco was forced to suspend operations due to a citywide power outage caused by a fire at a PG&E substation. Some vehicles were stranded on streets due to non-functional traffic signals and communication failures. This incident highlights the high dependence of autonomous systems on urban infrastructure (e.g., power, communication) and has sparked discussions about the emergency capabilities of autonomous vehicles in extreme situations.
Oracle Cloud AI Business Booms, Disclosure of $248 Billion Off-Balance-Sheet Commitments Raises Bubble Concerns
AI InfrastructureCloud ComputingAI Bubble
Oracle's latest financial report shows significant growth in its AI-related cloud services business. However, it also disclosed $248 billion in long-term off-balance-sheet commitments for data centers and cloud leases. These massive long-term obligations, not recorded as traditional debt, have raised market concerns about an AI infrastructure investment bubble and financial risks. Analysis suggests the AI infrastructure sector faces multiple challenges including high leverage, high capital expenditure, and demand uncertainty.
Claude Opus 4.5 Breaks Record for AI Long-Task Capability, Can Complete Nearly 5-Hour Tasks in a Single Session
Large ModelsAI Capability EvaluationAI Application
METR evaluations show that Anthropic's Claude Opus 4.5 model has achieved a breakthrough in long-task capabilities. Within a 50% confidence interval, it can reliably complete human-level professional tasks lasting 4 hours and 49 minutes, setting a new public benchmark record. This indicates AI can now handle some long-duration work requiring sustained focus and complex reasoning, pushing the practical application boundaries of AI in professional fields.
In 2025, AI-generated content surged, leading Merriam-Webster to select 'slop' as its Word of the Year, referring to low-quality, mass-produced AI content. Data shows that in May 2025, AI-generated content accounted for 48% of new online articles, with some platforms detecting 74.2% of new web pages containing AI content. Rising user concerns about content authenticity and trustworthiness have spurred demand for 'pre-AI internet' search tools and content filtering, posing new challenges for content platforms and regulators.
AI Safety Monitoring and Interpretability Become Core Issues for Large Model Deployment
AI SafetyInterpretabilityAI Governance
As AI model capabilities advance, how to monitor AI reasoning processes in real-time and detect anomalous behaviors has become an industry focus. Research from companies like OpenAI shows that chain-of-thought (CoT) monitoring is superior to output-only monitoring, but security monitoring faces challenges if the model intentionally conceals its reasoning. The industry calls for proactively maintaining reasoning interpretability during the training and deployment of large models to ensure AI systems are controllable and traceable.
Peak AI Infrastructure Investment: Data Centers, Power, and Water Emerge as New Bottlenecks
AI InfrastructureData CentersEnergy & Environment
Global AI infrastructure investment surged in 2025. Data center construction and compute expansion are driving a significant increase in power and water demand. Total investment in data center projects in the US and Asia has exceeded hundreds of billions of dollars. Individual AI campuses can consume hundreds of megawatts of electricity and millions of gallons of cooling water daily. Grid access, energy, and water shortages are becoming the physical limits to AI industry expansion. Companies are accelerating deployment of proprietary power sources, liquid cooling, and other new technologies to address these bottlenecks.
AI Regulation and Trust Framework Upgrades: US Federal Leadership, EU Policy Trends Pragmatic
AI RegulationAI GovernanceCompliance & Trust
In 2025, the US reinforced federal unification of AI regulation through an executive order, curbing fragmented state-level legislation and promoting AI risk management standardization. The EU relaxed some restrictions in the AI Act's implementation rules, emphasizing a balance between industrial competitiveness and safety. Security, traceability, and compliance have become core requirements for corporate AI procurement, and the AI 'Trust Stack' has become a prerequisite for deployment. Industry trends indicate AI governance is moving from concept to engineering and compliance.
AI Chip and Compute Ecosystem Competition Intensifies: Amazon's $10B Investment in OpenAI Challenges Nvidia
AI ChipsCompute InfrastructureIndustry Competition
Amazon plans to invest up to $10 billion in OpenAI and promote the adoption of its in-house Trainium AI chips, challenging Nvidia's dominance in the AI hardware field. The Trainium3/4 chips attract developers with high efficiency and cost advantages. Through massive capital expenditure and revolving financing models, Amazon aims to reshape the AI compute supply chain. Meanwhile, Nvidia continues to solidify its lead through ecosystem lock-in and new products, accelerating the evolution of the AI chip competitive landscape.
AI Long-Task Capability Breakthrough Accelerates Automation in Professional Fields
AI Capability BreakthroughIndustry AutomationLarge Model Application
Latest evaluations show that AI large models (e.g., Claude Opus 4.5) can now reliably complete complex tasks lasting nearly five hours. This marks a significant improvement in AI application capabilities for long-duration, complex reasoning, and sustained-focus professional scenarios. This capability breakthrough will accelerate automation and intelligent processes in knowledge-intensive industries such as law, finance, and healthcare.