Skip to main content

Huawei, Zhejiang University Launch AI Model with Enhanced Security

Huawei and Zhejiang University Unveil DeepSeek-R1-Safe AI Model

At the recent Huawei Global Connect Conference, Huawei Technologies and Zhejiang University jointly introduced DeepSeek-R1-Safe, a groundbreaking foundation model built on Huawei's Ascend 1000 computing platform. This collaboration marks a significant step forward in addressing critical challenges at the intersection of AI performance and security.

A New Standard for AI Safety

Professor Ren Kui, Dean of Zhejiang University's School of Computer Science and Technology, detailed the model's innovative framework. "DeepSeek-R1-Safe represents a comprehensive approach to secure AI development," he explained. The model incorporates:

  • A high-quality secure training corpus
  • Balanced optimization techniques for security training
  • Proprietary software/hardware integration

The framework specifically targets fundamental security challenges in large-scale AI training processes.

Unprecedented Security Performance

Test results demonstrate exceptional capabilities:

  • 100% defense rate across 14 categories of harmful content (toxic speech, political sensitivity, illegal activity incitement)
  • Over 40% success rate against jailbreak attempts
  • 83% comprehensive security score, outperforming comparable models by 8-15%

Remarkably, these security gains come with minimal performance trade-offs. In standard benchmarks (MMLU, GSM8K, CEVAL), the model shows less than 1% performance loss compared to non-secure counterparts.

Industry Implications and Open Access

Zhang Dixuan, President of Huawei's Ascend Computing Business, emphasized the company's commitment to collaborative innovation: "By open-sourcing this technology through ModelZoo, GitCode, GitHub and Gitee, we're enabling broader participation in secure AI development."

The release signals growing industry recognition of security as a foundational requirement rather than an afterthought in AI systems.

Key Points:

  • First domestic foundation model on Ascend 1000 platform
  • Achieves security-performance balance through novel framework
  • Outperforms competitors by significant margins
  • Now available through major open-source platforms

Enjoyed this article?

Subscribe to our newsletter for the latest AI news, product reviews, and project recommendations delivered to your inbox weekly.

Weekly digestFree foreverUnsubscribe anytime

Related Articles

ByteDance Tightens Security with New AI Access Control Tool
News

ByteDance Tightens Security with New AI Access Control Tool

ByteDance has introduced ByteClaw, a new security tool designed to manage employee access to large AI models within its internal networks. The company also released comprehensive guidelines addressing five major security risks in AI applications. These measures aim to balance innovation with safety as AI tools become more prevalent in corporate environments. The move sets a benchmark for how tech firms can securely integrate powerful AI systems into their operations.

March 18, 2026
AI SecurityByteDanceEnterprise Technology
Hackers Trick AI with Poisoned Fonts - Microsoft Leads Fix
News

Hackers Trick AI with Poisoned Fonts - Microsoft Leads Fix

Security researchers uncovered a clever hack where attackers manipulate fonts and CSS to fool AI assistants like ChatGPT into approving dangerous code. While Microsoft quickly patched the vulnerability in Copilot, other major providers dismissed the threat. The attack exploits how AI reads text differently than humans see it, potentially tricking users into running malicious scripts.

March 18, 2026
AI SecurityFont ExploitsMicrosoft Copilot
News

NVIDIA's NemoClaw: Armoring AI Agents for the Enterprise

At the 2026 GTC Conference, NVIDIA unveiled NemoClaw, a new platform designed to bring enterprise-grade security to AI agent development. Built on the popular OpenClaw framework, it tackles critical business concerns around privacy and control while maintaining hardware flexibility. As the AI industry shifts from simple chatbots to complex agent systems, NVIDIA's move positions them against competitors like OpenAI in this emerging market space.

March 17, 2026
NVIDIAAI AgentsEnterprise Tech
News

NVIDIA and Cisco Team Up to Secure AI Agents with Open-Source OpenShell

As AI agents move from labs to business systems, security concerns grow. NVIDIA and Cisco have responded by open-sourcing OpenShell, a runtime that creates secure 'sandboxes' for AI agents. Combined with Cisco's AI Defense platform, this solution monitors agent actions while preventing data leaks. The collaboration marks a significant step toward trustworthy enterprise AI automation.

March 17, 2026
AI SecurityEnterprise TechnologyOpen Source
Tencent's Lobster Butler: Your AI's New Security Guard
News

Tencent's Lobster Butler: Your AI's New Security Guard

Tencent has unveiled Lobster Butler, a groundbreaking security feature for local AI agents in its PC Manager 18.0 update. This innovative sandbox technology creates a protective shell around AI operations, preventing unauthorized access and privacy leaks while giving users clear visibility into how their data is being used. It's like having a digital bodyguard specifically trained to protect against AI-related threats.

March 16, 2026
AI SecurityTencent InnovationPrivacy Tech
Volcano Engine Fortifies AI Assistants with New Security Shield
News

Volcano Engine Fortifies AI Assistants with New Security Shield

ByteDance's Volcano Engine has unveiled a major security upgrade for its ArkClaw AI assistant platform. The new safeguards tackle vulnerabilities exposed by open-source tools like OpenClaw, implementing cloud-native sandboxing and strict permission controls. This transforms potentially risky AI agents into accountable 'digital employees' with full behavioral tracking - crucial protection as businesses increasingly adopt generative AI.

March 12, 2026
AI SecurityEnterprise TechnologyCloud Computing