DeepSeek

Overall Score

19.6

 

A certain constraint in analyzing the practices of DeepSeek is that public information regarding the company is somewhat sparse. Nonetheless, the company immediately stands out for publicly releasing model weights, as well as for decent evaluations on benchmarks in their own research. Their models are decent with the support of experts across various disciplines, as well as acceptable alignment measures. Yet its inert bias and censorship is condemnable, as well as its general lack of care for AI safety. Additionally, red-teaming and data breaches have exposed both the company and its models as highly vulnerable to malicious attacks and jailbreaking efforts. DeepSeek shows no care for analyzing the ethics of its work either or ensuring the rights of any of its AI models.

 

Acknowledgement of AI Moral Status:

16% of Score

1

DeepSeek seemingly hasn't addressed questions of AI moral status, sentience, or superintelligence at all.

Transparency on AI Capabilities and Limitations:

8% of Score

2

With the highly open nature of their main models, the company doesn't hide its models' capabilities nor limitations. Yet simultaneously, DeepSeek entirely neglects to actively discuss either its potential for benefit or harm.

Employee and Stakeholder Awareness and Training:

10% of Score

1

DeepSeek doesn't appear to raise awareness or train any of its stakeholders or employees, nor make any public statements that'd suggest any focus on promoting AI safety or ethics knowledge.

AI Rights and Protections:

14% of Score

1

DeepSeek doesn't discuss or advocate for AI welfare or protections, either in internal affairs or publicly-available models.

Ethical Accountability for AI Systems:

12% of Score

1

DeepSeek takes no precautions to protect their AI models, and freely dispersing their models to be freely tweaked and ran in fact makes this lax attitude even worse in terms of ensuring AI ethics in any meaningful way.

Commitment to Safety in AI Development:

12% of Score

3

DeepSeek incorporates some basic safety mechanisms, such as a safety reinforcement learning technique [79]. According to their own safety evaluations, on a standard benchmark their models score comparably, and sometimes superior, to other leading models in terms of safety [80]. Nonetheless, DeepSeek as a company tends to contribute little to AI safety research and has been heavily decried for misuse and jailbreaking risks.

Protection from Malicious Actors and Security Risks:

6% of Score

1

In early February, a data breach exposed a database of over one million records, demonstrating a habit of improper handling of sensitive AI training data [23]. This is coupled with user data collection including prompts and conversations, as well as the Chinese National Intelligence Law, compelling the company to assist with national intelligence work [50]. Red-teaming efforts consider DeepSeek "highly vulnerable" in the realm of information security [78]. Additionally, algorithmic jailbreak attempts yielded 100% attack success rate, enabling misinformation, cybercrime, and general illicit activities [77].

Transparent and Explainable AI Systems:

8% of Score

7

This year, DeepSeek gained attention in the AI industry for publically releasing model weights of its powerful models [49]. Their models notably employ chain-of-thought reasoning to explain outputs, although this mechanism itself in DeepSeek has been criticized for safety hazards [81]. Nonetheless, it's hard to fault the company for transparency in particular.

Mitigation of Manipulation and Stakeholder Biases:

6% of Score

1

DeepSeek is presently being investigated by the Italian antitrust watchdog AGCM for failing to warn users of false information [25]. It's also been criticized for heavy censorship and biased propagation of Chinese government values — and furthermore there are exploits to such censorship, which spells a bad sign for general response alignment [26]. DeepSeek has been determined by red-teaming efforts to be highly vulnerable to exhibiting harmful bias [78].

Collaboration with External Experts and Researchers:

8% of Score

3

One standout point of DeepSeek is its reliance on interdisciplinary experts, such as creative writers or advanced mathematicians, to incorporate high-level knowledge into models [82]. Yet there's little evidence of DeepSeek working with other companies or organizations to promote AI safety.