Security

CSO

Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek

Oh someone's in DeepShi...


China-based AI biz DeepSeek may have developed competitive, cost-efficient generative models, but its cybersecurity chops are another story.

Wiz, a New York-based infosec house, says that shortly after the DeepSeek R1 model gained widespread attention, it began investigating the machine-learning outfit's security posture. What Wiz found is that DeepSeek – which not only develops and distributes trained openly available models but also provides online access to those neural networks in the cloud – did not secure the database infrastructure of those services.

That means conversations with the online DeepSeek chatbot, and more data besides, were accessible from the public internet with no password required.

This database contained a significant volume of chat history, backend data and sensitive information

"Within minutes, we found a publicly accessible ClickHouse database linked to DeepSeek, completely open and unauthenticated, exposing sensitive data," the firm said in an advisory Wednesday. "It was hosted at oauth2callback.deepseek.com:9000 and dev.deepseek.com:9000.

"This database contained a significant volume of chat history, backend data and sensitive information, including log streams, API Secrets, and operational details."

To make matters worse, Wiz said, the exposure allowed for full control of the database and potential privilege escalation within the DeepSeek environment, without any authentication or barrier to external access.

Using ClickHouse's HTTP interface, security researchers were able to hit a /play endpoint and run arbitrary SQL queries from the browser. With the SHOW TABLES; query, they obtained a list of accessible datasets.

One of those tables, log_stream, is said to have contained all sorts of sensitive data within the million-plus log entries.

According to Wiz, this included timestamps, references to API endpoints, people's plaintext chat history, API keys, backend details, and operational metadata, among other things.

The researchers speculate depending on DeepSeek's ClickHouse configuration, an attacker could have potentially retrieved plaintext passwords, local files, and proprietary data simply with the appropriate SQL command – though they did not attempt such actions.

"The rapid adoption of AI services without corresponding security is inherently risky," Gal Nagli, a cloud security researcher at Wiz, told El Reg.

"While much of the attention around AI security is focused on futuristic threats, the real dangers often come from basic risks - like the accidental external exposure of databases. Protecting customer data must remain the top priority for security teams, and it is crucial that security teams work closely with AI engineers to safeguard data and prevent exposure."

According to Wiz, DeepSeek promptly fixed the issue when informed about it.

DeepSeek, which offers free web and app, and paid-for API access to its CCP-censored models, did not immediately respond to a request for comment.

Its privacy policy for its online services make it clear it logs and stores full usage information on its servers in China. The Android and iOS app is not available in Italy after the Euro nation's data-protection watchdog started asking pointed questions about the use of people's personal data. Ireland is also said to be investigating.

The biz also upset OpenAI in more ways than one; the US lab famous for scraping the internet for training data believes DeepSeek used OpenAI's GPT models to produce material to train DeepSeek's neural networks. ®

Send us news
71 Comments

How nice that state-of-the-art LLMs reveal their reasoning ... for miscreants to exploit

Blueprints shared for jail-breaking models that expose their chain-of-thought process

Microsoft warns Trump: Where the US won't sell AI tech, China will

Rule hamstringing our datacenters is 'gift' to Middle Kingdom, vice chair argues

UK's new thinking on AI: Unless it's causing serious bother, you can crack on

Plus: Keep calm and plug Anthropic's Claude into public services

We meet the protesters who want to ban Artificial General Intelligence before it even exists

STOP AI warns of doomsday scenario, demands governments pull the plug on advanced models

Does terrible code drive you mad? Wait until you see what it does to OpenAI's GPT-4o

Model was fine-tuned to write vulnerable software – then suggested enslaving humanity

Microsoft expands Copilot bug bounty targets, adds payouts for even moderate messes

Said bugs 'can have significant implications' – glad to hear that from Redmond

LLM aka Large Legal Mess: Judge wants lawyer fined $15K for using AI slop in filing

Plus: Anthropic rolls out Claude 3.7 Sonnet

Despite Wall Street jitters, AI hopefuls keep spending billions on AI infrastructure

Sunk cost fallacy? No, I just need a little more cash for this AGI thing I’ve been working on

Why AI benchmarks suck

Anyone remember when Volkswagen rigged its emissions results? Oh...

India's top telco plans cloud PCs for its 475 million subscribers

PLUS: China bans AI leaders from visiting USA; Acer data leak suspect cuffed; and more

Microsoft names alleged credential-snatching 'Azure Abuse Enterprise' operators

Crew helped lowlifes generate X-rated celeb deepfakes using Redmond's OpenAI-powered cloud – claim

Three charged in Singapore with alleged link to illicit shipments of Nvidia GPUs to China

Accused face up to 20 years in prison