Microsoft is investigating if DeepSeek illegally accessed OpenAI data to train its AI model

DragonSlayer101

Posts: 577   +3
Staff
In context: AI startup DeepSeek stunned the world with the release of its R1 AI model, which can mimic human reasoning at a level comparable to the best OpenAI models to date. While the company has received widespread acclaim for its achievement, it is now reportedly under investigation by Microsoft and OpenAI for allegedly accessing OpenAI's data output illegally to train its AI model.

According to unnamed sources cited by Bloomberg, the probe started last fall after Microsoft's security researchers discovered that a group linked to DeepSeek had accessed a large amount of data through OpenAI's API. While developers can legally pay for a license to use this data in their applications, Microsoft suspects that DeepSeek's actions violated OpenAI's terms of service.

Following the discovery, Microsoft informed OpenAI that it had detected an attempt by a DeepSeek-affiliated group to bypass restrictions on how much data a single party can access. The two companies then launched a joint investigation into the incident, which is now being treated as a potential data breach.

An OpenAI spokesperson declined to confirm the alleged data theft but told Reuters that the company employs cutting-edge "countermeasures" to protect its intellectual property. They added that OpenAI is working closely with the U.S. government to prevent its AI models from falling into the hands of foreign adversaries and competitors.

However, the White House's AI and crypto czar, David Sacks, was less restrained in his response to the report. In an interview with Fox News, he claimed there was "substantial evidence" that DeepSeek had accessed OpenAI's data in an unethical and possibly illegal manner. Microsoft refused to comment, while DeepSeek could not be reached for its reaction on the controversy.

With the launch of its all-conquering AI model, DeepSeek appears poised to challenge OpenAI, Google, and Meta in the field of AI research. However, if the allegations turns out to be accurate, it could spell trouble for the Chinese AI firm, which saw its new app overtake ChatGPT to become the top free app on both the App Store and Play Store in the U.S. this week.

DeepSeek also received unexpected praise from President Trump this week when he described the release of the R1 LLM as a "positive" development and an "asset" for the global tech sector. He noted that if DeepSeek delivers on its promise of accelerating AI training at a lower cost, "that's good (and) I view that as a positive."

Permalink to story:

 
OpenAI discovering what it feels like to have your data taken without permission is like a pirate getting mad that someone else is looting their treasure.

The irony aside, this raises an important question: If OpenAI is upset about this, does that mean they do believe there should be some kind of ownership or protection over training data? Because that’s not the argument they’ve been making in court.
 
AI is a collaborative effort, similar to Wikipedia. Therefore, there's no need to feel embarrassed if a more efficient model temporarily surpasses the current one. Neural network weights are not a form of data that can be copyrighted. Without copyright, you cannot license or enforce rules related to them.
Furthermore, the outputs generated by any neural network stem from synthetic intelligence that isn't legally recognized as having a personality. Consequently, these outputs are not subject to intellectual property rights and are by default in the public domain. This is because freedom of access to information is a fundamental right.
Training AI is a hobby for the ultra-wealthy, a side project and a path to make a gift to humanity. Due to its nature, AI isn't inherently suited for large-scale commercial exploitation.
I was writing a few weeks ago that a $200/month and $2,000/month will backfire if a better model becomes available for less...
 
All the pearl clutching has me in tears with laughter. As others have pointed out, Open AI doesn't have a leg to stand on. I'm sure the government puppets and their billionaire owners are seething at the thought of someone else stealing data better.
 
<<"Substantial evidence" of data theft, says Trump's AI and crypto czar>>

Says the orange turd who gutted cyber security in his first term and opened tech trade with china
 
The openAi legally used the copyright materials to train and assassinate the whistleblower and made it look like a suicide.

SO only you have the right to steal and use private data.
 
Training AI is a hobby for the ultra-wealthy, a side project and a path to make a gift to humanity. Due to its nature,....
I nearly fell off my chair laughing at this statement............."Ultra-wealthy.....gift to humanity"...........S**t, did you sit back and read your post at all? How could those words possibly belong in the same sentence? Oh...unless you happen to be Musky or Zuckerdick trolling...
 
Of course the U.S tech companies will try to find fault on them as that Chinese company showed them how they can have more efficient results with far less money.
That assuming they told the truth. If I know anything about communist or post-communist countries, it is that they are built on lies. Lies are not seen as a negative thing there, lies are woven deep into the fabric of society.
This is my biggest problem reading news from China and specially seeing people assuming 100% of what they say is true.
I do not know what it is specifically what makes countries that tasted communism to rely on lies so much, but I have studied how every bit of USSR was built around lies. Lies about harvests. Lies about population, I recall Stalin killed people who presented population growth numbers that were lower than he expected (the next executives of course showed tremendous population growth). They lied about everything like if it was their air. And when at the end of 80s Gorbachev started to allow a bit of truth to be told, the entire empire started melting like ice-cream on a summer day.
 
I know some people will take what I said wrong. But what I said comes from observing a communistic country not drastically different to what China looked like before it started to work with the West.
They both had mass starvation, concentration camps, 0 freedoms, bloody suppressions of protests. And a lot of lies.

Taking everything that comes from China as 100% truth is like sending sick Brad Pit money to help pay for treatment.
 
Really interesting now as cat is out of the bag
Deepseek R1 is open source - so every man and his horse can download it

people already have , they can rip out censorship etc attract FBI attention etc

Now every country can have AI cheaply - take open source models and improve them
So who will will big expensive Meta, Google, MS, Apple or the WWW with open source LLM

This is much bigger than people think - I imagine individuals and collectives can now build purpose built LLM for $3000 to $300 000

Don't like costs of big guys , want less restrictions.

To my TS readers - another money making opportunity
Start a KS to built an model for X purpose

Not sure how open source game playing Models are - start a KS to play the most common board games , with strech goals of Catan, Dominion etc etc
Start a Role playing KS LLM with a Dungeon Master etc

 
Back