Ad
Short

According to a study in Nature, Chinese AI company Deepseek trained its R1 language model for only $294,000. The run used 512 Nvidia H800 chips developed specifically for the Chinese market. Nvidia confirmed the chips were delivered before U.S. export restrictions took effect. In its technical paper, Deepseek also admitted to using A100 GPUs during preparation for a smaller prototype, after U.S. officials had earlier suspected the company of holding unauthorized H100s.

The figure does not include the much larger costs of training Deepseek’s underlying V3 foundation model. Estimates for that project vary widely, ranging from tens of millions to several hundred million dollars depending on the source.

Deepseek’s claim of unusually low training costs previously rattled global tech markets, triggering sharp declines in the share prices of major AI hardware and software companies.

Ad
Ad
Ad
Ad
Short

Wired reports that OpenAI is stepping back into robotics, with new hiring pointing toward work on humanoid machines.

According to job postings, the company is assembling a team focused on training robots through teleoperation and simulation. OpenAI is also seeking engineers specializing in sensing and prototyping. The listings describe the team’s mission as building "general-purpose robots" that could help push progress toward AGI.

It’s not confirmed that the effort targets humanoids, but signs point that way. New hires include Stanford researcher Chengshu Li, who worked on benchmarks for humanoid household robots. That makes it likely OpenAI’s robotics push could center on humanlike systems.

OpenAI shut down its robotics work in 2020, citing a lack of training data. But the company began posting robotics roles again in January, signaling a renewed focus on physical AI after a five-year pause.

Ad
Ad
Short

Google DeepMind has introduced a new language model called VaultGemma, designed with a focus on privacy. It is the largest open model to date trained from scratch with differential privacy, containing 1 billion parameters.

Normally, large language models can memorize parts of their training data, including sensitive information like names, addresses, or entire documents. Differential privacy avoids this by adding controlled random noise during training, making it statistically impossible to trace the model's outputs back to specific examples. In theory, even if VaultGemma were trained on confidential documents, those documents could not be reconstructed later.

According to Google, early tests confirm that the model does not reproduce training data. The tradeoff is performance: its output is roughly comparable to non-private LLMs released about five years ago.

The model weights are openly available on Hugging Face and Kaggle.

Google News