Read full article about: Microsoft's Bing team open-sources "Harrier" embedding model
Microsoft's Bing team (yes, really) has released "Harrier," an open-source embedding model. Harrier supports more than 100 languages, offers a 32,000-token context window, and was trained on over two billion examples plus synthetic data from GPT-5. According to the team, Harrier takes the top spot on the multilingual MTEB v2 benchmark and outperforms proprietary models from OpenAI and Amazon.
| Rank (Borda) | Model | Zero-shot | Active Params (B) | Total Params (B) | Embedding Dim | Max Tokens |
|---|---|---|---|---|---|---|
| 1 | harrier-oss-v1-27b | 78% | 25.6 | 27.0 | 5376 | 131072 |
| 2 | KaLM-Embedding-Gemma3-12B-2511 | 73% | 10.8 | 11.8 | 3840 | 32768 |
| 3 | llama-embed-nemotron-8b | 99% | 7.0 | 7.5 | 4096 | 32768 |
| 4 | Qwen3-Embedding-8B | 99% | 6.9 | 7.6 | 4096 | 32768 |
| 5 | gemini-embedding-001 | 99% | 3072 | 2048 | ||
| 6 | Qwen3-Embedding-4B | 99% | 3.6 | 4.0 | 2560 | 32768 |
| 7 | Octen-Embedding-8B | 99% | 6.9 | 7.6 | 4096 | 32768 |
| 8 | F2LLM-v2-14B | 88% | 13.2 | 14.0 | 5120 | 40960 |
| 9 | F2LLM-v2-8B | 88% | 6.9 | 7.6 | 4096 | 40960 |
| 10 | harrier-oss-v1-0.6b | 78% | 0.440 | 0.596 | 1024 | 32768 |
Alongside the full 27-billion-parameter model, the team released two smaller variants—0.6B and 270M—designed to run on less powerful hardware. All three models are available on Hugging Face under the MIT license. Going forward, the team plans to integrate the technology into Bing and into new grounding services for AI agents.
Embedding models handle the searching, retrieving, and organizing of information AI systems need for accurate answers. According to Microsoft, they're becoming increasingly critical as AI agents independently take on more complex, multi-step tasks.
Comment
Source: Microsoft Bing Blog
Read full article about: China actively targeting Taiwan's chip talent and technology, security report says
China is actively trying to poach Taiwan's semiconductor expertise and talent to circumvent international technology restrictions, according to a report from Taiwan's National Security Bureau cited by Reuters.
The report says China is using indirect channels to recruit talent, steal technology, and acquire controlled goods. Taiwan is home to TSMC, the world's largest contract chipmaker and a key supplier to Nvidia and Apple.
In the first quarter of 2026 alone, the report logged more than 170 million attempted cyberattacks on Taiwan's government network. The agency also warns that China could try to influence Taiwan's local elections later this year using deepfakes and fabricated polls.
Read full article about: Bezos' Project Prometheus hires xAI co-founder from OpenAI
Jeff Bezos' startup Project Prometheus has hired Kyle Kosic, a co-founder of Elon Musk's xAI who most recently worked at OpenAI, the Financial Times reports. Kosic led the infrastructure behind xAI's Colossus supercomputer and will continue working on AI infrastructure at Prometheus.
The startup, led by Bezos and former Google executive Vikram Bajaj, is building AI systems designed to understand the physical world with a focus on tasks in areas like engine design and engineering. Prometheus has already hired hundreds of employees across San Francisco, London, and Zurich. According to the FT, Bezos and Bajaj are looking to raise tens of billions of dollars for a permanent investment vehicle that would acquire stakes in companies across industries like aerospace and architecture.
Read full article about: OpenAI, Anthropic, and Google team up against unauthorized Chinese model copying
OpenAI, Anthropic, and Google have started working together to combat the unauthorized copying of their AI models by Chinese competitors, according to Bloomberg. The three companies are sharing information through the "Frontier Model Forum," founded in 2023, to detect so-called adversarial distillation. In distillation, the outputs of an existing AI model are used to train a cheaper copycat model. One of the first examples was Stanford's Alpaca model, which demonstrated the feasibility of the approach, but the practice has since become a real problem for US companies.
US authorities estimate that adversarial distillation costs American AI labs billions of dollars in lost revenue each year, Bloomberg reports. OpenAI had already warned Congress in February that Deepseek was using increasingly sophisticated methods to extract data from US models. Anthropic identified Deepseek, Moonshot, and Minimax as actors involved in the practice. The collaboration mirrors how the cybersecurity industry operates, where companies routinely share attack data with each other.
Read full article about: Meta plans to open-source parts of its new AI models
Meta is planning to release versions of its new AI models as open source, according to Axios. These would be the first models developed under the leadership of Alexandr Wang, who joined Meta in 2025 as part of a nearly $15 billion deal with Scale AI.
Unlike its approach with the Llama models, though, Meta plans to keep some components proprietary and review safety risks before releasing anything. The largest models won't be made publicly available either.
According to the report, Wang sees Meta as a counterweight to Anthropic and OpenAI, which focus more heavily on government and enterprise customers. Meta's strategy instead centers on consumer reach through WhatsApp, Facebook, and Instagram. Axios's sources say Meta already knows the new models won't match the competition in every area.
Meta employees compete for token consumption on an internal AI leaderboard
At Meta, employees compete for titles like “Token Legend,” “Model Connoisseur,” and “Cache Wizard” on an internal leaderboard that ranks AI token consumption. But burning through more tokens doesn’t automatically mean getting more done.
Read full article about: Anthropic signs multi-gigawatt TPU deal with Google and Broadcom
Anthropic has signed a deal with Google and Broadcom for multiple gigawatts of TPU computing capacity, set to come online starting in 2027. Most of the infrastructure will be built in the United States. The company points to surging demand as the reason for the expansion: its annualized revenue rate now exceeds $30 billion, up from roughly $9 billion at the end of 2025. The number of enterprise customers generating more than $1 million in annual revenue has doubled since February, surpassing 1,000.
Anthropic trains Claude on a mix of hardware: Amazon's AWS Trainium, Google's TPUs, and Nvidia's GPUs. This makes Claude the only one of the three major AI models available across all three major cloud platforms (AWS, Google Cloud, and Microsoft Azure). That said, Anthropic notes that Amazon remains its most important cloud partner.
OpenAI's safety brain drain finally gets an explanation and it's just Sam Altman's vibes
“My vibes don’t really fit.” In a new New Yorker profile based on over 100 interviews, Sam Altman explains why safety researchers keep leaving OpenAI and why shifting commitments others might call deception are just part of the job.