Boost Your iPhone’s Audio With This New DAC From iFi

The new iFi GO link 2 DAC is the most compact headphone amplifier that iFi has made to date. It’s 8% smaller and 29% lighter than the original version.

Final’s DX4000CL Are Closed-Back Headphones With An Open Secret

The new DX4000CL from Final offer a natural and richly textured tonal balance and are the next step in Final’s pursuit of closed-back headphone design.

Perplexity launches ‘Computer’ AI agent that coordinates 19 models, priced at $200 a month

Perplexity, the AI-powered search company valued at $20 billion, on Wednesday launched what it calls the most ambitious product in its three-year history: a multi-model agent orchestration platform called Computer that coordinates 19 different AI models to complete complex, long-running workflows entirely in the background.

The product, currently available only to Perplexity Max subscribers at $200 per month, is the company’s clearest articulation yet of a thesis it has been refining for more than a year: that AI models are not converging into general-purpose commodities but are instead specializing — and that the company best positioned to win the next era of AI is the one that can orchestrate all of them together.

“What has Perplexity been up to last two months? We’ve silently been working on the next big thing,” CEO Aravind Srinivas wrote on X, announcing that “Computer unifies every current capability of AI into a single system.” Srinivas said the system treats models as interchangeable tools rather than core products. “It’s multi-model by design,” he wrote. “When models specialise, they just become tools similar to the file system, CLI tools, connectors, browser, search.”

Computer arrives at a moment when the AI industry is grappling with a fundamental question: now that foundation models have become extraordinarily capable, who captures the value? The model makers — OpenAI, Anthropic, Google — or the companies that sit above them and turn raw intelligence into reliable, accurate products?

Perplexity is making a $20 billion bet on the latter.

Inside Computer: how Perplexity built a single interface that delegates work across Claude, Gemini, Grok and 16 other AI models

At its core, Computer functions as what Perplexity describes as “a general-purpose digital worker” — a system that can accept a high-level objective from a user, decompose it into subtasks, and delegate those subtasks to whichever AI model is best suited for each one. The Verge described it as existing “somewhere between OpenClaw and Claude Cowork,” referring to the viral open-source autonomous agent and Anthropic’s enterprise collaboration tool, respectively.

The platform’s central reasoning engine runs on Anthropic’s Claude Opus 4.6, which handles orchestration logic and coding tasks. Google’s Gemini powers deep research queries. Google’s Nano Banana generates images, and Veo 3.1 handles video. xAI’s Grok is deployed for lightweight, speed-sensitive tasks. OpenAI’s GPT-5.2 manages long-context recall and expansive web search. In total, the system coordinates 19 models on the backend, according to the company.

That model roster is not fixed. Perplexity says new models can be added as they demonstrate strength in specific domains, and the existing lineup will shift as models evolve. Users can also step into the orchestrator role themselves, manually assigning subtasks to particular models if they prefer.

What makes Computer distinct from existing agent tools is its combination of scope and accessibility. A user can describe a desired outcome — say, “Plan a weeklong trip to Japan, find flights under $1,200, and build a full itinerary with restaurant reservations” — and Computer will autonomously break that project into components, assign each to the right model, and work on it in the background. Perplexity says the system can operate quietly for extended periods, checking in with the user only when it genuinely needs input.

The enterprise data that convinced Perplexity no single AI model can do everything well

The intellectual foundation of Computer rests on data that Perplexity has been collecting across its enterprise customer base — data that, according to the company, no other AI company has access to at the same scale.

At a recent press briefing that VentureBeat attended with other reporters in San Francisco, Perplexity executives shared enterprise usage statistics that illustrated a dramatic shift in how businesses use AI models. In January 2025, more than 90 percent of enterprise tasks on the Perplexity platform were spread across just two models. By December 2025, no single model commanded more than 25 percent of usage across businesses and task types.

That shift, executives said, was driven partly by increasingly intelligent model routing on Perplexity’s side, and partly by a simple reality: models are getting better at different things, not the same things. A new frontier model emerged on average every 17.5 days in 2025, and each one brought distinct strengths rather than uniform improvement.

Claude, for instance, has emerged as the model of choice for software engineering tasks — a reputation so strong that even OpenClaw, the viral autonomous agent created by Austrian programmer Peter Steinberger (who was subsequently hired by OpenAI), was originally built on Claude’s code capabilities. But Claude’s strengths in coding do not translate to writing or creative generation, where Gemini tends to outperform. And in long-context retrieval and broad web search, GPT-5.2 holds advantages.

“What we’ve learned in this time is that they are not commoditizing. They’re specializing,” a senior Perplexity executive said at the briefing, characterizing Claude Opus 4.6 as “a terrible writer” while noting its coding prowess, and adding: “Everybody has job security on that one.”

This specialization dynamic creates what Perplexity sees as a structural advantage. A marketing team using Claude, executives argued, will generally produce worse results than one using Gemini. An engineering team using Gemini will underperform one using Claude. No company operates with only one type of team — and no single model can serve all of them equally well.

Why Perplexity says its cloud-based approach is safer than OpenClaw’s local-access model

Computer’s launch arrives in the immediate wake of OpenClaw, the open-source autonomous agent that went viral earlier this month and prompted OpenAI to hire its creator. OpenClaw captured the imagination of the AI community by demonstrating what a fully autonomous agent could accomplish when given broad access to a user’s entire digital ecosystem — files, email, messaging apps, API keys, and more.

But it also demonstrated the risks. In a widely shared incident this week, Meta AI security researcher Summer Yue posted screenshots on X of her frantic attempts to stop OpenClaw from deleting her entire email inbox — a process the agent had initiated and was refusing to halt. “I had to RUN to my Mac Mini like I was diffusing a bomb,” Yue wrote.

Perplexity has been vocal about why Computer runs entirely in the cloud rather than accessing a user’s local machine — an approach taken by rivals like Anthropic’s Claude and OpenAI’s Operator.

The company argues that local access creates unnecessary risk, comparing it to malware in how easily it can damage data or expose sensitive information. Computer instead operates inside what Perplexity describes as a safe and secure development sandbox, meaning security failures are contained and cannot spread to a user’s primary network or device. The company also said it has run thousands of tasks internally using Computer, from publishing web copy to building apps.

The distinction extends to accessibility. Where OpenClaw requires terminal access, API key configuration, and a dedicated machine (typically a Mac mini), Computer is designed to be invoked from a phone, a Slack message, or the Perplexity app. 

At the press briefing, executives elaborated on the philosophy, positioning Computer’s browser agent capabilities — built on Perplexity’s Comet browser technology — as central to the product. One executive noted that Perplexity’s browser agent usage numbers are three to five times higher than ChatGPT’s agent numbers published by The Information in January, despite Perplexity’s much smaller user base.

Perplexity’s revenue grew faster than its user base in 2025, and the company says it hasn’t even started trying

Perplexity’s product ambitions are backed by a business that, by the company’s own metrics, is growing faster than its user base — and executives say the company has barely begun to focus on monetization.

At the press briefing, executives disclosed that Perplexity grew users by 3.7x in 2025 and revenue by 4.7x, meaning the company is extracting more value from its existing users over time. Consumer subscriptions remain the largest revenue component, but the enterprise business is ramping with what executives acknowledged is a remarkably lean operation.

“We only have five people on our enterprise sales team,” one executive said, before adding that the company’s revenue per employee working on deals may be unmatched in the industry. Another executive noted that 92 percent of the Fortune 500 have Perplexity usage — though that figure encompasses employees signing up with personal accounts and work email addresses for the consumer version, not necessarily formal enterprise contracts.

A common enterprise sales conversation, executives said, starts with: “Did you know that there’s already 3,000 of your employees using Perplexity, and they’re using the consumer version that doesn’t adhere to all of your security policies?”

Notably, Perplexity is not pursuing advertising revenue, even as competitors like OpenAI move toward ad-supported models. Executives said advertising is fundamentally misaligned with the company’s accuracy mission. “The challenge with ads is, you know, a user will just start doubting everything,” one executive said. The company confirmed it has taken no economics on its shopping integrations and expressed doubt that any shopping-based monetization would materialize this year.

On the question of an IPO, Srinivas indicated the company has “very good properties of a company that can go public” given its low capital expenditure and healthy margins, but stopped short of committing to a timeline. Another executive warned that “a lot of IPO talk is hype” and that “if you over promise and under deliver the market punches you severely.”

TestingCatalog also reported this week that a new “Usage and Credits” settings area has appeared in Perplexity’s development builds, which would let users purchase additional credits to extend usage — potentially easing backlash from subscribers who saw their Deep Research query limits cut from roughly 500 per day to as few as 20 per month between late 2025 and early 2026.

Four of the ‘Magnificent Seven’ tech giants are already using Perplexity’s search API in production

Perhaps the least-discussed but most strategically significant element of Perplexity’s story is its search API business — an infrastructure play that positions the company not just as a consumer product but as a foundational layer for the broader AI ecosystem.

At the press briefing, executives revealed that Perplexity launched its search API approximately four months ago and already has four of the “Mag Seven” — the seven largest technology companies by market capitalization — using it in production at significant scale. “You guys cover the Mag Seven, you know that they don’t turn on a feature in production unless they’ve run rigorous evals and compared it,” one executive told reporters.

This disclosure suggests that the world’s largest technology companies have evaluated Perplexity’s search index against alternatives and concluded it is better optimized for AI-native use cases — a fundamentally different optimization target than Google’s traditional index, which was designed for humans scanning lists of links.

“Everything in our index is optimized, not for a human to see 10 blue links,” one executive explained. “It’s for an AI to be able to take those snippets and consume it in this context window and then reason through it.”

The company also confirmed it has fully independent search infrastructure, no longer relying on any third-party APIs from Google or Bing for its index — a significant departure from its earlier years.

For Chinese open-source models, which Perplexity uses in its orchestration stack, the company runs all inference from its own U.S. data centers, post-training the models for accuracy, removing what executives described as “state-infused propaganda,” and building custom inference kernels. The company open-sourced its methodology for depropagandizing Chinese models for others to use as well.

The search API creates a powerful data flywheel, executives argued: Perplexity can observe which snippets its search ranker surfaces for a given query, then track which of those snippets the LLM actually uses in its final output. That feedback loop makes the next query on a similar topic smarter — an advantage that pure API search businesses like Exa cannot replicate because they lack the consumer product generating user queries and feedback.

Copyright lawsuits and legal battles continue to shadow Perplexity’s rapid growth

Perplexity’s ambitions are not without complications. The company faces active lawsuits from multiple publishers, and the legal landscape grew more contentious this week.

As Business Insider’s Melia Russell reported, Perplexity filed a motion on February 24 in its ongoing legal battle with Dow Jones (publisher of The Wall Street Journal) and the New York Post, alleging that the publishers “cherry-picked” responses from Perplexity’s search engine to support their copyright claims. The company said it identified hundreds of prompts the publishers submitted that “were clear attempts to induce copyright-infringing answers,” including one instance where a user allegedly hit the “retry” button more than 50 times.

At the press briefing, Perplexity executives framed the broader copyright debate in historical terms, noting that waves of lawsuits have accompanied every major technology shift since radio. They expressed confidence that AI companies will ultimately prevail, particularly on the question of whether underlying knowledge — as distinct from unique creative expression — can be freely accessed by AI systems. “Countries have copyright law for one reason: to promote innovation,” one executive said, noting that the law protects unique expression while keeping the underlying knowledge open.

On user agents specifically, executives argued that a user’s AI agent is legally and technologically an extension of the user, not an independent actor. In the Amazon lawsuit, which challenges Perplexity’s ability to act as a purchasing agent on behalf of users, one executive offered a pointed analogy: “What Amazon’s claiming is that you shouldn’t be able to have your personal shopper be employed by you. It needs to be employed by them. They want you to use Rufus.”

Executives also clarified the company’s approach to citations, noting that citing a source like The New York Times (which is currently suing the company) does not necessarily mean Perplexity crawled that publication directly. “We can get the summary of that somewhere else, but we cite, we always try to cite that original source,” one executive said. “So drive that traffic to the New York Times if somebody clicks instead of driving them to a summary.”

What Perplexity Computer means for the future of AI: orchestration versus the single-model ecosystem

Computer’s launch crystallizes a tension that has been building in the AI industry for months. The major model makers — OpenAI, Anthropic, Google — have been racing to build end-to-end products that keep users within their ecosystems. OpenAI’s Codex and ChatGPT, Anthropic’s Claude Code and Cowork, Google’s Gemini — all assume that one model family can handle the full range of user needs.

Perplexity is making the opposite bet: that the future belongs to the orchestration layer, not the model layer. It is a bet with historical parallels. In the early days of cloud computing, the companies that built the best abstraction layers above commodity infrastructure — not the infrastructure providers themselves — often captured outsized value. Perplexity is positioning itself as that abstraction layer for AI.

The risk, of course, is that model makers could restrict API access or degrade service to platform competitors. Srinivas has said he isn’t worried, noting that he received congratulations from Anthropic and Google after Computer’s launch and that model makers benefit when their systems are part of broader workflows. But the AI industry’s history of platform dynamics suggests this détente may not last forever.

For enterprise technology leaders evaluating their AI strategies, Computer raises a practical question: should organizations standardize on a single model provider’s ecosystem, accepting its limitations in exchange for simplicity? Or should they invest in multi-model orchestration, gaining access to the best capabilities across providers at the cost of additional complexity?

Perplexity is betting that as models continue to specialize and the gap between their respective strengths widens, the answer will become obvious. The company’s enterprise usage data — showing a market that went from two-model dominance to no-model dominance in just 12 months — suggests the shift is already underway.

Computer is currently available to Perplexity Max subscribers, with a rollout to Pro and Enterprise users planned in the coming weeks. The company has also announced a developer event on March 11, where it plans to share more details about its search API, ranking embeddings, and the infrastructure powering its orchestration stack.

Alibaba’s new open source Qwen3.5-Medium models offer Sonnet 4.5 performance on local computers

Alibaba’s now famed Qwen AI development team has done it again: a little more than a day ago, they released the Qwen3.5 Medium Model series consisting of four new large language models (LLMs) with support for agentic tool calling, three of which are available for commercial usage by enterprises and indie developers under the standard open source Apache 2.0 license:

  • Qwen3.5-35B-A3B

  • Qwen3.5-122B-A10B

  • Qwen3.5-27B

Developers can download them now on Hugging Face and ModelScope. A fourth model, Qwen3.5-Flash, appears to be proprietary and only available through the Alibaba Cloud Model Studio API, but still offers a strong advantage in cost compared to other models in the West (see pricing comparison table below).

But the big twist with the open source models is that they offer comparably high performance on third-party benchmark tests to similarly-sized proprietary models from major U.S. startups like OpenAI or Anthropic, actually beating OpenAI’s GPT-5-mini and Anthropic’s Claude Sonnet 4.5 — the latter model which was just released five months ago.

And, the Qwen team says it has engineered these models to remain highly accurate even when “quantized,” a process that reduces their footprint further by reducing the numbers by which the model’s settings are stored from many values to far fewer.

Crucially, this release brings “frontier-level” context windows to the desktop PC. The flagship Qwen3.5-35B-A3B can now exceed a 1 million token context length on consumer-grade GPUs with 32GB of VRAM. While not something everyone has access to, this is far less compute than many other comparably-performant options.

This leap is made possible by near-lossless accuracy under 4-bit weight and KV cache quantization, allowing developers to process massive datasets without server-grade infrastructure.

Technology: Delta force

At the heart of Qwen 3.5’s performance is a sophisticated hybrid architecture. While many models rely solely on standard Transformer blocks, Qwen 3.5 integrates Gated Delta Networks combined with a sparse Mixture-of-Experts (MoE) system.The technical specifications for the Qwen3.5-35B-A3B reveal a highly efficient design:

  • Parameter Efficiency: While the model houses 35 billion parameters in total, it only activates 3 billion for any given token.

  • Expert Diversity: The MoE layer utilizes 256 experts, with 8 routed experts and 1 shared expert helping to maintain performance while slashing inference latency.

  • Near-Lossless Quantization: The series maintains high accuracy even when compressed to 4-bit weights, significantly reducing the memory footprint for local deployment.

  • Base Model Release: In a move to support the research community, Alibaba has open-sourced the Qwen3.5-35B-A3B-Base model alongside the instruct-tuned versions.

Product: Intelligence that ‘thinks’ first

Qwen 3.5 introduces a native “Thinking Mode” as its default state. Before providing a final answer, the model generates an internal reasoning chain—delimited by <think> tags—to work through complex logic.The product lineup is tailored for varying hardware environments:

  • Qwen3.5-27B: Optimized for high efficiency, supporting a context length of over 800K tokens.

  • Qwen3.5-Flash: The production-grade hosted version, featuring a default 1 million token context length and built-in official tools.

  • Qwen3.5-122B-A10B: Designed for server-grade GPUs (80GB VRAM), this model supports 1M+ context lengths while narrowing the gap with the world’s largest frontier models.

Benchmark results validate this architectural shift. The 35B-A3B model notably surpasses much larger predecessors, such as Qwen3-235B, as well as the aforementioned proprietary GPT-5 mini and Sonnet 4.5 in categories including knowledge (MMMLU) and visual reasoning (MMMU-Pro).

Pricing and API integration

For those not hosting their own weights, Alibaba Cloud Model Studio provides a competitive API for Qwen3.5-Flash.

  • Input: $0.1 per 1M tokens

  • Output: $0.4 per 1M tokens

  • Cache Creation: $0.125 per 1M tokens

  • Cache Read: $0.01 per 1M tokens

The API also features a granular Tool Calling pricing model, with Web Search at $10 per 1,000 calls and Code Interpreter currently offered for a limited time at no cost.

This makes Qwen3.5-Flash among the most affordable to run over API among all the major LLMs in the world. See a table comparing them below:

Model

Input

Output

Total Cost

Source

Qwen 3 Turbo

$0.05

$0.20

$0.25

Alibaba Cloud

Qwen3.5-Flash

$0.10

$0.40

$0.50

Alibaba Cloud

deepseek-chat (V3.2-Exp)

$0.28

$0.42

$0.70

DeepSeek

deepseek-reasoner (V3.2-Exp)

$0.28

$0.42

$0.70

DeepSeek

Grok 4.1 Fast (reasoning)

$0.20

$0.50

$0.70

xAI

Grok 4.1 Fast (non-reasoning)

$0.20

$0.50

$0.70

xAI

MiniMax M2.5

$0.15

$1.20

$1.35

MiniMax

MiniMax M2.5-Lightning

$0.30

$2.40

$2.70

MiniMax

Gemini 3 Flash Preview

$0.50

$3.00

$3.50

Google

Kimi-k2.5

$0.60

$3.00

$3.60

Moonshot

GLM-5

$1.00

$3.20

$4.20

Z.ai

ERNIE 5.0

$0.85

$3.40

$4.25

Baidu

Claude Haiku 4.5

$1.00

$5.00

$6.00

Anthropic

Qwen3-Max (2026-01-23)

$1.20

$6.00

$7.20

Alibaba Cloud

Gemini 3 Pro (≤200K)

$2.00

$12.00

$14.00

Google

GPT-5.2

$1.75

$14.00

$15.75

OpenAI

Claude Sonnet 4.5

$3.00

$15.00

$18.00

Anthropic

Gemini 3 Pro (>200K)

$4.00

$18.00

$22.00

Google

Claude Opus 4.6

$5.00

$25.00

$30.00

Anthropic

GPT-5.2 Pro

$21.00

$168.00

$189.00

OpenAI

What it means for enterprise technical leaders and decision-makers

With the launch of the Qwen3.5 Medium Models, the rapid iteration and fine-tuning once reserved for well-funded labs is now accessible for on-premise development at many non-technical firms, effectively decoupling sophisticated AI from massive capital expenditure.

Across the organization, this architecture transforms how data is handled and secured. The ability to ingest massive document repositories or hour-scale videos locally allows for deep institutional analysis without the privacy risks of third-party APIs.

By running these specialized “Mixture-of-Experts” models within a private firewall, organizations can maintain sovereign control over their data while utilizing native “thinking” modes and official tool-calling capabilities to build more reliable, autonomous agents.

Early adopters on Hugging Face have specifically lauded the model’s ability to “narrow the gap” in agentic scenarios where previously only the largest closed models could compete.

This shift toward architectural efficiency over raw scale ensures that AI integration remains cost-conscious, secure, and agile enough to keep pace with evolving operational needs.

New York Sues Valve Over Loot Boxes, Which It Claims Are Gambling

Valve is facing a massive lawsuit over the use of loot boxes in its games, which could have major ramifications for CS2 markets.

Panasonic Unveils 2026 Range Of OLED And LCD TVs

It’s business as (almost) usual for Panasonic despite – or perhaps because of! – its new Skyworth partnership.

The AI Megacycle: Five Forces Reshaping The Venture Market In 2026

AI’s 2026 megacycle is concentrating capital, compressing startup growth, reshaping IPO and forcing new monetization models that reward durable, AI native advantages.

Antigravity A1 drone review

Our thoughts on the world’s first dedicated 360-degree drone — the Antigravity A1.

The Cheapest 2026 EV Is Here: How The 2027 Chevy Bolt Price Stacks Up

The 2027 Chevy Bolt is the cheapest EV in 2026 and a strong contender for the best bang-for-the-buck electric vehicle. And adds to General Motors large stable of EVs.

Galaxy S26 Ultra Interview: The Dawn Of The ‘AI Phone’

Samsung’s Richard Chang and Kadesh Beckford explain how the new Galaxy S26 Ultra’s reach, openness, and confidence make it Samsung’s next-generation AI Phone.