# Upgrades to the Parallel Search & Extract APIs

Tags:Benchmarks
Reading time: 5 min
Search & Extract Benchmarks

Today, we’re announcing major upgrades to the quality and capabilities of our Search and Extract APIs, and making both generally available. These enhancements deliver industry-leading accuracy and token efficiency for web search across diverse use cases, including company research, coding agents, multilingual queries, finance, and broad internet research.

## Highlights

  • - New streamlined search modes catering to interactive and background agents
  • - Pareto optimal performance on challenging public benchmarks
  • - Leading performance on benchmarks focused on specialized knowledge work
  • - Superior compression of webpage context using the Extract API
  • - Improvements to global index coverage and multi-lingual capabilities

## The best web search for AI agents in interactive and background use cases

As AI agents have become better at handling knowledge work that needs the web, two main product patterns have emerged: interactive workflows and long-running background tasks.

Interactive workflows support real-time collaboration between a user and an agent. They are shaped by latency requirements and frequently focus on targeted web research or fetching specific information.

Background tasks work more independently on ambiguous, multi-step questions spanning several topics. These searches typically require deeper retrieval and orchestration of multiple tool calls.

The Search API now offers two modes designed for these distinct use cases.

**Basic mode** is optimized for quick retrieval. It delivers P50 latency under one second and P90 under two seconds, making it a good fit for interactive, human-in-the-loop experiences.

**Advanced mode** is built for depth. It spends more time querying, reranking, and compressing results across both general-purpose and specialized indexes. Traditional search APIs often force agents into a series of sequential searches, which increases latency and expands the context window. Advanced mode resolves more in a single call.

Search ModeMedian LatencyBest for
Basic1sSingle-hop, latency-sensitive agents, e.g., human-in-loop chat, interactive experiences
Advanced3sMulti-hop background agents that can tolerate extra latency for better depth, e.g., code review agents, deep research
BrowsecompHumanity's Last ExamSealQA-Seal0WebWalkerFreshQA
379,38TAVILY42% / 973CPMEXA40% / 1160CPMPARALLEL BASIC53% / 600CPMPARALLEL ADVANCED51% / 379CPM

COST (CPM)

ACCURACY (%)

Loading chart...

CPM: USD per 1000 requests. Cost is shown on a Log scale.

Others
Parallel
Benchmark comparison across Cost (CPM) and Accuracy (%). CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.


**Testing dates**

April 19-21, 2026

### Browsecomp

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Others   | Tavily            | 973        | 42           |
| Others   | Exa               | 1160       | 40           |
| Parallel | Parallel Basic    | 600        | 53           |
| Parallel | Parallel Advanced | 379        | 51           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.


**Testing dates**

April 19-21, 2026

### Humanity's Last Exam

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Parallel | Parallel Basic    | 451        | 58           |
| Parallel | Parallel Advanced | 315        | 56           |
| Others   | Exa               | 522        | 57           |
| Others   | Tavily            | 538        | 54           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.


**Testing dates**

April 19-21, 2026

### SealQA

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Parallel | Parallel Basic    | 258        | 45           |
| Parallel | Parallel Advanced | 191        | 41           |
| Others   | Tavily            | 243        | 45           |
| Others   | Exa               | 326        | 41           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to MAX_TOOL_CALLS=25 tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.

**Testing dates**

April 19-21, 2026

### WebWalker

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Others   | Exa               | 210        | 74           |
| Others   | Tavily            | 202        | 71           |
| Parallel | Parallel Advanced | 101        | 73           |
| Parallel | Parallel Basic    | 155        | 71           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.


**Testing dates**

April 19-21, 2026

### FreshQA

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Parallel | Parallel Advanced | 49         | 79           |
| Parallel | Parallel Basic    | 90         | 77           |
| Others   | Exa               | 84         | 78           |
| Others   | Tavily            | 89         | 78           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.


**Testing dates**

April 19-21, 2026

### FRAMES

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Parallel | Parallel Advanced | 93         | 87           |
| Parallel | Parallel Basic    | 165        | 84           |
| Others   | Exa               | 169        | 87           |
| Others   | Tavily            | 189        | 83           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

We evaluated search providers against five open benchmarks covering complementary aspects of agentic search: BrowseComp (hard multi-hop questions that require navigating the live web), Frames (multi-document factoid reasoning), FreshQA (time-sensitive questions where the correct answer depends on recent web information), HLE (Humanity's Last Exam — expert-level academic questions spanning math, science, and humanities), SealQA (ambiguity-robust factoid QA with intentionally misleading snippets), WebWalker (tasks designed around following links across pages to find an answer).

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.


**Testing dates**

April 19-21, 2026

## The best web search API for specialized knowledge work

We believe the best web search API for agents is a strong general-purpose system with the scale, depth, and ranking quality to perform well across many kinds of knowledge work.

Our API is built on that foundation. It offers broad coverage of the web while going deep in the domains that matter most, so agents can retrieve high-quality results for both general research and specialized tasks. Whether the job is debugging against current library documentation or researching a prospect’s funding history and tech stack, the same strengths apply: a large index, deep retrieval, and strong ranking.

Advanced mode is designed for the most demanding workflows, where agents benefit from deeper retrieval and more precise ranking across both broad web content and domain-relevant sources.

On our internal dataset focused on company search, Parallel’s Search API outperforms alternatives in finding highly relevant candidates that meet the specified search criteria.

Illustration demonstrating deep research API concepts, web search capabilities, or AI agent integration features
![](https://cdn.sanity.io/images/5hzduz3y/production/488a6088450e4127de0e02a9ac77361201f5f56d-2000x1056.jpg)

Company Benchmark Methodology

**Dataset**

Queries targeting companies across categories, filtering by sector, funding stage, geography, investors, revenue, and team composition. Queries were generated using an LLM.

**Sample queries**

The benchmark queries are intentionally complex — the kind of multi-constraint searches that break general-purpose web search:

  • - "Biotech and solar energy startups serving healthcare providers, backed by IVP"
  • - "Series B SaaS companies focused on manufacturers and industrial production."

**Evaluation setup:**

250 queries, 10 results per query, single-graded by GPT-5.4-mini. The LLM judge graded every returned URL on whether it represents a genuine match for the query objective.

We measure **precision** which represents the fraction of returned results that are relevant — the metric that matters most for candidate generation, where downstream workflows depend on signal quality.

Similarly, for queries generated by coding agents as part of development workflows, Parallel’s Search API provides the most relevant code snippets and documents.

Coding
154,74PARALLEL STANDARD82% / 154CPMPARALLEL BASIC81% / 269CPMEXA80% / 331CPMTAVILY75% / 352CPM

COST (CPM)

ACCURACY (%)

Loading chart...

CPM: USD per 1000 requests. Cost is shown on a Log scale.

Parallel
Others
Benchmark comparison across Cost (CPM) and Accuracy (%). CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

A proprietary coding dataset derived from production queries to Parallel’s search API.

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report the accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.

**Testing dates**

April 19-21, 2026

### Coding

| Series   | Model             | Cost (CPM) | Accuracy (%) |
| -------- | ----------------- | ---------- | ------------ |
| Parallel | Parallel Standard | 154        | 82           |
| Parallel | Parallel Basic    | 269        | 81           |
| Others   | Exa               | 331        | 80           |
| Others   | Tavily            | 352        | 75           |

CPM: USD per 1000 requests. Cost is shown on a Log scale.

**Dataset**

A proprietary coding dataset derived from production queries to Parallel’s search API.

**Evaluation methodology**

Every task is run through a shared deep-research harness: a single GPT-5.4 agent is given two tools (web search and web fetch) with an iterative budget of up to `MAX_TOOL_CALLS=25` tool calls per question. The agent plans sub-queries, fans out searches, fetches specific pages when snippets are insufficient, and returns an answer when it exhausts the number of allowed tool calls or has sufficient information to answer the question. Each answer is then LLM-graded by GPT-5.4. We report the accuracy of the final answer.

We measure accuracy and overall cost, which includes LLM token costs and tool call costs.

**Testing dates**

April 19-21, 2026

## Combine Search with Extract for the deepest, most token-efficient multi-hop agent

Search and Extract are designed to work together. Search finds the right sources with high precision. Extract turns those sources into concise, task-relevant context that an agent can use efficiently.

The Parallel Extract API is especially useful when the source is long or dense, such as a 200-page SEC filing, an investor document, or a complex technical page. Passing the full source into context increases token usage and makes downstream reasoning less focused.

The Parallel Extract API solves this by letting the agent specify an extraction objective. After Search identifies the right source, Extract returns only the content most relevant to that objective. The result is a more efficient retrieval pipeline: Search provides breadth and precision, while Extract compresses large pages and documents into high-signal context for the next reasoning step.

Together, they give agents the best of both: high-quality retrieval and high-quality compression in a single workflow.

On FinanceBench[FinanceBench](https://docs.patronus.ai/docs/research_and_differentiators/financebench), built around real-world workflows across SEC filings, earnings reports, and multi-page investor documents, the extract API consistently succeeds at getting the highest recall for any number of tokens returned.

Illustration demonstrating deep research API concepts, web search capabilities, or AI agent integration features
![](https://cdn.sanity.io/images/5hzduz3y/production/8d607881c601e54f00329d647527800d4d58ef8c-2000x1706.jpg)

Finance Benchmark Methodology

**Dataset**

The eval sampled 136 questions from the finance bench QA dataset.

 

**Evaluation setup**

For each evaluation question, the harness runs the extract APIs against the question's URLs and objective, then judges the response using an LLM-as-judge based on the reference answer.

 

**Evaluation**

1. Combine all extracted content returned by the engine into a single excerpt.

2. LLM judge (GPT 5.4 mini) determines whether the combined excerpt contains enough information to correctly answer the question, given a known reference answer. The judge returns a short chain of reasoning plus a binary verdict: *sufficient* or *insufficient*.

3. Aggregate recall across the dataset is the fraction of questions the extracted content could answer. If the engine returns no content, the question is scored zero without calling the judge.

## We’ve expanded Parallel Search to more corners of the world

Sophisticated AI products serve users globally. A GTM agent may prospect across markets, a compliance agent may track filings across jurisdictions, and a research assistant may serve users in many languages. These workflows require a search layer that can retrieve fresh, relevant content across languages and regions with consistent quality.

The Parallel Search API now supports this natively, giving agents a stronger foundation for global workflows.

  • - **Global index coverage: **Our index now spans web pages from 30+ countries, including Korean, Chinese, Japanese, Spanish, Indian, German, Arabic, Portuguese, and Russian language content.
  • - **Multilingual query input: **Submit queries in any language. The Search API handles retrieval across languages without additional configuration.
  • - **Location parameter: **Set a target location to refine results to a specific region.
Illustration demonstrating deep research API concepts, web search capabilities, or AI agent integration features
![](https://cdn.sanity.io/images/5hzduz3y/production/754f1331567a72c8b01834653f864fe239982d98-2192x1538.jpg)

Multi-Lingual Benchmark Methodology

**Dataset**

Starting from OpenAI's **SimpleQA**, each question was translated into all 26 languages from Apple's **MKQA** benchmark. Translations were carried out using LLMs and were spot checked by human annotators fluent in the respective languages. The reference answer is also translated/localized per language, giving us paired queries where the _same fact_ is asked across Latin-European, Cyrillic, CJK, Semitic (RTL), and SE Asian scripts.

 

**Evaluation methodology**

100 questions from each language; each query is sent to the search provider as-is, and the results are fed into an LLM for synthesis. The final answer is graded using an LLM-as-a-judge.

 

We report 2 metrics:

  • - **The fraction of answers marked as correct**, as reported by the judge. The grader sees the question, the gold answer, and the predicted answer, and returns CORRECT / INCORRECT / NOT_ATTEMPTED. We grade with an LLM rather than string EM/F1 because the agent's predicted answers are full sentences in many scripts — "The 2010 recipient was Michio Sugeno" in English, ميتشيو سوغينو transliterated variants in Arabic, etc. What we care about is whether the agent got the answer right, which the LLM judge captures directly.
  • - **Cross-lingual retrieval ability**: Because every question is paired across all 26 languages, the drop in accuracy when switching from English to non-English languages for the same question isolates the provider's cross-lingual retrieval quality. A small drop means the index and retrieval pipeline handle non-English queries nearly as well as English; a large drop means the provider is English-centric and falls off when asked to cross script families.

Parallel builds web infrastructure for AI. We give AI agents structured, grounded access to the open web, enabling agents to find, extract, monitor, and reason over information at a scale and quality no human workflow can match. Our infrastructure is powered by a rapidly growing proprietary index of the global internet, built from the ground up for AI consumption.

Fortune 100s and leading frontier AI companies, including Harvey, Manus, Modal, Starbridge, and Profound, rely on Parallel for grounding, fact-checking, contract monitoring, and high-quality content generation.

Get started on our Developer Platform[Developer Platform](https://platform.parallel.ai/) or read the documentation[documentation](https://docs.parallel.ai/).

Parallel avatar

By Parallel

April 21, 2026

## Related Posts60

How Finch is scaling plaintiff law with AI agents that research like associates

- [How Finch is scaling plaintiff law with AI agents that research like associates](https://parallel.ai/blog/case-study-finch)

Tags:Case Study
Reading time: 3 min
Genpact and Parallel Web Systems Partner to Drive Tangible Efficiency from AI Systems

- [Genpact and Parallel Web Systems Partner to Drive Tangible Efficiency from AI Systems](https://parallel.ai/blog/genpact-parallel-partnership)

Tags:Partnership
Reading time: 4 min
Genpact & Parallel

- [How Genpact helps top US insurers cut contents claims processing times in half with Parallel ](https://parallel.ai/blog/case-study-genpact)

Tags:Case Study
Reading time: 4 min
DeepSearchQA: Parallel Task API benchmarks deepresearch

- [A new deep research frontier on DeepSearchQA with the Task API Harness](https://parallel.ai/blog/deepsearchqa-taskapi-harness)

Tags:Benchmarks
Reading time: 7 min
How Modal saves tens of thousands annually by building in-house GTM pipelines with Parallel

- [How Modal saves tens of thousands annually by building in-house GTM pipelines with Parallel](https://parallel.ai/blog/case-study-modal)

Tags:Case Study
Reading time: 4 min
Opendoor and Parallel Case Study

- [How Opendoor uses Parallel as the enterprise grade web research layer powering its AI-native real estate operations](https://parallel.ai/blog/case-study-opendoor)

Tags:Case Study
Reading time: 6 min
Introducing stateful web research agents with multi-turn conversations

- [Introducing stateful web research agents with multi-turn conversations](https://parallel.ai/blog/task-api-interactions)

Tags:Product Release
Reading time: 3 min
Parallel is now live on Tempo via the Machine Payments Protocol (MPP)

- [Parallel is live on Tempo, now available natively to agents with the Machine Payments Protocol](https://parallel.ai/blog/tempo-stripe-mpp)

Tags:Partnership
Reading time: 4 min
Kepler | Parallel Case Study

- [How Parallel helped Kepler build AI that finance professionals can actually trust](https://parallel.ai/blog/case-study-kepler)

Tags:Case Study
Reading time: 5 min
Introducing the Parallel CLI

- [Introducing the Parallel CLI](https://parallel.ai/blog/parallel-cli)

Tags:Product Release
Reading time: 3 min
Profound + Parallel Web Systems

- [How Profound helps brands win AI Search with high-quality web research and content creation powered by Parallel](https://parallel.ai/blog/case-study-profound)

Tags:Case Study
Reading time: 4 min
How Harvey is expanding legal AI internationally with Parallel

- [How Harvey is expanding legal AI internationally with Parallel](https://parallel.ai/blog/case-study-harvey)

Tags:Case Study
Reading time: 3 min
Tabstack + Parallel Case Study

- [How Tabstack by Mozilla enables agents to navigate the web with Parallel’s best-in-class web search](https://parallel.ai/blog/case-study-tabstack)

Tags:Case Study
Reading time: 5 min
Parallel | Vercel

- [Parallel Web Tools and Agents now available across Vercel AI Gateway, AI SDK, and Marketplace](https://parallel.ai/blog/vercel)

Tags:Product Release
Reading time: 3 min
Product release: Authenticated page access for the Parallel Task API

- [Authenticated page access for the Parallel Task API](https://parallel.ai/blog/authenticated-page-access)

Tags:Product Release
Reading time: 3 min
Introducing structured outputs for the Monitor API

- [Introducing structured outputs for the Monitor API](https://parallel.ai/blog/structured-outputs-monitor)

Tags:Product Release
Reading time: 3 min
Product release: Research Models with Basis for the Parallel Chat API

- [Introducing research models with Basis for the Parallel Chat API](https://parallel.ai/blog/research-models-chat)

Tags:Product Release
Reading time: 2 min
Parallel + Cerebras

- [Build a real-time fact checker with Parallel and Cerebras](https://parallel.ai/blog/cerebras-fact-checker)

Tags:Cookbook
Reading time: 5 min
DeepSearch QA: Task API

- [Parallel Task API achieves state-of-the-art accuracy on DeepSearchQA](https://parallel.ai/blog/deepsearch-qa)

Tags:Benchmarks
Reading time: 3 min
Product release: Granular Basis

- [Introducing Granular Basis for the Task API](https://parallel.ai/blog/granular-basis-task-api)

Tags:Product Release
Reading time: 3 min
How Amp’s coding agents build better software with Parallel Search

- [How Amp’s coding agents build better software with Parallel Search](https://parallel.ai/blog/case-study-amp)

Tags:Case Study
Reading time: 3 min
Latency improvements on the Parallel Task API

- [Latency improvements on the Parallel Task API ](https://parallel.ai/blog/task-api-latency)

Tags:Product Release
Reading time: 3 min
Product release: Extract

- [Introducing Parallel Extract](https://parallel.ai/blog/introducing-parallel-extract)

Tags:Product Release
Reading time: 2 min
FindAll API - Product Release

- [Introducing Parallel FindAll](https://parallel.ai/blog/introducing-findall-api)

Tags:Product Release,Benchmarks
Reading time: 4 min
Product release: Monitor API

- [Introducing Parallel Monitor](https://parallel.ai/blog/monitor-api)

Tags:Product Release
Reading time: 3 min
Parallel raises $100M Series A to build web infrastructure for agents

- [Parallel raises $100M Series A to build web infrastructure for agents](https://parallel.ai/blog/series-a)

Tags:Fundraise
Reading time: 3 min
How Macroscope reduced code review false positives with Parallel

- [How Macroscope reduced code review false positives with Parallel](https://parallel.ai/blog/case-study-macroscope)

Reading time: 2 min
Product release - Parallel Search API

- [Introducing Parallel Search](https://parallel.ai/blog/introducing-parallel-search)

Tags:Benchmarks
Reading time: 7 min
Benchmarks: SealQA: Task API

- [Parallel processors set new price-performance standard on SealQA benchmark](https://parallel.ai/blog/benchmarks-task-api-sealqa)

Tags:Benchmarks
Reading time: 3 min
Introducing LLMTEXT, an open source toolkit for the llms.txt standard

- [Introducing LLMTEXT, an open source toolkit for the llms.txt standard](https://parallel.ai/blog/LLMTEXT-for-llmstxt)

Tags:Product Release
Reading time: 7 min
Starbridge + Parallel

- [How Starbridge powers public sector GTM with state-of-the-art web research](https://parallel.ai/blog/case-study-starbridge)

Tags:Case Study
Reading time: 4 min
Building a market research platform with Parallel Deep Research

- [Building a market research platform with Parallel Deep Research](https://parallel.ai/blog/cookbook-market-research-platform-with-parallel)

Tags:Cookbook
Reading time: 4 min
How Lindy brings state-of-the-art web research to automation flows

- [How Lindy brings state-of-the-art web research to automation flows](https://parallel.ai/blog/case-study-lindy)

Tags:Case Study
Reading time: 3 min
Introducing the Parallel Task MCP Server

- [Introducing the Parallel Task MCP Server](https://parallel.ai/blog/parallel-task-mcp-server)

Tags:Product Release
Reading time: 4 min
Introducing the Core2x Processor for improved compute control on the Task API

- [Introducing the Core2x Processor for improved compute control on the Task API](https://parallel.ai/blog/core2x-processor)

Tags:Product Release
Reading time: 2 min
How Day AI merges private and public data for business intelligence

- [How Day AI merges private and public data for business intelligence](https://parallel.ai/blog/case-study-day-ai)

Tags:Case Study
Reading time: 4 min
Full Basis framework for all Task API Processors

- [Full Basis framework for all Task API Processors](https://parallel.ai/blog/full-basis-framework-for-task-api)

Tags:Product Release
Reading time: 2 min
Building a real-time streaming task manager with Parallel

- [Building a real-time streaming task manager with Parallel](https://parallel.ai/blog/cookbook-sse-task-manager-with-parallel)

Tags:Cookbook
Reading time: 5 min
How Gumloop built a new AI automation framework with web intelligence as a core node

- [How Gumloop built a new AI automation framework with web intelligence as a core node](https://parallel.ai/blog/case-study-gumloop)

Tags:Case Study
Reading time: 3 min
Introducing the TypeScript SDK

- [Introducing the TypeScript SDK](https://parallel.ai/blog/typescript-sdk)

Tags:Product Release
Reading time: 1 min
Building a serverless competitive intelligence platform with MCP + Task API

- [Building a serverless competitive intelligence platform with MCP + Task API](https://parallel.ai/blog/cookbook-competitor-research-with-reddit-mcp)

Tags:Cookbook
Reading time: 6 min
Introducing Parallel Deep Research reports

- [Introducing Parallel Deep Research reports](https://parallel.ai/blog/deep-research-reports)

Tags:Product Release
Reading time: 2 min
BrowseComp / DeepResearch: Task API

- [A new pareto-frontier for Deep Research price-performance](https://parallel.ai/blog/deep-research-benchmarks)

Tags:Benchmarks
Reading time: 4 min
Building a Full-Stack Search Agent with Parallel and Cerebras

- [Building a Full-Stack Search Agent with Parallel and Cerebras](https://parallel.ai/blog/cookbook-search-agent)

Tags:Cookbook
Reading time: 5 min
Webhooks for the Parallel Task API

- [Webhooks for the Parallel Task API](https://parallel.ai/blog/webhooks)

Tags:Product Release
Reading time: 2 min
Introducing Parallel: Web Search Infrastructure for AIs

- [Introducing Parallel: Web Search Infrastructure for AIs ](https://parallel.ai/blog/introducing-parallel)

Tags:Benchmarks,Product Release
Reading time: 6 min
Introducing SSE for Task Runs

- [Introducing SSE for Task Runs](https://parallel.ai/blog/sse-for-tasks)

Tags:Product Release
Reading time: 2 min
A new line of advanced Processors: Ultra2x, Ultra4x, and Ultra8x

- [A new line of advanced Processors: Ultra2x, Ultra4x, and Ultra8x ](https://parallel.ai/blog/new-advanced-processors)

Tags:Product Release
Reading time: 2 min
Introducing Auto Mode for the Parallel Task API

- [Introducing Auto Mode for the Parallel Task API](https://parallel.ai/blog/task-api-auto-mode)

Tags:Product Release
Reading time: 1 min
A linear dithering of a search interface for agents

- [A state-of-the-art search API purpose-built for agents](https://parallel.ai/blog/search-api-benchmark)

Tags:Benchmarks
Reading time: 3 min
Parallel Search MCP Server in Devin

- [Parallel Search MCP Server in Devin](https://parallel.ai/blog/parallel-search-mcp-in-devin)

Tags:Product Release
Reading time: 2 min
Introducing Tool Calling via MCP Servers

- [Introducing Tool Calling via MCP Servers](https://parallel.ai/blog/mcp-tool-calling)

Tags:Product Release
Reading time: 2 min
Introducing the Parallel Search MCP Server

- [Introducing the Parallel Search MCP Server ](https://parallel.ai/blog/search-mcp-server)

Tags:Product Release
Reading time: 2 min
Starting today, Source Policy is available for both the Parallel Task API and Search API - giving you granular control over which sources your AI agents access and how results are prioritized.

- [Introducing Source Policy](https://parallel.ai/blog/source-policy)

Tags:Product Release
Reading time: 1 min
The Parallel Task Group API

- [The Parallel Task Group API](https://parallel.ai/blog/task-group-api)

Tags:Product Release
Reading time: 1 min
State of the Art Deep Research APIs

- [State of the Art Deep Research APIs](https://parallel.ai/blog/deep-research)

Tags:Benchmarks
Reading time: 3 min
Introducing the Parallel Search API

- [Parallel Search API is now available in alpha](https://parallel.ai/blog/parallel-search-api)

Tags:Product Release
Reading time: 2 min
Introducing the Parallel Chat API - a low latency web research API for web based LLM completions. The Parallel Chat API returns completions in text and structured JSON format, and is OpenAI Chat Completions compatible.

- [Introducing the Parallel Chat API ](https://parallel.ai/blog/chat-api)

Tags:Product Release
Reading time: 1 min
Parallel Web Systems introduces Basis with calibrated confidences - a new verification framework for AI web research and search API outputs that sets a new industry standard for transparent and reliable deep research.

- [Introducing Basis with Calibrated Confidences ](https://parallel.ai/blog/introducing-basis-with-calibrated-confidences)

Tags:Product Release
Reading time: 4 min
The Parallel Task API is a state-of-the-art system for automated web research that delivers the highest accuracy at every price point.

- [Introducing the Parallel Task API](https://parallel.ai/blog/parallel-task-api)

Tags:Product Release,Benchmarks
Reading time: 4 min
![Company Logo](https://parallel.ai/parallel-logo-540.png)

Contact

  • hello@parallel.ai[hello@parallel.ai](mailto:hello@parallel.ai)

Products

  • Search API[Search API](https://docs.parallel.ai/search/search-quickstart)
  • Extract API[Extract API](https://docs.parallel.ai/extract/extract-quickstart)
  • Task API[Task API](https://docs.parallel.ai/task-api/task-quickstart)
  • FindAll API[FindAll API](https://docs.parallel.ai/findall-api/findall-quickstart)
  • Chat API[Chat API](https://docs.parallel.ai/chat-api/chat-quickstart)
  • Monitor API[Monitor API](https://docs.parallel.ai/monitor-api/monitor-quickstart)

Resources

  • About[About](https://parallel.ai/about)
  • Pricing[Pricing](https://parallel.ai/pricing)
  • Docs[Docs](https://docs.parallel.ai)
  • Blog[Blog](https://parallel.ai/blog)
  • Changelog[Changelog](https://docs.parallel.ai/resources/changelog)
  • Careers[Careers](https://jobs.ashbyhq.com/parallel)

Info

  • Terms of Service[Terms of Service](https://parallel.ai/terms-of-service)
  • Customer Terms[Customer Terms](https://parallel.ai/customer-terms)
  • Privacy[Privacy](https://parallel.ai/privacy-policy)
  • Acceptable Use[Acceptable Use](https://parallel.ai/acceptable-use-policy)
  • Trust Center[Trust Center](https://trust.parallel.ai/)
  • Report Security Issue[Report Security Issue](mailto:security@parallel.ai)
LinkedIn[LinkedIn](https://www.linkedin.com/company/parallel-web/about/)Twitter[Twitter](https://x.com/p0)GitHub[GitHub](https://github.com/parallel-web)
All Systems Operational
![SOC 2 Compliant](https://parallel.ai/soc2.svg)

Parallel Web Systems Inc. 2026