Tool Integration

DataForSEO + Claude: Enterprise-Scale Local SEO Data

Bulk keyword queries at volume pricing, custom data pipelines, raw API access. When query volume exceeds standard credit models.

Get on GitHub

DataForSEO Tool Integration: Enterprise-Scale Data Operations

LocalSEOData’s credit model works well for standard practitioner workflows. DataForSEO’s bulk pricing model becomes more efficient when:

  • Running 1,000+ keyword queries per session
  • Building custom data pipelines
  • Operating at agency scale with high monthly query volume

What This Integration Unlocks

Bulk data operations — Enterprise-scale keyword research, SERP pulls, rank tracking.

Raw API access — Build custom pipelines for automated reporting systems.

Cost efficiency at volume — Bulk pricing beats per-query pricing above certain thresholds.

SERP data types beyond standard — Specialized endpoints for edge cases.

Dispatch Routing

Routes to dataforseo-tool when:

  • Query volume exceeds standard credit model efficiency
  • Building custom data pipelines
  • Bulk keyword research at 500+ terms
  • Practitioner explicitly requests DataForSEO

For standard-volume workflows, LocalSEOData remains the default.

Three Workflows

Workflow 1 — Bulk Keyword Research:

Prompt: "Using DataForSEO, pull search volume and keyword difficulty for
this list of 800 local legal keywords across Phoenix, Scottsdale, and Tempe."

Output:
Bulk Keyword Data — 800 keywords × 3 cities = 2,400 data points

Sample output structure:
keyword                              | city       | MSV  | KD   | CPC
"personal injury lawyer"             | Phoenix    | 720  | 68   | $89
"personal injury lawyer"             | Scottsdale | 210  | 52   | $76
"personal injury lawyer"             | Tempe      | 140  | 48   | $71
"car accident attorney"              | Phoenix    | 590  | 64   | $82
[continues for all 2,400 combinations]

Export: CSV ready for analysis
Cost: Bulk pricing applied (~$0.002/query vs. standard $0.01)

Workflow 2 — Custom Reporting Pipeline Design:

Prompt: "Design a DataForSEO API pipeline that automatically pulls weekly
local pack data for 50 client locations and outputs to our reporting system."

Output:
PIPELINE DESIGN: Weekly Local Pack Monitoring — 50 Locations

API calls per cycle:
- 50 × local_pack calls (one per location)
- 50 × 3 keywords average = 150 total queries/week

Endpoint: /v3/serp/google/organic/live/regular
Parameters per call:
{
  "keyword": "[client keyword]",
  "location_code": [geo code],
  "language_code": "en",
  "device": "mobile",
  "os": "android"
}

Response parsing:
Extract: local_pack → items[0-2] → {title, rating, reviews, position}

Output format: JSON → your reporting database
Schedule: Cron every Monday 6am
Error handling: Retry 3x, log failures, alert if >10% fail rate

Estimated monthly cost: ~$300 at current volume

Workflow 3 — Enterprise SERP Intelligence:

Prompt: "Pull full SERP data for our top 30 keywords across all 25 markets
using DataForSEO. I need organic, local pack, and ads in one pull."

Output:
Batch request: 30 keywords × 25 markets = 750 SERP pulls

Data returned per query:
- Organic positions 1-10 with URL, title, description
- Local pack composition (all businesses)
- Ad positions and ad copy
- SERP features present (AI Overview, PAA, etc.)

Processing time: ~15 minutes for full batch
Export: Structured JSON, CSV conversion available

Setup

  1. DataForSEO account at dataforseo.com
  2. API credentials: login (email) + password (acts as API key)
  3. In Claude Code: Settings → MCP Servers → Add Server
    • Name: DataForSEO
    • Uses HTTP Basic auth: username + password
  4. Verify connection

First Prompt After Setup

"Ping DataForSEO and confirm the connection is active."

Skill Documentation

For technical details on how this skill works, what data it pulls, and complete prompt reference, see the full skill documentation.

All 36 skills. Free. Open source. Get on GitHub →