IP Geolocation Data Quality in 2026: How to Actually Evaluate Accuracy, VPN Coverage, and False Positives
Every IP geolocation vendor claims 99% accuracy and global coverage. But when you deploy their API in production, you find city-level results that place New York users in Philadelphia, VPN flags on your own employees, and blind spots across entire regions. This post gives you a framework to evaluate data quality before you commit — with concrete testing methods, not vendor marketing slides.
Data Quality Evaluation: What to Measure
Why Vendor Accuracy Claims Mean Almost Nothing
When a vendor advertises "99.9% accuracy," that number usually refers to country-level resolution on a curated test set of known IPs. It does not tell you how accurate their city-level results are for your specific traffic mix, how well they detect residential proxies used in your region, or how many legitimate VPN users they will misclassify.
Accuracy is not a single number. It is a matrix of dimensions that vary by geography, IP type, and use case. The vendor whose overall accuracy is highest may perform worse than a competitor on the specific dimension that matters most to your application.
Five Dimensions of IP Intelligence Data Quality
Before you sign a contract or write integration code, evaluate vendors across these five dimensions. Each one affects a different part of your application, and the relative importance depends on your use case.
| Dimension | What It Measures | Who Cares Most |
|---|---|---|
| Geolocation precision | Country, region, city, and coordinate accuracy across your traffic geographies | Analytics, personalization, content localization |
| VPN/proxy detection coverage | Commercial VPN, residential proxy, Tor, and datacenter IP detection rates | Fraud prevention, account security, ad verification |
| False-positive rate | How often legitimate traffic gets flagged as suspicious or misclassified | All teams — directly impacts user experience and revenue |
| ASN/ISP intelligence depth | ASN accuracy, organization name, carrier identification, connection type | Network operations, B2B enrichment, traffic analysis |
| Coverage and freshness | How many IPs are covered, how recently the data was updated, IPv6 support | All teams — gaps create blind spots in your data |
How to Build Your Own Accuracy Test
Do not rely on vendor-provided benchmarks. Build your own ground-truth dataset from IPs whose real location you can verify independently. Here is the methodology:
1Assemble a Ground-Truth Dataset
Collect IPs with independently verified locations. Good sources include your own office and datacenter IPs, known server infrastructure, employee remote connections (with consent), and CDN edge node IP ranges published by providers like Cloudflare and AWS. Target at least 500-1,000 unique IPs spanning the regions where your users actually connect from.
2Query Every Vendor with the Same Dataset
Send the same IP list to each vendor you are evaluating. Record the country, region, city, latitude, longitude, ASN, ISP name, VPN flag, and threat level for each response. Use the vendor's production API, not a demo endpoint — demo environments sometimes return curated results.
# Test script: Query an IP against the geolocation API curl -s "https://ip-info.app/api/v1/geolocate/8.8.8.8" \ -H "x-api-key: YOUR_API_KEY" | python3 -m json.tool # Expected response fields to evaluate: # countryCode, city.name, city.latitude, city.longitude, # asn, aso, isVPN, isProxy, isTor, threatLevel, # city.accuracy_radius, registeredCountryCode
3Score Results by Geography and IP Type
Do not calculate a single accuracy percentage. Instead, break results down by country or region, by IP type (fixed-line, mobile, datacenter, VPN), and by resolution level (country, region, city). This breakdown reveals where each vendor is strong and where they have gaps that matter for your traffic.
# TypeScript: Calculate accuracy by region
interface TestResult {
ip: string;
knownCountry: string;
vendorCountry: string;
knownCity?: string;
vendorCity?: string;
}
function scoreByRegion(results: TestResult[]) {
const regions: Record<string, { total: number; correct: number }> = {};
for (const r of results) {
const region = r.knownCountry;
if (!regions[region]) regions[region] = { total: 0, correct: 0 };
regions[region].total++;
if (r.knownCountry === r.vendorCountry) {
regions[region].correct++;
}
}
// Sort by volume to prioritize your largest markets
return Object.entries(regions)
.sort((a, b) => b[1].total - a[1].total)
.map(([region, data]) => ({
region,
accuracy: ((data.correct / data.total) * 100).toFixed(1) + '%',
sampleSize: data.total,
}));
}4Measure VPN False Positives Separately
This is where most evaluations go wrong. Take your known-corporate IPs and employee VPN exit nodes, query each vendor, and count how many get flagged as suspicious. A vendor that catches 95% of VPNs but flags 15% of your legitimate enterprise traffic will cause more damage than one that catches 85% of VPNs with a 3% false-positive rate.
VPN and Proxy Detection: What Accuracy Actually Means
VPN detection is not binary. The landscape includes commercial VPNs (NordVPN, ExpressVPN), corporate VPNs and SD-WAN platforms, mobile carrier NAT pools that look like proxy traffic, consumer privacy networks (Apple Private Relay, Cloudflare WARP), Tor exit nodes, and residential proxy networks. Each category has different detection difficulty and different false-positive risk.
| Category | Detection Difficulty | FP Risk | Impact If Missed |
|---|---|---|---|
| Commercial VPN | Low — well-known IP ranges | Low | High for fraud, low for compliance |
| Residential proxy | High — IPs rotate frequently | Medium | Critical for payment fraud |
| Corporate VPN | Medium — overlap with datacenter IPs | High | Blocks real enterprise users |
| Privacy relay | High — Apple/Cloudflare infrastructure | High | 23%+ of traffic on some sites |
| Tor exit nodes | Low — published directory | Low | High for security-sensitive apps |
When you evaluate vendors, ask for detection rates broken down by these categories, not a single aggregate number. A vendor that detects 99% of commercial VPNs but only 60% of residential proxies is not necessarily worse than one that detects 95% of both — it depends on whether your fraud exposure comes from VPN users or proxy networks.
Regional Accuracy Gaps That Catch Teams Off Guard
IP geolocation accuracy is not uniform. It follows the distribution of internet infrastructure investment, BGP visibility, and the density of ground-truth measurement networks. Here is what to expect across major regions:
North America & Western Europe(High)
Country-level: 95-99%. City-level: 80-95%. Dense ISP infrastructure and measurement coverage make this the most reliable region for all vendors.
East Asia & Pacific(High to Medium)
Country-level: 90-98%. City-level: 70-90%. Strong in Japan, South Korea, Australia. More variable in Southeast Asia and Pacific Islands.
Latin America(Medium)
Country-level: 85-95%. City-level: 60-80%. Good coverage in Brazil and Mexico; weaker in Central America and smaller markets.
Middle East & Africa(Variable)
Country-level: 75-95%. City-level: 40-70%. GCC countries are well-covered; Sub-Saharan Africa has the largest accuracy gaps.
These ranges are typical across the industry, not specific to any single vendor. The point is that if your application serves users in Africa or Southeast Asia, you need to test those regions specifically — a vendor who excels in North America may not be the best choice for your actual traffic distribution.
Buyer's Scorecard: What to Ask Every Vendor
Use this framework when evaluating IP intelligence providers. These are the questions that separate marketing claims from operational reality:
| Evaluation Criterion | Red Flag | Green Flag |
|---|---|---|
| Accuracy claims | Single "99%" number with no methodology | Breakdowns by region, IP type, and resolution |
| VPN detection | No FP rate disclosure, only "detection rate" | Category-specific rates (VPN, proxy, Tor, relay) |
| Data freshness | No update cadence or stale database | Documented update frequency with change logs |
| IPv6 coverage | IPv4 only or limited IPv6 | Full IPv4 + IPv6 with comparable accuracy |
| Bulk processing | No batch API or very low throughput | Batch endpoint with documented throughput |
| SLA and uptime | No SLA or only "best effort" | 99.9%+ SLA with documented incident history |
| Compliance posture | No SOC 2, ISO 27001, or GDPR documentation | SOC 2 Type II, ISO 27001 certified |
Test Data Quality With Your Own Traffic
The live demo supports IPv4 and IPv6 queries. Run your ground-truth IPs through it and compare the results against whatever you are using today. No account required for the demo.
Frequently Asked Questions
How accurate is IP geolocation at the city level?▼
City-level accuracy varies significantly by region. Urban areas in the US and Western Europe typically achieve 80-95% accuracy, while parts of Africa, Southeast Asia, and rural areas may drop below 50%. Accuracy also depends on whether the IP is from a fixed-line ISP, mobile carrier, or corporate network. Always test with your own traffic distribution, not a vendors headline number.
What is a typical false-positive rate for VPN detection?▼
VPN false-positive rates typically range from 3-8% across commercial providers. Corporate VPNs, mobile carrier gateways, and consumer privacy services like Apple Private Relay are the most common sources of false positives. A good vendor publishes their FP rate by category and lets you tune sensitivity thresholds rather than forcing a binary block/allow decision.
How do I test IP geolocation accuracy myself?▼
Start with a ground-truth dataset of IPs whose locations you know with confidence — for example, IPs from your own office networks, employee VPNs, or server infrastructure. Query each vendor with the same dataset and compare the returned country, city, and coordinates against your known values. Measure accuracy at the country, region, and city levels separately, since they degrade at different rates.
Does IPv6 geolocation accuracy differ from IPv4?▼
IPv6 geolocation has historically been less accurate than IPv4 because IPv6 address allocation patterns are more complex and the data sources are less mature. However, as global IPv6 adoption exceeds 43%, vendors are investing heavily in IPv6 accuracy. When evaluating vendors, test both address families against your traffic mix.