High reliability, strong isolation, production-grade stability.
NodeRoute runs a security-first scraping platform designed to maximize success rates, protect customer workloads, prevent abuse, and stay stable even at very high scale — across every plan.
Platform security principles
Our infrastructure is built around four core principles that protect customers and keep the platform stable.
1) Stability first
Prevent overload, abuse, and cascading system failure with strict controls.
2) Isolation by design
Strong separation between customers across queues, workers, and data paths.
3) Automatic failover
No single proxy/provider dependency — routing adapts when conditions change.
4) Secure data handling
Externalized storage + encrypted pipelines to protect outputs end-to-end.
Security isn’t a plan upgrade
These principles apply across all tiers — limits scale with volume, but the security model stays consistent.
Traffic control & server protection
We enforce strict rate, concurrency, and job-size constraints to protect target sites, customer workloads, and platform stability.
Rate limiting
Every plan enforces request & usage ceilings to keep performance stable.
- Requests per minute limits
- Maximum concurrent jobs
- Maximum job size
- Monthly credit limits
Concurrency control
We cap parallelism to prevent resource exhaustion and queue congestion.
- Max concurrent jobs per user
- Parallel browser executions
- High-intensity scraping bursts
Job size limits
Each plan enforces max crawl size to avoid unbounded workloads.
- Starter: 10k URLs
- Growth: 100k URLs
- Scale: 1M URLs
Intelligent retry & failover protection
When targets throttle or networks fail, the platform adapts automatically — without hammering target servers.
Smart retry logic
Automatic retries with exponential backoff for transient failures and interruptions.
- Network failures
- Temporary blocks
- Timeout events
- CAPTCHA interruptions
Circuit breaker system
Provider-level circuit breakers stop cascading failures when a route degrades.
- Auto-disables failing proxy providers/regions/paths
- Routes traffic to healthy alternatives
- Restores only after stability is verified
Multi-provider routing architecture
NodeRoute supports multi-provider pools and dynamically routes traffic based on real-time signals.
Queue & worker infrastructure
Distributed execution built for crash recovery, retries, and tenant isolation.
BullMQ + Redis
High-throughput queues with durability and recovery built-in.
- Guaranteed job execution
- Horizontal scaling
- Retry persistence
- Crash recovery
- Dead-letter queues
Worker isolation
Jobs execute in isolated worker pools with strict resource controls.
- CPU & memory constraints
- Automatic worker recycling
- Per-job execution limits
Execution safety controls
Hard limits keep jobs from stalling or running forever.
- Hard execution timeouts
- Browser execution caps
- Script watchdog timers
Secure data storage architecture
Large datasets are stored in external object storage with strong durability and encrypted transfers.
Externalized object storage
We do not store large scraped datasets directly on VPS servers. Outputs are delivered to object storage providers.
- Cloudflare R2
- Backblaze B2
- Wasabi
- AWS S3
Data isolation & secure transfer
Designed to minimize cross-tenant risk and protect data in transit.
- Per-customer storage buckets
- Separate access tokens & API credentials
- Encrypted transfers in transit
- Token-based access controls
Platform-level abuse prevention
Automated detection protects both customers and the web from misconfigured or malicious workloads.
Protection mechanisms
We continuously monitor patterns and apply automated controls when behavior looks unsafe.
Burst detection
Detects sudden request spikes and throttles safely.
Abnormal patterns
Flags infinite loops, broken scripts, and exploit attempts.
Emergency termination
Stops jobs when necessary to protect platform stability.
Plan-based security controls
Limits scale by plan to match volume, while keeping the same core safety model.
| Control |
Starter
Safe validation
|
Growth
Sustained high volume
|
Scale
Large pipelines
|
|---|---|---|---|
| Rate limit | 60 requests/min | 300 requests/min | 1,200 requests/min |
| Concurrent jobs | 3 | 10 | 50 |
| Max job size | 10k URLs | 100k URLs | 1M URLs |
| Retry & failover | Basic retry + failover | Smart retries + multi-provider failover | Advanced routing + dynamic failover |
| Isolation & safety controls | ✅ | ✅ | ✅ |
| Abuse prevention | ✅ | ✅ | ✅ |
Operational reliability
Built to keep running through crashes, provider issues, and high load.
Automatic recovery
- Job auto-resume after crashes
- Worker auto-restart
- Queue state recovery
Monitoring
- Real-time health monitoring
- Proxy performance tracking
- Latency monitoring
- Failure rate analysis
Why this matters for your business
You don’t overload your systems or target websites, you don’t lose data, you avoid proxy lock-in, and you can scale safely as volume grows.
Contact
Need a security review, custom limits, or a reliability walkthrough? Send your requirements — we’ll respond fast.
What to include
- Target sites + geo requirements
- Daily volume (URLs/SKUs) + frequency
- Need JS rendering / CAPTCHA support?
- Data delivery: S3, R2, webhooks
What you’ll get
A clear, practical plan for stable throughput without creating risk for you or your targets.
- Recommended tier + scaling path
- Safe rate + concurrency guidance
- Routing strategy (providers + regions)
- Data isolation & storage setup