From palantir-pack
Implements Palantir Foundry API rate limiting with exponential backoff, jittered retries for 429 errors, request queuing, and token bucket limiting to optimize throughput.
npx claudepluginhub jeremylongshore/claude-code-plugins-plus-skills --plugin palantir-packThis skill is limited to using the following tools:
Handle Foundry API rate limits with exponential backoff, request queuing, and monitoring. Foundry rate limits vary by endpoint and enrollment tier.
Diagnose and fix Palantir Foundry API errors (401/403, ObjectTypeNotFound, DatasetNotFound) and transform failures (AnalysisException, OutOfMemory) with copy-paste Python solutions.
Implements Apollo.io API rate limiting with sliding window limiter, backoff on 429s, and header parsing for per-endpoint minute/burst limits.
Manages Apify API rate limits using apify-client retries, batching, and PQueue for concurrency control to handle 429 errors and bulk operations.
Share bugs, ideas, or general feedback.
Handle Foundry API rate limits with exponential backoff, request queuing, and monitoring. Foundry rate limits vary by endpoint and enrollment tier.
foundry-platform-sdk installedFoundry rate limits are per-user and per-endpoint. Key limits:
| Endpoint Category | Typical Limit | Burst |
|---|---|---|
| Ontology reads | 100 req/s | 200 |
| Ontology writes (Actions) | 50 req/s | 100 |
| Dataset reads | 50 req/s | 100 |
| Search queries | 20 req/s | 50 |
Rate limit headers returned:
X-RateLimit-Limit — max requests per windowX-RateLimit-Remaining — requests left in windowRetry-After — seconds to wait (on 429)import time
import random
import foundry
def retry_foundry_call(fn, *args, max_retries=5, base_delay=1.0, **kwargs):
"""Retry Foundry API calls with jittered exponential backoff."""
for attempt in range(max_retries + 1):
try:
return fn(*args, **kwargs)
except foundry.ApiError as e:
if attempt == max_retries:
raise
if e.status_code not in (429, 500, 502, 503):
raise # Non-retryable error
delay = base_delay * (2 ** attempt) + random.uniform(0, 0.5)
retry_after = getattr(e, "retry_after", None)
if retry_after:
delay = max(delay, float(retry_after))
print(f" Retry {attempt+1}/{max_retries} in {delay:.1f}s (HTTP {e.status_code})")
time.sleep(delay)
# Usage
employees = retry_foundry_call(
client.ontologies.OntologyObject.list,
ontology="my-company", object_type="Employee", page_size=100,
)
import asyncio
from collections import deque
class FoundryRateLimiter:
"""Token bucket rate limiter for batch Foundry operations."""
def __init__(self, max_per_second: int = 50):
self.max_per_second = max_per_second
self.tokens = max_per_second
self._last_refill = time.monotonic()
def _refill(self):
now = time.monotonic()
elapsed = now - self._last_refill
self.tokens = min(self.max_per_second, self.tokens + elapsed * self.max_per_second)
self._last_refill = now
def acquire(self):
self._refill()
if self.tokens < 1:
wait = (1 - self.tokens) / self.max_per_second
time.sleep(wait)
self._refill()
self.tokens -= 1
limiter = FoundryRateLimiter(max_per_second=40) # 80% of limit
def rate_limited_call(fn, *args, **kwargs):
limiter.acquire()
return retry_foundry_call(fn, *args, **kwargs)
def batch_update_objects(client, ontology, action_type, items, batch_size=10):
"""Apply actions in rate-limited batches."""
results = []
for i in range(0, len(items), batch_size):
batch = items[i:i+batch_size]
for item in batch:
result = rate_limited_call(
client.ontologies.Action.apply,
ontology=ontology,
action_type=action_type,
parameters=item,
)
results.append({"item": item, "status": result.validation})
print(f" Processed {min(i+batch_size, len(items))}/{len(items)}")
return results
| HTTP Code | Meaning | Action |
|---|---|---|
| 429 | Rate limited | Wait Retry-After seconds, then retry |
| 500 | Server error | Retry with backoff |
| 502/503 | Gateway error | Retry with backoff |
| 400/403/404 | Client error | Do not retry — fix the request |
For security best practices, see palantir-security-basics.