From bamboohr-pack
Optimizes BambooHR API performance with custom reports to eliminate N+1 queries, incremental sync via changed-since, caching, and connection pooling. For slow responses or sync throughput issues.
npx claudepluginhub jeremylongshore/claude-code-plugins-plus-skills --plugin bamboohr-packThis skill is limited to using the following tools:
Optimize BambooHR API performance through request reduction, caching, incremental sync, and connection pooling. The biggest wins come from eliminating N+1 query patterns using custom reports and the changed-since endpoint.
Audits BambooHR API usage, reduces calls via caching and patterns, and monitors to prevent rate limits in integrations.
Automates BambooHR HR tasks like listing/searching employees, tracking changes, managing time-off via Rube MCP and Composio toolkit. Requires active connection; search tools first for schemas.
Automates BambooHR HR tasks like listing/searching employees, tracking changes, time-off balances, benefits, dependents, and updates via Rube MCP and Composio toolkit. For HR data workflows.
Share bugs, ideas, or general feedback.
Optimize BambooHR API performance through request reduction, caching, incremental sync, and connection pooling. The biggest wins come from eliminating N+1 query patterns using custom reports and the changed-since endpoint.
The single biggest performance improvement: use POST /reports/custom instead of individual employee GETs.
// BAD: 501 API calls for 500 employees
const dir = await client.getDirectory(); // 1 call
for (const emp of dir.employees) {
await client.getEmployee(emp.id, ['salary', 'hireDate']); // 500 calls
}
// GOOD: 1 API call for all employees with all needed fields
const report = await client.customReport([
'firstName', 'lastName', 'department', 'jobTitle',
'hireDate', 'workEmail', 'status', 'location',
'supervisor', 'employeeNumber',
]);
// 1 call, returns all employees with all fields
Performance impact: 500x reduction in API calls. Custom reports return all active employees in one request.
import { readFileSync, writeFileSync } from 'fs';
const LAST_SYNC_FILE = '.bamboohr-last-sync';
async function incrementalSync(client: BambooHRClient): Promise<string[]> {
// Read last sync timestamp
let lastSync: string;
try {
lastSync = readFileSync(LAST_SYNC_FILE, 'utf-8').trim();
} catch {
lastSync = new Date(Date.now() - 24 * 60 * 60 * 1000).toISOString(); // Default: 24h ago
}
// GET /employees/changed/?since=... — returns only changed employee IDs
const changed = await client.request<{
employees: Record<string, { id: string; lastChanged: string }>;
}>('GET', `/employees/changed/?since=${lastSync}`);
const changedIds = Object.keys(changed.employees || {});
console.log(`${changedIds.length} employees changed since ${lastSync}`);
if (changedIds.length === 0) return [];
// Fetch only changed employees' details
// For large sets, use custom report with filter; for small sets, individual GETs
if (changedIds.length > 20) {
// Bulk: use custom report (returns all, then filter client-side)
const report = await client.customReport([
'firstName', 'lastName', 'department', 'status',
]);
const changedData = report.employees.filter(e =>
changedIds.includes(e.id?.toString()),
);
// Process changedData...
} else {
// Small set: individual GETs are fine
for (const id of changedIds) {
const emp = await client.getEmployee(id, ['firstName', 'lastName', 'department', 'status']);
// Process emp...
}
}
// Save sync timestamp
writeFileSync(LAST_SYNC_FILE, new Date().toISOString());
return changedIds;
}
Also available for table data:
// GET /employees/changed/tables/{tableName}?since=...
const changedJobs = await client.request<any>(
'GET', `/employees/changed/tables/jobInfo?since=${lastSync}`,
);
// Returns { employees: { "123": { lastChanged: "..." }, ... } }
import { LRUCache } from 'lru-cache';
// BambooHR directory data changes infrequently — cache aggressively
const cache = new LRUCache<string, any>({
max: 500,
ttl: 5 * 60 * 1000, // 5 minutes for directory data
});
async function cachedRequest<T>(
key: string,
fetcher: () => Promise<T>,
ttlMs?: number,
): Promise<T> {
const cached = cache.get(key) as T | undefined;
if (cached) {
console.log(`Cache hit: ${key}`);
return cached;
}
const result = await fetcher();
cache.set(key, result, { ttl: ttlMs });
return result;
}
// Usage
const directory = await cachedRequest(
'directory',
() => client.getDirectory(),
5 * 60 * 1000, // Cache for 5 min
);
// Single employee — shorter cache
const employee = await cachedRequest(
`employee:${id}`,
() => client.getEmployee(id, fields),
60 * 1000, // Cache for 1 min
);
Redis caching for multi-instance deployments:
import Redis from 'ioredis';
const redis = new Redis(process.env.REDIS_URL);
async function redisCached<T>(
key: string,
fetcher: () => Promise<T>,
ttlSec = 300,
): Promise<T> {
const cached = await redis.get(`bamboohr:${key}`);
if (cached) return JSON.parse(cached);
const result = await fetcher();
await redis.setex(`bamboohr:${key}`, ttlSec, JSON.stringify(result));
return result;
}
// Invalidate on webhook
async function invalidateCache(employeeId: string) {
await redis.del(`bamboohr:employee:${employeeId}`);
await redis.del('bamboohr:directory'); // Directory includes this employee
}
import { Agent } from 'https';
// Reuse TCP connections for BambooHR API calls
const keepAliveAgent = new Agent({
keepAlive: true,
maxSockets: 5, // Max 5 parallel connections
maxFreeSockets: 2,
timeout: 30_000,
keepAliveMsecs: 10_000,
});
// Pass to fetch via undici or node-fetch
// For native fetch in Node 20+, connection pooling is automatic
import DataLoader from 'dataloader';
// Batch individual employee GETs into a custom report
const employeeLoader = new DataLoader<string, Record<string, string>>(
async (ids) => {
// One custom report instead of N individual GETs
const report = await client.customReport([
'id', 'firstName', 'lastName', 'department', 'jobTitle',
]);
const byId = new Map(report.employees.map(e => [e.id, e]));
return ids.map(id => byId.get(id) || new Error(`Employee ${id} not found`));
},
{
maxBatchSize: 100,
batchScheduleFn: cb => setTimeout(cb, 50), // Batch window: 50ms
cache: true,
},
);
// Usage — automatically batched into one API call
const [emp1, emp2, emp3] = await Promise.all([
employeeLoader.load('1'),
employeeLoader.load('2'),
employeeLoader.load('3'),
]);
class BambooHRMetrics {
private requests: { duration: number; status: number; endpoint: string }[] = [];
record(endpoint: string, status: number, durationMs: number) {
this.requests.push({ duration: durationMs, status, endpoint });
// Keep last 1000 requests
if (this.requests.length > 1000) this.requests.shift();
}
summary() {
const durations = this.requests.map(r => r.duration).sort((a, b) => a - b);
const errors = this.requests.filter(r => r.status >= 400);
return {
totalRequests: this.requests.length,
errorRate: (errors.length / Math.max(this.requests.length, 1) * 100).toFixed(1) + '%',
p50: durations[Math.floor(durations.length * 0.5)] || 0,
p95: durations[Math.floor(durations.length * 0.95)] || 0,
p99: durations[Math.floor(durations.length * 0.99)] || 0,
topEndpoints: this.topEndpoints(),
};
}
private topEndpoints() {
const counts = new Map<string, number>();
for (const r of this.requests) {
counts.set(r.endpoint, (counts.get(r.endpoint) || 0) + 1);
}
return [...counts.entries()].sort((a, b) => b[1] - a[1]).slice(0, 5);
}
}
| Optimization | Before | After | Improvement |
|---|---|---|---|
| Custom reports vs N+1 | 501 calls | 1 call | 500x |
| Incremental sync | Full pull | Delta only | 10-100x |
| Directory caching (5 min) | Every request | 1/5 min | 50x |
| Connection pooling | New conn/request | Reused | 2-3x latency |
| Issue | Cause | Solution |
|---|---|---|
| Cache stampede | All caches expire simultaneously | Stagger TTLs with jitter |
| Stale data | Cache TTL too long | Invalidate on webhook events |
| DataLoader timeout | Custom report too slow | Reduce batch size |
| Memory pressure | LRU cache too large | Set max entries limit |
For cost optimization, see bamboohr-cost-tuning.