Applications
Symptoms related to application workloads and compute specifications.
Compute Spec
- CPU Avg Increase Version Changed High - Sustained increase in average CPU usage after version change
- Has DB Connections Increase High Instance - Has instances with high database connections increase
- Has DB Query Duration Increase High Instance - Has instances with high database query duration increase
- Has Frequent Crash High Instance - Has instances with high crash frequency
- Has Frequent OOM Kill High Instance - Has instances with high OOM kill frequency
- Has GC Time High Instance - Has instances with high garbage collection time
- Has Java Heap Utilization High Instance - Has instances with high Java heap utilization
- Has Mutex Wait Time High Instance - Has instances with high mutex wait time
- Has Redis Connections Utilization High Instance - Has instances with high Redis connections utilization
- Memory Avg Increase Version Changed High - Sustained increase in average memory usage after version change
Workload
- .Net Unhandled Exception In Log - .NET unhandled exception detected in logs
- Apache Max Workers In Log - Apache MaxRequestWorkers limit reached detected in logs
- Cache Size High - High cache size
- Cassandra Tombstone In Log - Excessive tombstone scanning detected in logs
- Circuit Breaker Open In Log - Circuit breaker open event detected in logs
- Command Latency High - High command latency
- Connection Pool Exhausted In Log - Connection pool exhaustion detected in logs
- Connections High - High number of connections
- DB Connections Utilization High - High database connections utilization
- DB Query Duration High - High database query duration
- Disk Full In Log - Disk full (ENOSPC) error detected in logs
- Elasticsearch Red In Log - Elasticsearch cluster RED status detected in logs
- File Descriptor Utilization High - High file descriptor utilization
- File Limit In Log - File descriptor limit reached detected in logs
- GC Time High - High garbage collection time
- Go Goroutine Leak In Log - Go goroutine leak detected in logs
- Go Max Procs Too High - GOMAXPROCS value is too high
- Go Nil Pointer Panic In Log - Go nil pointer panic detected in logs
- HAProxy Max Conn In Log - HAProxy backend unavailable event detected in logs
- High Error Log Count - High rate of ERROR or CRITICAL log lines
- High Kafka Data Logs Disk Utilization - High Kafka broker data log disk utilization
- High Kafka Message Rate - High Kafka message rate
- High Kafka Offline Partitions - Kafka has one or more offline partitions
- High Kafka Under Replicated Partitions - Kafka has under-replicated partitions
- High MongoDB Connection Usage - High MongoDB active connection utilization
- High MongoDB Open Cursor Count - High number of open MongoDB cursors
- High MongoDB Replica Lag - MongoDB replication lag is high
- High Postgres Checkpoint Write Time - Postgres checkpoint write time above 5 seconds, disk I/O is a bottleneck
- High Postgres Connection Utilization - Postgres active sessions above 85% of max_connections, new connections will soon be rejected
- High Postgres Deadlock Rate - Postgres deadlock rate above 0.1 seconds, concurrent conflicting transactions
- High Postgres Lock Wait Count - More than 5 sessions waiting on a lock, lock contention is occurring
- High Redis Cache Miss Rate - Redis cache miss ratio above 30%, most requests are falling through to the database
- High Redis Memory Fragmentation - Redis memory fragmentation ratio above 1.5, significant allocator overhead
- High Redis Replication Lag - Redis replica is more than 10 seconds behind the master
- High Redis Server Memory Usage - Redis server memory utilization above 85%, evictions are likely or imminent
- Java GC Pause In Log - Java GC allocation failure or to-space exhaustion detected in logs
- Java Heap Utilization High - High Java heap utilization
- Java Null Pointer In Log - Java NullPointerException detected in logs
- Java OOM In Log - Java OutOfMemoryError detected in logs
- Java Stack Overflow In Log - Java StackOverflowError detected in logs
- Java Thread Exhaustion In Log - Java thread pool rejection detected in logs
- Kafka ISR Shrink In Log - Kafka ISR shrinkage event detected in logs
- Low Postgres Cache Hit Rate - Postgres cache miss rate above 10%, queries are reading from disk instead of shared_buffers
- Memory Utilization High - High memory utilization
- Mutex Wait Time High - High mutex wait time
- MySQL Deadlock In Log - MySQL InnoDB deadlock detected in logs
- Nginx Upstream Timeout In Log - Nginx upstream connection timeout detected in logs
- Nginx Worker Conn In Log - Nginx worker connections exhausted detected in logs
- Postgres Deadlock In Log - PostgreSQL deadlock detected in logs
- Postgres Idle-in-Transaction Sessions Accumulating - More than 3 sessions in idle-in-transaction, they hold locks and block vacuum
- Postgres Query Memory Spill - Postgres is creating temporary files at a rate above 0.5 seconds, queries are spilling to disk
- Python Unhandled Exception In Log - Python unhandled exception detected in logs
- Queries Queued High - High number of queued queries
- RabbitMQ Alarm In Log - RabbitMQ memory or disk resource alarm detected in logs
- Rate Limit In Log - Rate limit exceeded detected in logs
- Redis Blocked Clients Accumulating - More than 5 clients are blocked on BLPOP/BRPOP commands
- Redis Connections Utilization High - High Redis connections utilization
- Redis Max Memory In Log - Redis maxmemory limit exceeded detected in logs
- Redis Rejecting Connections - Redis is rejecting new connections at a rate above 0.1 seconds, maxclients has been reached
- Redis Slow Commands Queued - Redis slow command log is growing, O(N) commands being run against large data structures
- Request Duration High - High request duration/latency
- Request Error Rate High - High request error rate
- Server Errors High - High server error count
- Throttled High - High throttled request count
- Transaction Duration High - High transaction duration
- Transaction Error High - High transaction error rate
- Transaction IDs Congested High - High transaction ID congestion
- User Errors High - High user error count