There is a difference between installing monitoring and actually making it useful for an operations team.
This note leaned more toward the second half: not just getting Prometheus into the cluster, but wiring the alerting and service exposure in a way that makes the stack usable outside the cluster too.
1. Install the Base Stack
The initial deployment was the usual kube-prometheus flow:
| |
Then the monitoring services were exposed:
| |
2. Add Alertmanager Configuration
The note also included the operational step that often gets delayed: actually wiring alert delivery.
The generalized shape looked like this:
| |
The original note contained a live webhook secret and internal channel details, so those are intentionally replaced here.
3. Extend the Prometheus RBAC
Like the other monitoring note in this batch, this one needed more permissions than the default cluster role had.
The useful part to preserve is the pattern:
- check what Prometheus is trying to scrape
- compare that to the current role
- add the missing
get,list, and sometimeswatchverbs for the relevant resources
That is not glamorous work, but it is the sort of thing you end up doing in real clusters.
4. Treat Monitoring as a Real Service
The practical lesson from this note is that monitoring is not “done” when the pods are Running.
It is done when:
- Prometheus can actually see the resources you care about
- Grafana is reachable where operators need it
- Alertmanager is wired to something a human will actually see
That is a different definition of success, and it is usually the more useful one.