I’m just so exhausted these days. We have formal SLA’s, but its not like they’re ever followed. After all, Customer X needs to be notified within 5 minutes of any anomalous events in their cluster, and Customer Y is our biggest customer, so we give them the white glove treatment.

Yadda yadda, bla bla. So on and so forth, almost every customer has some exception/difference in SLAs.

I was hired on to be an SRE, but I’m just a professional dashboard starer at this point. The amount of times I’ve been alerted in the middle of the night because CPU was running high for 5 minutes is too damn high. Just so I can apologize to Mr. Customer that they maybe had a teensy slowdown during that time.

If I try to get us back to fundamentals and suggest we should only alert on impact, not short lived anomalies, there is some surface level agreement, but everyone seems to think “well we might miss something, so we need to keep it”.

It’s like we’re trying to prevent outages by monitoring for potential issues rather than actually making our system more robust and automate-able.

How do I convince these people that this isn’t sustainable? That trying to “catch” incidents before they happen is a fools errand. It’s like that chart about the “war on drugs” where it shows exponential expense growth as you try to prevent ALL drug usage (which is impossible). Yet this tech company seems to think we should be trying to prevent all outages with excessive monitoring.

And that doesn’t even get into the bonkers agreements we make with customers to agree to do a deep dive research on why 2 different environments have a response time that differs by 1ms.

Or the agreements that force us to complete customer provided training - while not assessing how much training we already committed to. It’s entirely normal to do 3-4x HIPAA / PCI / Compliance trainings when everyone else in the org only has to do one set of those.

I’m at a point where I’m considering moving on. This job just isn’t sustainable and there’s no interest in the org to make it sustainable.

But perhaps one of y’all managed to fix something similar in their org with a few key conversations and some effort? What other things could I try as a sort of final “Hail Mary” before looking to greener pastures?

  • th3raid0r@programming.devOP
    link
    fedilink
    arrow-up
    2
    arrow-down
    2
    ·
    edit-2
    5 months ago
    1. They want to be notified of anything that could potentially slow down their system. So any anomaly. The catch being is that they constantly change patterns because they introduce new workloads weekly - which wouldn’t be a problem if they could better communicate their forecasts. And that’s just one of a few dozen customers - again all with unique cluster configuration and needs.

    2. Yeah, it sucks. The first year was pretty great and we had a fully integrated and unified managed services team where we were getting some great automation done. Then they split the team in half in order to focus on a different flavor of our product (with an entirely new backend) and left folks who were newer (myself included) with maintaining the old product. We were even told that we should be doing minimal maintenance on the thing as the new product would be the new norm. Then once upper management remembered how contracts work, they decided we needed to support 3 new platforms without growing the team. All while onboarding new customers and growing the environment count. We’re now in operational overload after some turnover that was backfilled with offshore support that has a very minimal presence.

    3. I have tried championing this, but I don’t expect an ableist, masculinity shaming person like you to understand a call for social pointers on how to “manage up”.

    “Man Up” - good lord, way to be an ass.