13 March 2026

When something goes wrong – what boards learn about your cloud and cyber decisions

This series of blogs covers how organisations fall short when building resilient IT.  You can read the others online here and here.

The systems are back online. 

Customers can transact again. The dashboards look stable. Your teams have finally taken a breath. 

And then the questions start. 

The board wants a root cause analysis. Finance wants impact numbers. Risk wants clarity on exposure. Legal wants to understand if there are regulatory implications. Communications wants to know what can be said externally. 

This is the moment that really matters. Not the outage itself, but the review afterwards. 

Because once disruption has been contained, the conversation shifts. It stops being purely technical and starts becoming forensic. 

Most organisations follow a structured post-incident process, separated into stages: Containment, Recovery, Analysis, Notification and Evaluation (CRANE). Containment stabilises the situation. Recovery restores service. Analysis identifies what happened. Notification fulfils regulatory and stakeholder obligations. 

But it’s the final stage, Evaluation, that carries real weight at board level. 

That’s when leadership steps back and asks harder questions about judgement, governance and preparedness. 

  • Where did it fail? 

  • Why did it fail? 

  • How much did it cost us? 

  • Could it have been avoided? 

  • What are we doing to make sure it doesn’t happen again? 

  • Are we still vulnerable?  

And in that room, your past decisions about cloud architecture, cybersecurity investment, network setup, third-party providers and recovery strategy are no longer strategic initiatives. They are evidence. 

Boards don’t review tools. They review outcomes. 

It’s tempting to assume boards care about architecture, vendors or whether the roadmap was followed. 

In reality, they care about money, reputation and risk. 

  • How much revenue was lost? 

  • How much will remediation cost? 

  • Whether customer trust has been weakened. 

  • Whether shareholders value has been affected. 

  • Will it happen again? 

Speed, coordination and clarity matter, but largely because slow, fragmented recovery increases financial exposure and reputational damage. Regulators may focus closely on governance mechanics and operational discipline, but they also assess the real-world impact of incidents, particularly the level of harm caused to customers and individuals. Boards focus on commercial impact and reputational risk. 

If recovery took longer than your declared objectives or the impact tolerances you anticipated, the question becomes why. If the incident escalated, the question becomes whether earlier intervention, in the form of stronger monitoring, clearer recovery design or better integrated cloud and security controls, could have changed the outcome. If communication felt uncertain, the question becomes whether leadership had control. 

The board is not judging your tech stack. It is judging your judgement. 

When disruption hits, your board isn’t reviewing your architecture. They’re reviewing the consequences of your decisions. 

Disruption exposes how your strategy behaves under pressure 

Every major incident tests the decisions you’ve already made. 

It reveals whether resilience and recovery were designed deliberately alongside performance and growth, or whether they were assumed to exist without being properly engineered and tested. It shows whether your security controls support rapid containment and recovery, or whether configuration gaps, access dependencies and cross-team processes introduce friction at exactly the wrong moment. It highlights whether your network gives you visibility and prioritisation when demand spikes, or whether it becomes an afterthought until it fails. 

When everything is running as expected, each of these areas can look mature:

  • You design your network to prioritise critical traffic and handle peak demand appropriately. 

  • You optimise your cloud environment for performance and cost. 

  • You invest in patching, monitoring and strengthening your security controls. 
     

That effort matters. But those optimisations shouldn’t be tested in isolation, and testing your playbook for resilience means testing whether these interdependencies hold up. 

Under pressure, those interdependencies come to light. 

  • If recovery pathways assume connectivity is immediately available, but network prioritisation is focused on stabilising live services first, containment and restoration actions take longer than planned. 

  • If disaster recovery plans have not been tested recently against current architecture, restoration can become a voyage of discovery at exactly the wrong time, with teams clarifying dependencies while the business is still exposed. 

  • If cloud and security monitoring operate in separate silos, time is lost building a coherent picture of what is happening. That delay slows decision-making and extends impact. 

In hindsight, it becomes obvious where coordination faltered. That’s the uncomfortable part. What looked sensible in isolation can feel fragile once tested by reality. 

“We invested” rarely satisfies the room 

Many organisations underestimate how long recovery can actually take. In real-world ransomware incidents, companies often face 1–3 weeks of disruption before full restoration, not just a few hours of IT work. In other studies, average recovery timelines from serious breaches have stretched to more than seven months, far longer than executives expected. That gap between expectation and reality can magnify financial impact, erode trust and turn inevitable scrutiny into a strategic problem, especially when you’ve assured boards or regulators of your readiness.  

Spending has increased across industries. So have breach costs. 

The gap isn’t usually about buying the wrong technology. It’s about whether those technologies and the teams who oversee them operate as a coordinated system when you’re under pressure. 

Without testing your recovery plans across cloud, security and network domains together, factoring in how these systems are interconnected, you are relying on assumptions. And assumptions do not age well in hindsight. 

In financial services, scrutiny has layers 

If you operate in financial services, this dynamic is even sharper. 

Cloud, cybersecurity and network functions often maintain separate risk registers. Each reports upward. Each tracks its own exposures. Each demonstrates mitigation plans. 

But incidents don’t respect those boundaries. 

When disruption hits, the board doesn’t review three risk registers. It reviews one outcome. 

Under UK operational resilience expectations, firms are required to demonstrate that important business services can remain within defined impact tolerances. This moves the focus away from documentation and toward performance. Can you actually recover in the real world, not just describe how you would? 

Even where customer impact is limited, regulatory scrutiny can intensify if your response appears uncoordinated or poorly evidenced. Post-incident reviews do not just examine what happened. They examine how well you were prepared for it. 

Siloed governance can feel comprehensive until an incident reveals the seams. 

In retail, the pressure is immediate and commercial 

In retail, the pressure is different but no less intense. 

Network controls are a preventative measure, not an afterthought. They are designed and implemented long before an incident occurs to ensure stores, warehouses, offices and digital platforms stay connected when pressure rises. 

Your network underpins far more than your website. It connects tills, handheld devices, headsets and tablets in-store. It links distribution centres to inventory systems. It supports office locations and customer service teams. At the same time, your cloud platforms are driving online sales and customer transactions. Networks can be architected with intelligent traffic routing and prioritisation so critical services remain available, even during failover or restoration activities. 

Prevention really is better than cure. The network is the last thing you want to lose, because everything else depends on it. That is why resilience testing should include scenarios where cloud failover, recovery traffic and peak trading demand collide. If you have not tested whether your network can handle that combination, you are relying on assumption. 

During peak trading, even small performance issues translate quickly into lost revenue and damaged trust. Research shows that 53% of mobile visitors abandon a page if it takes more than three seconds to load, and bounce rates increase by 123% as load times extend from one to ten seconds. 

Prevention is no longer the only measure of success 

Boards and regulators increasingly accept that you cannot eliminate all IT failures. 

What they are watching for is how you respond: 

  • Did you detect quickly? 

  • How effective was the incident response plan? 

  • Was ownership clear from the outset? 

  • Did teams and suppliers operate as one? 

  • Were decisions taken confidently, or did they stall due to guesswork or hesitation? 

  • Can you now explain those decisions clearly and defensibly? 

A good incident response plan, tested regularly through realistic simulations, should cover all of this. 

The board will not ask whether you had a plan. They will ask whether it worked. 

Response is the proof point of strategy. Did you do everything reasonably possible to limit harm before the incident, and everything necessary to control it afterwards.

What good looks like when you replay the incident 

When you examine an incident from a board perspective, the difference between preparedness and improvisation becomes clear. 

  • Ownership is clear and teams are empowered to take action and make decisions. 

  • Recovery pathways are understood in advance. 

  • Security containment supports restoration rather than complicating it. 

  • Network monitoring tools enable visibility to make traffic prioritisation and failovers seamless. 

  • Teams and partners move in a coordinated rhythm. 

That kind of response rarely happens by accident. It is designed deliberately and rehearsed under realistic conditions with incident response playbooks and tabletop simulations. 

The strongest organisations prepare for board scrutiny by treating their recovery capability as a strategic decision, not an operational afterthought. 

The best plan is one that has been tested, improved and regularly re-tested, long before it is needed. 

The question worth asking now 

If you had to defend a major disruption tomorrow, would you be able to explain not only what happened, but why recovery was designed the way it was? 

You don’t control unlimited budget. You do control how clearly you articulate the risk of underinvestment, and how honestly you align recovery capability with declared tolerance. 

When the board asks “could this have been avoided?”, they are often asking whether they were fully informed of the risk. 

  • Could you clearly describe the recovery pathway? 

  • Would you be comfortable defending the trade-offs made under pressure? 

  • Would you know exactly where the coordination of multiple teams held up, or where it didn’t? 

If those answers feel uneasy, this is your early warning. 

Pressure-testing your IT resilience ensures that when disruption inevitably arrives, you are prepared not just to recover quickly, but to justify your past decisions and demonstrate lessons learnt, and that it will result in better recovery performance when it next happens. 

We work with you to modernise and transform your IT, so that you can reduce the risk and impact of IT incidents and recover quickly when the worst happens. 

Because when something goes wrong, hindsight is unforgiving. 

Foresight is a choice. 

If you want to know how your strategy would stand up in a real board review, test it before you have to defend it. 

We work with CIOs, CISOs and risk leaders to pressure-test recovery across cloud, security and networks together, identifying where coordination holds and where it doesn’t. 

Before your next incident turns into a governance conversation, talk to us.