Your business continuity plan is sitting in a shared folder right now. The executive summary is reasonably current. The data center section got updated last year. And then there's the telecom section—probably unchanged since 2021, with carrier phone numbers that route to deprecated departments and circuit IDs that don't match your current infrastructure.
This gap exists almost everywhere. Telecom disaster recovery gets less attention than other business continuity domains because outages feel random, recovery feels out of your control (you're dependent on carriers), and most plans haven't been stress-tested. But when your primary ISP fails at 9 AM on a Tuesday, you'll learn very quickly what your actual RTO and RPO are, and whether your "redundancy" is real or theoretical.
The good news: telecom DR is both manageable and testable. It requires discipline and specificity, not exotic technology. This guide walks you through the components that actually hold up in outages.
The Failure Modes That Actually Happen
Start by knowing what you're building resilience against. Real telecom outages cluster around a small set of root causes:
Fiber cuts dominate. They account for roughly 70% of telecom outages in most regions. A utility contractor hits a conduit, a tree falls on an aerial line, or scheduled fiber maintenance goes wrong. The outage is often widespread—a single fiber cut can affect hundreds of customers on the same route. These are the most common, most predictable, and yet least planned-for failure mode in most continuity plans.
Carrier PoP failures take out an entire point of presence. This might be power loss, cooling failure, equipment failure, or software bugs in core routing equipment. When a major carrier's PoP in your region goes down, every circuit terminating there is affected simultaneously. Regional carriers have better redundancy than they used to, but PoP failures still happen 2–3 times per year in most large metros.
Data center power outages at the termination point cascade into your telecom. Your ISP's demarc is in a data center. If that facility loses power and generator failover is slow or incomplete, your circuits go dark even if the carrier's backbone is fine. This is less common than fiber cuts but longer-duration when it happens.
Building inaccessibility is increasingly common post-pandemic. Physical access restrictions due to fire, flooding, HVAC failure, or security incidents prevent staff from accessing the phone system, demarc, or local router. You have working circuits but can't reach the equipment that terminates them.
DNS failures take down SIP trunks and internet calling. If your SRV records are stale, your SIP provider is unreachable, or your registrar is compromised, VoIP stops working while voice carrier circuits might still be live. This is a common failure mode for hybrid environments where you're bridging legacy TDM with SIP.
BGP route hijacking and intentional route withdrawals can redirect your traffic to unexpected destinations. This is rarer for most organizations than the above, but it happens. Knowing your BGP AS number, your carrier's AS number, and having monitoring in place is baseline defense.
Notice what's missing from this list: "all our vendors failed simultaneously" and "we lost all carriers everywhere." You're building resilience against partial, localized failures—which is both the realistic threat and the one you can actually defend against.
Defining RTO and RPO for Telecom
RTO (Recovery Time Objective) for telecom means: how long can calls be down before material business impact occurs? RPO (Recovery Point Objective) means: how much call history, voicemail, or session data can you lose?
Most organizations haven't set these numbers specifically for telecom. They've set organizational RTO (usually "4 hours" or "1 hour") and assumed it applies to voice. It doesn't. Voice has different failure modes and different recovery mechanisms than data systems.
For call-critical roles (customer service, sales, emergency response), realistic RTO is 15–30 minutes. You can absorb brief silence. You cannot accept customers getting a "carrier unavailable" message for an hour. Once a call can't complete, customer churn and transaction loss accelerates sharply after 30 minutes.
For standard office staff, RTO is usually 60–90 minutes. Employees can use mobile devices for critical calls in the interim. Beyond 90 minutes, productivity loss mounts and staff become frustrated.
For internal-only roles (accounting, HR, IT), RTO can be 2–4 hours. These calls are necessary but not directly customer-facing.
For RPO: how much call history can you lose? If your phone system goes down and you're recovered to the previous day's backup, you've lost a full day of call logs and voicemail. For compliance-heavy industries, this might be unacceptable. For others, 24 hours of call data loss is acceptable. Know your requirement. Document it. Build your backup schedule around it.
Building Redundant Telecom
Dual ISPs are table stakes. These must be on different carriers using different physical paths. "Different carriers" doesn't mean you ordered both from Comcast—it means Comcast and Verizon, or Comcast and a regional fiber provider. Check the physical routing. If both paths cross the same fiber conduit for the last mile, you don't have redundancy; you have the same fiber cut taking out both.
Request from your carriers: the physical path documentation. Ask specifically which conduits, which poles, which fiber strands carry your circuit. If they won't provide it, push back. You can't build intelligent redundancy without understanding the actual physical layout.
4G/5G LTE as a backup ISP is underutilized and highly effective. A dedicated cellular hotspot (not tethering from an employee's phone) provides true redundancy. Real-world bandwidth: 4G LTE averages 20–50 Mbps download, 10–20 Mbps upload. 5G mmWave in deployed areas hits 100–400 Mbps down, 20–100 Mbps up, though coverage is spotty and range is limited. Even on 4G, you can run VoIP adequately (voice calls use roughly 100 Kbps), and basic data traffic (email, messaging) remains usable. LTE fails over when your primary ISP is down, and failover can be automatic with SD-WAN.
SD-WAN for automatic failover removes manual intervention from recovery. When your primary link is down, SD-WAN automatically routes traffic to your secondary ISP or LTE link. This is not magic—SD-WAN detects link failure through health checks—but it is automatic, fast (seconds), and doesn't require someone to manually reconfigure routing. Most modern SD-WAN platforms (Meraki, Fortinet, Cisco) support this out of the box.
Cloud phone number forwarding is a hidden gem. If you lose access to your on-premise phone system, cloud-based number forwarding (provided by your carrier or a third-party service) can route inbound calls to mobile devices, remote offices, or a temporary call center. Set this up now, while you have time. During an outage, you won't.
DID failover to mobile. Direct Inbound Dial numbers can be configured to forward to mobile devices during outages. Verify this works: call your own DID number and confirm it rings your cell phone. This is simple, often forgotten, and invaluable when your main office is inaccessible.
Maintaining Call Continuity During Outages
If your organization uses cloud-based UCaaS (Microsoft Teams, RingCentral, Zoom Phone, etc.), you have an advantage: the phone system is already geographically distributed. If your office is inaccessible but your internet is restored via LTE or remote access, your cloud phone system continues working. The caveat: this assumes your network is actually restored and your SIP trunks can reach the cloud provider. If your ISP is down and you have no LTE failover, no amount of cloud-based redundancy helps.
For hybrid environments (on-premise PBX with SIP trunks), the same principle applies: calls route through the cloud. But you need SIP failover configured. If your primary SIP trunk provider is unreachable, does your PBX automatically failover to a secondary trunk provider? Most systems support this, but it must be configured and tested. A common failure: PBX failover is configured, but the backup SIP provider uses different codecs or authentication methods, so failover attempts fail silently.
Auto-attendant DR modes are worth setting up explicitly. During an outage, you may not need a full phone system—you need callers to reach a human or get meaningful information. Configure your auto-attendant to:
- Route all calls to a specific department or person during outages.
- Play a message explaining the outage and provide alternative contact methods.
- Forward to a mobile number or remote call center.
During an outage, your first question will be: "Is this our problem or the carrier's?" If you can't reach your phone system admin, you can't make any changes. Ensure remote access to your phone system is available and tested. This might mean:
- VPN access to the PBX management interface.
- Out-of-band management (cellular modem on the PBX itself).
- Cloud-based PBX management portal.
Testing Your Telecom DR Plan
A telecom DR plan that hasn't been tested is a document. A tested plan is insurance.
Quarterly failover tests are the minimum. Each quarter, initiate a controlled failure:
- Disable your primary ISP (or have your carrier isolate it) and verify traffic fails over to your secondary ISP or LTE link.
- Test SIP trunk failover: confirm calls route through your secondary SIP provider if the primary is down.
- Test DID forwarding to mobile: call your main number and verify it rings your cell phone.
- Verify remote access to the PBX: connect via VPN and confirm you can access the management interface.
Tabletop exercises test your team's response, not just the technology. Scenario: "Your main office building is flooded. Your primary ISP is down. Customers are calling. What happens in the first 30 minutes?" Run through decision trees: Who makes the failover decision? How do you communicate it? How long until calls are rerouting? These conversations, held in advance, are worth their weight during an actual outage.
Common things that fail in tests:
- Number portability assumptions. You planned to port your main number to a backup carrier, but you've never verified that the number can actually be ported. (Spoiler: some numbers are locked by legacy carriers.)
- Mobile forwarding rules are stale. The cell phones you configured forwarding to belong to staff members who left the company two years ago.
- Remote access credentials are expired. The VPN account for PBX admin access hasn't been used in 18 months and the password is lost.
- SIP trunk authentication is broken. The backup SIP trunk uses a different authentication method (TLS vs. digest) and the PBX isn't configured for it.
- Failover is slow or manual. Your backup ISP is configured, but switching to it requires manual intervention. This is not failover; this is a manual recovery option, and it's much slower.
Documentation That Matters
Your continuity plan is only as useful as your team's ability to execute it during a stressful outage. Documentation must be:
- Accessible offline. Your network is down. You can't access the wiki. Print critical information or store it on a local drive.
- Current. Outdated escalation numbers and stale circuit IDs actively harm recovery. Audit documentation quarterly.
- Specific and actionable. "Call the ISP" is not helpful. "Contact Comcast Enterprise Support: 1-800-555-XXXX, account 123456, ask for the Network Operations Center. If transfers to NOC take more than 3 minutes, escalate to the account manager: Direct: 503-555-XXXX, mobile: 503-999-XXXX."
What to document for each carrier:
- Direct escalation phone numbers (not the 800 line—that routes to support agents in a call queue).
- Circuit IDs for every location.
- Responsible account manager contact info (direct number and mobile).
- Emergency contact for after-hours support (many carriers have this; yours might not be configured).
- Physical demarc location (address, floor, room).
- Port numbers on the carrier's equipment (the physical port your circuit terminates on).
PBX documentation:
- Admin access credentials (stored in a credential vault, not in the plan).
- Remote access method (VPN address, account, instructions).
- Dial plan for failover: which button activates the auto-attendant DR mode?
- SIP trunk primary and secondary provider info (hostnames, authentication method, codec requirements).
- Mobile forwarding numbers for each on-call role.
Credential management: Don't embed passwords in your continuity plan. Use a credential vault (Vault, 1Password, Bitwarden) that your team can access during an outage. Test access quarterly. One common failure: the person who set up the vault left the company and took the master password.
The most common DR planning gap: Carrier escalation paths with wrong or outdated numbers. You've documented "Call Comcast support: 1-800-COMCAST" in your continuity plan. During an outage, you reach the general support line, get transferred to residential support, wait 45 minutes, and learn they don't handle your account. Meanwhile, 30 minutes of your RTO has already elapsed. Instead, document the direct number for your account team, your account manager's cell phone, and the Enterprise Support escalation path. Test these numbers quarterly. Update them immediately when staff changes occur at the carrier.
Frequently Asked Questions
How often should we actually test our telecom DR plan?
Quarterly is the recommended minimum. This means four tests per year, one per quarter. Each test should include at least: failover to secondary ISP, verification of SIP trunk failover, and testing of DID forwarding to mobile. If you have multiple office locations, rotate which location you test each quarter. If you test more frequently, even better—but quarterly is the threshold where you'll catch configuration drift and staff changes.
What bandwidth can we actually expect from 4G LTE for business continuity?
4G LTE typically delivers 20–50 Mbps download and 10–20 Mbps upload in real-world deployments. This is sufficient for VoIP (voice calls use ~100 Kbps), email, and instant messaging. It's not sufficient for large file transfers or high-bandwidth applications. 5G mmWave, where available, provides 100–400 Mbps down and 20–100 Mbps up, but coverage is limited to specific areas and requires line-of-sight to the tower. For business continuity, assume 4G availability in your area and 5G as a bonus if it's deployed nearby. Use a dedicated cellular hotspot (with a business data plan), not employee phone tethering, so you have guaranteed priority and access during emergencies.
If we use Microsoft Teams or RingCentral, does our phone system automatically work during a building outage?
Not automatically, but conditionally. If your cloud phone system can reach the internet (via your backup ISP or LTE failover), calls will continue to route to employees' devices. But this requires: (1) your backup ISP is actually available and failover happened, (2) your SIP trunks can reach the cloud provider's servers, and (3) your employees have network access from wherever they are (home, mobile, etc.). If any of these conditions is not met, cloud phones fail just like on-premise phones. The advantage: you don't have to be in your office to use them. The requirement: your network redundancy must be working.
What should we actually do in the first 30 minutes of a telecom outage?
First 30 seconds: Verify the problem is real. Ping the demarc, check ISP status page, call the carrier's automated status line (many carriers have this). First 5 minutes: Activate failover (secondary ISP, LTE, backup SIP trunk). If failover is automatic, this is already happening. Verify it worked. First 15 minutes: Communicate. Email staff, post status update, notify customers if their service is affected. First 30 minutes: Begin diagnosis. If this is a carrier issue (PoP failure, fiber cut), they own recovery and you should have escalation in progress. If this is your issue (equipment failure, configuration), begin remediation. Document timeline.
Do carrier SLAs matter for disaster recovery planning?
Only marginally. Carrier SLAs typically promise 99.9% uptime, which sounds good until you do the math: 99.9% uptime = 43 minutes of acceptable downtime per month. When a fiber cut happens, the SLA doesn't restore your service faster; it just means you might get a service credit after the outage is over. Build your DR plan around the assumption that when a carrier fails, they will take their SLA recovery time to restore service. Your redundancy should bridge that gap. SLAs are useful as a baseline for comparing carriers and understanding their commit, but they should not be your recovery strategy.
4G/5G LTE failover is the most underused and highly effective disaster recovery tool available to most organizations. It's inexpensive ($50–100/month for a business data plan), requires minimal configuration, and provides genuine redundancy when your primary ISP fails. Yet the majority of organizations we audit have no LTE failover in their continuity plan, or they've attempted it with employee phone tethering (which is unreliable and doesn't scale). A dedicated cellular hotspot, integrated into your network via SD-WAN or simple failover routing, costs almost nothing and has single-handedly prevented extended outages in dozens of implementations we've seen. If your continuity plan doesn't include LTE failover, this is the highest-ROI addition you can make.
Next Steps: Getting Your Plan From Paper to Practice
A telecom disaster recovery plan is only as strong as your team's ability to execute it. Start here:
- Audit your current redundancy. Do you actually have dual ISPs on different physical paths? Document the paths.
- Set RTO and RPO explicitly. Decide: how long can voice be down? How much call history can you lose? Document the decision.
- Document carrier escalation paths. Get direct numbers, not 800 numbers. Get your account manager's mobile. Verify the numbers work.
- Schedule your first failover test. Quarterly tests going forward.
- Add LTE failover if you don't have it. A dedicated cellular hotspot is low-cost, high-value insurance.
- Test remote access to your phone system. Verify VPN access works, credentials are current, and admin can manage the system remotely.
Telecom disaster recovery isn't glamorous, and it's not flashy. But when a fiber cut hits on a Monday morning and your calls start failing, a well-executed DR plan is the difference between a 15-minute recovery and a 4-hour outage. That's the operational reality. Build it now.