NSA Didn't Break First. Our Assumptions Did.
5G NSA · LTE · architecture assumptions . 7 min read
There is a common narrative that early 5G NSA deployments struggled because the architecture was transitional. That framing misses what actually happened. NSA did not introduce new problems. It made existing ones visible at a scale that could no longer be ignored.
The LTE anchor had always been treated as the stable, predictable layer. That assumption held when LTE carried LTE traffic and nothing else. Once NR was layered on top, small inconsistencies in the anchor became amplified in ways that no amount of NR-side tuning could fix.
Where the assumption failures actually surfaced
Fig 1 -- NSA layer interaction: three failure classes tied to LTE anchor assumptions, not NR behavior
Each of these failure classes was present in the LTE network before NSA. They had been managed, worked around, or accepted as within-threshold. NSA changed the cost of each one, because NR availability and user-plane quality were now directly dependent on anchor stability.
Anchor reselection timing -- what it looked like in the data
Pre-NSA: anchor cell reselection visible as brief HO event, KPI within target
Post-NSA: same event triggers NR SCell release and re-addition
Re-addition delay: 280-450ms
During re-addition: user plane on LTE only (no NR throughput)
Session impact: invisible in voice KPIs, significant in data experience
Aggregate NR availability: appeared healthy
Per-session NR utilization: 60-70% of session time, not 90%+
Why 2024 traffic made it harder
Applications no longer behaved in long, steady sessions. Short bursts, background sync, and latency-sensitive flows forced the network to make more decisions per second, especially at state boundaries. The scheduler, the mobility layer, and the anchor selection logic were all making independent decisions at higher frequency than the tuning model anticipated.
Traffic diversity impact on NSA decision rate -- 2024 vs 2020:
2020 traffic profile (at NSA launch):
Dominant pattern: sustained video, large file transfers
Session duration: minutes
State transitions per session: low
Scheduler decisions per session: high but predictable
2024 traffic profile:
Dominant pattern: short bursts, app sync, real-time feeds
Session duration: seconds to tens of seconds
State transitions per session: high
Scheduler decisions per session: compressed into shorter windows
Effect on NSA:
More anchor reselection events per device per hour
More NR SCell add/release cycles
More opportunities for timer misalignment to produce visible impact
Higher sensitivity to LTE scheduler congestion spikes
What the effective fixes actually targeted
The most impactful changes were not the ones that pushed NR performance harder. They were the ones that reduced unnecessary variability in the anchor layer. Tuning NSA turned out to be mostly a discipline exercise in the LTE configuration.
Fix category
What was changed
Why it worked
Timer alignment
T310, T311, SCell deactivation timer synchronized across anchor and secondary
Eliminated timer-driven SCell releases that were not triggered by actual radio conditions
State churn reduction
RRC inactivity timer extended at high-activity cells; C-DRX cycle aligned with burst pattern
Reduced unnecessary idle-to-connected transitions that triggered SCell re-addition cycles
Anchor selection
Anchor cell candidate list filtered by load and mobility history, not coverage alone
Devices landed on anchors less likely to reselect during active NR sessions
Throughput vs consistency
Aggressive CA and NR MIMO configurations relaxed at high-churn cells
Slightly lower peak throughput, significantly higher session consistency
None of these were high-visibility changes. They did not show up in vendor feature lists or release notes. They were the kind of unglamorous alignment work that only becomes necessary once you are measuring what the user actually experienced rather than what the network reported.
How validation had to change
Fig 2 -- Validation scope: the shift from lab feature checks to live network behavioral testing
Lab success and vendor certification confirmed that a feature functioned. They did not confirm how it behaved when combined with real device population, live mobility, and actual scheduler contention simultaneously. Those conditions are what users experience. They had to be what validation tested.
NSA readiness validation criteria -- what changed:
Previous standard:
NR SCell addition success rate: above threshold
Throughput in test drive: meets minimum
Feature enabled in config: confirmed
Vendor acceptance: passed
Updated criteria:
NR session continuity during anchor reselection events
(SCell available within 200ms of re-addition trigger)
NR SCell utilization per session: p50 above 80% of session time
State transition rate at busy hour: within timer design range
Anchor reselection rate during active NR sessions: below 3%
All measured under mixed broadband + IoT + background sync load
NSA did not fail because it was a transitional architecture. It was hard because it demanded discipline across layers that legacy tuning models treated independently. The LTE anchor assumptions, the mobility decisions, the scheduler priorities -- all of them were designed before NR dependence existed. Adjusting for that dependence was the real work of NSA optimization.
That discipline carried forward into SA readiness work and everything that followed. The lesson was not specific to NSA. Any time a new capability depends on the behavior of an existing layer that was not designed with that dependency in mind, the assumptions in that existing layer become the constraint. Finding them requires measuring the right things. Fixing them is usually quieter than expected.
5G NSA · LTE · RAN Optimization · Architecture · Performance Engineering · OSS Analytics · Telecommunications