5G/NR - URLLC  

 

 

 

URLLC (Ultra Reliable Low Latency)

As the name implies, this is to guarantee the traffic with Low Latency and Very High Reliability at the same time. It has to satisfy the most challenging two requirement (Latency and Reliability) simultaneously.

How Low and How Reliable it should be ?

According to RP-191584 and TR 38.913-7.9, this requirement is specified as follows : (I think this is for single hop (e.g, between UE and gNB) probably at the layer of MAC/RLC)

  • Reliability : up to 10^-6
  • Latency : 0.5 ~1 ms

In reality, I think more important performance indicator is end-to-end latency for each application and use case. Following table is the summary of the latency requirement for each application/use cases that I collected from various references listed in the reference section.

Application

Latency

Service Area

Description

VR

 

 

 

 

 

360 VR Video

20 ~ 40 ms

 

4K (3845x2160), 60 FPS, 20~40 Mbps

 

 

90 ~ 130 ms

 

8K (7680x4320), 90 FPS, 90~130 Mbps

 

 

~10 ms

 

12K 3D(11520x6480),120 FPS, 500~700 Mbps

 

CG VR Video

~20 ms

 

2K (2560x1440), 70 FPS, 30~50 Mbps

 

 

~16 ms

 

4K (3840x1920), 90 FPS, 50~200 Mbps

 

 

~10 ms

 

8K (7680x3840), 120 FPS, 200~800 Mbps

 

 

 

 

 

V2X

3~10 ms (E2E)

few meters

 

 

Remote Driving

5 ms (E2E)

 

1 Mbps/DL, 20 Mbps/UL, speed ~250 km/h

 

Collective Perception

3 ms (E2E)

200 m

Exchange of real-time information among vehicles

 

 

10 ms (E2E)

500 m

 

 

 

50 ms (E2E)

1000 m

 

Industrial Automation

 

 

 

 

Motion Control

0.5 ~ 2 ms (E2E)

 

 

 

Discrete Automation

10 ms (E2E)

 

~10 Mbps

 

Process Automation

100 ms (E2E)

 

1~100 Mbps

 

 

 

 

 

 

<  22.261 v17.11 Table 7.6.1-1 KPI Table for additional high data rate and low latency service >

What is it for ?

Why we need this kind of challenging feature ? It seems that they have some use case as follows in mind when they investigated on this. Following is the use case specified in RP-191584

  • Release 15 enabled use case improvements
    • Such as AR/VR (Entertainment industry)
  • New Release 16 use cases with higher requirements
    • Factory automation
    • Transport Industry, including the remote driving use case
    • Electrical Power Distribution

How to Reduce Latency ?

To reduce the end-to-end latency, we need to reduce the latency both on RAN side and on Corenetwork side. Let's briefly think of how we can reduce the latency on each side. Some common technologies to reduce the latency on RAN and Corenetwork side can be summarized in illustration as shown below. RAN side latency reduction is relatively well described in 3GPP specification (mostly in Release 16), but Core network side technology seems to be more up to implementation. Personally I think Core network latency would play more role in terms of end-to-end latency and it would take a while and huge investment to restructure and optimize overall architecture of the core network.

As the name implies, URLLC is not only for Latency Reduction. Reliability improvement is also important to meet the super high standards of URLLC. I think this part would be even more challenging than reducing the latency. A common trick to increase the reliablity on RAN side is to use MCS table for low code rate and repetitive data transmission. Unfortunately this would result in throughput reduction, but it is not easy to think of any other means to increase reliability without sacrificing some throughput performance.  In addition, we would come up with some solutions to increase the reliability on coreside (i.e, reliability over coreside data path), but as of now (Mar 2021) I don't have any specify knowledge on how to improve core side reliability. I will update on this as I learn further.

What cause the delay ?

As the term implies, the critical component of URLLC (Ultra-Reliable Low Latency Communication) is about delay… more strictly, it's all about reducing delay. In order to reduce delay, we need to understand the source of the delay in the first place.

Why is minimizing latency so crucial?  Well, in many next-generation applications, even milliseconds matter. Think of remote surgery, where a surgeon needs real-time feedback from robotic instruments, or autonomous vehicles, where split-second decisions can be the difference between safety and danger.  In these scenarios, traditional network communication just won't cut it.

URLLC aims to tackle these delays head-on, employing a range of techniques to ensure lightning-fast and incredibly reliable communication. In this blog post, we'll dive deep into the world of URLLC, exploring its key technologies, applications, and the challenges that lie ahead.

In this note, we will look into some of the major source of delays. The initial list came from Ultra-Reliable Low-Latency in 5G: A Close Reality or a Distant Goal? (Arman Maghsoudnia) and further details will be added as I learn more.

Source of the dealy

There are the different sources of latency that hinder the achievement of Ultra-Reliable Low-Latency Communication (URLLC) in 5G. Each source contributes uniquely to the overall latency, and their interplay is critical in determining system performance.

URLLC is a key feature introduced in 5G to support mission-critical applications such as autonomous vehicles, remote surgery, industrial automation, and augmented reality, which demand extremely low latency (as low as 0.5ms) and high reliability (99.999%). Despite years of theoretical research and advancements, practical implementations of URLLC often fall short of meeting these stringent requirements due to bottlenecks in system design, protocol inefficiencies, and hardware/software limitations

Processing Latency

  • Processing latency refers to the time required for decision-making and data processing within the different layers of the 5G system. This latency is distributed across the User Equipment (UE), the gNB (next-generation NodeB), and the core network.
  • Examples:
    • UE Side: When a UE sends a Scheduling Request (SR), it first processes the data through layers such as the Application (APP), Service Data Adaptation Protocol (SDAP), Packet Data Convergence Protocol (PDCP), Radio Link Control (RLC), Medium Access Control (MAC), and Physical (PHY) layers. Each layer adds a certain amount of processing time.
      • Example: Encrypting data in the PDCP layer or reassembling packets in the RLC layer introduces delays.
    • gNB Side: The gNB receives and processes the uplink (UL) data through the PHY, MAC, and higher layers. This includes decoding the PHY layer, scheduling resource allocation in the MAC layer, and reassembling data in the RLC layer.
      • Example: Table 2 in the document shows the mean processing times for gNB layers, with the RLC queue (RLC-q) contributing the highest latency at 484.20 µs on average.
  • Key Insights:
    • Software-based implementations can exacerbate processing latency due to non-deterministic Operating System (OS) scheduling.
    • Resource-heavy operations like encryption, modulation, and demodulation consume significant time.

Protocol Latency

  • Protocol latency arises from mechanisms and configurations required by the 5G protocols. These include scheduling, resource allocation, and the inherent delays in the transmission procedure.
    • Examples:
      • Scheduling Delays: In grant-based uplink transmissions, the UE must send an SR and wait for the gNB to allocate resources (UL Grant) before transmitting data. This two-step process significantly increases latency.
    • Example: In a TDD system, the UE has to wait for the next uplink slot before transmitting the SR, and again for the next slot to send the UL data after receiving the grant. This introduces protocol-induced delays of up to one TDD period (e.g., 0.5 ms for a DDDU configuration).
  • Slot Duration: Protocol constraints dictate that data must fit within predefined time slots. If a packet arrives just after a slot is allocated, it must wait until the next available slot.
    • Example: In the FR1 band (<6 GHz), a minimum slot duration of 0.25 ms is required for low-latency configurations. However, packets arriving just after a scheduling event experience an additional 0.25 ms delay.
  • Key Insights:
    • Grant-free uplink transmissions can reduce protocol latency by pre-allocating resources, but they are less scalable for networks with many UEs.
    • Mini-slot configurations provide finer-grained scheduling but increase signaling overhead and network coordination complexity.

Radio Latency

  • Radio latency encompasses delays in the radio hardware and its interaction with the PHY layer. It includes analog-to-digital (A/D) and digital-to-analog (D/A) conversions, queuing delays on interface buses, and transmission delays.
    • Examples:
      • A/D and D/A Conversion: Converting signals from analog to digital (or vice versa) at the Radio Head (RH) introduces latency.
    • Example:  The RH(radio head) in the testbed (connected via USB) added 500 µs of radio latency, making it impossible to meet URLLC requirements in certain configurations.
  • Bus Transmission Delays: The time required to transfer data over interfaces like PCIe, Ethernet, or USB affects latency.
    • Example: USB 3.0 performs better than USB 2.0 for submitting radio samples but still experiences spikes in latency due to OS scheduling delays.
  • Queuing Delays: Queuing data on the radio interface or during inter-layer communication can bottleneck performance.
    • Example: The queuing delay at the gNB’s RLC layer averages 484.20 µs, as shown in Table 2.
  • Key Insights:
    • Radio latency is highly dependent on the hardware interface. SDR-based systems offer flexibility but introduce higher latencies compared to ASIC-based implementations.
    • Sub-millisecond latency is challenging in non-line-of-sight scenarios, especially in mmWave bands.

Deterministic vs Non-deterministic latencies

Latency in 5G systems can be broadly understood in terms of deterministic and non-deterministic latencies, which represent the predictability of delays during data transmission and processing.

Deterministic Latency

Deterministic latency refers to delays that are predictable and consistent. These latencies occur in processes or components where timing behavior is predefined and not subject to significant variation.

Characteristics:

  • Predictability: Deterministic latency is consistent for a given operation or under specific conditions.
  • Controllable: Can be managed and optimized through system design, such as scheduling algorithms or hardware configurations.
  • Examples in 5G:
    • Slot Duration: The time required to transmit data in one slot is fixed based on the numerology and configuration (e.g., 0.25ms, 0.5ms).
    • PHY Layer Processing: Tasks such as encoding and modulation have predictable durations when executed on dedicated hardware like ASICs.
    • Scheduling Cycles: Periodic scheduling in the MAC layer, where resources are allocated in well-defined time slots.

Non-Deterministic Latency

Non-deterministic latency, on the other hand, arises from processes or components where delays are variable and unpredictable. These latencies are often caused by dynamic factors, such as system load, shared resources, or environmental conditions.

  • Characteristics:
    • Variability: Non-deterministic latency fluctuates, making it harder to predict.
    • Dependency on External Factors: Influenced by factors like OS scheduling, hardware interfaces, and wireless channel conditions.
    • Examples in 5G:
      • OS Scheduling Delays: In software-based implementations, the operating system may prioritize other tasks, delaying critical 5G processes.
      • Queuing Delays: Data packets waiting in buffers (e.g., at the RLC or MAC layer) experience variable delays based on traffic load and resource availability.
      • Radio Conditions: Wireless channel conditions, such as fading, interference, or mobility, introduce unpredictable delays

PHY/MAC Enhancement

Even though I haven't seen much of real implementation(UE and Network) and real network deployment in Release 15 as of now(Jan 2021), there are features in Release 15 that is introduced with URLLC in mind. The list of these features are well summarized in 5GAmericas white paper as follows :

As you may guess, we would need a lot of PHY/MAC enhancement to make this possible. Those enhancement is described in RP-191584 and 38.824 V16.0.0 (2019-03). As you see, it requires the enhancement across all physical channels and MAC scheduling.

  • PDCCH enhancements
    • DCI format(s) with configurable sizes for some fields, with a minimum DCI size targeting a reduction of 10~16 bits relative to Rel-15 DCI format 0_0/1_0 and a maximum DCI size that can be larger than Rel-15 DCI format 0_0/1_0, and provide the possibility to align with the size of the DCI format 0_0/1_0 (including possible zero padding if any)
    • Increased PDCCH monitoring capability on at least the maximum number of non-overlapped CCEs per slot for channel estimation for at least one SCS subject to restrictions including, but not necessary limited to, those identified in TR 38.824. Enhancements for PDCCH monitoring capability on the maximum number of monitored PDCCH candidates per slot (with potential restrictions) can be further considered.
  • UCI enhancements
    • More than one PUCCH for HARQ-ACK transmission within a slot
    • At least two HARQ-ACK codebooks simultaneously constructed, intended for supporting different service types for a UE
  • PUSCH enhancements for both grant-based PUSCH and configured grant based PUSCH
    • For a transport block, one dynamic UL grant or one configured grant schedules two or more PUSCH repetitions that can be in one slot, or across slot boundary in consecutive available slots
  • Scheduling/HARQ enhancement
    • Out-of-order HARQ-ACK associated with PDSCHs with different HARQ process IDs
    • Out-of-order PUSCH scheduling associated with different HARQ process IDs, including overlapping PUSCHs and non-overlapping PUSCHs in time-domain
    • Methods to handle DL data/data resource conflicts for overlapping PDSCHs in time-domain, scheduled by dynamic DL assignments
  • Inter UE Tx prioritization/multiplexing Enhancement
    • UL cancelation scheme
    • Enhanced UL power control scheme 
  • UL configured grant transmission Enhancement
    • Multiple active configured grant type 1 and type 2 configurations for a given BWP of a serving cell

Presentations on YouTube

[1] URLLC Standardization and Use Cases | ITN WindMill Project

[2] Can todays Internet protocols deliver URLLC?

[3] Teaser: 5G Remote Surgery - Fact or Fiction?

[4] The Okay, But How? Show, Episode 2: 5G and low latency

[5] Wireless Access for URLLC | ITN WindMill Project

[6] neXt Curve Webcast: URLLC Myths and Realities

[7] Deterministic Networking for Real-Time Systems (Using TSN) - Henrik Austad, Cisco Systems

[8] AWS Wavelength - Edge Computing for 5G Networks

[9] What is Edge Computing and its Impact on 5G?

[10] Keynote: Network of Tomorrow - 5G and Edge - Rajesh Gadiyar, Intel

[11] 5G Connected Edge Cloud - Keynote by Andre Fuetsch, AT&T

[12] Aether - Managed 5G Edge Cloud Platform

[13] 5G network slicing: automation, assurance and optimization of 5G transport slices (Nokia)

[14] AREA Webinar | The Challenges of Enterprise AR Implementation

[15] Technical Challenges of Effective VR 360 Delivery

[16] B5: Overcoming the Technical Challenges of Delivering VR Video

[17] Why Virtual Reality Is STRUGGLING

[18] AR VR Women - Silicon Valley @ Nvidia - Technical Challenges for VR and AR - 2016 06 15

[19] R&S Thirty-Five: C-V2X and end-to-end application testing

[20] Intermediate: Vehicle to Everything (V2X) Introduction  

[21] V2X Technology and Deployment : As of Sep. 2020

[22] What is V2X - V2X Communciation Explained Part 1

[23] DSRC or 5G?

[24] How Can C-V2X Create an Environment that Improves Quality of Life for Everyone? (GSMA)  

[25] Dr. Javier Gozalvez - V2X Networks for Connected and Automated Driving - IEEE LCN 2019 Keynote (IEEE)

[26] What is TSN Part 1: Where We Are Today

[27] What is TSN Part 2: What is Time-Sensitive Networking?

[28] What is TSN Part 3: Why is Time-Sensitive Networking Important for Automation?

[29] 5G and the Future of Connected Cars

Video Demo :

[1] Taking factory automation to the next level with Audi through URLLC

[2] Field Trials on 5G mMTC and uRLLC in Japan

[3] 5G Smart Factories

[4] Nokia 5G Demonstration Video 5G: driving the automation of everything

[5] Part 10: 5G Use cases - 5G for Absolute Beginners

[6] Powering safer vehicle production with 5G

[7] 5G for Robotics: Ultra-Low Latency Control of Distributed Robotic Systems

[8] Low-latency communication for Industrial IoT Demo

[9] Inside Ericsson King's College London 5G Tactile Internet Showroom

[10] Achieving industrial automation protocols with 5G URLLC (Ericsson)

[11] 5G: A Game Changer for Augmented Reality, Virtual Reality and Mixed Reality (Verizon)

[12] 5G enabled VR GLASSES are COMING in 2021 | HAND TRACKING SUPPORT! | HP REVERB G2 images 'Leaked'

[13] NTT DoCoMo demos VR control of robots over 5G

[14] Best THEATER Apps For The Oculus Quest

[15] SKTelecom 5G]SK Telecom Uses 5G AR to Bring Fire-Breathing Dragon to Baseball Park (SKTelecom)

[16] IMR Wireless VR streaming over 5G

[17] Oculus Quest XR2 - 5G, 8K Resolution, and SEVEN Cameras!?

[18] XR - The Merging of Augmented Reality AR, Virtual Reality VR and Mixed Reality in 2020

[19] Why Microsoft Uses Virtual Reality Headsets To Train Workers

[20] Inside Imogen Heaps cutting-edge VR concert | The Future of Music with Dani Deahl

[21] Wave | Virtual Concerts (Mixed Reality (XR))

[22] Experience VR concerts and move around freely! - NOYS VR gameplay

[23] CES 2021 brings INSANE new VR Technology

[24] HoloLens 2 AR Headset: On Stage Live Demonstration

[25] HoloLens 2: inside Microsoft's new headset

[26] What the heck is MEC?

[27] Demo: Network Slicing with Blue Planet 5G Automation

[28] Ericsson showed live end to end 5G network slicing at Mobile World Congress LA

[29] ZTE 5G End to End Network Slicing Demo

[30] Oculus Quest 2 Review - 100 Hours Later

[31] The Evolution of Virtual Reality by 2025

[32] V2X Technology and Deployment : As of Sep. 2020

[33] 5G C-V2X Technology Evolution Simulation (Qualcomm)

[34] Introducing Cellular V2X (Qualcomm)  

[35] Honda V2X Communications and automated driving (Honda)

[36] Cellular V2X paving the path to 5G (Qualcomm)

[37] 5G and the Future of Connected Cars

[38] Autotalks in a major live C-V2X testing on the streets of Shanghai  

[39] Boneworks - Next Gen VR Gameplay!

[40] Is Your Internet FAST Enough?

[41] What Is the Best Internet Speed for Gaming? [Simple Guide]

[42] Testing Google Stadia at 3 Different Connection Speeds!

[43] Oculus Quest Virtual Desktop Ultimate Low Latency Setup - Best Hardware For 20ms Latency   

[44] Virtual Desktop Settings: Ultimate Performance, Graphics & Low Latency: Streamer App & Streaming Tab

[45] Wireless VR Sim Racing is here! (90hz low latency step by step guide for Oculus Quest 2)

[46] Qualcomm Technologies and Siemens Drive the Future of Industrial Manufacturing with 5G (Qualcomm)

[47] Qualcomm Snapdragon XR2 5G Platform turns the ordinary into extraordinary XR (Qualcomm)

[48] Qualcomm 5G NR Industrial IoT Demo (Qualcomm)

[49] Qualcomm, Bosch highlight 5G as an industrial enabler (Qualcomm)

Papers/Whitepaper