6G Network Slicing for Low-Latency AI-Edge Deployments

Authors

  • Hari Krishnan Independent Researcher Perambur, Chennai, India (IN) – 600011 Author

DOI:

https://doi.org/10.63345/081cja89

Keywords:

6G Network Slicing, Low-Latency, AI-Edge Computing, Slice Orchestration, Predictive Slicing

Abstract

The evolution toward sixth-generation (6G) wireless networks represents a paradigm shift in how connectivity, computation, and intelligence converge to enable a new class of applications. At the forefront of this evolution are low-latency AI-edge deployments—scenarios in which artificial intelligence (AI) inference and decision-making occur at or near edge devices, necessitating end-to-end communication delays in the sub-millisecond range. Traditional 5G network slicing approaches, which statically or reactively allocate resources, often struggle to meet these stringent latency requirements under highly dynamic traffic patterns. This manuscript presents a comprehensive investigation of advanced 6G network slicing strategies tailored specifically for low-latency AI-edge use cases. We begin by contextualizing 6G’s technical enablers—such as terahertz (THz) communications, reconfigurable intelligent surfaces (RIS), and integrated sensing, computing, and communication (ISCC)—and their implications for supporting ultra-reliable, low-latency communications (URLLC). Next, we review the state of the art in network slicing, highlighting the limitations of static and reactive paradigms and the promise of AI-driven predictive orchestration. We then describe our dual-pronged methodological framework, which blends analytical latency modeling with large-scale, event-driven simulation to evaluate slice instantiation times, queuing delays, and end-to-end packet latency across three distinct slicing schemes: static, reactive, and predictive. A detailed statistical analysis—comprising thousands of inference request samples under Poisson and bursty arrival processes—reveals that predictive slicing reduces mean end-to-end latency by approximately 35% relative to static slicing and by 11% relative to reactive slicing, while also tightening latency variance and 95th percentile tail behavior. Importantly, we quantify orchestration overheads and resource utilization efficiencies, demonstrating that the modest control-plane cost of predictive models is offset by substantial improvements in latency and resource elasticity.

Downloads

Download data is not yet available.

Downloads

Additional Files

Published

2025-01-03

Issue

Section

Original Research Articles

How to Cite

6G Network Slicing for Low-Latency AI-Edge Deployments. (2025). World Journal of Future Technologies in Computer Science and Engineering (WJFTCSE), 1(1), Jan (40-48_. https://doi.org/10.63345/081cja89

Similar Articles

51-60 of 66

You may also start an advanced similarity search for this article.