Analyzing Symbolic Properties for DRL Agents in Systems and Networking

2026-04-06Networking and Internet Architecture

Networking and Internet ArchitectureArtificial IntelligenceMachine Learning
AI summary

The authors explore how to better check the behavior of deep reinforcement learning (DRL) agents in systems like video streaming and network control. Instead of looking at single example inputs, they focus on 'symbolic properties' that cover a range of inputs and describe expected behaviors more broadly. They introduce a method called diffRL that uses existing tools to analyze these symbolic properties more efficiently. Their experiments show this broader approach finds more meaningful issues and helps understand how DRL models change during training and with different sizes. Overall, the authors demonstrate a practical way to verify DRL agents over wide input ranges rather than just specific points.

Deep Reinforcement LearningSymbolic PropertiesVerificationAdaptive Video StreamingWireless Resource ManagementCongestion ControlMonotonicityRobustnessDNN VerificationdiffRL
Authors
Mohammad Zangooei, Jannis Weil, Amr Rizk, Mina Tahmasbi Arashloo, Raouf Boutaba
Abstract
Deep reinforcement learning (DRL) has shown remarkable performance on complex control problems in systems and networking, including adaptive video streaming, wireless resource management, and congestion control. For safe deployment, however, it is critical to reason about how agents behave across the range of system states they encounter in practice. Existing verification-based methods in this domain primarily focus on point properties, defined around fixed input states, which offer limited coverage and require substantial manual effort to identify relevant input-output pairs for analysis. In this paper, we study symbolic properties, that specify expected behavior over ranges of input states, for DRL agents in systems and networking. We present a generic formulation for symbolic properties, with monotonicity and robustness as concrete examples, and show how they can be analyzed using existing DNN verification engines. Our approach encodes symbolic properties as comparisons between related executions of the same policy and decomposes them into practically tractable sub-properties. These techniques serve as practical enablers for applying existing verification tools to symbolic analysis. Using our framework, diffRL, we conduct an extensive empirical study across three DRL-based control systems, adaptive video streaming, wireless resource management, and congestion control. Through these case studies, we analyze symbolic properties over broad input ranges, examine how property satisfaction evolves during training, study the impact of model size on verifiability, and compare multiple verification backends. Our results show that symbolic properties provide substantially broader coverage than point properties and can uncover non-obvious, operationally meaningful counterexamples, while also revealing practical solver trade-offs and limitations.