A Unified Memory Perspective for Probabilistic Trustworthy AI

2026-03-26Machine Learning

Machine LearningArtificial IntelligenceHardware ArchitectureEmerging Technologies
AI summary

The authors explain that trustworthy AI often relies on probability-based computing, which mixes regular data use with lots of random sampling. They show this combined process can be seen as a special case of random sampling, helping analyze how memory systems handle data and randomness together. This approach highlights that higher random sampling demand makes memory less efficient and can limit system performance. The authors propose new ways to evaluate memory systems and explore modern designs that mix data storage and random sampling for better AI hardware.

probabilistic computationstochastic samplingdeterministic data accessmemory systemsentropy-limited operationcompute-in-memoryhardware efficiencytrustworthy AIdata-access efficiencyrandomness generation
Authors
Xueji Zhao, Likai Pei, Jianbo Liu, Kai Ni, Ningyuan Cao
Abstract
Trustworthy artificial intelligence increasingly relies on probabilistic computation to achieve robustness, interpretability, security and privacy. In practical systems, such workloads interleave deterministic data access with repeated stochastic sampling across models, data paths and system functions, shifting performance bottlenecks from arithmetic units to memory systems that must deliver both data and randomness. Here we present a unified data-access perspective in which deterministic access is treated as a limiting case of stochastic sampling, enabling both modes to be analyzed within a common framework. This view reveals that increasing stochastic demand reduces effective data-access efficiency and can drive systems into entropy-limited operation. Based on this insight, we define memory-level evaluation criteria, including unified operation, distribution programmability, efficiency, robustness to hardware non-idealities and parallel compatibility. Using these criteria, we analyze limitations of conventional architectures and examine emerging probabilistic compute-in-memory approaches that integrate sampling with memory access, outlining pathways toward scalable hardware for trustworthy AI.