Abstract
Two remote senders observe X and Y, respectively, and can noiselessly send information via a common relay node to a receiver that observes Z. The receiver wants to compute a function f (X, Y, Z) of these possibly related observations, without error. We study the average number of bits that need to be conveyed to that end by each sender to the relay and by the relay to the receiver, in the limit of multiple instances. We relate these quantities to the entropy region of a probabilistic graph with respect to a Cartesian representation of its vertex set, which we define as a natural extension of graph entropy. General properties and bounds for the graph entropy region are derived, and mapped back to special cases of the distributed computing setup.
Original language | English |
---|---|
Article number | 6729087 |
Pages (from-to) | 3435-3449 |
Number of pages | 15 |
Journal | IEEE Transactions on Information Theory |
Volume | 60 |
Issue number | 6 |
DOIs | |
State | Published - Jun 2014 |
Keywords
- Distributed source coding
- graph entropy
- zero-error information theory
All Science Journal Classification (ASJC) codes
- Information Systems
- Computer Science Applications
- Library and Information Sciences