Abstract
In data-intensive applications, it is advantageous to perform partial processing close to the data, and communicate intermediate results to a central processor, instead of the data itself. When the communication or computation medium is noisy, the resulting degradation in computation quality at the central processor must be mitigated. We study this problem for the setup of binary classification performed by an ensemble of base functions communicating real-valued confidence levels. We propose a noise-mitigation solution that optimizes the transmission gains and aggregation coefficients of the base functions. Toward that, we formulate a post-training gradient-based optimization algorithm that minimizes the error probability given the training dataset and the noise parameters. We further derive lower and upper bounds on the optimized error probability, and show empirical results that demonstrate the enhanced performance achieved by our approach on real data.
| Original language | English |
|---|---|
| Pages (from-to) | 1067-1080 |
| Number of pages | 14 |
| Journal | IEEE Journal on Selected Areas in Communications |
| Volume | 41 |
| Issue number | 4 |
| DOIs | |
| State | Published - 1 Apr 2023 |
Keywords
- Classification algorithms
- Gaussian noise
- Hardware
- Noise measurement
- Optimization
- Performance evaluation
- Reliability
- Training
- boosting
- classification algorithms
- distributed computing
- inference algorithms
- machine learning
All Science Journal Classification (ASJC) codes
- Computer Networks and Communications
- Electrical and Electronic Engineering
Fingerprint
Dive into the research topics of 'Ensemble Classification With Noisy Real-Valued Base Functions'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver