a linear inference on some local data given a network-side matrix. The computation is performed at the network edge over a number of edge servers. We propose a coding scheme that provides information-theoretic privacy against colluding (honest-but-curious) edge servers, while minimizing the overall latency—comprising upload, computation, download, and decoding latency—in the presence of straggling servers. The proposed …
We consider a mobile edge computing scenario where a number of devices want to perform a linear inference on some local data given a network-side matrix . The computation is performed at the network edge over a number of edge servers. We propose a coding scheme that provides information-theoretic privacy against colluding (honest-but-curious) edge servers, while minimizing the overall latency—comprising upload, computation, download, and decoding latency—in the presence of straggling servers. The proposed scheme exploits Shamir’s secret sharing to yield data privacy and straggler mitigation, combined with replication to provide spatial diversity for the download. We also propose two variants of the scheme that further reduce latency. For a considered scenario with 9 edge servers, the proposed scheme reduces the latency by 8% compared to the nonprivate scheme recently introduced by Zhang and Simeone, while providing privacy against an honest-but-curious edge server.