Results 1 - 10
of
174
LT Codes
, 2002
"... We introduce LT codes, the first rateless erasure codes that are very efficient as the data length grows. ..."
Abstract
-
Cited by 568 (2 self)
- Add to MetaCart
(Show Context)
We introduce LT codes, the first rateless erasure codes that are very efficient as the data length grows.
Informed Content Delivery Across Adaptive Overlay Networks
, 2002
"... Overlay networks have emerged as a powerful and highly flexible method for delivering content. We study how to optimize through-put of large, multipoint transfers across richly connected overlay networks, focusing on the question of what to put in each transmit-ted packet. We first make the case for ..."
Abstract
-
Cited by 247 (8 self)
- Add to MetaCart
Overlay networks have emerged as a powerful and highly flexible method for delivering content. We study how to optimize through-put of large, multipoint transfers across richly connected overlay networks, focusing on the question of what to put in each transmit-ted packet. We first make the case for transmitting encoded content in this scenario, arguing for the digital fountain approach which en-ables end-hosts to efficiently restitute the original content of size n from a subset of any n symbols from a large universe of encoded symbols. Such an approach affords reliability and a substantial de-gree of application-level flexibility, as it seamlessly tolerates packet loss, connection migration, and parallel transfers. However, since the sets of symbols acquired by peers are likely to overlap substan-tially, care must be taken to enable them to collaborate effectively. We provide a collection of useful algorithmic tools for efficient es-timation, summarization, and approximate reconciliation of sets of symbols between pairs of collaborating peers, all of which keep messaging complexity and computation to a minimum. Through simulations and experiments on a prototype implementation, we demonstrate the performance benefits of our informed content de-livery mechanisms and how they complement existing overlay net-work architectures.
Using redundancy to cope with failures in a delay tolerant network
- in Proceedings of ACM SIGCOMM
, 2005
"... We consider the problem of routing in a delay tolerant net-work (DTN) in the presence of path failures. Previous work on DTN routing has focused on using precisely known network dy-namics, which does not account for message losses due to link failures, buffer overruns, path selection errors, unsched ..."
Abstract
-
Cited by 146 (4 self)
- Add to MetaCart
(Show Context)
We consider the problem of routing in a delay tolerant net-work (DTN) in the presence of path failures. Previous work on DTN routing has focused on using precisely known network dy-namics, which does not account for message losses due to link failures, buffer overruns, path selection errors, unscheduled de-lays, or other problems. We show how to split, replicate, and erasure code message fragments over multiple delivery paths to optimize the probability of successful message delivery. We provide a formulation of this problem and solve it for two cases: a 0/1 (Bernoulli) path delivery model where messages are ei-ther fully lost or delivered, and a Gaussian path delivery model where only a fraction of a message may be delivered. Ideas from the modern portfolio theory literature are borrowed to solve the underlying optimization problem. Our approach is directly relevant to solving similar problems that arise in replica place-ment in distributed file systems and virtual node placement in DHTs. In three different simulated DTN scenarios covering a wide range of applications, we show the effectiveness of our ap-proach in handling failures.
Resilient Multicast using Overlays
- In Proc. of ACM Sigmetrics
, 2003
"... (PRM): a multicast data recovery scheme that improves data delivery ratios while maintaining low end-to-end latencies. PRM has both a proactive and a reactive components; in this paper we describe how PRM can be used to improve the performance of application-layer multicast protocols especially when ..."
Abstract
-
Cited by 126 (10 self)
- Add to MetaCart
(Show Context)
(PRM): a multicast data recovery scheme that improves data delivery ratios while maintaining low end-to-end latencies. PRM has both a proactive and a reactive components; in this paper we describe how PRM can be used to improve the performance of application-layer multicast protocols especially when there are high packet losses and host failures. Through detailed analysis in this paper, we show that this loss recovery technique has efficient scaling properties—the overheads at each overlay node asymptotically decrease to zero with increasing group sizes. As a detailed case study, we show how PRM can be applied to the NICE application-layer multicast protocol. We present detailed simulations of the PRM-enhanced NICE protocol for 10 000 node Internet-like topologies. Simulations show that PRM achieves a high delivery ratio ( 97%) with a low latency bound (600 ms) for environments with high end-to-end network losses (1%–5%) and high topology change rates (5 changes per second) while incurring very low overheads ( 5%). Index Terms—Multicast, networks, overlays, probabilistic forwarding, protocols, resilience. I.
Bayesian compressive sensing via belief propagation
- IEEE Trans. Signal Processing
, 2010
"... Compressive sensing (CS) is an emerging field based on the revelation that a small collection of linear projections of a sparse signal contains enough information for stable, sub-Nyquist signal acquisition. When a statistical characterization of the signal is available, Bayesian inference can comple ..."
Abstract
-
Cited by 125 (19 self)
- Add to MetaCart
Compressive sensing (CS) is an emerging field based on the revelation that a small collection of linear projections of a sparse signal contains enough information for stable, sub-Nyquist signal acquisition. When a statistical characterization of the signal is available, Bayesian inference can complement conventional CS methods based on linear programming or greedy algorithms. We perform approximate Bayesian inference using belief propagation (BP) decoding, which represents the CS encoding matrix as a graphical model. Fast encoding and decoding is provided using sparse encoding matrices, which also improve BP convergence by reducing the presence of loops in the graph. To decode a length-N signal containing K large coefficients, our CS-BP decoding algorithm uses O(K log(N)) measurements and O(N log 2 (N)) computation. Finally, sparse encoding matrices and the CS-BP decoding algorithm can be modified to support a variety of signal models and measurement noise. 1
Recent and Emerging Topics in Wireless Industrial Communications: A Selection
, 2007
"... In this paper we discuss a selection of promising and interesting research areas in the design of protocols and systemsforwirelessindustrialcommunications.Wehaveselected topicsthathaveeitheremergedashottopicsintheindustrial communicationscommunityinthelastfewyears(likewireless sensornetworks),orwhi ..."
Abstract
-
Cited by 96 (1 self)
- Add to MetaCart
(Show Context)
In this paper we discuss a selection of promising and interesting research areas in the design of protocols and systemsforwirelessindustrialcommunications.Wehaveselected topicsthathaveeitheremergedashottopicsintheindustrial communicationscommunityinthelastfewyears(likewireless sensornetworks),orwhichcouldbeworthwhileresearchtopicsin thenextfewyears(forexamplecooperativediversitytechniques for error control, cognitive radio/opportunistic spectrum access for mitigation of external interferences).
Digital Fountains: A Survey and Look Forward
, 2004
"... We survey constructions and applications of digital fountains, an abstraction of erasure coding for network communication. Digital fountains effectively change the standard paradigm where a user receives an ordered stream of packets to one where a user must simply receive enough packets in order to ..."
Abstract
-
Cited by 89 (0 self)
- Add to MetaCart
We survey constructions and applications of digital fountains, an abstraction of erasure coding for network communication. Digital fountains effectively change the standard paradigm where a user receives an ordered stream of packets to one where a user must simply receive enough packets in order to obtain the desired data. Obviating the need for ordered data simplifies data delivery, especially when the data is large or is to be distributed to a large number of users. We also examine barriers to the adoption of digital fountains and discuss whether they can be overcome.
Push-to-peer video-on-demand system: Design and evaluation
- In UMass Computer Science Techincal Report 2006–59
, 2006
"... Number: CR-PRL-2006-11-0001 ..."
(Show Context)
Compressed Sensing Reconstruction via Belief Propagation
, 2006
"... Compressed sensing is an emerging field that enables to reconstruct sparse or compressible signals from a small number of linear projections. We describe a specific measurement scheme using an LDPC-like measurement matrix, which is a real-valued analogue to LDPC techniques over a finite alphabet. We ..."
Abstract
-
Cited by 55 (7 self)
- Add to MetaCart
(Show Context)
Compressed sensing is an emerging field that enables to reconstruct sparse or compressible signals from a small number of linear projections. We describe a specific measurement scheme using an LDPC-like measurement matrix, which is a real-valued analogue to LDPC techniques over a finite alphabet. We then describe the reconstruction details for mixture Gaussian signals. The technique can be extended to additional compressible signal models. 1
Improving Fairness, Efficiency, and Stability in HTTP-based Adaptive Video Streaming with FESTIVE
"... Many commercial video players rely on bitrate adaptation logic to adapt the bitrate in response to changing network conditions. Past measurement studies have identified issues with today’s commercial players with respect to three key metrics—efficiency, fairness, and stability—when multiple bitrate- ..."
Abstract
-
Cited by 39 (5 self)
- Add to MetaCart
(Show Context)
Many commercial video players rely on bitrate adaptation logic to adapt the bitrate in response to changing network conditions. Past measurement studies have identified issues with today’s commercial players with respect to three key metrics—efficiency, fairness, and stability—when multiple bitrate-adaptive players share a bottleneck link. Unfortunately, our current understanding of why these effects occur and how they can be mitigated is quite limited. In this paper, we present a principled understanding of bitrate adaptation and analyze several commercial players through the lens of an abstract player model. Through this framework, we identify the root causes of several undesirable interactions that arise as a consequence of overlaying the video bitrate adaptation over HTTP. Building on these insights, we develop a suite of techniques that can systematically guide the tradeoffs between stability, fairness and efficiency and thus lead to a general framework for robust video adaptation. We pick one concrete instance from this design space and show that it significantly outperforms today’s commercial players on all three key metrics across a range of experimental scenarios.