• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

A practical analysis of low-density parity-check erasure codes for wide-area storage applications (2004)

by J Plank, M Thomason
Venue:In Proc. of the IEEE/IFIP DSN’04
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 47
Next 10 →

Ubiquitous access to distributed data in large-scale sensor networks through decentralized erasure codes

by Alexandros G. Dimakis, Vinod Prabhakaran, Kannan Ramchandran , 2005
"... Consider a large-scale wireless sensor network of n nodes, where a fraction k out of n generate data packets of global interest. Assuming that the individual nodes have limited storage and computational capabilities, we address the problem of how to enable ubiquitous access to the distributed data p ..."
Abstract - Cited by 90 (7 self) - Add to MetaCart
Consider a large-scale wireless sensor network of n nodes, where a fraction k out of n generate data packets of global interest. Assuming that the individual nodes have limited storage and computational capabilities, we address the problem of how to enable ubiquitous access to the distributed data packets. Specifically, we assume that each node can store at most one data packet, and study the problem of diffusing the data so that by querying any k nodes, it is possible to retrieve all the k data packets of interest (with high probability). We introduce a class of erasure codes and show how to solve this problem efficiently in a completely distributed and robust way. Specifically we show that we can efficiently diffuse the data by “prerouting” only O(ln n) packets per data node to randomly selected storage nodes. By using the proposed scheme, the distributed data becomes available “at the fingertips” of a potential data collector located anywhere in the network.
(Show Context)

Citation Context

...tems [18], and redundant disk arrays [3]. Also, LDPC codes and more recently fountain codes [19] were proposed as alternatives with randomized construction and faster encoding and decoding times. See =-=[22]-=- for a practical investigation on using these codes for distributed storage. The key feature of our problem that makes these codes unsuitable is the fact that the data is distributed. We could convert...

Optimizing Cauchy Reed-Solomon codes for fault-tolerant network storage applications

by James S. Plank, Lihao Xu - In NCA-06: 5th IEEE International Symposium on Network Computing Applications , 2006
"... NOTE: NCA’s page limit is rather severe: 8 pages. As a result, the final paper is pretty much a hatchet job of the original submission. I would recommend reading the technical report version of this paper, because it presents the material with some accompanying tutorial material, and is easier to re ..."
Abstract - Cited by 49 (12 self) - Add to MetaCart
NOTE: NCA’s page limit is rather severe: 8 pages. As a result, the final paper is pretty much a hatchet job of the original submission. I would recommend reading the technical report version of this paper, because it presents the material with some accompanying tutorial material, and is easier to read. The technical report is available at:
(Show Context)

Citation Context

...two ways – either by the number of (redundant) coding nodes required to achieve a baseline of faulttolerance [16, 15], or by the average number of failures tolerated by a given number of coding nodes =-=[21, 30, 27]-=-. Regardless of the evaluation methodology, space optimality may be achieved when the number of coding nodes is equal to the number of failures that may be tolerated. These codes are called Maximum Di...

Data Persistence in Large-scale Sensor Networks with Decentralized Fountain Codes

by Yunfeng Lin, et al.
"... It may not be feasible for sensor networks monitoring nature and inaccessible geographical regions to include powered sinks with Internet connections. We consider the scenario where sinks are not present in large-scale sensor networks, and unreliable sensors have to collectively resort to storing s ..."
Abstract - Cited by 41 (2 self) - Add to MetaCart
It may not be feasible for sensor networks monitoring nature and inaccessible geographical regions to include powered sinks with Internet connections. We consider the scenario where sinks are not present in large-scale sensor networks, and unreliable sensors have to collectively resort to storing sensed data over time on themselves. At a time of convenience, such cached data from a small subset of live sensors may be collected by a centralized (possibly mobile) collector. In this paper, we propose a decentralized algorithm using fountain codes to guarantee the persistence and reliability of cached data on unreliable sensors. With fountain codes, the collector is able to recover all data as long as a sufficient number of sensors are alive. We use random walks to disseminate data from a sensor to a random subset of sensors in the network. Our algorithms take advantage of the low decoding complexity of fountain codes, as well as the scalability of the dissemination process via random walks. We have proposed two algorithms based on random walks. Our theoretical analysis and simulation-based studies have shown that, the first algorithm maintains the same level of fault tolerance as the original centralized fountain code, while introducing lower overhead than naive random-walk based implementation in the dissemination process. Our second algorithm has lower level of fault tolerance than the original centralized fountain code, but consumes much lower dissemination cost.

A Hybrid Routing Approach for Opportunistic Networks

by Ling-jyh Chen, Chen-hung Yu, Tony Sun, Yung-chih Chen, Hao-hua Chu - In Proc. of ACM SIGCOMM Workshop on Challenged Networks , 2006
"... With wireless networking technologies extending into the fabrics of our working and operating environments, proper handling of intermittent wireless connectivity and network disruptions is of significance. As the sheer number of potential opportunistic application continues to surge (i.e. wireless s ..."
Abstract - Cited by 32 (7 self) - Add to MetaCart
With wireless networking technologies extending into the fabrics of our working and operating environments, proper handling of intermittent wireless connectivity and network disruptions is of significance. As the sheer number of potential opportunistic application continues to surge (i.e. wireless sensor networks, underwater sensor networks, pocket switched networks, transportation networks, and etc.), the design for an effective routing scheme that considers and accommodates the various intricate behaviors observed in an opportunistic network is of interest and remained desirable. While previous solutions use either replication or coding techniques to address the challenges in opportunistic networks, the tradeoff of these two techniques only make them ideal under certain network scenarios. In this paper, we propose a hybrid scheme, named H-EC, to deal with a wide variety of opportunistic network cases. H-EC is designed to fully combine the robustness of erasure coding based routing techniques, while preserving the performance advantages of replication techniques. We evaluate H-EC against other similar strategies in terms of delivery ratio and latency, and find that H-EC offers robustness in worst-case delay performance cases while achieving good performance in small delay performance cases. We also discuss the traffic overhead issues associated with H-EC as compared to other schemes, and present several strategies that can potentially alleviate the traffic overhead of H-EC schemes.
(Show Context)

Citation Context

...original data [27]. Two most popular examples of erasure coding algorithms are ReedSolomon coding and Low-Density Parity-Check (LDPC) based coding (e.g., Gallager codes, Tornado codes, and IRA codes) =-=[20, 21]-=-. These algorithms differ in the encoding/decoding efficiency, replication factor, and the minimum number of code blocks to reconstruct a message. The selection of the proper erasure coding algorithm ...

Determining fault tolerance of XOR-based erasure codes efficiently

by Jay J. Wylie, Ram Swaminathan - In Proceedings of the 2007 International Conference on Dependable Systems and Networks (DSN , 2007
"... We propose a new fault tolerance metric for XOR-based erasure codes: the minimal erasures list (MEL). A minimal erasure is a set of erasures that leads to irrecoverable data loss and in which every erasure is necessary and sufficient for this to be so. The MEL is the enumeration of all minimal erasu ..."
Abstract - Cited by 28 (2 self) - Add to MetaCart
We propose a new fault tolerance metric for XOR-based erasure codes: the minimal erasures list (MEL). A minimal erasure is a set of erasures that leads to irrecoverable data loss and in which every erasure is necessary and sufficient for this to be so. The MEL is the enumeration of all minimal erasures. An XOR-based erasure code has an irregular structure that may permit it to tolerate faults at and beyond its Hamming distance. The MEL completely describes the fault tolerance of an XOR-based erasure code at and beyond its Hamming distance; it is therefore a useful metric for comparing the fault tolerance of such codes. We also propose an algorithm that efficiently determines the MEL of an erasure code. This algorithm uses the structure of the erasure code to efficiently determine the MEL. We show that, in practice, the number of minimal erasures for a given code is much less than the total number of sets of erasures that lead to data loss: in our empirical results for one corpus of codes, there were over 80 times fewer minimal erasures. We use the proposed algorithm to identify the most fault tolerant XOR-based erasure code for all possible systematic erasure codes with up to seven data symbols and up to seven parity symbols. 1.
(Show Context)

Citation Context

...iciency and fault tolerance. In practice, the exact degree of fault tolerance such codes provide in storage systems is not yet well understood, although there is much recent activity towards this end =-=[14, 13, 6, 4, 5, 7]-=-. To completely understand the fault tolerance of an XORbased erasure code, we must enumerate every set of erasures that leads to data loss. This is necessary because of the irregular structure of suc...

Flat XOR-based erasure codes in storage systems: Constructions, efficient recovery, and tradeoffs

by Kevin M. Greenan, Xiaozhou Li, Jay J. Wylie
"... Abstract—Large scale storage systems require multi-disk fault tolerant erasure codes. Replication and RAID extensions that protect against two- and three-disk failures offer a stark tradeoff between how much data must be stored, and how much data must be read to recover a failed disk. Flat XOR-codes ..."
Abstract - Cited by 22 (2 self) - Add to MetaCart
Abstract—Large scale storage systems require multi-disk fault tolerant erasure codes. Replication and RAID extensions that protect against two- and three-disk failures offer a stark tradeoff between how much data must be stored, and how much data must be read to recover a failed disk. Flat XOR-codes—erasure codes in which parity disks are calculated as the XOR of some subset of data disks—offer a tradeoff between these extremes. In this paper, we describe constructions of two novel flat XOR-code, Stepped Combination and HD-Combination codes. We describe an algorithm for flat XOR-codes that enumerates recovery equations, i.e., sets of disks that can recover a failed disk. We also describe two algorithms for flat XOR-codes that generate recovery schedules, i.e., sets of recovery equations that can be used in concert to achieve efficient recovery. Finally, we analyze the key storage properties of many flat XOR-codes and of MDS codes such as replication and RAID 6 to show the cost-benefit tradeoff gap that flat XOR-codes can fill. I.
(Show Context)

Citation Context

...c constructions are known (e.g., [19], [20], [21], [22], [23]). The performance, fault tolerance, and reliability of small LDPC codes is also not well understood, though there is much progress (e.g., =-=[24]-=-, [25], [26], [27], [28], [29], [11]). In this paper we provide many constructions of flat XOR-codes, whichinsomesensearesmallLDPC codes, and an analysis to put the storage properties of such construc...

Reliability for networked storage nodes

by Kk Rao, James Lee Hafner, Richard A. Golding - Research Report RJ-10358, IBM Almaden Research , 2006
"... High-end enterprise storage has traditionally consisted of monolithic systems with customized hardware, multiple redundant components and paths, and no single point of failure. Distributed storage systems realized through networked storage nodes offer several advantages over monolithic systems such ..."
Abstract - Cited by 19 (2 self) - Add to MetaCart
High-end enterprise storage has traditionally consisted of monolithic systems with customized hardware, multiple redundant components and paths, and no single point of failure. Distributed storage systems realized through networked storage nodes offer several advantages over monolithic systems such as lower cost and increased scalability. In order to achieve reliability goals associated with enterprise-class storage systems, redundancy will have to be distributed across the collection of nodes to tolerate both node and drive failures. In this paper, we present alternatives for distributing this redundancy, and models to determine the reliability of such systems. We specify a reliability target and determine the configurations that meet this target. Further, we perform sensitivity analyses where selected parameters are varied to observe their effect on reliability. 1.
(Show Context)

Citation Context

...ata systems respectively. The availability analysis presented in [8] is intended only to gain insights into the factors affecting availability rather than to derive accurate predictions. Plank et al. =-=[10]-=- describe LDPC erasure codes in the context of peer-to-peer and widely distributed storage systems, and Goodson et al. [5] describe erasurecoded storage that tolerates Byzantine failures. Both these p...

Small parity-check erasure codes - exploration and observations

by James S. Plank, Adam L. Buchsbaum, Rebecca L. Collins, Michael G. Thomason - In DSN-05: International Conference on Dependable Systems and Networks , 2005
"... Erasure codes have profound uses in wide- and mediumarea storage applications. While infinite-size codes have been developed with optimal properties, there remains a need to develop small codes with optimal properties. In this paper, we provide a framework for exploring very small codes, and we use ..."
Abstract - Cited by 18 (1 self) - Add to MetaCart
Erasure codes have profound uses in wide- and mediumarea storage applications. While infinite-size codes have been developed with optimal properties, there remains a need to develop small codes with optimal properties. In this paper, we provide a framework for exploring very small codes, and we use this framework to derive optimal and near-optimal ones for discrete numbers of data bits and coding bits. These codes have heretofore been unknown and unpublished, and should be useful in practice. We also use our exploration to make observations about upper bounds for these codes, in order to gain a better understanding of them and to spur future derivations of larger, optimal and near-optimal codes. 1
(Show Context)

Citation Context

... optimal value of one. For small values of n and m (less than 1000), there is little theory, and recent work has shown that the techniques developed for asymptotics do not fare well for small n and m =-=[PT04]-=-. The purpose of this paper is to start closing this hole in the theory. Rather than concentrate on large values of n and m, we concentrate on very small values, using enumeration and heuristics to de...

Partial network coding: Theory and application for continuous sensor data collection

by Dan Wang, Qian Zhang, Jiangchuan Liu - in 14 th IEEE International Workshop on Quality of Service , 2006
"... Abstract — Wireless sensor networks have been widely used for surveillance in harsh environments. In many such applications, the environmental data are continuously sensed, and data collection by a server is only performed occasionally. Hence, the sensor nodes have to temporarily store the data, and ..."
Abstract - Cited by 18 (3 self) - Add to MetaCart
Abstract — Wireless sensor networks have been widely used for surveillance in harsh environments. In many such applications, the environmental data are continuously sensed, and data collection by a server is only performed occasionally. Hence, the sensor nodes have to temporarily store the data, and provide easy and on-hand access for the most updated data when the server approaches. Given the expensive server-to-sensor communications, the large amount of sensors and the limited storage space at each tiny sensor, continuous data collection becomes a challenging problem. In this paper, we present partial network coding (PNC) as a generic tool for the above applications. PNC generalizes the existing network coding (NC) paradigm, an elegant solution for ubiquitous data distribution and collection. Yet, PNC enables efficient storage replacement for continuous data, which is a major deficiency of the conventional NC. We prove that the performance of PNC is quite close to NC, except for a sublinear overhead on storage and communications. We then address a set of practical concerns toward PNC-based continuous data collection in sensor networks. Its feasibility and superiority are further demonstrated through simulation results. I.
(Show Context)

Citation Context

...egments and builds C coded segments, C ≥ N. If any N out of C coded segments are collected, the original data segments can be decoded [8][18]. A practical investigation of these codes can be found in =-=[22]-=-. As mentioned before, these centralized operations are not suitable for our application environment that involves a large quantity of tiny sensors. An alternative is linear network coding [1][28], wh...

Downloading Replicated, Wide-Area Files - A Framework and Empirical Evaluation

by Rebecca L. Collins, James S. Plank - Proceedings of the 3rd IEEE International Symposium on Network Computing and Applications (NCA 2004 , 2004
"... ..."
Abstract - Cited by 13 (2 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

... timeouts as a complementary failover mechanism in the future. Finally, erasure codes have arisen as a viable alternative to replication for both caching and fault-tolerance in wide-area file systems =-=[5, 14, 19, 21]-=-. In future work, we intend to see how these downloading algorithms apply to file systems based on erasure codes, what additional considerations apply, and what the performance impact is. 7 Acknowledg...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University