DMCA
Flat Datacenter Storage (2012)
Cached
Download Links
Venue: | In Proc. OSDI |
Citations: | 30 - 2 self |
Citations
4465 | Chord: A scalable Peer-To-Peer lookup service for internet applications
- Stoica, Morris, et al.
- 2003
(Show Context)
Citation Context ...a tract and uses deterministic placement to eliminate the need for an additional manager service to respond to requests for the location of individual files within the storage system. DHTs like Chord =-=[31]-=- use techniques such as consistent hashing [20] to eliminating the need for centralized coordination when locating data. However, under churn, a request within a DHT might be routed to several differe... |
1499 | The Google file system
- Ghemawat, Gobioff, et al.
(Show Context)
Citation Context ...ow does a writer know where to send data? How does a reader find data that has been previously written? Many systems solve this problem using a metadata server that stores the location of data blocks =-=[14, 30]-=-. Writers contact the metadata server to find out where to write a new block; the metadata server picks a data server, durably stores that decision and returns it to the writer. Readers contact the me... |
933 | Scale and Performance in a Distributed File System.
- Howard, Kazar, et al.
- 1988
(Show Context)
Citation Context ...to accelerate RAID recovery, while FDS ensures that all disks participate in recovery by striping tracts among all disks in the cluster. Distributed file systems like Frangiapani [32], GPFS [28], AFS =-=[18]-=- and NFS [29] export a remote storage model across a shared, hierarchical name-space. These systems must contend with strong consistency guarantees, and the vagaries of remote, shared access to a POSI... |
761 | Dryad: distributed data- parallel programs from sequential building blocks
- ISARD, BUDIU, et al.
- 2007
(Show Context)
Citation Context ...aramount. Software developers accustomed to treating the network as an abstraction are forced to think in terms of “rack locality.” New programming models (e.g., MapReduce [13], Hadoop [1], and Dryad =-=[19]-=-) emerged to help exploit locality, preferentially moving computation to data rather than viceversa. They effectively expose a cluster’s aggregate disk throughput for tasks with high reduction factors... |
521 | GPFS: A shared-disk file system for large computing clusters.
- Schmuck, Haskin
- 2002
(Show Context)
Citation Context ...y servers to accelerate RAID recovery, while FDS ensures that all disks participate in recovery by striping tracts among all disks in the cluster. Distributed file systems like Frangiapani [32], GPFS =-=[28]-=-, AFS [18] and NFS [29] export a remote storage model across a shared, hierarchical name-space. These systems must contend with strong consistency guarantees, and the vagaries of remote, shared access... |
473 | Serverless network file systems.
- Anderson, Dahlin, et al.
- 1995
(Show Context)
Citation Context ...cts to stripe data across the cluster. Finally, FDS fully distributes the blob metadata into the cluster using metadata tracts, which further minimizes the need to centrally administer the store. xFS =-=[6]-=- also proposed distributing file system metadata among storage nodes through distributed metadata managers. A replicated manager map determined which manager was responsible for the location of a part... |
460 | VL2: A Scalable and Flexible Data Center Network
- Greenberg, Hamilton, et al.
- 2009
(Show Context)
Citation Context ...y constraint, itself rooted in the datacenter bandwidth shortage. When bandwidth was scarce, these sacrifices were necessary to achieve the best performance. However, recently developed CLOS networks =-=[16, 15, 24]-=-—large numbers of small commodity switches with redundant interconnections—have made it economical to build non-oversubscribed full bisection bandwidth networks at the scale of a datacenter for the fi... |
361 | Thekkath. Petal: Distributed Virtual Disks
- Lee, A
- 1996
(Show Context)
Citation Context ...only on fast access to blob storage, FDS provides weak consistency guarantees with very high performance. Among many others, systems such as Swift [11], Zebra [17], GPFS [28], Panasas [35], and Petal =-=[21]-=- stripe files, blocks, or logs across file servers to improve read and write throughput for traditional hierarchical file systems. FDS follows in the footsteps of these systems by using the tract loca... |
343 | The hadoop distributed file system.
- Shvachko, Kuang, et al.
- 2010
(Show Context)
Citation Context ...ow does a writer know where to send data? How does a reader find data that has been previously written? Many systems solve this problem using a metadata server that stores the location of data blocks =-=[14, 30]-=-. Writers contact the metadata server to find out where to write a new block; the metadata server picks a data server, durably stores that decision and returns it to the writer. Readers contact the me... |
320 | Frangipani: A Scalable Distributed File System
- Thekkath, Mann, et al.
- 1997
(Show Context)
Citation Context ... across many servers to accelerate RAID recovery, while FDS ensures that all disks participate in recovery by striping tracts among all disks in the cluster. Distributed file systems like Frangiapani =-=[32]-=-, GPFS [28], AFS [18] and NFS [29] export a remote storage model across a shared, hierarchical name-space. These systems must contend with strong consistency guarantees, and the vagaries of remote, sh... |
302 | The zebra striped network file system.
- Hartman, Ousterhout
- 1995
(Show Context)
Citation Context ...ss to a POSIX compliant API. By focusing only on fast access to blob storage, FDS provides weak consistency guarantees with very high performance. Among many others, systems such as Swift [11], Zebra =-=[17]-=-, GPFS [28], Panasas [35], and Petal [21] stripe files, blocks, or logs across file servers to improve read and write throughput for traditional hierarchical file systems. FDS follows in the footsteps... |
223 | Hedera: Dynamic Flow Scheduling for Data Center Networks,” in NSDI,
- Al-Fares, Radhakrishnan, et al.
- 2010
(Show Context)
Citation Context .... One drawback of ECMP is that full bisection bandwidth is not guaranteed, but only stochastically likely across multiple flows. Long-lived, high-bandwidth flows are known to be problematic with ECMP =-=[3]-=-. FDS, however, was designed to use a large number of short-lived (tract-sized) flows to a large, pseudo-random sequence of destinations. This was done in part to satisfy the stochastic requirement of... |
117 | Scalable performance of the panasas parallel file system.
- Welch, Unangst, et al.
- 2008
(Show Context)
Citation Context ...ure recovery back to stable storage. Further, RAMCloud distributes data to disk purely for reasons of fault tolerance, while FDS replication is used both for fault tolerance and availability. Panasas =-=[35]-=- uses RAID 5 to stripe files, rather than blocks, across many servers to accelerate RAID recovery, while FDS ensures that all disks participate in recovery by striping tracts among all disks in the cl... |
113 | Chain replication for supporting high throughput and availability.
- Renesse, Schneider
- 2004
(Show Context)
Citation Context ...common; for example, clients of the Google File System [14] must handle garbage entries in files. However, if strong consistency guarantees are desired, FDS could be modified to use chain replication =-=[33]-=- to provide strong consistency guarantees for all updates to individual tracts. Tractservers may also be inconsistent during failure recovery. A tractserver recently assigned to a TLT entry will not h... |
110 | Web caching with consistent hashing.
- KARGER, LEIGHTON, et al.
- 1999
(Show Context)
Citation Context ...minate the need for an additional manager service to respond to requests for the location of individual files within the storage system. DHTs like Chord [31] use techniques such as consistent hashing =-=[20]-=- to eliminating the need for centralized coordination when locating data. However, under churn, a request within a DHT might be routed to several different servers before finding an up-to-date locatio... |
94 | Swift: Using distributed disk striping to provide high I/O data rates.
- Cabrera, Long
- 1991
(Show Context)
Citation Context ... shared access to a POSIX compliant API. By focusing only on fast access to blob storage, FDS provides weak consistency guarantees with very high performance. Among many others, systems such as Swift =-=[11]-=-, Zebra [17], GPFS [28], Panasas [35], and Petal [21] stripe files, blocks, or logs across file servers to improve read and write throughput for traditional hierarchical file systems. FDS follows in t... |
87 |
Reining in the outliers in map-reduce clusters using mantri
- Ananthanarayanan, Kandula, et al.
- 2010
(Show Context)
Citation Context ...k to worker can be done at very short timescales. This enables FDS to mitigate stragglers—a significant bottleneck in large systems because a task is not complete until its slowest worker is complete =-=[5]-=-. Hadoop- and MapReduce-style clusters that primarily process data locally are very sensitive to machines that are slow due to factors such as misbehaving hardware, jobs running concurrently, hotspots... |
83 | Fast crash recovery in ramcloud.
- Ongaro, Rumble, et al.
- 2011
(Show Context)
Citation Context ...0 disks recovers 92GB of data lost from a single disk in only 6.2s, and 655GB lost from 7 disks on a failed machine in 33.7s. Though the broad approach of FDS’ failure recovery is similar to RAMCloud =-=[26]-=-, RAMCloud recovers data to DRAM and uses replication only for fault tolerance. FDS uses replication both for availability and fault tolerance while recovering data back to stable storage. Such fast f... |
69 | Portland: A scalable fault-tolerant layer 2 data center network fabric.
- Mysore, Pamboris, et al.
- 2009
(Show Context)
Citation Context ...y constraint, itself rooted in the datacenter bandwidth shortage. When bandwidth was scarce, these sacrifices were necessary to achieve the best performance. However, recently developed CLOS networks =-=[16, 15, 24]-=-—large numbers of small commodity switches with redundant interconnections—have made it economical to build non-oversubscribed full bisection bandwidth networks at the scale of a datacenter for the fi... |
60 | BTowards a next generation data center architecture: Scalability and commoditization,[ in
- Greenberg, Lahiri, et al.
- 2008
(Show Context)
Citation Context ...y constraint, itself rooted in the datacenter bandwidth shortage. When bandwidth was scarce, these sacrifices were necessary to achieve the best performance. However, recently developed CLOS networks =-=[16, 15, 24]-=-—large numbers of small commodity switches with redundant interconnections—have made it economical to build non-oversubscribed full bisection bandwidth networks at the scale of a datacenter for the fi... |
44 | Winning a 60 Second Dash with a Yellow Elephant.
- ’Malley, Murthy
- 2014
(Show Context)
Citation Context ...(OSDI ’12) USENIX AssociationSystem Computers Data Disks Sort Size Time Implied Disk Throughput MinuteSort—Daytona class (general purpose) FDS, 2012 256 1,033 1,401GB 59s 46MB/s Yahoo!, Hadoop, 2009 =-=[25]-=- 1,408 5,632 500GB 59s 3MB/s Yahoo!, Hadoop, 2009 [25] 1,408 5,632 1,000GB 62s 5.7MB/s (unofficial 1TB run) MinuteSort—Indy class (benchmark-specific optimizations allowed) FDS, 2012 256 1,033 1,470GB... |
32 | Tritonsort: A balanced large-scale sorting system.
- RASMUSSEN, PORTER, et al.
- 2011
(Show Context)
Citation Context ...adoop, 2009 [25] 1,408 5,632 1,000GB 62s 5.7MB/s (unofficial 1TB run) MinuteSort—Indy class (benchmark-specific optimizations allowed) FDS, 2012 256 1,033 1,470GB 59.4s 47.9MB/s UCSD TritonSort, 2011 =-=[27]-=- 66 1,056 1,353GB 59.2s 43.3MB/s Table 2: Comparison of FDS MinuteSort results with the previously standing records. In accordance with sort benchmark rules, all reported times are the median of 15 ru... |
30 |
Run-time adaptation in River
- Arpaci-Dusseau
- 2003
(Show Context)
Citation Context ...scale full bisection bandwidth networks. Other full bisection networks exist, such as Infiniband [23], but at a cost and scale that limited them to supercomputing and HPC environments. Finally, River =-=[8]-=- used a distributed queue to dynamically adjust the assignment of work to applications nodes at run time in data flow computations. Similarly, FDS applications use dynamic work allocation to choose wh... |
27 |
GFS: Evolution on fast-forward.
- MCKUSICK, QUINLAN
- 2010
(Show Context)
Citation Context ...data server becomes a centralized scaling and performance bottleneck. In a recent interview, Google architects described the GFS metadata server as a limiting factor in terms of scale and performance =-=[22]-=-. Additionally, a desire to reduce the size of a chunk from 64MB was limited by the proportional increase in the number of chunks in the system. FDS uses deterministic placement to eliminate the scali... |
3 |
Data center: Load balancing data center services solutions reference nework design
- Systems
- 2004
(Show Context)
Citation Context ...ovide one, they come with a significant performance penalty because networks at datacenter scales have historically been oversubscribed. Individual machines were typically attached in a tree topology =-=[12]-=-; for cost efficiency, links near the root had significantly less capacity than the aggregate capacity below them. Core oversubscription ratios of hundreds to one were common, which meant communicatio... |
2 | and etc. Minutesort with flat datacenter storage
- Apacible, Draves
- 2012
(Show Context)
Citation Context ... as assuming 100-byte records and uniformly distributed 10-byte keys. In April 2012, our FDS-based sort application set the world record for sort in both the Indy and Daytona categories of MinuteSort =-=[7]-=-, which measures the amount of data that can be sorted in 60 seconds. Using a cluster of 1,033 disks and 256 computers (136 for tractservers, 120 for the application), our Daytona-class app sorted 10 ... |
2 | Solving TCP incast in cluster storage systems
- Vasudevan, Shah, et al.
- 2009
(Show Context)
Citation Context ...n seen during reads, high packet loss can occur as queues fill during bursts. The reaction of standard TCP to such losses can have a devastating effect on performance. This is sometimes called incast =-=[34]-=-. Schemes such as DCTCP [4] ameliorate incast in concert with routers’ explicit congestion notification (ECN). However, because our network has full bisection bandwidth, collisions mostly occur at the... |
1 |
A measure of transaction processing power
- Bitton, Brown, et al.
- 1985
(Show Context)
Citation Context ...ny big-data applications. Its load pattern is similar to other common tasks such as distributed database joins and large matrix operations. This has made it an important benchmark since at least 1985 =-=[9]-=-. A group informally sponsored by SIGMOD curates an annual disk-to-disk sort performance competition with divisions for speed, cost efficiency, and energy efficiency [2]. Each has sub-divisions for ge... |
1 |
The Amazon Simple Storage Service (Amazon S3
- Borthakur
(Show Context)
Citation Context ...ystem. At the scale of large, big-data clusters that routinely exceed thousands of computers, this “flat” model of storage is still highly desirable. While some blob storage systems such as Amazon S3 =-=[10]-=- provide one, they come with a significant performance penalty because networks at datacenter scales have historically been oversubscribed. Individual machines were typically attached in a tree topolo... |
1 |
Building a Scalable Storage with InfiniBand
- Mellanox
- 2012
(Show Context)
Citation Context ...hout loss of performance. PortLand [24] and VL2 [15] make it economically feasible to build datacenter-scale full bisection bandwidth networks. Other full bisection networks exist, such as Infiniband =-=[23]-=-, but at a cost and scale that limited them to supercomputing and HPC environments. Finally, River [8] used a distributed queue to dynamically adjust the assignment of work to applications nodes at ru... |