• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 11 - 20 of 1,553
Next 10 →

Evaluation of Fault-Tolerance Latency from Real-Time Application's Perspectives

by Hagbae Kim, Kang G. Shin - IEEE Trans. on Computers , 1994
"... The Fault-Tolerance Latency (FTL) defined as the time required by all sequential steps taken to recover from an error is important to the design and evaluation of fault-tolerant computers used in safety-critical real-time control systems. To meet timing constraints or avoid dynamic failure, the late ..."
Abstract - Cited by 2 (1 self) - Add to MetaCart
The Fault-Tolerance Latency (FTL) defined as the time required by all sequential steps taken to recover from an error is important to the design and evaluation of fault-tolerant computers used in safety-critical real-time control systems. To meet timing constraints or avoid dynamic failure

Tolerating Latency in Software Distributed Shared Memory Systems Through Non-Binding Prefetching

by Charles Quoc Cuong Chan , 1998
"... A key obstacle to achieving high performance on software distributed shared memory (DSM) systems is their high memory latencies. Software-controlled prefetching tolerates memory latency by overlapping computation with communication. This thesis proposes and evaluates an implementation of software-co ..."
Abstract - Cited by 1 (1 self) - Add to MetaCart
A key obstacle to achieving high performance on software distributed shared memory (DSM) systems is their high memory latencies. Software-controlled prefetching tolerates memory latency by overlapping computation with communication. This thesis proposes and evaluates an implementation of software

Comparative Evaluation of Latency Reducing and Tolerating Techniques

by Anoop Gupta, John Hennessy, Kourosh Gharachorloo, Todd Mowry, Wolf-dietrich Weber - In Proceedings of the 18th Annual International Symposium on Computer Architecture , 1991
"... Techniques that can cope with the large latency of memory accesses are essential for achieving high processor utilization in large-scale shared-memory multiprocessors. In this paper, we consider four architectural techniques that address the latency problem: (i) hardware coherent caches, (ii) relaxe ..."
Abstract - Cited by 112 (6 self) - Add to MetaCart
Techniques that can cope with the large latency of memory accesses are essential for achieving high processor utilization in large-scale shared-memory multiprocessors. In this paper, we consider four architectural techniques that address the latency problem: (i) hardware coherent caches, (ii

Tolerating branch predictor latency on SMT

by Ayose Falcón, Oliverio J. Santana, Alex Ramírez, Mateo Valero - In Proceedings of the 5th International Symposium on High Performance Computing (ISHPC-V , 2003
"... Abstract. Simultaneous Multithreading (SMT) tolerates latency by executing instructions from multiple threads. If a thread is stalled, resources can be used by other threads. However, fetch stall conditions caused by multi-cycle branch predictors prevent SMT to achieve all its potential performance, ..."
Abstract - Cited by 2 (2 self) - Add to MetaCart
Abstract. Simultaneous Multithreading (SMT) tolerates latency by executing instructions from multiple threads. If a thread is stalled, resources can be used by other threads. However, fetch stall conditions caused by multi-cycle branch predictors prevent SMT to achieve all its potential performance

Tolerating Memory Latency through Software-Controlled Pre-Execution in Simultaneous Multithreading Processors

by Chi-Keung Luk - In Proceedings of the 28th Annual International Symposium on Computer Architecture , 2001
"... Hardly predictable data addresses in many irregular applications have rendered prefetching ineffective. In many cases, the only accurate way to predict these addresses is to directly execute the code that generates them. As multithreaded architectures become increasingly popular, one attractive appr ..."
Abstract - Cited by 174 (0 self) - Add to MetaCart
Hardly predictable data addresses in many irregular applications have rendered prefetching ineffective. In many cases, the only accurate way to predict these addresses is to directly execute the code that generates them. As multithreaded architectures become increasingly popular, one attractive approach is to use idle threads on these machines to perform pre-execution---essentially a combined act of speculative address generation and prefetching--- to accelerate the main thread. In this paper, we propose such a pre-execution technique for simultaneous multithreading (SMT) processors. By using software to control pre-execution, we are able to handle some of the most important access patterns that are typically difficult to prefetch. Compared with existing work on pre-execution, our technique is significantly simpler to implement (e.g., no integration of pre-execution results, no need of shortening programs for pre-execution, and no need of special hardware to copy register values upon thread spawns). Consequently, only minimal extensions to SMT machines are required to support our technique. Despite its simplicity, our technique offers an average speedup of 24% in a set of irregular applications, which is a 19% speedup over state-of-the-art software-controlled prefetching.

APRIL: A Processor Architecture for Multiprocessing

by Anant Agarwal, Beng-Hong Lim, David Kranz, John Kubiatowicz - IN PROCEEDINGS OF THE 17TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE , 1990
"... Processors in large-scale multiprocessors must be able to tolerate large communication latencies and synchronization delays. This paper describes the architecture of a rapid-context-switching processor called APRIL with support for fine-grain threads and synchronization. APRIL achieves high single-t ..."
Abstract - Cited by 283 (25 self) - Add to MetaCart
Processors in large-scale multiprocessors must be able to tolerate large communication latencies and synchronization delays. This paper describes the architecture of a rapid-context-switching processor called APRIL with support for fine-grain threads and synchronization. APRIL achieves high single

Load Latency Tolerance In Dynamically Scheduled Processors

by Srikanth T. Srinivasan, Alvin R. Lebeck - JOURNAL OF INSTRUCTION LEVEL PARALLELISM , 1998
"... This paper provides a quantitative evaluation of load latency tolerance in a dynamically scheduled processor. To determine the latency tolerance of each memory load operation, our simulations use flexible load completion policies instead of a fixed memory hierarchy that dictates the latency. Alth ..."
Abstract - Cited by 76 (2 self) - Add to MetaCart
This paper provides a quantitative evaluation of load latency tolerance in a dynamically scheduled processor. To determine the latency tolerance of each memory load operation, our simulations use flexible load completion policies instead of a fixed memory hierarchy that dictates the latency

Latency Tolerance For Dynamic Processors

by James E. Bennett, Michael J. Flynn , 1996
"... While a number of dynamically scheduled processors have recently been brought to market, work on hardware techniques for memory latency tolerance has mostly targeted statically scheduled processors. This paper attempts to remedy this situation by examining the applicability of hardware latency toler ..."
Abstract - Cited by 3 (1 self) - Add to MetaCart
While a number of dynamically scheduled processors have recently been brought to market, work on hardware techniques for memory latency tolerance has mostly targeted statically scheduled processors. This paper attempts to remedy this situation by examining the applicability of hardware latency

Vertical Handoffs in Wireless Overlay Networks

by Mark Stemm , 1996
"... We present extensions to a traditional cellular [Ses95] handoff system to handle the simultaneous operation of multiple wireless network interfaces. This new system allows mobile users to roam in a "Wireless Overlay Network" structure consisting of room-size, building-size, and wide-area d ..."
Abstract - Cited by 252 (4 self) - Add to MetaCart
-area data networks. In this structure, the user can connect to the wired network through multiple wireless subnets, and offers the best possible connectivity given the user's geographic location and local wireless connectivity. We present the basic handoff system and show that the handoff latency

Memory Latency: to Tolerate or to Reduce

by Amol Bakshi, Jean-luc Gaudiot, Wen-yen Lin, Manil Makhija, Viktor K. Prasanna, Wonwoo Ro, Chulho Shin - In 12th Symposium on Computer Architecture and High Performance Computing. Invited paper , 2000
"... It has become a truism that the gap between processor speed and memory access latency is continuing to increase at a rapid rate. This paper presents some of the architecture strategies which are used to bridge this gap. They are mostly of two kinds: memory latency reducing approaches such as employe ..."
Abstract - Cited by 2 (0 self) - Add to MetaCart
such as employed in caches and HiDISC (Hierarchical Decoupled Architecture) or memory latency tolerating schemes such as SMT (Simultaneous Multithreading) or ISSC (I-structure software cache). Yet a third technique reduces the latency by integrating on the same chip processor and DRAM. Finally, algorithmic
Next 10 →
Results 11 - 20 of 1,553
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University