• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 2,045
Next 10 →

Amortized Efficiency of List Update and Paging Rules

by Daniel D. Sleator, Robert E. Tarjan , 1985
"... In this article we study the amortized efficiency of the “move-to-front” and similar rules for dynamically maintaining a linear list. Under the assumption that accessing the ith element from the front of the list takes 0(i) time, we show that move-to-front is within a constant factor of optimum amo ..."
Abstract - Cited by 824 (8 self) - Add to MetaCart
paging, a setting in which the access cost is not convex. The paging rule corresponding to move-to-front is the “least recently used” (LRU) replacement rule. We analyze the amortized complexity of LRU, showing that its efficiency differs from that of the offline paging rule (Belady’s MIN algorithm) by a

Data Preparation for Mining World Wide Web Browsing Patterns

by Robert Cooley, Bamshad Mobasher, Jaideep Srivastava - KNOWLEDGE AND INFORMATION SYSTEMS , 1999
"... The World Wide Web (WWW) continues to grow at an astounding rate in both the sheer volume of tra#c and the size and complexity of Web sites. The complexity of tasks such as Web site design, Web server design, and of simply navigating through a Web site have increased along with this growth. An i ..."
Abstract - Cited by 567 (43 self) - Add to MetaCart
. An important input to these design tasks is the analysis of how a Web site is being used. Usage analysis includes straightforward statistics, such as page access frequency, as well as more sophisticated forms of analysis, such as finding the common traversal paths through a Web site. Web Usage Mining

Extracting structured data from web pages

by Arvind Arasu - In ACM SIGMOD , 2003
"... Many web sites contain a large collection of “structured” web pages. These pages encode data from an underlying structured source, and are typically generated dynamically. An example of such a collection is the set of book pages in Amazon. There are two important characteristics of such a collection ..."
Abstract - Cited by 310 (0 self) - Add to MetaCart
Many web sites contain a large collection of “structured” web pages. These pages encode data from an underlying structured source, and are typically generated dynamically. An example of such a collection is the set of book pages in Amazon. There are two important characteristics of such a

Open information extraction from the web

by Michele Banko, Michael J Cafarella, Stephen Soderland, Matt Broadhead, Oren Etzioni - IN IJCAI , 2007
"... Traditionally, Information Extraction (IE) has focused on satisfying precise, narrow, pre-specified requests from small homogeneous corpora (e.g., extract the location and time of seminars from a set of announcements). Shifting to a new domain requires the user to name the target relations and to ma ..."
Abstract - Cited by 373 (39 self) - Add to MetaCart
and to manually create new extraction rules or hand-tag new training examples. This manual labor scales linearly with the number of target relations. This paper introduces Open IE (OIE), a new extraction paradigm where the system makes a single data-driven pass over its corpus and extracts a large set

found at the ENTCS Macro Home Page. Rules and Strategies in Java

by Pierre-etienne Moreau, Antoine Reilles
"... this file with prentcsmacro.sty for your meeting, ..."
Abstract - Add to MetaCart
this file with prentcsmacro.sty for your meeting,

2009 10th International Conference on Document Analysis and Recognition Page Rule-Line Removal using Linear Subspaces in Monochromatic Handwritten Arabic Documents ∗

by Wael Abd-almageed, Jayant Kumar, David Doermann
"... In this paper we present a novel method for removing page rule lines in monochromatic handwritten Arabic documents using subspace methods with minimal effect on the quality of the foreground text. We use moment and histogram properties to extract features that represent the characteristics of the un ..."
Abstract - Add to MetaCart
In this paper we present a novel method for removing page rule lines in monochromatic handwritten Arabic documents using subspace methods with minimal effect on the quality of the foreground text. We use moment and histogram properties to extract features that represent the characteristics

Rules and Regulations of The Association

by unknown authors
"... page ..."
Abstract - Add to MetaCart
page

Expertise Networks in Online Communities: Structure and Algorithms

by Jun Zhang, Mark S. Ackerman, Lada Adamic - In Proceedings of WWW2007
"... Web-based communities have become important places for people to seek and share expertise. We find that networks in these communities typically differ in their topology from other online networks such as the World Wide Web. Systems targeted to augment web-based communities by automatically identifyi ..."
Abstract - Cited by 181 (9 self) - Add to MetaCart
identifying users with expertise, for example, need to adapt to the underlying interaction dynamics. In this study, we analyze the Java Forum, a large online help-seeking community, using social network analysis methods. We test a set of network-based ranking algorithms, including PageRank and HITS

Generating Finite-State Transducers For Semi-Structured Data Extraction From The Web

by Chun-nan Hsu, Ming-Tzung Dung , 1998
"... Integrating a large number of Web information sources may significantly increase the utility of the World-Wide Web. A promising solution to the integration is through the use of a Web Information mediator that provides seamless, transparent access for the clients. Information mediators need wrappers ..."
Abstract - Cited by 178 (3 self) - Add to MetaCart
and typos. This paper presents SoftMealy, a novel wrapper representation formalism. This representation is based on a finite-state transducer (FST) and contextual rules. This approach can wrap a wide range of semistructured Web pages because FSTs can encode each different attribute permutation as a path. A

Information Extraction from HTML: Application of a General Machine Learning Approach

by Dayne Freitag - In Proceedings of the Fifteenth National Conference on Artificial Intelligence , 1998
"... Because the World Wide Web consists primarily of text, information extraction is central to any effort that would use the Web as a resource for knowledge discovery. We show how information extraction can be cast as a standard machine learning problem, and argue for the suitability of relational lear ..."
Abstract - Cited by 175 (6 self) - Add to MetaCart
in learning extraction patterns. Instead, structural and other information is supplied as input in the form of an extensible token-oriented feature set. We demonstrate the effectiveness of this approach by adapting SRV for use in learning extraction rules for a domain consisting of university course
Next 10 →
Results 1 - 10 of 2,045
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University