Results 1  10
of
3,002,037
Graphs over Time: Densification Laws, Shrinking Diameters and Possible Explanations
, 2005
"... How do real graphs evolve over time? What are “normal” growth patterns in social, technological, and information networks? Many studies have discovered patterns in static graphs, identifying properties in a single snapshot of a large network, or in a very small number of snapshots; these include hea ..."
Abstract

Cited by 534 (48 self)
 Add to MetaCart
, and we observe some surprising phenomena. First, most of these graphs densify over time, with the number of edges growing superlinearly in the number of nodes. Second, the average distance between nodes often shrinks over time, in contrast to the conventional wisdom that such distance parameters should
Preference Parameters And Behavioral Heterogeneity: An Experimental Approach In The Health And Retirement Study
, 1997
"... This paper reports measures of preference parameters relating to risk tolerance, time preference, and intertemporal substitution. These measures are based on survey responses to hypothetical situations constructed using an economic theorist's concept of the underlying parameters. The individual ..."
Abstract

Cited by 524 (12 self)
 Add to MetaCart
This paper reports measures of preference parameters relating to risk tolerance, time preference, and intertemporal substitution. These measures are based on survey responses to hypothetical situations constructed using an economic theorist's concept of the underlying parameters
Toward the next generation of recommender systems: A survey of the stateoftheart and possible extensions
 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING
, 2005
"... This paper presents an overview of the field of recommender systems and describes the current generation of recommendation methods that are usually classified into the following three main categories: contentbased, collaborative, and hybrid recommendation approaches. This paper also describes vario ..."
Abstract

Cited by 1420 (21 self)
 Add to MetaCart
various limitations of current recommendation methods and discusses possible extensions that can improve recommendation capabilities and make recommender systems applicable to an even broader range of applications. These extensions include, among others, an improvement of understanding of users and items
The Dantzig Selector: Statistical Estimation When p Is Much Larger Than n
, 2007
"... In many important statistical applications, the number of variables or parameters p is much larger than the number of observations n. Suppose then that we have observations y = Xβ + z, where β ∈ Rp is a parameter vector of interest, X is a data matrix with possibly far fewer rows than columns, n ≪ p ..."
Abstract

Cited by 877 (14 self)
 Add to MetaCart
In many important statistical applications, the number of variables or parameters p is much larger than the number of observations n. Suppose then that we have observations y = Xβ + z, where β ∈ Rp is a parameter vector of interest, X is a data matrix with possibly far fewer rows than columns, n
A View Of The Em Algorithm That Justifies Incremental, Sparse, And Other Variants
 Learning in Graphical Models
, 1998
"... . The EM algorithm performs maximum likelihood estimation for data in which some variables are unobserved. We present a function that resembles negative free energy and show that the M step maximizes this function with respect to the model parameters and the E step maximizes it with respect to the d ..."
Abstract

Cited by 988 (18 self)
 Add to MetaCart
estimation problem. A variant of the algorithm that exploits sparse conditional distributions is also described, and a wide range of other variant algorithms are also seen to be possible. 1. Introduction The ExpectationMaximization (EM) algorithm finds maximum likelihood parameter estimates in problems
The Infinite Hidden Markov Model
 Machine Learning
, 2002
"... We show that it is possible to extend hidden Markov models to have a countably infinite number of hidden states. By using the theory of Dirichlet processes we can implicitly integrate out the infinitely many transition parameters, leaving only three hyperparameters which can be learned from data. Th ..."
Abstract

Cited by 629 (41 self)
 Add to MetaCart
We show that it is possible to extend hidden Markov models to have a countably infinite number of hidden states. By using the theory of Dirichlet processes we can implicitly integrate out the infinitely many transition parameters, leaving only three hyperparameters which can be learned from data
New empirical relationships among magnitude, rupture length, rupture width, rupture area, and surface
, 1994
"... Abstract Source parameters for historical earthquakes worldwide are compiled to develop a series of empirical relationships among moment magnitude (M), surface rupture length, subsurface rupture length, downdip rupture width, rupture area, and maximum and average displacement per event. The resultin ..."
Abstract

Cited by 524 (0 self)
 Add to MetaCart
Abstract Source parameters for historical earthquakes worldwide are compiled to develop a series of empirical relationships among moment magnitude (M), surface rupture length, subsurface rupture length, downdip rupture width, rupture area, and maximum and average displacement per event
Results 1  10
of
3,002,037