Institution
Carnegie Mellon University
Education•Pittsburgh, Pennsylvania, United States•
About: Carnegie Mellon University is a education organization based out in Pittsburgh, Pennsylvania, United States. It is known for research contribution in the topics: Computer science & Robot. The organization has 36317 authors who have published 104359 publications receiving 5975734 citations. The organization is also known as: CMU & Carnegie Mellon.
Papers published on a yearly basis
Papers
More filters
••
11 May 2003TL;DR: The random-pairwise keys scheme is presented, which perfectly preserves the secrecy of the rest of the network when any node is captured, and also enables node-to-node authentication and quorum-based revocation.
Abstract: Key establishment in sensor networks is a challenging problem because asymmetric key cryptosystems are unsuitable for use in resource constrained sensor nodes, and also because the nodes could be physically compromised by an adversary. We present three new mechanisms for key establishment using the framework of pre-distributing a random set of keys to each node. First, in the q-composite keys scheme, we trade off the unlikeliness of a large-scale network attack in order to significantly strengthen random key predistribution's strength against smaller-scale attacks. Second, in the multipath-reinforcement scheme, we show how to strengthen the security between any two nodes by leveraging the security of other links. Finally, we present the random-pairwise keys scheme, which perfectly preserves the secrecy of the rest of the network when any node is captured, and also enables node-to-node authentication and quorum-based revocation.
3,125 citations
••
TL;DR: This paper shows that the accuracy of learned text classifiers can be improved by augmenting a small number of labeled training documents with a large pool of unlabeled documents, and presents two extensions to the algorithm that improve classification accuracy under these conditions.
Abstract: This paper shows that the accuracy of learned text classifiers can be improved by augmenting a small number of labeled training documents with a large pool of unlabeled documents. This is important because in many text classification problems obtaining training labels is expensive, while large quantities of unlabeled documents are readily available.
We introduce an algorithm for learning from labeled and unlabeled documents based on the combination of Expectation-Maximization (EM) and a naive Bayes classifier. The algorithm first trains a classifier using the available labeled documents, and probabilistically labels the unlabeled documents. It then trains a new classifier using the labels for all the documents, and iterates to convergence. This basic EM procedure works well when the data conform to the generative assumptions of the model. However these assumptions are often violated in practice, and poor performance can result. We present two extensions to the algorithm that improve classification accuracy under these conditions: (1) a weighting factor to modulate the contribution of the unlabeled data, and (2) the use of multiple mixture components per class. Experimental results, obtained using text from three different real-world tasks, show that the use of unlabeled data reduces classification error by up to 30%.
3,123 citations
••
University of Colorado Boulder1, Carnegie Mellon University2, Paul Scherrer Institute3, University at Albany, SUNY4, University of California, Berkeley5, Swiss Federal Laboratories for Materials Science and Technology6, University of California, Davis7, State University of New York System8, University of Eastern Finland9, Finnish Meteorological Institute10, University of Helsinki11, Stockholm University12, Texas A&M University13, Max Planck Society14, University of Tokyo15, University of New Hampshire16, National Oceanic and Atmospheric Administration17
TL;DR: A unifying model framework describing the atmospheric evolution of OA that is constrained by high–time-resolution measurements of its composition, volatility, and oxidation state is presented, which can serve as a basis for improving parameterizations in regional and global models.
Abstract: Organic aerosol (OA) particles affect climate forcing and human health, but their sources and evolution remain poorly characterized. We present a unifying model framework describing the atmospheric evolution of OA that is constrained by high-time-resolution measurements of its composition, volatility, and oxidation state. OA and OA precursor gases evolve by becoming increasingly oxidized, less volatile, and more hygroscopic, leading to the formation of oxygenated organic aerosol (OOA), with concentrations comparable to those of sulfate aerosol throughout the Northern Hemisphere. Our model framework captures the dynamic aging behavior observed in both the atmosphere and laboratory: It can serve as a basis for improving parameterizations in regional and global models.
3,104 citations
••
TL;DR: In this article, an alternative approach, which relies on the assumption that areas of true neural activity will tend to stimulate signal changes over contiguous pixels, is presented, which can improve statistical power by as much as fivefold over techniques that rely solely on adjusting per pixel false positive probabilities.
Abstract: The typical functional magnetic resonance (fMRI) study presents a formidable problem of multiple statistical comparisons (i.e., > 10,000 in a 128 x 128 image). To protect against false positives, investigators have typically relied on decreasing the per pixel false positive probability. This approach incurs an inevitable loss of power to detect statistically significant activity. An alternative approach, which relies on the assumption that areas of true neural activity will tend to stimulate signal changes over contiguous pixels, is presented. If one knows the probability distribution of such cluster sizes as a function of per pixel false positive probability, one can use cluster-size thresholds independently to reject false positives. Both Monte Carlo simulations and fMRI studies of human subjects have been used to verify that this approach can improve statistical power by as much as fivefold over techniques that rely solely on adjusting per pixel false positive probabilities.
3,094 citations
••
TL;DR: Computer tutors based on a set of pedagogical principles derived from the ACT theory of cognition have been developed for teaching students to do proofs in geometry and to write computer programs in the language LISP.
Abstract: Cognitive psychology, artificial intelligence, and computer technology have advanced to the point where it is feasible to build computer systems that are as effective as intelligent human tutors Computer tutors based on a set of pedagogical principles derived from the ACT theory of cognition have been developed for teaching students to do proofs in geometry and to write computer programs in the language LISP
3,092 citations
Authors
Showing all 36645 results
Name | H-index | Papers | Citations |
---|---|---|---|
Yi Chen | 217 | 4342 | 293080 |
Rakesh K. Jain | 200 | 1467 | 177727 |
Robert C. Nichol | 187 | 851 | 162994 |
Michael I. Jordan | 176 | 1016 | 216204 |
Jasvinder A. Singh | 176 | 2382 | 223370 |
J. N. Butler | 172 | 2525 | 175561 |
P. Chang | 170 | 2154 | 151783 |
Krzysztof Matyjaszewski | 169 | 1431 | 128585 |
Yang Yang | 164 | 2704 | 144071 |
Geoffrey E. Hinton | 157 | 414 | 409047 |
Herbert A. Simon | 157 | 745 | 194597 |
Yongsun Kim | 156 | 2588 | 145619 |
Terrence J. Sejnowski | 155 | 845 | 117382 |
John B. Goodenough | 151 | 1064 | 113741 |
Scott Shenker | 150 | 454 | 118017 |