scispace - formally typeset
Search or ask a question
Institution

University of Massachusetts Amherst

EducationAmherst Center, Massachusetts, United States
About: University of Massachusetts Amherst is a education organization based out in Amherst Center, Massachusetts, United States. It is known for research contribution in the topics: Population & Galaxy. The organization has 37274 authors who have published 83965 publications receiving 3834996 citations. The organization is also known as: UMass Amherst & Massachusetts State College.


Papers
More filters
Posted Content
TL;DR: This article introduced a new type of deep contextualized word representation that models both complex characteristics of word use (e.g., syntax and semantics), and how these uses vary across linguistic contexts (i.e., to model polysemy).
Abstract: We introduce a new type of deep contextualized word representation that models both (1) complex characteristics of word use (e.g., syntax and semantics), and (2) how these uses vary across linguistic contexts (i.e., to model polysemy). Our word vectors are learned functions of the internal states of a deep bidirectional language model (biLM), which is pre-trained on a large text corpus. We show that these representations can be easily added to existing models and significantly improve the state of the art across six challenging NLP problems, including question answering, textual entailment and sentiment analysis. We also present an analysis showing that exposing the deep internals of the pre-trained network is crucial, allowing downstream models to mix different types of semi-supervision signals.

1,696 citations

Proceedings ArticleDOI
01 Jan 2016
TL;DR: This work introduces and study an RNN-based variational autoencoder generative model that incorporates distributed latent representations of entire sentences that allows it to explicitly model holistic properties of sentences such as style, topic, and high-level syntactic features.
Abstract: The standard recurrent neural network language model (RNNLM) generates sentences one word at a time and does not work from an explicit global sentence representation. In this work, we introduce and study an RNN-based variational autoencoder generative model that incorporates distributed latent representations of entire sentences. This factorization allows it to explicitly model holistic properties of sentences such as style, topic, and high-level syntactic features. Samples from the prior over these sentence representations remarkably produce diverse and well-formed sentences through simple deterministic decoding. By examining paths through this latent space, we are able to generate coherent novel sentences that interpolate between known sentences. We present techniques for solving the difficult learning problem presented by this model, demonstrate its effectiveness in imputing missing words, explore many interesting properties of the model's latent sentence space, and present negative results on the use of the model in language modeling.

1,690 citations

Journal ArticleDOI
26 Aug 2000-Langmuir
TL;DR: In this article, a series of silicon surfaces were prepared by photolithography and hydrophobized using silanization reagents, and water droplets were pinned on surfaces containing square posts with larger dimensions.
Abstract: We discuss dynamic hydrophobicity from the perspective of the force required to move a water droplet on a surface and argue that the structure of the three-phase contact line is important. We studied the wettability of a series of silicon surfaces that were prepared by photolithography and hydrophobized using silanization reagents. Hydrocarbon, siloxane, and fluorocarbon surfaces were prepared. The surfaces contain posts of different sizes, shapes, and separations. Surfaces containing square posts with X−Y dimensions of 32 μm and less exhibited ultrahydrophobic behavior with high advancing and receding water contact angles. Water droplets moved very easily on these surfaces and rolled off of slightly tilted surfaces. Contact angles were independent of the post height from 20 to 140 μm and independent of surface chemistry. Water droplets were pinned on surfaces containing square posts with larger dimensions. Increasing the distance between posts and changing the shape of the posts from square to staggered ...

1,690 citations

Journal ArticleDOI
TL;DR: An examination of the performance of the tests when the correct model has a quadratic term but a model containing only the linear term has been fit shows that the Pearson chi-square, the unweighted sum-of-squares, the Hosmer-Lemeshow decile of risk, the smoothed residual sum- of-Squares and Stukel's score test, have power exceeding 50 per cent to detect moderate departures from linearity.
Abstract: Recent work has shown that there may be disadvantages in the use of the chi-square-like goodness-of-fit tests for the logistic regression model proposed by Hosmer and Lemeshow that use fixed groups of the estimated probabilities. A particular concern with these grouping strategies based on estimated probabilities, fitted values, is that groups may contain subjects with widely different values of the covariates. It is possible to demonstrate situations where one set of fixed groups shows the model fits while the test rejects fit using a different set of fixed groups. We compare the performance by simulation of these tests to tests based on smoothed residuals proposed by le Cessie and Van Houwelingen and Royston, a score test for an extended logistic regression model proposed by Stukel, the Pearson chi-square and the unweighted residual sum-of-squares. These simulations demonstrate that all but one of Royston's tests have the correct size. An examination of the performance of the tests when the correct model has a quadratic term but a model containing only the linear term has been fit shows that the Pearson chi-square, the unweighted sum-of-squares, the Hosmer-Lemeshow decile of risk, the smoothed residual sum-of-squares and Stukel's score test, have power exceeding 50 per cent to detect moderate departures from linearity when the sample size is 100 and have power over 90 per cent for these same alternatives for samples of size 500. All tests had no power when the correct model had an interaction between a dichotomous and continuous covariate but only the continuous covariate model was fit. Power to detect an incorrectly specified link was poor for samples of size 100. For samples of size 500 Stukel's score test had the best power but it only exceeded 50 per cent to detect an asymmetric link function. The power of the unweighted sum-of-squares test to detect an incorrectly specified link function was slightly less than Stukel's score test. We illustrate the tests within the context of a model for factors associated with low birth weight.

1,666 citations

Journal ArticleDOI
TL;DR: The new version of the ConSurf web server is presented, providing an easier and more intuitive step-by-step interface, while offering the user more flexibility during the process, and calculates the evolutionary rates for nucleic acid sequences.
Abstract: It is informative to detect highly conserved positions in proteins and nucleic acid sequence/structure since they are often indicative of structural and/or functional importance. ConSurf (http://consurf.tau.ac.il) and ConSeq (http://conseq.tau.ac.il) are two well-established web servers for calculating the evolutionary conservation of amino acid positions in proteins using an empirical Bayesian inference, starting from protein structure and sequence, respectively. Here, we present the new version of the ConSurf web server that combines the two independent servers, providing an easier and more intuitive step-by-step interface, while offering the user more flexibility during the process. In addition, the new version of ConSurf calculates the evolutionary rates for nucleic acid sequences. The new version is freely available at: http://consurf.tau.ac.il/.

1,665 citations


Authors

Showing all 37601 results

NameH-indexPapersCitations
George M. Whitesides2401739269833
Joan Massagué189408149951
David H. Weinberg183700171424
David L. Kaplan1771944146082
Michael I. Jordan1761016216204
James F. Sallis169825144836
Bradley T. Hyman169765136098
Anton M. Koekemoer1681127106796
Derek R. Lovley16858295315
Michel C. Nussenzweig16551687665
Alfred L. Goldberg15647488296
Donna Spiegelman15280485428
Susan E. Hankinson15178988297
Bernard Moss14783076991
Roger J. Davis147498103478
Network Information
Related Institutions (5)
Cornell University
235.5K papers, 12.2M citations

96% related

University of Illinois at Urbana–Champaign
225.1K papers, 10.1M citations

96% related

University of Minnesota
257.9K papers, 11.9M citations

96% related

University of Wisconsin-Madison
237.5K papers, 11.8M citations

95% related

University of Toronto
294.9K papers, 13.5M citations

94% related

Performance
Metrics
No. of papers from the Institution in previous years
YearPapers
2023103
2022536
20213,983
20203,858
20193,712
20183,385