scispace - formally typeset
Search or ask a question
Author

Ron Weiss

Bio: Ron Weiss is an academic researcher from Massachusetts Institute of Technology. The author has contributed to research in topics: Synthetic biology & Speech synthesis. The author has an hindex of 82, co-authored 292 publications receiving 89189 citations. Previous affiliations of Ron Weiss include French Institute for Research in Computer Science and Automation & Google.


Papers
More filters
Journal ArticleDOI
TL;DR: In this article, the authors describe control systems approaches for achieving context-aware devices that are robust to context effects, and then consider cell fate programing as a case study to explore the potential impact of contextaware devices for regenerative medicine applications.
Abstract: The rise of systems biology has ushered a new paradigm: the view of the cell as a system that processes environmental inputs to drive phenotypic outputs. Synthetic biology provides a complementary approach, allowing us to program cell behavior through the addition of synthetic genetic devices into the cellular processor. These devices, and the complex genetic circuits they compose, are engineered using a design-prototype-test cycle, allowing for predictable device performance to be achieved in a context-dependent manner. Within mammalian cells, context effects impact device performance at multiple scales, including the genetic, cellular, and extracellular levels. In order for synthetic genetic devices to achieve predictable behaviors, approaches to overcome context dependence are necessary. Here, we describe control systems approaches for achieving context-aware devices that are robust to context effects. We then consider cell fate programing as a case study to explore the potential impact of context-aware devices for regenerative medicine applications.

26 citations

Proceedings Article
01 Jan 2020
TL;DR: This article proposed a mixture invariant training (MixIT) method that requires only single-channel acoustic mixtures, where training examples are constructed by mixing together existing mixtures and the model separates them into a variable number of latent sources, such that the separated sources can be remixed to approximate the original mixtures.
Abstract: In recent years, rapid progress has been made on the problem of single-channel sound separation using supervised training of deep neural networks. In such supervised approaches, a model is trained to predict the component sources from synthetic mixtures created by adding up isolated ground-truth sources. Reliance on this synthetic training data is problematic because good performance depends upon the degree of match between the training data and real-world audio, especially in terms of the acoustic conditions and distribution of sources. The acoustic properties can be challenging to accurately simulate, and the distribution of sound types may be hard to replicate. In this paper, we propose a completely unsupervised method, mixture invariant training (MixIT), that requires only single-channel acoustic mixtures. In MixIT, training examples are constructed by mixing together existing mixtures, and the model separates them into a variable number of latent sources, such that the separated sources can be remixed to approximate the original mixtures. We show that MixIT can achieve competitive performance compared to supervised methods on speech separation. Using MixIT in a semi-supervised learning setting enables unsupervised domain adaptation and learning from large amounts of real world data without ground-truth source waveforms. In particular, we significantly improve reverberant speech separation performance by incorporating reverberant mixtures, train a speech enhancement system from noisy mixtures, and improve universal sound separation by incorporating a large amount of in-the-wild data.

26 citations

Journal ArticleDOI
TL;DR: A similar device that can solve a satisfiability problem in any form is devised that occupies linear space and operates in quadratic time, while a previously described competing device is built in quad ratic space and operating in quadRatic time or greater.
Abstract: Previously, we described ways to implement the functions AND and OR in a DNA computer consisting of microreactors with attached heating elements that control annealing of DNA Based on these findings, we have devised a similar device that can solve a satisfiability problem in any form The device occupies linear space and operates in quadratic time, while a previously described competing device is built in quadratic space and operates in quadratic time or greater Reducing the number of reactors in a DNA computer reduces the loss of DNA through binding to the surfaces of the system

25 citations

Journal ArticleDOI
TL;DR: The proposed algorithm effectively combines information derived from low level perceptual cues, similar to those used by the human auditory system, with higher level information related to speaker identity to derive an EM algorithm for finding the maximum likelihood parameters of the joint model.

24 citations


Cited by
More filters
28 Jul 2005
TL;DR: PfPMP1)与感染红细胞、树突状组胞以及胎盘的单个或多个受体作用,在黏附及免疫逃避中起关键的作�ly.
Abstract: 抗原变异可使得多种致病微生物易于逃避宿主免疫应答。表达在感染红细胞表面的恶性疟原虫红细胞表面蛋白1(PfPMP1)与感染红细胞、内皮细胞、树突状细胞以及胎盘的单个或多个受体作用,在黏附及免疫逃避中起关键的作用。每个单倍体基因组var基因家族编码约60种成员,通过启动转录不同的var基因变异体为抗原变异提供了分子基础。

18,940 citations

Proceedings ArticleDOI
13 Aug 2016
TL;DR: XGBoost as discussed by the authors proposes a sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning to achieve state-of-the-art results on many machine learning challenges.
Abstract: Tree boosting is a highly effective and widely used machine learning method. In this paper, we describe a scalable end-to-end tree boosting system called XGBoost, which is used widely by data scientists to achieve state-of-the-art results on many machine learning challenges. We propose a novel sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning. More importantly, we provide insights on cache access patterns, data compression and sharding to build a scalable tree boosting system. By combining these insights, XGBoost scales beyond billions of examples using far fewer resources than existing systems.

14,872 citations

Journal ArticleDOI
01 Apr 1998
TL;DR: This paper provides an in-depth description of Google, a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and looks at the problem of how to effectively deal with uncontrolled hypertext collections where anyone can publish anything they want.
Abstract: In this paper, we present Google, a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext. Google is designed to crawl and index the Web efficiently and produce much more satisfying search results than existing systems. The prototype with a full text and hyperlink database of at least 24 million pages is available at http://google.stanford.edu/. To engineer a search engine is a challenging task. Search engines index tens to hundreds of millions of web pages involving a comparable number of distinct terms. They answer tens of millions of queries every day. Despite the importance of large-scale search engines on the web, very little academic research has been done on them. Furthermore, due to rapid advance in technology and web proliferation, creating a web search engine today is very different from three years ago. This paper provides an in-depth description of our large-scale web search engine -- the first such detailed public description we know of to date. Apart from the problems of scaling traditional search techniques to data of this magnitude, there are new technical challenges involved with using the additional information present in hypertext to produce better search results. This paper addresses this question of how to build a practical large-scale system which can exploit the additional information present in hypertext. Also we look at the problem of how to effectively deal with uncontrolled hypertext collections where anyone can publish anything they want.

14,696 citations

Proceedings Article
11 Nov 1999
TL;DR: This paper describes PageRank, a mathod for rating Web pages objectively and mechanically, effectively measuring the human interest and attention devoted to them, and shows how to efficiently compute PageRank for large numbers of pages.
Abstract: The importance of a Web page is an inherently subjective matter, which depends on the readers interests, knowledge and attitudes. But there is still much that can be said objectively about the relative importance of Web pages. This paper describes PageRank, a mathod for rating Web pages objectively and mechanically, effectively measuring the human interest and attention devoted to them. We compare PageRank to an idealized random Web surfer. We show how to efficiently compute PageRank for large numbers of pages. And, we show how to apply PageRank to search and to user navigation.

14,400 citations

Proceedings ArticleDOI
TL;DR: This paper proposes a novel sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning and provides insights on cache access patterns, data compression and sharding to build a scalable tree boosting system called XGBoost.
Abstract: Tree boosting is a highly effective and widely used machine learning method. In this paper, we describe a scalable end-to-end tree boosting system called XGBoost, which is used widely by data scientists to achieve state-of-the-art results on many machine learning challenges. We propose a novel sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning. More importantly, we provide insights on cache access patterns, data compression and sharding to build a scalable tree boosting system. By combining these insights, XGBoost scales beyond billions of examples using far fewer resources than existing systems.

13,333 citations