scispace - formally typeset
Search or ask a question
Author

Albert-László Barabási

Bio: Albert-László Barabási is an academic researcher from Northeastern University. The author has contributed to research in topics: Complex network & Network science. The author has an hindex of 152, co-authored 438 publications receiving 200119 citations. Previous affiliations of Albert-László Barabási include Budapest University of Technology and Economics & Lawrence Livermore National Laboratory.


Papers
More filters
Patent
10 Nov 2016
TL;DR: In this article, the authors proposed a method for determining a disease state of a patient based on generating personalized biomarker expression perturbation profiles for a plurality of individual subjects with a disease and creating a disease module by combining representations of biomarkers from the personalized profiles.
Abstract: The disclosed methods and systems allow for a systematic quantification of the heterogeneity of disease states between different subjects on a molecular (e.g., gene or protein expression) level. One example embodiment of the invention is a method for determining a disease state of a patient. The method includes generating personalized biomarker expression perturbation profiles for a plurality of individual subjects with a disease. The profiles include representations of biomarker expressions that are perturbed beyond a threshold amount. The method also includes creating a disease module by combining representations of biomarkers from the personalized profiles. The disease module includes a network of representations of biomarkers having perturbations associated with the disease. The method also includes accessing biomarker data for the patient from a sample obtained from the patient and determining the disease state of the patient based on a comparison of the biomarker data and the disease module.

2 citations

Journal ArticleDOI
10 Oct 1996-EPL
TL;DR: In this paper, the interaction of two nonequilibrium conservative fields was analyzed for the growth of semiconductors. But the coupling between the surfactant thickness and the interface height cannot account for the experimentally observed layered growth, implying that reduced diffusion of the embedded atoms is a key mechanism in the growth.
Abstract: We present an analytical study of the interaction of two nonequilibrium conservative fields. Due to the conservative character of the relaxation mechanism, the scaling exponents can be obtained exactly using dynamic renormalization group. We apply our results to surfactant-mediated growth of semiconductors. We find that the coupling between the surfactant thickness and the interface height cannot account for the experimentally observed layered growth, implying that reduced diffusion of the embedded atoms is a key mechanism in surfactant-mediated growth.

2 citations

Journal ArticleDOI
TL;DR: In this article, the authors show how computer simulations can give unique information on the growth of nanostructures and thin films, specifically they can predict the morphologies and the island size distributions corresponding to different growth mechanisms.
Abstract: We show how computer simulations can give unique information on the growth of nanostructures and thin films. Specifically, they can predict the morphologies and the island size distributions corresponding to different growth mechanisms. This information cannot be obtained from other approaches such as mean-field mathematical theories or scaling analysis. Special attention is given to the effects of small cluster mobility on experimental results.

2 citations

Book ChapterDOI
01 Jan 1995

1 citations

Journal ArticleDOI
TL;DR: This Article contains an error in the order of the Figures, which were published as Figures 1, 2, 3 and 4 respectively.
Abstract: Scientific Reports 7: Article number: 39978; published online: 05 January 2017; updated: 09 March 2017 This Article contains an error in the order of the Figures. Figures 1, 2, 3 and 4 were published as Figures 2, 4, 1 and 3 respectively. The correct Figures appear below. The Figure legends are correct.

1 citations


Cited by
More filters
Journal ArticleDOI
15 Oct 1999-Science
TL;DR: A model based on these two ingredients reproduces the observed stationary scale-free distributions, which indicates that the development of large networks is governed by robust self-organizing phenomena that go beyond the particulars of the individual systems.
Abstract: Systems as diverse as genetic networks or the World Wide Web are best described as networks with complex topology. A common property of many large networks is that the vertex connectivities follow a scale-free power-law distribution. This feature was found to be a consequence of two generic mechanisms: (i) networks expand continuously by the addition of new vertices, and (ii) new vertices attach preferentially to sites that are already well connected. A model based on these two ingredients reproduces the observed stationary scale-free distributions, which indicates that the development of large networks is governed by robust self-organizing phenomena that go beyond the particulars of the individual systems.

33,771 citations

01 May 1993
TL;DR: Comparing the results to the fastest reported vectorized Cray Y-MP and C90 algorithm shows that the current generation of parallel machines is competitive with conventional vector supercomputers even for small problems.
Abstract: Three parallel algorithms for classical molecular dynamics are presented. The first assigns each processor a fixed subset of atoms; the second assigns each a fixed subset of inter-atomic forces to compute; the third assigns each a fixed spatial region. The algorithms are suitable for molecular dynamics models which can be difficult to parallelize efficiently—those with short-range forces where the neighbors of each atom change rapidly. They can be implemented on any distributed-memory parallel machine which allows for message-passing of data between independently executing processors. The algorithms are tested on a standard Lennard-Jones benchmark problem for system sizes ranging from 500 to 100,000,000 atoms on several parallel supercomputers--the nCUBE 2, Intel iPSC/860 and Paragon, and Cray T3D. Comparing the results to the fastest reported vectorized Cray Y-MP and C90 algorithm shows that the current generation of parallel machines is competitive with conventional vector supercomputers even for small problems. For large problems, the spatial algorithm achieves parallel efficiencies of 90% and a 1840-node Intel Paragon performs up to 165 faster than a single Cray C9O processor. Trade-offs between the three algorithms and guidelines for adapting them to more complex molecular dynamics simulations are also discussed.

29,323 citations

Book
08 Sep 2000
TL;DR: This book presents dozens of algorithms and implementation examples, all in pseudo-code and suitable for use in real-world, large-scale data mining projects, and provides a comprehensive, practical look at the concepts and techniques you need to get the most out of real business data.
Abstract: The increasing volume of data in modern business and science calls for more complex and sophisticated tools. Although advances in data mining technology have made extensive data collection much easier, it's still always evolving and there is a constant need for new techniques and tools that can help us transform this data into useful information and knowledge. Since the previous edition's publication, great advances have been made in the field of data mining. Not only does the third of edition of Data Mining: Concepts and Techniques continue the tradition of equipping you with an understanding and application of the theory and practice of discovering patterns hidden in large data sets, it also focuses on new, important topics in the field: data warehouses and data cube technology, mining stream, mining social networks, and mining spatial, multimedia and other complex data. Each chapter is a stand-alone guide to a critical topic, presenting proven algorithms and sound implementations ready to be used directly or with strategic modification against live data. This is the resource you need if you want to apply today's most powerful data mining techniques to meet real business challenges. * Presents dozens of algorithms and implementation examples, all in pseudo-code and suitable for use in real-world, large-scale data mining projects. * Addresses advanced topics such as mining object-relational databases, spatial databases, multimedia databases, time-series databases, text databases, the World Wide Web, and applications in several fields. *Provides a comprehensive, practical look at the concepts and techniques you need to get the most out of real business data

23,600 citations

28 Jul 2005
TL;DR: PfPMP1)与感染红细胞、树突状组胞以及胎盘的单个或多个受体作用,在黏附及免疫逃避中起关键的作�ly.
Abstract: 抗原变异可使得多种致病微生物易于逃避宿主免疫应答。表达在感染红细胞表面的恶性疟原虫红细胞表面蛋白1(PfPMP1)与感染红细胞、内皮细胞、树突状细胞以及胎盘的单个或多个受体作用,在黏附及免疫逃避中起关键的作用。每个单倍体基因组var基因家族编码约60种成员,通过启动转录不同的var基因变异体为抗原变异提供了分子基础。

18,940 citations

Journal ArticleDOI
TL;DR: In this paper, a simple model based on the power-law degree distribution of real networks was proposed, which was able to reproduce the power law degree distribution in real networks and to capture the evolution of networks, not just their static topology.
Abstract: The emergence of order in natural systems is a constant source of inspiration for both physical and biological sciences. While the spatial order characterizing for example the crystals has been the basis of many advances in contemporary physics, most complex systems in nature do not offer such high degree of order. Many of these systems form complex networks whose nodes are the elements of the system and edges represent the interactions between them. Traditionally complex networks have been described by the random graph theory founded in 1959 by Paul Erdohs and Alfred Renyi. One of the defining features of random graphs is that they are statistically homogeneous, and their degree distribution (characterizing the spread in the number of edges starting from a node) is a Poisson distribution. In contrast, recent empirical studies, including the work of our group, indicate that the topology of real networks is much richer than that of random graphs. In particular, the degree distribution of real networks is a power-law, indicating a heterogeneous topology in which the majority of the nodes have a small degree, but there is a significant fraction of highly connected nodes that play an important role in the connectivity of the network. The scale-free topology of real networks has very important consequences on their functioning. For example, we have discovered that scale-free networks are extremely resilient to the random disruption of their nodes. On the other hand, the selective removal of the nodes with highest degree induces a rapid breakdown of the network to isolated subparts that cannot communicate with each other. The non-trivial scaling of the degree distribution of real networks is also an indication of their assembly and evolution. Indeed, our modeling studies have shown us that there are general principles governing the evolution of networks. Most networks start from a small seed and grow by the addition of new nodes which attach to the nodes already in the system. This process obeys preferential attachment: the new nodes are more likely to connect to nodes with already high degree. We have proposed a simple model based on these two principles wich was able to reproduce the power-law degree distribution of real networks. Perhaps even more importantly, this model paved the way to a new paradigm of network modeling, trying to capture the evolution of networks, not just their static topology.

18,415 citations