Institution
University of California, San Diego
Education•San Diego, California, United States•
About: University of California, San Diego is a education organization based out in San Diego, California, United States. It is known for research contribution in the topics: Population & Poison control. The organization has 83317 authors who have published 204524 publications receiving 12315489 citations. The organization is also known as: UCSD & UC San Diego.
Topics: Population, Poison control, Signal transduction, Cancer, Gene
Papers published on a yearly basis
Papers
More filters
••
TL;DR: In this paper, the Hartree and Hartree-Fock equations are applied to a uniform electron gas, where the exchange and correlation portions of the chemical potential of the gas are used as additional effective potentials.
Abstract: From a theory of Hohenberg and Kohn, approximation methods for treating an inhomogeneous system of interacting electrons are developed. These methods are exact for systems of slowly varying or high density. For the ground state, they lead to self-consistent equations analogous to the Hartree and Hartree-Fock equations, respectively. In these equations the exchange and correlation portions of the chemical potential of a uniform electron gas appear as additional effective potentials. (The exchange portion of our effective potential differs from that due to Slater by a factor of $\frac{2}{3}$.) Electronic systems at finite temperatures and in magnetic fields are also treated by similar methods. An appendix deals with a further correction for systems with short-wavelength density oscillations.
47,477 citations
••
TL;DR: In this article, the ground state of an interacting electron gas in an external potential was investigated and it was proved that there exists a universal functional of the density, called F[n(mathrm{r})], independent of the potential of the electron gas.
Abstract: This paper deals with the ground state of an interacting electron gas in an external potential $v(\mathrm{r})$. It is proved that there exists a universal functional of the density, $F[n(\mathrm{r})]$, independent of $v(\mathrm{r})$, such that the expression $E\ensuremath{\equiv}\ensuremath{\int}v(\mathrm{r})n(\mathrm{r})d\mathrm{r}+F[n(\mathrm{r})]$ has as its minimum value the correct ground-state energy associated with $v(\mathrm{r})$. The functional $F[n(\mathrm{r})]$ is then discussed for two situations: (1) $n(\mathrm{r})={n}_{0}+\stackrel{\ifmmode \tilde{}\else \~{}\fi{}}{n}(\mathrm{r})$, $\frac{\stackrel{\ifmmode \tilde{}\else \~{}\fi{}}{n}}{{n}_{0}}\ensuremath{\ll}1$, and (2) $n(\mathrm{r})=\ensuremath{\phi}(\frac{\mathrm{r}}{{r}_{0}})$ with $\ensuremath{\phi}$ arbitrary and ${r}_{0}\ensuremath{\rightarrow}\ensuremath{\infty}$. In both cases $F$ can be expressed entirely in terms of the correlation energy and linear and higher order electronic polarizabilities of a uniform electron gas. This approach also sheds some light on generalized Thomas-Fermi methods and their limitations. Some new extensions of these methods are presented.
38,160 citations
•
01 Jan 1943TL;DR: Combinations involving trigonometric and hyperbolic functions and power 5 Indefinite Integrals of Special Functions 6 Definite Integral Integral Functions 7.Associated Legendre Functions 8 Special Functions 9 Hypergeometric Functions 10 Vector Field Theory 11 Algebraic Inequalities 12 Integral Inequality 13 Matrices and related results 14 Determinants 15 Norms 16 Ordinary differential equations 17 Fourier, Laplace, and Mellin Transforms 18 The z-transform
Abstract: 0 Introduction 1 Elementary Functions 2 Indefinite Integrals of Elementary Functions 3 Definite Integrals of Elementary Functions 4.Combinations involving trigonometric and hyperbolic functions and power 5 Indefinite Integrals of Special Functions 6 Definite Integrals of Special Functions 7.Associated Legendre Functions 8 Special Functions 9 Hypergeometric Functions 10 Vector Field Theory 11 Algebraic Inequalities 12 Integral Inequalities 13 Matrices and related results 14 Determinants 15 Norms 16 Ordinary differential equations 17 Fourier, Laplace, and Mellin Transforms 18 The z-transform
27,354 citations
••
TL;DR: The relationship between co-integration and error correction models, first suggested in Granger (1981), is here extended and used to develop estimation procedures, tests, and empirical examples.
Abstract: The relationship between co-integration and error correction models, first suggested in Granger (1981), is here extended and used to develop estimation procedures, tests, and empirical examples. If each element of a vector of time series x first achieves stationarity after differencing, but a linear combination a'x is already stationary, the time series x are said to be co-integrated with co-integrating vector a. There may be several such co-integrating vectors so that a becomes a matrix. Interpreting a'x,= 0 as a long run equilibrium, co-integration implies that deviations from equilibrium are stationary, with finite variance, even though the series themselves are nonstationary and have infinite variance. The paper presents a representation theorem based on Granger (1983), which connects the moving average, autoregressive, and error correction representations for co-integrated systems. A vector autoregression in differenced variables is incompatible with these representations. Estimation of these models is discussed and a simple but asymptotically efficient two-step estimator is proposed. Testing for co-integration combines the problems of unit root tests and tests with parameters unidentified under the null. Seven statistics are formulated and analyzed. The critical values of these statistics are calculated based on a Monte Carlo simulation. Using these critical values, the power properties of the tests are examined and one test procedure is recommended for application. In a series of examples it is found that consumption and income are co-integrated, wages and prices are not, short and long interest rates are, and nominal GNP is co-integrated with M2, but not M1, M3, or aggregate liquid assets.
27,170 citations
••
TL;DR: Back-propagation repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector, which helps to represent important features of the task domain.
Abstract: We describe a new learning procedure, back-propagation, for networks of neurone-like units. The procedure repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector. As a result of the weight adjustments, internal ‘hidden’ units which are not part of the input or output come to represent important features of the task domain, and the regularities in the task are captured by the interactions of these units. The ability to create useful new features distinguishes back-propagation from earlier, simpler methods such as the perceptron-convergence procedure1.
23,814 citations
Authors
Showing all 84160 results
Name | H-index | Papers | Citations |
---|---|---|---|
Robert Langer | 281 | 2324 | 326306 |
Zhong Lin Wang | 245 | 2529 | 259003 |
Michael Karin | 236 | 704 | 226485 |
Eugene Braunwald | 230 | 1711 | 264576 |
Fred H. Gage | 216 | 967 | 185732 |
Robert J. Lefkowitz | 214 | 860 | 147995 |
Peter Libby | 211 | 932 | 182724 |
Peer Bork | 206 | 697 | 245427 |
Rob Knight | 201 | 1061 | 253207 |
Ronald M. Evans | 199 | 708 | 166722 |
Carlo M. Croce | 198 | 1135 | 189007 |
Lewis C. Cantley | 196 | 748 | 169037 |
John C. Reed | 190 | 891 | 164382 |
Gad Getz | 189 | 520 | 247560 |
Scott M. Grundy | 187 | 841 | 231821 |