Institution
University of Colorado Colorado Springs
Education•Colorado Springs, Colorado, United States•
About: University of Colorado Colorado Springs is a education organization based out in Colorado Springs, Colorado, United States. It is known for research contribution in the topics: Population & Poison control. The organization has 6664 authors who have published 10872 publications receiving 323416 citations. The organization is also known as: UCCS & University of Colorado at Colorado Springs.
Topics: Population, Poison control, Thin film, Capacitor, Ferroelectricity
Papers published on a yearly basis
Papers
More filters
••
Bose Corporation1, Cardiff University2, Durham University3, Adam Mickiewicz University in Poznań4, Max Planck Society5, Katholieke Universiteit Leuven6, Delft University of Technology7, Georgia Institute of Technology8, Kaiserslautern University of Technology9, Beihang University10, École Polytechnique Fédérale de Lausanne11, Saratov State University12, Paris-Sorbonne University13, The Catholic University of America14, University of Notre Dame15, University of Münster16, Emory University17, Polytechnic University of Milan18, Dresden University of Technology19, Helmholtz-Zentrum Dresden-Rossendorf20, University of Exeter21, Donetsk National University22, SRM University23, National University of Singapore24, University of Greifswald25, Kyoto University26, Tohoku University27, Federico Santa María Technical University28, University of Santiago, Chile29, University of Perugia30, Université Paris-Saclay31, University of Manitoba32, University of Colorado Colorado Springs33, University of Tokyo34, University of Groningen35, Technische Universität München36, Technical University of Dortmund37, University of Vienna38, Aalto University39, University of California, Riverside40, Intel41, University of Duisburg-Essen42, University of Oldenburg43
TL;DR: The Roadmap on Magnonics as mentioned in this paper is a collection of 22 sections written by leading experts in this field who review and discuss the current status but also present their vision of future perspectives.
Abstract: Magnonics is a rather young physics research field in nanomagnetism and nanoscience that addresses the use of spin waves (magnons) to transmit, store, and process information. After several papers and review articles published in the last decade, with a steadily increase in the number of citations, we are presenting the first Roadmap on Magnonics. This a collection of 22 sections written by leading experts in this field who review and discuss the current status but also present their vision of future perspectives. Today, the principal challenges in applied magnonics are the excitation of sub-100 nm wavelength magnons, their manipulation on the nanoscale and the creation of sub-micrometre devices using low-Gilbert damping magnetic materials and the interconnections to standard electronics. In this respect, magnonics offers lower energy consumption, easier integrability and compatibility with CMOS structure, reprogrammability, shorter wavelength, smaller device features, anisotropic properties, negative group velocity, non-reciprocity and efficient tunability by various external stimuli to name a few. Hence, despite being a young research field, magnonics has come a long way since its early inception. This Roadmap represents a milestone for future emerging research directions in magnonics and hopefully it will be followed by a series of articles on the same topic.
188 citations
••
TL;DR: A class of adaptive algorithms designed for use with IIR digital filters which offer a much reduced computational load for basically the same performance, and have their basis in the theory of hyperstability, which yields HARF, a hyperstable adaptive recursive filtering algorithm which has provable convergence properties.
Abstract: The concept of adaptation in digital filtering has proven to be a powerful and versatile means of signal processing in applications where precise a priori filter design is impractical. Adaptive filters have traditionally been implemented with FIR structures, making their analysis fairly straightforward but leading to high computation cost in many cases of practical interest (e.g, sinusoid enhancement). This paper introduces a class of adaptive algorithms designed for use with IIR digital filters which offer a much reduced computational load for basically the same performance. These algorithms have their basis in the theory of hyperstability, a concept historically associated with the analysis of closed-loop nonlinear time-varying control systems. Exploiting this theory yields HARF, a hyperstable adaptive recursive filtering algorithm which has provable convergence properties. A simplified version of the algorithm, called SHARF, is then developed which retains provable convergence at low convergence rates and is well suited to real-time applications. In this paper both HARF and SHARF are described and some background into the meaning and utility of hyperstability is given, in addition, computer simulations are presented for two practical applications of IIR adaptive filters: noise and multi-path cancellation.
187 citations
•
01 Jan 2018TL;DR: This paper introduces a new evaluation metric that focuses on comparing the performance of multiple approaches in scenarios where such unseen classes or unknowns are encountered, and develops novel loss functions that train networks using negative samples from some classes.
Abstract: Agnostophobia, the fear of the unknown, can be experienced by deep learning engineers while applying their networks to real-world applications. Unfortunately, network behavior is not well defined for inputs far from a networks training set. In an uncontrolled environment, networks face many instances that are not of interest to them and have to be rejected in order to avoid a false positive. This problem has previously been tackled by researchers by either a) thresholding softmax, which by construction cannot return "none of the known classes", or b) using an additional background or garbage class. In this paper, we show that both of these approaches help, but are generally insufficient when previously unseen classes are encountered. We also introduce a new evaluation metric that focuses on comparing the performance of multiple approaches in scenarios where such unseen classes or unknowns are encountered. Our major contributions are simple yet effective Entropic Open-Set and Objectosphere losses that train networks using negative samples from some classes. These novel losses are designed to maximize entropy for unknown inputs while increasing separation in deep feature space by modifying magnitudes of known and unknown samples. Experiments on networks trained to classify classes from MNIST and CIFAR-10 show that our novel loss functions are significantly better at dealing with unknown inputs from datasets such as Devanagari, NotMNIST, CIFAR-100 and SVHN.
187 citations
•
03 Jun 1986TL;DR: In this paper, a method for transferring data from a microprocessor located in a transaction card through a card reader by emulating a prerecorded magnetic stripe on a conventional transaction car such as a credit or debit card is presented.
Abstract: A device and method for transferring data from a microprocessor located in a transaction card through a card reader by emulating a prerecorded magnetic stripe on a conventional transaction car such as a credit or debit card. Data is sequentially produced by the microprocessor within the card and applied to a magnetic field generator which produces magnetic fields that emulate prerecorded data on a conventional magnetic stripe of a transaction card. This allows transfer of data from a microprocessor to standard card readers without the necessity of substantially modifying the card reader device. Circuitry is also provided for detecting the position and speed of movement of the card through the card reader to ensure that all of the data is transmitted from the microprocessor to the magnetic field generators within the scanning time of the card across read head of the card reader.
187 citations
••
TL;DR: Triage is the process of determining which requirements a product should satisfy given the time and resources available, and three product development case studies and 14 recommendations for practicing this neglected art are presented.
Abstract: Driven by an increasingly competitive market, companies add features and compress schedules for the delivery of every product, often creating a complete mismatch of requirements and resources that results in products failing to satisfy customer needs. Triage is the process of determining which requirements a product should satisfy given the time and resources available. The author presents three product development case studies and 14 recommendations for practicing this neglected art.
187 citations
Authors
Showing all 6706 results
Name | H-index | Papers | Citations |
---|---|---|---|
Jeff Greenberg | 105 | 542 | 43600 |
James F. Scott | 99 | 714 | 58515 |
Martin Wikelski | 89 | 420 | 25821 |
Neil W. Kowall | 89 | 279 | 34943 |
Ananth Dodabalapur | 85 | 394 | 27246 |
Tom Pyszczynski | 82 | 246 | 30590 |
Patrick S. Kamath | 78 | 466 | 31281 |
Connie M. Weaver | 77 | 473 | 30985 |
Alejandro Lucia | 75 | 680 | 23967 |
Michael J. McKenna | 70 | 356 | 16227 |
Timothy J. Craig | 69 | 458 | 18340 |
Sheldon Solomon | 67 | 150 | 23916 |
Michael H. Stone | 65 | 370 | 16355 |
Christopher J. Gostout | 65 | 334 | 13593 |
Edward T. Ryan | 60 | 303 | 11822 |