scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Phd by thesis

01 Apr 1988-Nature (Nature Publishing Group)-Vol. 332, Iss: 6166, pp 676-676
TL;DR: In this paper, a sedimentological core and petrographic characterisation of samples from eleven boreholes from the Lower Carboniferous of Bowland Basin (Northwest England) is presented.
Abstract: Deposits of clastic carbonate-dominated (calciclastic) sedimentary slope systems in the rock record have been identified mostly as linearly-consistent carbonate apron deposits, even though most ancient clastic carbonate slope deposits fit the submarine fan systems better. Calciclastic submarine fans are consequently rarely described and are poorly understood. Subsequently, very little is known especially in mud-dominated calciclastic submarine fan systems. Presented in this study are a sedimentological core and petrographic characterisation of samples from eleven boreholes from the Lower Carboniferous of Bowland Basin (Northwest England) that reveals a >250 m thick calciturbidite complex deposited in a calciclastic submarine fan setting. Seven facies are recognised from core and thin section characterisation and are grouped into three carbonate turbidite sequences. They include: 1) Calciturbidites, comprising mostly of highto low-density, wavy-laminated bioclast-rich facies; 2) low-density densite mudstones which are characterised by planar laminated and unlaminated muddominated facies; and 3) Calcidebrites which are muddy or hyper-concentrated debrisflow deposits occurring as poorly-sorted, chaotic, mud-supported floatstones. These

Content maybe subject to copyright    Report

Citations
More filters
ReportDOI
01 Nov 1990
TL;DR: This report will establish methods for performing a domain analysis and describe the products of the domain analysis process to illustrate the application of domain analysis to a representative class of software systems.
Abstract: : Successful Software reuse requires the systematic discovery and exploitation of commonality across related software systems. By examining related software systems and the underlying theory of the class of systems they represent, domain analysis can provide a generic description of the requirements of that class of systems and a set of approaches for their implementation. This report will establish methods for performing a domain analysis and describe the products of the domain analysis process. To illustrate the application of domain analysis to a representative class of software systems, this report will provide a domain analysis of window management system software.

4,420 citations

Journal ArticleDOI
TL;DR: The straw person model (SPM) as mentioned in this paper has been proposed to explain the orientation effects of active galactic nuclei (AGN) and quasars in the line of sight (LOS) images.
Abstract: Because the critical central regions of Active Galactic Nuclei (AGN) and quasars are strongly nonspherical but spatially unresolved, orientation effects have been the source of much confusion. In fact, it now appears that much of the variety in AGN types is just the result of varying orientation relative to the line of sight. We can define an extreme hypothesis,, the straw person model (SPM), in which there are two basic types of AGN: the radio quiets and the radio louds. For each type there is a range in intrinsic luminosity, and the luminosity controls some properties such as the Fanaroff and Riley classes. However, at a given intrinsic luminosity, all other properties such as spectroscopic classification and VLBI component speeds are ascribed to orientation. This model is only a caricature of the unification idea, and is already ruled out on many grounds, but it will be useful for organizing the discussion. I’ll describe what I consider to be convincing evidence that orientation effects are important and widespread. The true situation may be in some sense half way between the SPM and the hypothesis that orientation doesn’t affect classification at aIl. To us optimists, the orienration cup is half full rather than half empty. Although it is too soon to say for sure, the hypothesis that most objects’ classifications would be different if seen from other directions is a tenable one today.

4,005 citations


Cites methods from "Phd by thesis"

  • ...Kay (1990) and Tran et al (1992) have fo und more Seyfert 2s with broad permitted lines in the polarized fl ux, but the FC polarizations are generally much less than that of NGC 1068....

    [...]

Journal ArticleDOI
TL;DR: This Review assembles the current knowledge on the isolation of microfibrillated cellulose from wood and its application in nanocomposites; the preparation of nanocrystalline cellulose and its use as a reinforcing agent; and the biofabrication of bacterial nanocellulose, as well as its evaluation as a biomaterial for medical implants.
Abstract: Cellulose fibrils with widths in the nanometer range are nature-based materials with unique and potentially useful features. Most importantly, these novel nanocelluloses open up the strongly expanding fields of sustainable materials and nanocomposites, as well as medical and life-science devices, to the natural polymer cellulose. The nanodimensions of the structural elements result in a high surface area and hence the powerful interaction of these celluloses with surrounding species, such as water, organic and polymeric compounds, nanoparticles, and living cells. This Review assembles the current knowledge on the isolation of microfibrillated cellulose from wood and its application in nanocomposites; the preparation of nanocrystalline cellulose and its use as a reinforcing agent; and the biofabrication of bacterial nanocellulose, as well as its evaluation as a biomaterial for medical implants.

3,452 citations


Cites background or methods from "Phd by thesis"

  • ...b) Spheres formed by agitated cultivation with a shaking rate of 80–100 rpm; diameter: 2–3 mm, smooth surface.([181]) c) Tubes...

    [...]

  • ...a) Film prepared in a PP container under static conditions; dimensions: 25 25 cm(2), thickness: 200 mm.([181]) b) Spheres formed by agitated cultivation with a shaking rate of 80–100 rpm; diameter: 2–3 mm, smooth surface....

    [...]

Journal ArticleDOI
TL;DR: The Arcade Learning Environment (ALE) as discussed by the authors is a platform for evaluating the development of general, domain-independent AI technology, which provides an interface to hundreds of Atari 2600 game environments, each one different, interesting, and designed to be a challenge for human players.
Abstract: In this article we introduce the Arcade Learning Environment (ALE): both a challenge problem and a platform and methodology for evaluating the development of general, domain-independent AI technology. ALE provides an interface to hundreds of Atari 2600 game environments, each one different, interesting, and designed to be a challenge for human players. ALE presents significant research challenges for reinforcement learning, model learning, model-based planning, imitation learning, transfer learning, and intrinsic motivation. Most importantly, it provides a rigorous testbed for evaluating and comparing approaches to these problems. We illustrate the promise of ALE by developing and benchmarking domain-independent agents designed using well-established AI techniques for both reinforcement learning and planning. In doing so, we also propose an evaluation methodology made possible by ALE, reporting empirical results on over 55 different games. All of the software, including the benchmark agents, is publicly available.

2,429 citations

Journal ArticleDOI
TL;DR: The solid electrolyte interphase (SEI) is a protecting layer formed on the negative electrode of Li-ion batteries as a result of electrolyte decomposition, mainly during the first cycle as discussed by the authors.

2,386 citations

References
More filters
30 Sep 1996
TL;DR: A variant of clausal normal form table aux that takes advantage of the central idea of (positive) hyper resolution, namely to resolve away all negative literals of a clause in a single inference step to enable new efficient proof procedures for full first order theories as variants of table ux calculi.
Abstract: This paper introduces a variant of clausal normal form table aux that we call “hyper tableaux”. Hyper tableaux keep many desirabl e features of analytic tableaux while taking advantage of the central idea fr om (positive) hyper resolution, namely to resolve away all negative literals of a clause in a single inference step. Another feature of the proposed calculus is th e extensive use of universally quantified variables. This enables new efficient fo rward-chaining proof procedures for full first order theories as variants of table ux calculi.

192 citations

Proceedings ArticleDOI
07 Nov 2001
TL;DR: The feasibility of automatically finding places in the program that are candidates for specific refactoring, using an invariant pattern matcher for several common refactorings, is demonstrated and applied to an existing Java code base.
Abstract: Program refactoring-transforming a program to improve readability, structure, performance, abstraction, maintainability, or other features-is not applied in practice as much as might be desired One deterrent is the cost of detecting candidates for refactoring and of choosing the appropriate refactoring transformation This paper demonstrates the feasibility of automatically finding places in the program that are candidates for specific refactorings The approach uses program invariants: when a particular pattern of invariant relationships appears at a program point, a specific refactoring is applicable Since most programs lack explicit invariants, an invariant detection tool called Daikon is used to infer the required invariants We developed an invariant pattern matcher for several common refactorings and applied it to an existing Java code base Numerous refactorings were detected, and one of the developers of the code base assessed their efficacy

191 citations

Journal ArticleDOI
TL;DR: In this article, the authors discuss challenges, opportunities, and early demonstrations of nanophotonic EOMs attempting to address the weak light-matter interaction (LMI) challenge, and the early benchmarks suggest that densely integrated high-performance photonic integrated circuits.
Abstract: As electronic device feature sizes scale-down, the power consumed due to onchip communications as compared to computations will increase dramatically; likewise, the available bandwidth per computational operation will continue to decrease. Integrated photonics can offer savings in power and potential increase in bandwidth for onchip networks. Classical diffraction-limited photonics currently utilized in photonic integrated circuits (PIC) is characterized by bulky and inefficient devices compared to their electronic counterparts due to weak light–matter interactions (LMI). Performance critical for the PIC is electro-optic modulators (EOM), whose performances depend inherently on enhancing LMIs. Current EOMs based on diffraction-limited optical modes often deploy ring resonators and are consequently bulky, photon-lifetime modulation limited, and power inefficient due to large electrical capacitances and thermal tuning requirements. In contrast, wavelength-scale EOMs are potentially able to surpass fundamental restrictions set by classical (i.e. diffraction-limited) devices via (a) high-index modulating materials, (b) nonresonant field and density-of-states enhancements such as found in metal optics, and (c) synergistic onchip integration schemes. This manuscript discusses challenges, opportunities, and early demonstrations of nanophotonic EOMs attempting to address this LMI challenge, and early benchmarks suggest that nanophotonic building blocks allow for densely integrated high-performance photonic integrated circuits.

190 citations

Journal ArticleDOI
A. Aab1, P. Abreu2, Marco Aglietta3, Eun-Joo Ahn4  +470 moreInstitutions (67)
TL;DR: In this paper, a single-flavor limit to the diffuse flux of ultrahigh energy neutrinos with an E-2 spectrum in the energy range 1.0×1017eV-2.
Abstract: Neutrinos in the cosmic ray flux with energies near 1 EeV and above are detectable with the Surface Detector array (SD) of the Pierre Auger Observatory. We report here on searches through Auger data from 1 January 2004 until 20 June 2013. No neutrino candidates were found, yielding a limit to the diffuse flux of ultrahigh energy neutrinos that challenges the Waxman-Bahcall bound predictions. Neutrino identification is attempted using the broad time structure of the signals expected in the SD stations, and is efficiently done for neutrinos of all flavors interacting in the atmosphere at large zenith angles, as well as for "Earth-skimming" neutrino interactions in the case of tau neutrinos. In this paper the searches for downward-going neutrinos in the zenith angle bins 60°-75°and 75°-90°as well as for upward-going neutrinos, are combined to give a single limit. The 90% C.L. single-flavor limit to the diffuse flux of ultrahigh energy neutrinos with an E-2 spectrum in the energy range 1.0×1017eV-2.5×1019eV is Eν2dNν/dEν<6.4×10-9GeVcm-2s-1sr-1.

189 citations

Proceedings Article
25 Jul 2015
TL;DR: This paper investigates the intersection of reinforcement learning and expert demonstrations, leveraging the theoretical guarantees provided by reinforcement learning, and using expert demonstrations to speed up this learning by biasing exploration through a process called reward shaping.
Abstract: Reinforcement learning describes how a learning agent can achieve optimal behaviour based on interactions with its environment and reward feedback. A limiting factor in reinforcement learning as employed in artificial intelligence is the need for an often prohibitively large number of environment samples before the agent reaches a desirable level of performance. Learning from demonstration is an approach that provides the agent with demonstrations by a supposed expert, from which it should derive suitable behaviour. Yet, one of the challenges of learning from demonstration is that no guarantees can be provided for the quality of the demonstrations, and thus the learned behavior. In this paper, we investigate the intersection of these two approaches, leveraging the theoretical guarantees provided by reinforcement learning, and using expert demonstrations to speed up this learning by biasing exploration through a process called reward shaping. This approach allows us to leverage human input without making an erroneous assumption regarding demonstration optimality. We show experimentally that this approach requires significantly fewer demonstrations, is more robust against suboptimality of demonstrations, and achieves much faster learning than the recently developed HAT algorithm.

189 citations