scispace - formally typeset
Search or ask a question

Showing papers by "Instituto Superior Técnico published in 2020"


Journal ArticleDOI
Richard J. Abbott1, T. D. Abbott2, Sheelu Abraham3, Fausto Acernese4  +1334 moreInstitutions (150)
TL;DR: In this paper, the authors reported the observation of a compact binary coalescence involving a 222 −243 M ⊙ black hole and a compact object with a mass of 250 −267 M ⋆ (all measurements quoted at the 90% credible level) The gravitational-wave signal, GW190814, was observed during LIGO's and Virgo's third observing run on 2019 August 14 at 21:10:39 UTC and has a signal-to-noise ratio of 25 in the three-detector network.
Abstract: We report the observation of a compact binary coalescence involving a 222–243 M ⊙ black hole and a compact object with a mass of 250–267 M ⊙ (all measurements quoted at the 90% credible level) The gravitational-wave signal, GW190814, was observed during LIGO's and Virgo's third observing run on 2019 August 14 at 21:10:39 UTC and has a signal-to-noise ratio of 25 in the three-detector network The source was localized to 185 deg2 at a distance of ${241}_{-45}^{+41}$ Mpc; no electromagnetic counterpart has been confirmed to date The source has the most unequal mass ratio yet measured with gravitational waves, ${0112}_{-0009}^{+0008}$, and its secondary component is either the lightest black hole or the heaviest neutron star ever discovered in a double compact-object system The dimensionless spin of the primary black hole is tightly constrained to ≤007 Tests of general relativity reveal no measurable deviations from the theory, and its prediction of higher-multipole emission is confirmed at high confidence We estimate a merger rate density of 1–23 Gpc−3 yr−1 for the new class of binary coalescence sources that GW190814 represents Astrophysical models predict that binaries with mass ratios similar to this event can form through several channels, but are unlikely to have formed in globular clusters However, the combination of mass ratio, component masses, and the inferred merger rate for this event challenges all current models of the formation and mass distribution of compact-object binaries

913 citations


Journal ArticleDOI
R. Abbott1, T. D. Abbott2, Sheelu Abraham3, Fausto Acernese4  +1332 moreInstitutions (150)
TL;DR: It is inferred that the primary black hole mass lies within the gap produced by (pulsational) pair-instability supernova processes, with only a 0.32% probability of being below 65 M⊙, which can be considered an intermediate mass black hole (IMBH).
Abstract: On May 21, 2019 at 03:02:29 UTC Advanced LIGO and Advanced Virgo observed a short duration gravitational-wave signal, GW190521, with a three-detector network signal-to-noise ratio of 14.7, and an estimated false-alarm rate of 1 in 4900 yr using a search sensitive to generic transients. If GW190521 is from a quasicircular binary inspiral, then the detected signal is consistent with the merger of two black holes with masses of 85_{-14}^{+21} M_{⊙} and 66_{-18}^{+17} M_{⊙} (90% credible intervals). We infer that the primary black hole mass lies within the gap produced by (pulsational) pair-instability supernova processes, with only a 0.32% probability of being below 65 M_{⊙}. We calculate the mass of the remnant to be 142_{-16}^{+28} M_{⊙}, which can be considered an intermediate mass black hole (IMBH). The luminosity distance of the source is 5.3_{-2.6}^{+2.4} Gpc, corresponding to a redshift of 0.82_{-0.34}^{+0.28}. The inferred rate of mergers similar to GW190521 is 0.13_{-0.11}^{+0.30} Gpc^{-3} yr^{-1}.

876 citations


Journal ArticleDOI
T. Aoyama1, Nils Asmussen2, M. Benayoun3, Johan Bijnens4  +146 moreInstitutions (64)
TL;DR: The current status of the Standard Model calculation of the anomalous magnetic moment of the muon is reviewed in this paper, where the authors present a detailed account of recent efforts to improve the calculation of these two contributions with either a data-driven, dispersive approach, or a first-principle, lattice approach.

801 citations


Journal ArticleDOI
T. Aoyama1, Nils Asmussen2, M. Benayoun3, Johan Bijnens4  +146 moreInstitutions (64)
TL;DR: The current status of the Standard Model calculation of the anomalous magnetic moment of the muon has been reviewed in this paper, where the authors present a detailed account of recent efforts to improve the calculation of these two contributions with either a data-driven, dispersive approach, or a first-principle, lattice-QCD approach.
Abstract: We review the present status of the Standard Model calculation of the anomalous magnetic moment of the muon. This is performed in a perturbative expansion in the fine-structure constant $\alpha$ and is broken down into pure QED, electroweak, and hadronic contributions. The pure QED contribution is by far the largest and has been evaluated up to and including $\mathcal{O}(\alpha^5)$ with negligible numerical uncertainty. The electroweak contribution is suppressed by $(m_\mu/M_W)^2$ and only shows up at the level of the seventh significant digit. It has been evaluated up to two loops and is known to better than one percent. Hadronic contributions are the most difficult to calculate and are responsible for almost all of the theoretical uncertainty. The leading hadronic contribution appears at $\mathcal{O}(\alpha^2)$ and is due to hadronic vacuum polarization, whereas at $\mathcal{O}(\alpha^3)$ the hadronic light-by-light scattering contribution appears. Given the low characteristic scale of this observable, these contributions have to be calculated with nonperturbative methods, in particular, dispersion relations and the lattice approach to QCD. The largest part of this review is dedicated to a detailed account of recent efforts to improve the calculation of these two contributions with either a data-driven, dispersive approach, or a first-principle, lattice-QCD approach. The final result reads $a_\mu^\text{SM}=116\,591\,810(43)\times 10^{-11}$ and is smaller than the Brookhaven measurement by 3.7$\sigma$. The experimental uncertainty will soon be reduced by up to a factor four by the new experiment currently running at Fermilab, and also by the future J-PARC experiment. This and the prospects to further reduce the theoretical uncertainty in the near future-which are also discussed here-make this quantity one of the most promising places to look for evidence of new physics.

420 citations


Journal ArticleDOI
Richard J. Abbott1, T. D. Abbott2, Sheelu Abraham3, Fausto Acernese4  +1329 moreInstitutions (150)
TL;DR: The GW190521 signal is consistent with a binary black hole (BBH) merger source at redshift 0.13-0.30 Gpc-3 yr-1.8 as discussed by the authors.
Abstract: The gravitational-wave signal GW190521 is consistent with a binary black hole (BBH) merger source at redshift 0.8 with unusually high component masses, 85-14+21 M o˙ and 66-18+17 M o˙, compared to previously reported events, and shows mild evidence for spin-induced orbital precession. The primary falls in the mass gap predicted by (pulsational) pair-instability supernova theory, in the approximate range 65-120 M o˙. The probability that at least one of the black holes in GW190521 is in that range is 99.0%. The final mass of the merger (142-16+28 M o˙) classifies it as an intermediate-mass black hole. Under the assumption of a quasi-circular BBH coalescence, we detail the physical properties of GW190521's source binary and its post-merger remnant, including component masses and spin vectors. Three different waveform models, as well as direct comparison to numerical solutions of general relativity, yield consistent estimates of these properties. Tests of strong-field general relativity targeting the merger-ringdown stages of the coalescence indicate consistency of the observed signal with theoretical predictions. We estimate the merger rate of similar systems to be 0.13-0.11+0.30 Gpc-3 yr-1. We discuss the astrophysical implications of GW190521 for stellar collapse and for the possible formation of black holes in the pair-instability mass gap through various channels: via (multiple) stellar coalescences, or via hierarchical mergers of lower-mass black holes in star clusters or in active galactic nuclei. We find it to be unlikely that GW190521 is a strongly lensed signal of a lower-mass black hole binary merger. We also discuss more exotic possible sources for GW190521, including a highly eccentric black hole binary, or a primordial black hole binary.

347 citations


Journal ArticleDOI
TL;DR: Physiological modelling of first-pass metabolism using data from a robotic instrument that fluidically links relevant organ chips predicts human pharmacokinetic parameters for orally administered nicotine and for intravenously injected cisplatin.
Abstract: Analyses of drug pharmacokinetics (PKs) and pharmacodynamics (PDs) performed in animals are often not predictive of drug PKs and PDs in humans, and in vitro PK and PD modelling does not provide quantitative PK parameters. Here, we show that physiological PK modelling of first-pass drug absorption, metabolism and excretion in humans-using computationally scaled data from multiple fluidically linked two-channel organ chips-predicts PK parameters for orally administered nicotine (using gut, liver and kidney chips) and for intravenously injected cisplatin (using coupled bone marrow, liver and kidney chips). The chips are linked through sequential robotic liquid transfers of a common blood substitute by their endothelium-lined channels (as reported by Novak et al. in an associated Article) and share an arteriovenous fluid-mixing reservoir. We also show that predictions of cisplatin PDs match previously reported patient data. The quantitative in-vitro-to-in-vivo translation of PK and PD parameters and the prediction of drug absorption, distribution, metabolism, excretion and toxicity through fluidically coupled organ chips may improve the design of drug-administration regimens for phase-I clinical trials.

243 citations


Journal ArticleDOI
TL;DR: In this article, Wiedenhofer et al. evaluate empirical studies of decoupling related to final/useful energy, exergy, use of material resources, as well as CO2 and total GHG emissions.
Abstract: Strategies toward ambitious climate targets usually rely on the concept of "decoupling"; that is, they aim at promoting economic growth while reducing the use of natural resources and GHG emissions. GDP growth coinciding with absolute reductions in emissions or resource use is denoted as "absolute decoupling", as opposed to "relative decoupling", where resource use or emissions increase less so than does GDP. Based on the bibliometric mapping in part I (Wiedenhofer et al., this issue), we synthesize the evidence emerging from the selected 835 peer-reviewed articles. We evaluate empirical studies of decoupling related to final/useful energy, exergy, use of material resources, as well as CO2 and total GHG emissions. We find that relative decoupling is frequent for material use as well as GHG and CO2 emissions but not for useful exergy, a quality-based measure of energy use. Primary energy can be decoupled from GDP largely to the extent to which the conversion of primary energy to useful exergy is improved. Examples of absolute long-term decoupling are rare, but recently some industrialized countries have decoupled GDP from both production- and, weaklier, consumption-based CO2 emissions. We analyze policies or strategies in the decoupling literature by classifying them into three groups: (1) Green growth, if sufficient reductions of resource use or emissions were deemed possible without altering the growth trajectory. (2) Degrowth, if reductions of resource use or emissions were given priority over GDP growth. (3) Others, e.g. if the role of energy for GDP growth was analyzed without reference to climate change mitigation. We conclude that large rapid absolute reductions of resource use and GHG emissions cannot be achieved through observed decoupling rates, hence decoupling needs to be complemented by sufficiency-oriented strategies and strict enforcement of absolute reduction targets. More research is needed on interdependencies between wellbeing, resources and emissions.

230 citations


Journal ArticleDOI
TL;DR: In this paper, a run-of-river hydropower plant located in a river of the Tagus basin characterised by a pluvial winter flow regime was investigated.

221 citations


Journal ArticleDOI
TL;DR: In this article, the authors provide an overview of the current state and future trends of the offshore wind farms worldwide along with the technological challenges, especially the wind farm layout and the main components.

212 citations


Journal ArticleDOI
TL;DR: In this article, the authors present updated clinical practice guidelines for hereditary diffuse gastric cancer (HDGC) from the International Gastric Cancer Linkage Consortium (IGCLC), which recognise the emerging evidence of variability in Gastric cancer risk between families with HDGC, and increased experience of managing long-term sequelae of total gastrectomy in young patients.
Abstract: Hereditary diffuse gastric cancer (HDGC) is an autosomal dominant cancer syndrome that is characterised by a high prevalence of diffuse gastric cancer and lobular breast cancer. It is largely caused by inactivating germline mutations in the tumour suppressor gene CDH1, although pathogenic variants in CTNNA1 occur in a minority of families with HDGC. In this Policy Review, we present updated clinical practice guidelines for HDGC from the International Gastric Cancer Linkage Consortium (IGCLC), which recognise the emerging evidence of variability in gastric cancer risk between families with HDGC, the growing capability of endoscopic and histological surveillance in HDGC, and increased experience of managing long-term sequelae of total gastrectomy in young patients. To redress the balance between the accessibility, cost, and acceptance of genetic testing and the increased identification of pathogenic variant carriers, the HDGC genetic testing criteria have been relaxed, mainly through less restrictive age limits. Prophylactic total gastrectomy remains the recommended option for gastric cancer risk management in pathogenic CDH1 variant carriers. However, there is increasing confidence from the IGCLC that endoscopic surveillance in expert centres can be safely offered to patients who wish to postpone surgery, or to those whose risk of developing gastric cancer is not well defined.

208 citations


Journal ArticleDOI
12 Nov 2020-Foods
TL;DR: Investigation of nanofiltration for the recovery of phenolic compounds, with bioactive capacity like antioxidant, from red grape pomace extract finds the CA400-22 exhibited the highest permeate flux, the lowest rejection coefficients towards the reference solutes and the best performance in terms of separation between sugars and Phenolic compounds.
Abstract: The winemaking process generates a large amount of residues such as vine shots, stalks, grape pomace, and wine lees, which were only recently considered for exploitation of their valuable compounds. The purpose of this work was to investigate the performance of nanofiltration for the recovery of phenolic compounds, with bioactive capacity like antioxidant, from red grape pomace extract. Four membranes were compared in this study—three cellulose acetate (CA series: lab-prepared by phase inversion) and one commercial (NF90). All membranes were characterized for their hydraulic permeability and rejection coefficients to reference solutes like saccharose, glucose, raffinose, polyethylene glycol, sodium chloride, and sodium sulfate. Permeation flowrates and rejection coefficients towards total phenolics content, antioxidant activity, proanthocyanidins, glucose and fructose were measured in the nanofiltration of grape pomace extract using selected operating conditions. Among the investigated membranes, the CA400-22 exhibited the highest permeate flux (50.58 L/m2 h at 20 bar and 25 °C), low fouling index (of about 23%), the lowest rejection coefficients towards the reference solutes and the best performance in terms of separation between sugars and phenolic compounds. Indeed, the observed rejections for glucose and fructose were 19% and 12%, respectively. On the other hand, total phenolics content and proanthocyanidins were rejected for 73% and 92%, respectively.

Journal ArticleDOI
TL;DR: An 'interrogator' that employs liquid-handling robotics, custom software and an integrated mobile microscope for the automated culture, perfusion, medium addition, fluidic linking, sample collection and in situ microscopy imaging of up to ten organ chips inside a standard tissue-culture incubator is described.
Abstract: Organ chips can recapitulate organ-level (patho)physiology, yet pharmacokinetic and pharmacodynamic analyses require multi-organ systems linked by vascular perfusion. Here, we describe an 'interrogator' that employs liquid-handling robotics, custom software and an integrated mobile microscope for the automated culture, perfusion, medium addition, fluidic linking, sample collection and in situ microscopy imaging of up to ten organ chips inside a standard tissue-culture incubator. The robotic interrogator maintained the viability and organ-specific functions of eight vascularized, two-channel organ chips (intestine, liver, kidney, heart, lung, skin, blood-brain barrier and brain) for 3 weeks in culture when intermittently fluidically coupled via a common blood substitute through their reservoirs of medium and endothelium-lined vascular channels. We used the robotic interrogator and a physiological multicompartmental reduced-order model of the experimental system to quantitatively predict the distribution of an inulin tracer perfused through the multi-organ human-body-on-chips. The automated culture system enables the imaging of cells in the organ chips and the repeated sampling of both the vascular and interstitial compartments without compromising fluidic coupling.

Journal ArticleDOI
Georges Aad1, Brad Abbott2, Dale Charles Abbott3, A. Abed Abud4  +2962 moreInstitutions (199)
TL;DR: A search for heavy neutral Higgs bosons is performed using the LHC Run 2 data, corresponding to an integrated luminosity of 139 fb^{-1} of proton-proton collisions at sqrt[s]=13‬TeV recorded with the ATLAS detector.
Abstract: A search for heavy neutral Higgs bosons is performed using the LHC Run 2 data, corresponding to an integrated luminosity of 139 fb^{-1} of proton-proton collisions at sqrt[s]=13 TeV recorded with the ATLAS detector. The search for heavy resonances is performed over the mass range 0.2-2.5 TeV for the τ^{+}τ^{-} decay with at least one τ-lepton decaying into final states with hadrons. The data are in good agreement with the background prediction of the standard model. In the M_{h}^{125} scenario of the minimal supersymmetric standard model, values of tanβ>8 and tanβ>21 are excluded at the 95% confidence level for neutral Higgs boson masses of 1.0 and 1.5 TeV, respectively, where tanβ is the ratio of the vacuum expectation values of the two Higgs doublets.

Journal ArticleDOI
TL;DR: How AR has been applied, which industries are most interested in the technology, how the technology has been developed to meet industry needs, as well as the benefits and challenges of AR are described.

Journal ArticleDOI
TL;DR: A vascularized human bone-marrow-on-a-chip improves the maintenance of patient-derived CD34 + cells, and recapitulates clinically relevant aspects of bone marrow injury as well as key haematopoietic defects of patients with a rare genetic disorder.
Abstract: The inaccessibility of living bone marrow (BM) hampers the study of its pathophysiology under myelotoxic stress induced by drugs, radiation or genetic mutations. Here, we show that a vascularized human BM-on-a-chip (BM chip) supports the differentiation and maturation of multiple blood cell lineages over 4 weeks while improving CD34+ cell maintenance, and that it recapitulates aspects of BM injury, including myeloerythroid toxicity after clinically relevant exposures to chemotherapeutic drugs and ionizing radiation, as well as BM recovery after drug-induced myelosuppression. The chip comprises a fluidic channel filled with a fibrin gel in which CD34+ cells and BM-derived stromal cells are co-cultured, a parallel channel lined by human vascular endothelium and perfused with culture medium, and a porous membrane separating the two channels. We also show that BM chips containing cells from patients with the rare genetic disorder Shwachman–Diamond syndrome reproduced key haematopoietic defects and led to the discovery of a neutrophil maturation abnormality. As an in vitro model of haematopoietic dysfunction, the BM chip may serve as a human-specific alternative to animal testing for the study of BM pathophysiology. A vascularized human bone-marrow-on-a-chip improves the maintenance of patient-derived CD34+ cells, and recapitulates clinically relevant aspects of bone marrow injury as well as key haematopoietic defects of patients with a rare genetic disorder.

Journal ArticleDOI
TL;DR: This work reviews existing case studies of traditional constructed wetlands and new integrated technologies for greywater treatment and reuse, with a specific focus on their treatment performance as a function of hydraulic operating parameters, to understand if the application of NBS can represent a valid alternative to conventional treatment technologies.

Journal ArticleDOI
TL;DR: YEASTRACT+ tools were upgraded to enable predictions based on orthologous regulatory associations described for other yeast species, including two new tools for cross-species transcription regulation comparison, based on multi-species promoter and TF regulatory network analyses.
Abstract: The YEASTRACT+ information system (http://YEASTRACT-PLUS.org/) is a wide-scope tool for the analysis and prediction of transcription regulatory associations at the gene and genomic levels in yeasts of biotechnological or human health relevance. YEASTRACT+ is a new portal that integrates the previously existing YEASTRACT (http://www.yeastract.com/) and PathoYeastract (http://pathoyeastract.org/) databases and introduces the NCYeastract (Non-Conventional Yeastract) database (http://ncyeastract.org/), focused on the so-called non-conventional yeasts. The information in the YEASTRACT database, focused on Saccharomyces cerevisiae, was updated. PathoYeastract was extended to include two additional pathogenic yeast species: Candida parapsilosis and Candida tropicalis. Furthermore, the NCYeastract database was created, including five biotechnologically relevant yeast species: Zygosaccharomyces baillii, Kluyveromyces lactis, Kluyveromyces marxianus, Yarrowia lipolytica and Komagataella phaffii. The YEASTRACT+ portal gathers 289 706 unique documented regulatory associations between transcription factors (TF) and target genes and 420 DNA binding sites, considering 247 TFs from 10 yeast species. YEASTRACT+ continues to make available tools for the prediction of the TFs involved in the regulation of gene/genomic expression. In this release, these tools were upgraded to enable predictions based on orthologous regulatory associations described for other yeast species, including two new tools for cross-species transcription regulation comparison, based on multi-species promoter and TF regulatory network analyses.

Journal ArticleDOI
TL;DR: Three Dimensional printing can be used to generate multilayered vascularized human skin grafts that can potentially overcome the limitations of graft survival observed in current avascular skin substitutes.
Abstract: Multilayered skin substitutes comprising allogeneic cells have been tested for the treatment of nonhealing cutaneous ulcers. However, such nonnative skin grafts fail to permanently engraft because they lack dermal vascular networks important for integration with the host tissue. In this study, we describe the fabrication of an implantable multilayered vascularized bioengineered skin graft using 3D bioprinting. The graft is formed using one bioink containing human foreskin dermal fibroblasts (FBs), human endothelial cells (ECs) derived from cord blood human endothelial colony-forming cells (HECFCs), and human placental pericytes (PCs) suspended in rat tail type I collagen to form a dermis followed by printing with a second bioink containing human foreskin keratinocytes (KCs) to form an epidermis. In vitro, KCs replicate and mature to form a multilayered barrier, while the ECs and PCs self-assemble into interconnected microvascular networks. The PCs in the dermal bioink associate with EC-lined vascular structures and appear to improve KC maturation. When these 3D printed grafts are implanted on the dorsum of immunodeficient mice, the human EC-lined structures inosculate with mouse microvessels arising from the wound bed and become perfused within 4 weeks after implantation. The presence of PCs in the printed dermis enhances the invasion of the graft by host microvessels and the formation of an epidermal rete. Impact Statement Three Dimensional printing can be used to generate multilayered vascularized human skin grafts that can potentially overcome the limitations of graft survival observed in current avascular skin substitutes. Inclusion of human pericytes in the dermal bioink appears to improve both dermal and epidermal maturation.

Journal ArticleDOI
TL;DR: In this article, a brief review on the different approaches available, focusing on the FAO56 framework for coping with the effects of soil salinity on crop ET and yields, is presented.

Proceedings ArticleDOI
14 Jun 2020
TL;DR: 3DRegNet as discussed by the authors is a novel deep learning architecture for the registration of 3D scans, which uses a set of point correspondences to address the following two challenges: (i) classification of the point correspondence into inliers/outliers, and (ii) regression of the motion parameters that align the scans into a common reference frame.
Abstract: We present 3DRegNet, a novel deep learning architecture for the registration of 3D scans. Given a set of 3D point correspondences, we build a deep neural network to address the following two challenges: (i) classification of the point correspondences into inliers/outliers, and (ii) regression of the motion parameters that align the scans into a common reference frame. With regard to regression, we present two alternative approaches: (i) a Deep Neural Network (DNN) registration and (ii) a Procrustes approach using SVD to estimate the transformation. Our correspondence-based approach achieves a higher speedup compared to competing baselines. We further propose the use of a refinement network, which consists of a smaller 3DRegNet as a refinement to improve the accuracy of the registration. Extensive experiments on two challenging datasets demonstrate that we outperform other methods and achieve state-of-the-art results. The code is available.

Journal ArticleDOI
TL;DR: The computed results in this paper are more conformity with statistical data for that the error of predicted failure rate of this study is 4.5%, which can be compared with that of 13% concluded by fault tree analysis.

Journal ArticleDOI
TL;DR: In this paper, a general formalism for multiple moduli fields and their associated modular symmetries was developed, and applied to an example based on three moduli field with finite modular symmetry, where the symmetry is broken by two bitriplet scalars to the diagonal subgroup.
Abstract: We develop a general formalism for multiple moduli and their associated modular symmetries. We apply this formalism to an example based on three moduli with finite modular symmetries ${S}_{4}^{A}$, ${S}_{4}^{B}$, and ${S}_{4}^{C}$, associated with two right-handed neutrinos and the charged lepton sector, respectively. The symmetry is broken by two bitriplet scalars to the diagonal ${S}_{4}$ subgroup. The low energy effective theory involves the three independent moduli fields ${\ensuremath{\tau}}_{A}$, ${\ensuremath{\tau}}_{B}$, and ${\ensuremath{\tau}}_{C}$, which preserve the residual modular subgroups ${Z}_{3}^{A}$, ${Z}_{2}^{B}$, and ${Z}_{3}^{C}$, in their respective sectors, leading to trimaximal ${\mathrm{TM}}_{1}$ lepton mixing, consistent with current data, without flavons.

Journal ArticleDOI
TL;DR: The paper aims to uncover a number of characteristics of the FRAM research, both in terms of the method’s application and of the authors contributing to its development, as well as proposing potential future research directions.

Journal ArticleDOI
TL;DR: An improved and sound taxonomy for a set of key CD terms, namely, parametric, generative, and algorithmic design is proposed, based on an extensive literature review from which different definitions by various authors were collected, analyzed, and compared.
Abstract: Computation-based approaches in design have emerged in the last decades and rapidly became popular among architects and other designers. Design professionals and researchers adopted different terminologies to address these approaches. However, some terms are used ambiguously and inconsistently, and different terms are commonly used to express the same concept. This paper discusses computational design (CD) and proposes an improved and sound taxonomy for a set of key CD terms, namely, parametric, generative, and algorithmic design, based on an extensive literature review from which different definitions by various authors were collected, analyzed, and compared.

Journal ArticleDOI
TL;DR: In this paper, the authors present a mathematical tool for distinguishing open quantum systems that are chaotic from those that are exactly solvable, filling an important gap in understanding dissipation and decoherence in scenarios relevant to quantum-based technologies.
Abstract: Mathematical tools for distinguishing open quantum systems that are chaotic from those that are exactly solvable fill an important gap in understanding dissipation and decoherence in scenarios relevant to quantum-based technologies.

Journal ArticleDOI
TL;DR: In this article, the authors present the state of the practice of asphalt rubber (AR) mixes in pavement technology, presenting some benefits and challenges of these paving materials, as well as a discussion on the performance of AR solutions and on life-cycle assessment and emissions, energy consumption and costs, contributing to circular economy.

Journal ArticleDOI
TL;DR: In this article, a computerized microtomograph was used to evaluate closed porosity, volume of limestone aggregate fraction and volume of mortar of the multi-recycled aggregate concrete, in order to answer the question: how many times it is possible to recycle concrete?

Journal ArticleDOI
TL;DR: New magnetic iron oxide (Fe3O4)/activated charcoal (AC)/β-cyclodextrin (CD)/sodium alginate (Alg) polymer nanocomposite materials were prepared by direct mixing of the polymer matrix with the nanofillers and showed excellent regeneration capacity.
Abstract: New magnetic iron oxide (Fe3O4)/activated charcoal (AC)/β-cyclodextrin (CD)/sodium alginate (Alg) polymer nanocomposite materials were prepared by direct mixing of the polymer matrix with the nanofillers. The obtained materials were utilized as nano-adsorbents for the elimination of methylene blue (MB), a hazardous water-soluble cationic dye, from aqueous solutions, and showed excellent regeneration capacity. The formation of the nanocomposites was followed by high-resolution transmission electron microscopy (HRTEM), scanning electron microscopy (SEM) equipped with energy dispersive X-ray spectrometry (EDX), Fourier-transform infrared spectroscopy (FTIR), vibrating sample magnetometer (VSM), X-ray diffraction (XRD) and adsorption of N2 at -196 °C. The rate of adsorption was investigated varying several factors, namely contact time, pH, amount of adsorbent and MB concentration on the adsorption process. Studies dealing with equilibrium and kinetics were carried out in batch conditions. The obtained results indicated that the removal rate of MB was 99.53% in 90 min. Langmuir's isotherm fitted better to the equilibrium data of MB. Fe3O4/AC/CD/Alg polymer beads shows amazing adsorption capacities in the elimination of cationic dyes (2.079 mg/g for polymer gel beads and 10.63 mg g-1 for dry powder beads), in comparison to other adsorbent materials. The obtained adsorbent is spherical with hydrophobic cross-linked surface properties that enable an easy recovery without any significant weight loss of in the adsorbent used.

Proceedings ArticleDOI
01 May 2020
TL;DR: This paper presents a security analysis of popular TrustZone-assisted TEE systems (targeting Cortex-A processors) developed by Qualcomm, Trustonic, Huawei, Nvidia, and Linaro, and identified several critical vulnerabilities across existing systems.
Abstract: Hundreds of millions of mobile devices worldwide rely on Trusted Execution Environments (TEEs) built with Arm TrustZone for the protection of security-critical applications (e.g., DRM) and operating system (OS) components (e.g., Android keystore). TEEs are often assumed to be highly secure; however, over the past years, TEEs have been successfully attacked multiple times, with highly damaging impact across various platforms. Unfortunately, these attacks have been possible by the presence of security flaws in TEE systems. In this paper, we aim to understand which types of vulnerabilities and limitations affect existing TrustZone-assisted TEE systems, what are the main challenges to build them correctly, and what contributions can be borrowed from the research community to overcome them. To this end, we present a security analysis of popular TrustZone-assisted TEE systems (targeting Cortex-A processors) developed by Qualcomm, Trustonic, Huawei, Nvidia, and Linaro. By studying publicly documented exploits and vulnerabilities as well as by reverse engineering the TEE firmware, we identified several critical vulnerabilities across existing systems which makes it legitimate to raise reasonable concerns about the security of commercial TEE implementations.

Journal ArticleDOI
TL;DR: A nonlocal sparse tensor factorization approach for the semiblind fusion of HSI and MSI, called the NLSTF_SMBF, which is blind with respect to the point spread function (PSF) of the hyperspectral sensor and copes with spatially variant PSFs.
Abstract: Combining a high-spatial-resolution multispectral image (HR-MSI) with a low-spatial-resolution hyperspectral image (LR-HSI) has become a common way to enhance the spatial resolution of the HSI. The existing state-of-the-art LR-HSI and HR-MSI fusion methods are mostly based on the matrix factorization, where the matrix data representation may be hard to fully make use of the inherent structures of 3-D HSI. We propose a nonlocal sparse tensor factorization approach, called the NLSTF_SMBF, for the semiblind fusion of HSI and MSI. The proposed method decomposes the HSI into smaller full-band patches (FBPs), which, in turn, are factored as dictionaries of the three HSI modes and a sparse core tensor. This decomposition allows to solve the fusion problem as estimating a sparse core tensor and three dictionaries for each FBP. Similar FBPs are clustered together, and they are assumed to share the same dictionaries to make use of the nonlocal self-similarities of the HSI. For each group, we learn the dictionaries from the observed HR-MSI and LR-HSI. The corresponding sparse core tensor of each FBP is computed via tensor sparse coding. Two distinctive features of NLSTF_SMBF are that: 1) it is blind with respect to the point spread function (PSF) of the hyperspectral sensor and 2) it copes with spatially variant PSFs. The experimental results provide the evidence of the advantages of the NLSTF_SMBF method over the existing state-of-the-art methods, namely, in semiblind scenarios.