scispace - formally typeset
Search or ask a question

Showing papers by "Ross Anderson published in 2021"


Proceedings ArticleDOI
01 Sep 2021
TL;DR: Sponge examples as mentioned in this paper exploit carefully crafted sponge examples, which are inputs designed to maximize energy consumption and latency, to drive machine learning (ML) systems towards their worst-case performance.
Abstract: The high energy costs of neural network training and inference led to the use of acceleration hardware such as GPUs and TPUs. While such devices enable us to train large-scale neural networks in datacenters and deploy them on edge devices, their designers' focus so far is on average-case performance. In this work, we introduce a novel threat vector against neural networks whose energy consumption or decision latency are critical. We show how adversaries can exploit carefully-crafted sponge examples, which are inputs designed to maximise energy consumption and latency, to drive machine learning (ML) systems towards their worst-case performance. Sponge examples are, to our knowledge, the first denial-of-service attack against the ML components of such systems. We mount two variants of our sponge attack on a wide range of state-of-the-art neural network models, and find that language models are surprisingly vulnerable. Sponge examples frequently increase both latency and energy consumption of these models by a factor of 30×. Extensive experiments show that our new attack is effective across different hardware platforms (CPU, GPU and an ASIC simulator) on a wide range of different language tasks. On vision tasks, we show that sponge examples can be produced and a latency degradation observed, but the effect is less pronounced. To demonstrate the effectiveness of sponge examples in the real world, we mount an attack against Microsoft Azure's translator and show an increase of response time from 1ms to 6s (6000×). We conclude by proposing a defense strategy: shifting the analysis of energy consumption in hardware from an average-case to a worst-case perspective.

12 citations


Posted Content
TL;DR: This paper explore a large class of adversarial examples that can be used to attack text-based models in a black-box setting without making any human-perceptible visual modification to inputs.
Abstract: Several years of research have shown that machine-learning systems are vulnerable to adversarial examples, both in theory and in practice. Until now, such attacks have primarily targeted visual models, exploiting the gap between human and machine perception. Although text-based models have also been attacked with adversarial examples, such attacks struggled to preserve semantic meaning and indistinguishability. In this paper, we explore a large class of adversarial examples that can be used to attack text-based models in a black-box setting without making any human-perceptible visual modification to inputs. We use encoding-specific perturbations that are imperceptible to the human eye to manipulate the outputs of a wide range of Natural Language Processing (NLP) systems from neural machine-translation pipelines to web search engines. We find that with a single imperceptible encoding injection -- representing one invisible character, homoglyph, reordering, or deletion -- an attacker can significantly reduce the performance of vulnerable models, and with three injections most models can be functionally broken. Our attacks work against currently-deployed commercial systems, including those produced by Microsoft and Google, in addition to open source models published by Facebook and IBM. This novel series of attacks presents a significant threat to many language processing systems: an attacker can affect systems in a targeted manner without any assumptions about the underlying model. We conclude that text-based NLP systems require careful input sanitization, just like conventional applications, and that given such systems are now being deployed rapidly at scale, the urgent attention of architects and operators is required.

8 citations


Posted Content
TL;DR: In this article, the authors present a novel class of training-time attacks that require no changes to the underlying dataset or model architecture, but instead only change the order in which data are supplied to the model.
Abstract: Machine learning is vulnerable to a wide variety of attacks. It is now well understood that by changing the underlying data distribution, an adversary can poison the model trained with it or introduce backdoors. In this paper we present a novel class of training-time attacks that require no changes to the underlying dataset or model architecture, but instead only change the order in which data are supplied to the model. In particular, we find that the attacker can either prevent the model from learning, or poison it to learn behaviours specified by the attacker. Furthermore, we find that even a single adversarially-ordered epoch can be enough to slow down model learning, or even to reset all of the learning progress. Indeed, the attacks presented here are not specific to the model or dataset, but rather target the stochastic nature of modern learning procedures. We extensively evaluate our attacks on computer vision and natural language benchmarks to find that the adversary can disrupt model training and even introduce backdoors.

6 citations


Journal ArticleDOI
TL;DR: There is a high rate of severe ED, both following PFUI and remaining after posterior urethroplasty, and rates of ejaculatory difficulty and patient perceived changes in penile length and curvature underscore the complex nature of the impact of these injuries on sexual function beyond simple erectile function.
Abstract: Background To evaluate erectile and sexual function after pelvic fracture urethral injury (PFUI) by performing a retrospective review of a large multi-center database. We hypothesized that most men will have erectile dysfunction (ED) and poor sexual function following PFUI, which will remain after posterior urethroplasty. Methods Using the Trauma and Urologic Reconstructive Networks of Surgeons (TURNS) database, we identified PFUI patients undergoing posterior urethroplasty. We excluded patients with incomplete demographic, surgical and/or questionnaire data. Sexual Health Inventory of Men (SHIM), Male Sexual Health Questionnaire (MSHQ), and subjective changes in penile curvature were collected before urethroplasty surgery and at follow-up. We performed descriptive statistics for erectile and ejaculatory function using STATA v12. Results We identified 92 men meeting inclusion criteria; median age was 41.7 years and BMI was 26.5. The mechanism of injury was blunt in all patients, and average distraction defect length was 2.3 cm (SD 1.0 cm). In the 38 patients who completed both pre and post-operative SHIM questionnaires, the mean SHIM score was 10.5 (SD 7.0), with 63% having severe ED (SHIM <12). The median follow-up was 5.6 months and the mean post-operative SHIM was 9.3 (SD 6.5), with 68% having severe ED. The mean change in SHIM score was -1.18 (SD 6.29) with 6 (16%) patients reporting de novo ED (≥5 point decrease in score). Of the men with pre-operative MSHQ data, 46/74 (62.1%) had difficulty with ejaculation, 25/35 (71%) had change in penile length, and 6/33 (18%) reported penile curvature. In men with post-operative MSHQ, 19/44 (43%) expressed difficulty with ejaculation, 23/32 (72%) had change in penile length, and 9/33 (27%) reported penile curvature. Conclusions There is a high rate of severe ED, both following PFUI and remaining after posterior urethroplasty. Additionally, rates of ejaculatory difficulty and patient perceived changes in penile length and curvature underscore the complex nature of the impact of these injuries on sexual function beyond simple erectile function.

2 citations


Posted Content
TL;DR: In this article, Markpainting is used as a manipulation alarm that becomes visible in the event of inpainting and is transferable to models that have different architectures or were trained on different datasets.
Abstract: Inpainting is a learned interpolation technique that is based on generative modeling and used to populate masked or missing pieces in an image; it has wide applications in picture editing and retouching. Recently, inpainting started being used for watermark removal, raising concerns. In this paper we study how to manipulate it using our markpainting technique. First, we show how an image owner with access to an inpainting model can augment their image in such a way that any attempt to edit it using that model will add arbitrary visible information. We find that we can target multiple different models simultaneously with our technique. This can be designed to reconstitute a watermark if the editor had been trying to remove it. Second, we show that our markpainting technique is transferable to models that have different architectures or were trained on different datasets, so watermarks created using it are difficult for adversaries to remove. Markpainting is novel and can be used as a manipulation alarm that becomes visible in the event of inpainting.

2 citations


Proceedings Article
06 Dec 2021
TL;DR: In this paper, the authors present a novel class of training-time attacks that require no changes to the underlying dataset or model architecture, but instead only change the order in which data are supplied to the model.
Abstract: Machine learning is vulnerable to a wide variety of attacks. It is now well understood that by changing the underlying data distribution, an adversary can poison the model trained with it or introduce backdoors. In this paper we present a novel class of training-time attacks that require no changes to the underlying dataset or model architecture, but instead only change the order in which data are supplied to the model. In particular, we find that the attacker can either prevent the model from learning, or poison it to learn behaviours specified by the attacker. Furthermore, we find that even a single adversarially-ordered epoch can be enough to slow down model learning, or even to reset all of the learning progress. Indeed, the attacks presented here are not specific to the model or dataset, but rather target the stochastic nature of modern learning procedures. We extensively evaluate our attacks on computer vision and natural language benchmarks to find that the adversary can disrupt model training and even introduce backdoors.

1 citations


Posted Content
TL;DR: Trojan source attacks as mentioned in this paper exploit subtleties in text-encoding standards such as Unicode to produce source code whose tokens are logically encoded in a different order from the one in which they are displayed, leading to vulnerabilities that cannot be perceived directly by human code reviewers.
Abstract: We present a new type of attack in which source code is maliciously encoded so that it appears different to a compiler and to the human eye. This attack exploits subtleties in text-encoding standards such as Unicode to produce source code whose tokens are logically encoded in a different order from the one in which they are displayed, leading to vulnerabilities that cannot be perceived directly by human code reviewers. 'Trojan Source' attacks, as we call them, pose an immediate threat both to first-party software and of supply-chain compromise across the industry. We present working examples of Trojan-Source attacks in C, C++, C#, JavaScript, Java, Rust, Go, and Python. We propose definitive compiler-level defenses, and describe other mitigating controls that can be deployed in editors, repositories, and build pipelines while compilers are upgraded to block this attack.

Posted Content
TL;DR: In this article, the authors argue that client-side scanning does not guarantee efficacious crime prevention nor prevents surveillance and that client side scanning can fail, can be evaded, and can be abused.
Abstract: Our increasing reliance on digital technology for personal, economic, and government affairs has made it essential to secure the communications and devices of private citizens, businesses, and governments. This has led to pervasive use of cryptography across society. Despite its evident advantages, law enforcement and national security agencies have argued that the spread of cryptography has hindered access to evidence and intelligence. Some in industry and government now advocate a new technology to access targeted data: client-side scanning (CSS). Instead of weakening encryption or providing law enforcement with backdoor keys to decrypt communications, CSS would enable on-device analysis of data in the clear. If targeted information were detected, its existence and, potentially, its source, would be revealed to the agencies; otherwise, little or no information would leave the client device. Its proponents claim that CSS is a solution to the encryption versus public safety debate: it offers privacy -- in the sense of unimpeded end-to-end encryption -- and the ability to successfully investigate serious crime. In this report, we argue that CSS neither guarantees efficacious crime prevention nor prevents surveillance. Indeed, the effect is the opposite. CSS by its nature creates serious security and privacy risks for all society while the assistance it can provide for law enforcement is at best problematic. There are multiple ways in which client-side scanning can fail, can be evaded, and can be abused.