Institution
Amazon.com
Company•Seattle, Washington, United States•
About: Amazon.com is a company organization based out in Seattle, Washington, United States. It is known for research contribution in the topics: Computer science & Service (business). The organization has 13363 authors who have published 17317 publications receiving 266589 citations.
Topics: Computer science, Service (business), Service provider, Context (language use), Virtual machine
Papers published on a yearly basis
Papers
More filters
•
12 Oct 2001TL;DR: A hybrid machine/human computing arrangement as discussed by the authors involves humans to assist a computer to solve particular tasks, allowing the computer to be more efficient in solving the tasks more efficiently, such as image or speech comparison.
Abstract: A hybrid machine/human computing arrangement which advantageously involves humans to assist a computer to solve particular tasks, allowing the computer to solve the tasks more efficiently. In one embodiment, a computer system decomposes a task, such as, for example, image or speech comparison, into subtasks for human performance, and requests the performances. The computer system programmatically conveys the request to a central coordinating server of the hybrid machine/human computing arrangement, which in turn dispatches the subtasks to personal computers operated by the humans. The humans perform the subtasks and provide the results back to the server, which receives the responses, and generates a result for the task based at least in part on the results of the human performances.
110 citations
••
TL;DR: This work constructs near-optimal heuristics for the reassignment for a large set of customer orders to minimize the total number of shipments and presents evidence of significant saving opportunities by testing the heuristic on order data from a major online retailer.
Abstract: When a customer orders online, an online retailer assigns the order to one or more of its warehouses and/or drop-shippers to minimize procurement and transportation costs based on the available current information. However, this assignment is necessarily myopic because it cannot account for any subsequent customer orders or future inventory replenishment. We examine the benefits of periodically reevaluating these real-time assignments. We construct near-optimal heuristics for the reassignment for a large set of customer orders to minimize the total number of shipments. Finally, we present evidence of significant saving opportunities by testing the heuristics on order data from a major online retailer.
110 citations
•
29 Mar 2006TL;DR: In this article, power management features of a reader device control an amount of electrical energy supplied to the reader device, including several power control categories that a user may select to affect the amount of power consumed by the reader.
Abstract: Power management features of a reader device control an amount of electrical energy supplied to the reader device. The power management features include several power control categories that a user may select to affect an amount of power consumed by the reader device. One power control category controls power consumption based on a genre of an electronic book being processed by the reader device. Another power control category controls power consumption based on a time of day an electronic book is being processed by the reader device. Power control instructions that control how much power a reader device consumes may also be generated based on a characteristic of a user's interaction with the device.
109 citations
•
TL;DR: The TabTransformer is a novel deep tabular data modeling architecture for supervised and semi-supervised learning built upon self-attention based Transformers that outperforms the state-of-the-art deep learning methods fortabular data by at least 1.0% on mean AUC, and matches the performance of tree-based ensemble models.
Abstract: We propose TabTransformer, a novel deep tabular data modeling architecture for supervised and semi-supervised learning. The TabTransformer is built upon self-attention based Transformers. The Transformer layers transform the embeddings of categorical features into robust contextual embeddings to achieve higher prediction accuracy. Through extensive experiments on fifteen publicly available datasets, we show that the TabTransformer outperforms the state-of-the-art deep learning methods for tabular data by at least 1.0% on mean AUC, and matches the performance of tree-based ensemble models. Furthermore, we demonstrate that the contextual embeddings learned from TabTransformer are highly robust against both missing and noisy data features, and provide better interpretability. Lastly, for the semi-supervised setting we develop an unsupervised pre-training procedure to learn data-driven contextual embeddings, resulting in an average 2.1% AUC lift over the state-of-the-art methods.
109 citations
••
01 Nov 2019TL;DR: This article proposed a debiasing algorithm based on residual fitting, which achieved significant gains over baseline models on two challenge test sets, while maintaining reasonable performance on the original test sets. This article proposed to learn a biased model that only uses features that are known to relate to dataset bias.
Abstract: Statistical natural language inference (NLI) models are susceptible to learning dataset bias: superficial cues that happen to associate with the label on a particular dataset, but are not useful in general, e.g., negation words indicate contradiction. As exposed by several recent challenge datasets, these models perform poorly when such association is absent, e.g., predicting that “I love dogs.” contradicts “I don’t love cats.”. Our goal is to design learning algorithms that guard against known dataset bias. We formalize the concept of dataset bias under the framework of distribution shift and present a simple debiasing algorithm based on residual fitting, which we call DRiFt. We first learn a biased model that only uses features that are known to relate to dataset bias. Then, we train a debiased model that fits to the residual of the biased model, focusing on examples that cannot be predicted well by biased features only. We use DRiFt to train three high-performing NLI models on two benchmark datasets, SNLI and MNLI. Our debiased models achieve significant gains over baseline models on two challenge test sets, while maintaining reasonable performance on the original test sets.
109 citations
Authors
Showing all 13498 results
Name | H-index | Papers | Citations |
---|---|---|---|
Jiawei Han | 168 | 1233 | 143427 |
Bernhard Schölkopf | 148 | 1092 | 149492 |
Christos Faloutsos | 127 | 789 | 77746 |
Alexander J. Smola | 122 | 434 | 110222 |
Rama Chellappa | 120 | 1031 | 62865 |
William F. Laurance | 118 | 470 | 56464 |
Andrew McCallum | 113 | 472 | 78240 |
Michael J. Black | 112 | 429 | 51810 |
David Heckerman | 109 | 483 | 62668 |
Larry S. Davis | 107 | 693 | 49714 |
Chris M. Wood | 102 | 795 | 43076 |
Pietro Perona | 102 | 414 | 94870 |
Guido W. Imbens | 97 | 352 | 64430 |
W. Bruce Croft | 97 | 426 | 39918 |
Chunhua Shen | 93 | 681 | 37468 |