Institution
Salesforce.com
About: Salesforce.com is a based out in . It is known for research contribution in the topics: User interface & Object (computer science). The organization has 2418 authors who have published 2775 publications receiving 63956 citations.
Topics: User interface, Object (computer science), Metadata, Cloud computing, Set (abstract data type)
Papers published on a yearly basis
Papers
More filters
•
TL;DR: Comparative Attention-Supervised Tuning (CAST) is proposed, which uses unsupervised saliency maps to intelligently sample crops, and to provide grounding supervision via a Grad-CAM attention loss to overcome contrastive SSL methods' limitations.
Abstract: Recent advances in self-supervised learning (SSL) have largely closed the gap with supervised ImageNet pretraining. Despite their success these methods have been primarily applied to unlabeled ImageNet images, and show marginal gains when trained on larger sets of uncurated images. We hypothesize that current SSL methods perform best on iconic images, and struggle on complex scene images with many objects. Analyzing contrastive SSL methods shows that they have poor visual grounding and receive poor supervisory signal when trained on scene images. We propose Contrastive Attention-Supervised Tuning(CAST) to overcome these limitations. CAST uses unsupervised saliency maps to intelligently sample crops, and to provide grounding supervision via a Grad-CAM attention loss. Experiments on COCO show that CAST significantly improves the features learned by SSL methods on scene images, and further experiments show that CAST-trained models are more robust to changes in backgrounds.
42 citations
•
04 Feb 2011TL;DR: In this paper, the authors provided mechanisms and methods for displaying data utilizing a selected source and visualization, which can enable enhanced data display, improved data display development, increased time savings, etc.
Abstract: In accordance with embodiments, there are provided mechanisms and methods for displaying data utilizing a selected source and visualization. These mechanisms and methods for displaying data utilizing a selected source and visualization can enable enhanced data display, improved data display development, increased time savings, etc.
42 citations
•
TL;DR: This article used CoS-E to train language models to automatically generate explanations that can be used during training and inference in a novel commonsense Auto-Generated Explanation (CAGE) framework.
Abstract: Deep learning models perform poorly on tasks that require commonsense reasoning, which often necessitates some form of world-knowledge or reasoning over information not immediately present in the input. We collect human explanations for commonsense reasoning in the form of natural language sequences and highlighted annotations in a new dataset called Common Sense Explanations (CoS-E). We use CoS-E to train language models to automatically generate explanations that can be used during training and inference in a novel Commonsense Auto-Generated Explanation (CAGE) framework. CAGE improves the state-of-the-art by 10% on the challenging CommonsenseQA task. We further study commonsense reasoning in DNNs using both human and auto-generated explanations including transfer to out-of-domain tasks. Empirical results indicate that we can effectively leverage language models for commonsense reasoning.
42 citations
•
11 Sep 2009TL;DR: In this article, the authors present a system for automating sharing data between subscribers of a multi-tenant database service, where users or customers associated with one organization that is a tenant of the system are enabled to share data objects such as leads, opportunities, accounts, contacts, cases, tasks and custom objects, (or other data objects) and other information with their business partners.
Abstract: Systems, methods, and apparatus for automating sharing data between subscribers of a multi-tenant database service. Users or customers associated with one organization that is a tenant of the multi-tenant database system are enabled to share data objects such as leads, opportunities, accounts, contacts, cases, tasks and custom objects, (or other data objects) and other information with their business partners (e.g., users or customers associated with a different organization that is a tenant) and get real-time updates on the shared data.
42 citations
•
TL;DR: This work introduces a simple and effective model-free method to learn from shaped distance-to-goal rewards on tasks where success depends on reaching a goal state and introduces an auxiliary distance-based reward based on pairs of rollouts to encourage diverse exploration.
Abstract: While using shaped rewards can be beneficial when solving sparse reward tasks, their successful application often requires careful engineering and is problem specific. For instance, in tasks where the agent must achieve some goal state, simple distance-to-goal reward shaping often fails, as it renders learning vulnerable to local optima. We introduce a simple and effective model-free method to learn from shaped distance-to-goal rewards on tasks where success depends on reaching a goal state. Our method introduces an auxiliary distance-based reward based on pairs of rollouts to encourage diverse exploration. This approach effectively prevents learning dynamics from stabilizing around local optima induced by the naive distance-to-goal reward shaping and enables policies to efficiently solve sparse reward tasks. Our augmented objective does not require any additional reward engineering or domain expertise to implement and converges to the original sparse objective as the agent learns to solve the task. We demonstrate that our method successfully solves a variety of hard-exploration tasks (including maze navigation and 3D construction in a Minecraft environment), where naive distance-based reward shaping otherwise fails, and intrinsic curiosity and reward relabeling strategies exhibit poor performance.
42 citations
Authors
Showing all 2418 results
Name | H-index | Papers | Citations |
---|---|---|---|
Philip S. Yu | 148 | 1914 | 107374 |
Michael R. Lyu | 89 | 696 | 33257 |
Silvio Savarese | 89 | 386 | 35975 |
Jiashi Feng | 77 | 426 | 21521 |
Richard Socher | 77 | 274 | 97703 |
Haibin Ling | 72 | 383 | 20858 |
Dragomir R. Radev | 69 | 288 | 20131 |
Irwin King | 67 | 476 | 19056 |
Steven C. H. Hoi | 66 | 375 | 15935 |
Xiaodan Liang | 61 | 318 | 14121 |
Caiming Xiong | 60 | 336 | 18037 |
Min-Yen Kan | 52 | 253 | 10207 |
Justin Yifu Lin | 48 | 302 | 13491 |
Hannaneh Hajishirzi | 42 | 181 | 7802 |
Larry S. Davis | 40 | 105 | 6960 |