scispace - formally typeset
Search or ask a question
Author

Sameep Mehta

Bio: Sameep Mehta is an academic researcher from IBM. The author has contributed to research in topics: Service (business) & Resource (project management). The author has an hindex of 22, co-authored 160 publications receiving 2093 citations. Previous affiliations of Sameep Mehta include Lady Hardinge Medical College & All India Institute of Medical Sciences.


Papers
More filters
Patent
Sreyash Kenkre1, Sameep Mehta1, Krishnasuri Narayanam1, Vinayaka Pandit1, Soujanya Soni1 
06 Feb 2012
TL;DR: In this paper, a description of a resource associated with a service of an entity can be captured, where the service can be associated with one or more resources, a constraint, and a demand.
Abstract: A description of a resource associated with a service of an entity can be captured. The service can be associated with one or more resources, a constraint, and a demand. The resource can be associated with one or more characteristics including a utility, a limited availability, and a consumption rate. The entity can be an organization or a system. An initial allocation problem associated with the resource can be formulated as a two phase problem. The first phase can be an optimization problem and the second phase can be a restricted allocation problem. The initial allocation problem can be associated with reconfiguring a previously established allocation of a baseline scenario. The optimization problem can be solved optimally or approximately to establish a favorable allocation. The favorable allocation can minimizes the reconfiguration cost of the reconfiguring. The baseline scenario can be a normal operation of the service.

18 citations

Proceedings ArticleDOI
30 Mar 2008
TL;DR: This work proposes and model a simple and comprehensive set of transformations that capture evolution of a single actor and interactions among multiple actors, and presents algorithms to rank each transformation and shows how ranking helps to infer important relationships between actors and stories in a corpus.
Abstract: The natural way to model a news corpus is as a directed graph where stories are linked to one another through a variety of relationships We formalize this notion by viewing each news story as a set of actors, and by viewing links between stories as transformations these actors go through We propose and model a simple and comprehensive set of transformations: create, merge, split, continue, and cease These transformations capture evolution of a single actor and interactions among multiple actors We present algorithms to rank each transformation and show how ranking helps us to infer important relationships between actors and stories in a corpus We demonstrate the effectiveness of our notions by experimenting on large news corpora

17 citations

Proceedings ArticleDOI
26 Oct 2010
TL;DR: A novel ranking technique that was developed in the context of an application that arose in a Service Delivery setting based on extension of eigen value methods and results on real-life, public domain datasets from the Internet Movie DataBase are presented.
Abstract: In this paper, we present a novel ranking technique that we developed in the context of an application that arose in a Service Delivery setting. We consider the problem of ranking agents of a service organization. The service agents typically need to interact with other service agents to accomplish the end goal of resolving customer requests. Their ranking needs to take into account two aspects: firstly, their importance in the network structure that arises as a result of their interactions, and secondly, the value generated by the interactions involving them. We highlight several other applications which have the common theme of ranking the participants of a value creation process based on the network structure of their interactions and the value generated by their interactions. We formally present the problem and describe the modeling technique which enables us to encode the value of interaction in the graph. Our ranking algorithm is based on extension of eigen value methods. We present experimental results on real-life, public domain datasets from the Internet Movie DataBase. This makes our experiments replicable and verifiable.

15 citations

Posted Content
TL;DR: An end-to-end system that can understand policies written in natural language, alert users to policy violations during data usage, and log each activity performed using the data in an immutable storage so that policy compliance or violation can be proven later is proposed.
Abstract: In consequential real-world applications, machine learning (ML) based systems are expected to provide fair and non-discriminatory decisions on candidates from groups defined by protected attributes such as gender and race. These expectations are set via policies or regulations governing data usage and decision criteria (sometimes explicitly calling out decisions by automated systems). Often, the data creator, the feature engineer, the author of the algorithm and the user of the results are different entities, making the task of ensuring fairness in an end-to-end ML pipeline challenging. Manually understanding the policies and ensuring fairness in opaque ML systems is time-consuming and error-prone, thus necessitating an end-to-end system that can: 1) understand policies written in natural language, 2) alert users to policy violations during data usage, and 3) log each activity performed using the data in an immutable storage so that policy compliance or violation can be proven later. We propose such a system to ensure that data owners and users are always in compliance with fairness policies.

14 citations

Proceedings Article
25 Jul 2015
TL;DR: Using data from the 2012 US presidential elections and the 2013 Philippines General elections, this work provides detailed experiments on methods that use granger causality to identify topics that were most "causal" for public opinion and which in turn give an interpretable insight into "elections topics" that weremost important.
Abstract: In recent times, social media has become a popular medium for many election campaigns. It not only allows candidates to reach out to a large section of the electorate, it is also a potent medium for people to express their opinion on the proposed policies and promises of candidates. Analyzing social media data is challenging as the text can be noisy, sparse and even multilingual. In addition, the information may not be completely trustworthy, particularly in the presence of propaganda, promotions and rumors. In this paper we describe our work for analyzing election campaigns using social media data. Using data from the 2012 US presidential elections and the 2013 Philippines General elections, we provide detailed experiments on our methods that use granger causality to identify topics that were most "causal" for public opinion and which in turn, give an interpretable insight into "elections topics" that were most important. Our system was deployed by the largest media organization in the Philippines during the 2013 General elections and using our work, the media house able to identify and report news stories much faster than competitors and reported higher TRP ratings during the election.

14 citations


Cited by
More filters
Journal ArticleDOI
09 Mar 2018-Science
TL;DR: A large-scale analysis of tweets reveals that false rumors spread further and faster than the truth, and false news was more novel than true news, which suggests that people were more likely to share novel information.
Abstract: We investigated the differential diffusion of all of the verified true and false news stories distributed on Twitter from 2006 to 2017. The data comprise ~126,000 stories tweeted by ~3 million people more than 4.5 million times. We classified news as true or false using information from six independent fact-checking organizations that exhibited 95 to 98% agreement on the classifications. Falsehood diffused significantly farther, faster, deeper, and more broadly than the truth in all categories of information, and the effects were more pronounced for false political news than for false news about terrorism, natural disasters, science, urban legends, or financial information. We found that false news was more novel than true news, which suggests that people were more likely to share novel information. Whereas false stories inspired fear, disgust, and surprise in replies, true stories inspired anticipation, sadness, joy, and trust. Contrary to conventional wisdom, robots accelerated the spread of true and false news at the same rate, implying that false news spreads more than the truth because humans, not robots, are more likely to spread it.

4,241 citations

01 Jan 2012

3,692 citations

21 Jan 2018
TL;DR: It is shown that the highest error involves images of dark-skinned women, while the most accurate result is for light-skinned men, in commercial API-based classifiers of gender from facial images, including IBM Watson Visual Recognition.
Abstract: The paper “Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification” by Joy Buolamwini and Timnit Gebru, that will be presented at the Conference on Fairness, Accountability, and Transparency (FAT*) in February 2018, evaluates three commercial API-based classifiers of gender from facial images, including IBM Watson Visual Recognition. The study finds these services to have recognition capabilities that are not balanced over genders and skin tones [1]. In particular, the authors show that the highest error involves images of dark-skinned women, while the most accurate result is for light-skinned men.

2,528 citations

Posted Content
TL;DR: This survey investigated different real-world applications that have shown biases in various ways, and created a taxonomy for fairness definitions that machine learning researchers have defined to avoid the existing bias in AI systems.
Abstract: With the widespread use of AI systems and applications in our everyday lives, it is important to take fairness issues into consideration while designing and engineering these types of systems. Such systems can be used in many sensitive environments to make important and life-changing decisions; thus, it is crucial to ensure that the decisions do not reflect discriminatory behavior toward certain groups or populations. We have recently seen work in machine learning, natural language processing, and deep learning that addresses such challenges in different subdomains. With the commercialization of these systems, researchers are becoming aware of the biases that these applications can contain and have attempted to address them. In this survey we investigated different real-world applications that have shown biases in various ways, and we listed different sources of biases that can affect AI applications. We then created a taxonomy for fairness definitions that machine learning researchers have defined in order to avoid the existing bias in AI systems. In addition to that, we examined different domains and subdomains in AI showing what researchers have observed with regard to unfair outcomes in the state-of-the-art methods and how they have tried to address them. There are still many future directions and solutions that can be taken to mitigate the problem of bias in AI systems. We are hoping that this survey will motivate researchers to tackle these issues in the near future by observing existing work in their respective fields.

1,571 citations