scispace - formally typeset
Search or ask a question

Showing papers by "My T. Thai published in 2019"


Proceedings ArticleDOI
07 Jul 2019
TL;DR: A new sharding paradigm is proposed, called OptChain, in which cross-shard transactions are minimized, resulting in almost twice faster confirmation time and throughput, and when combined with Omniledger sharding protocol, OptChain delivers a 6000 transactions per second throughput with 10.5s confirmation time.
Abstract: A major challenge in blockchain sharding protocols is that more than 95% transactions are cross-shard. Not only those cross-shard transactions degrade the system throughput but also double the confirmation time, and exhaust an already scarce network bandwidth. Are cross-shard transactions imminent for sharding schemes? In this paper, we propose a new sharding paradigm, called OptChain, in which cross-shard transactions are minimized, resulting in almost twice faster confirmation time and throughput. By treating transactions as a stream of nodes in an online graph, OptChain utilizes a lightweight and on-the-fly transaction placement method to group both related and soon-related transactions into the same shards. At the same time, OptChain maintains a temporal balance among shards to guarantee the high parallelism. Our comprehensive and large-scale simulation using Oversim P2P library confirms a significant boost in performance with up to 10 folds reduction in cross-shard transactions, more than twice reduction in confirmation time, and 50% increase in throughput. When combined with Omniledger sharding protocol, OptChain delivers a 6000 transactions per second throughput with 10.5s confirmation time.

76 citations


Proceedings ArticleDOI
07 Jul 2019
TL;DR: It is shown that Bitcoin is vulnerable to spatial, temporal, spatio-temporal, and logical partitioning attacks with an increased attack feasibility due to network dynamics.
Abstract: Bitcoin is the leading example of a blockchain application that facilitates peer-to-peer transactions without the need for a trusted intermediary. This paper considers possible attacks related to the decentralized network architecture of Bitcoin. We perform a data driven study of Bitcoin and present possible attacks based on spatial and temporal characteristics of its network. Towards that, we revisit the prior work, dedicated to the study of centralization of Bitcoin nodes over the Internet, through a fine-grained analysis of network distribution, and highlight the increasing centralization of the Bitcoin network over time. As a result, we show that Bitcoin is vulnerable to spatial, temporal, spatio-temporal, and logical partitioning attacks with an increased attack feasibility due to network dynamics. We verify our observations by simulating attack scenarios and the implications of each attack on the Bitcoin . We conclude with suggested countermeasures.

38 citations


Proceedings ArticleDOI
01 Jan 2019
TL;DR: Theoretical analysis and thorough evaluations show that the novel Heterogeneous Gaussian Mechanism notably improves the robustness of differentially private deep neural networks, compared with baseline approaches, under a variety of model attacks.
Abstract: In this paper, we propose a novel Heterogeneous Gaussian Mechanism (HGM) to preserve differential privacy in deep neural networks, with provable robustness against adversarial examples. We first relax the constraint of the privacy budget in the traditional Gaussian Mechanism from (0, 1] to (0, \infty), with a new bound of the noise scale to preserve differential privacy. The noise in our mechanism can be arbitrarily redistributed, offering a distinctive ability to address the trade-off between model utility and privacy loss. To derive provable robustness, our HGM is applied to inject Gaussian noise into the first hidden layer. Then, a tighter robustness bound is proposed. Theoretical analysis and thorough evaluations show that our mechanism notably improves the robustness of differentially private deep neural networks, compared with baseline approaches, under a variety of model attacks.

36 citations


Journal ArticleDOI
TL;DR: In this article, the authors propose TipTop, a cost-aware target viral marketing (CTVM) solution that scales to very large networks, such as Twitter, using integer programming.
Abstract: In this paper, we study the cost-aware target viral marketing (CTVM) problem, a generalization of influence maximization. CTVM asks for the most cost-effective users to influence the most relevant users. In contrast to the vast literature, we attempt to offer exact solutions. As the problem is NP-hard, thus, exact solutions are intractable, we propose TipTop, a $(1-\epsilon)$ -optimal solution for arbitrary $\epsilon >0$ that scales to very large networks, such as Twitter. At the heart of TipTop lies an innovative technique that reduces the number of samples as much as possible. This allows us to exactly solve CTVM on a much smaller space of generated samples using integer programming. Furthermore, TipTop lends a tool for researchers to benchmark their solutions against the optimal one in large-scale networks, which is currently not available.

32 citations


Journal ArticleDOI
TL;DR: This work aims to find the smallest set of nodes whose removal from a social network reduces the influence of misinformation greater than a given threshold, called the Targeted Misinformation Blocking problem, and shows that this problem is #P- hard under Linear Threshold and NP-hard under Independent Cascade diffusion models.
Abstract: Preventing misinformation spreading has recently become a critical topic due to an explosive growth of online social networks. Instead of focusing on blocking misinformation with a given budget as usually studied in the literatures, we aim to find the smallest set of nodes (minimize the budget) whose removal from a social network reduces the influence of misinformation (influence reduction) greater than a given threshold, called the Targeted Misinformation Blocking problem. We show that this problem is #P-hard under Linear Threshold and NP-hard under Independent Cascade diffusion models. We then propose several efficient algorithms, including approximation and heuristic algorithms to solve the problem. Experiments on real-world network topologies show the effectiveness and scalability of our algorithms that outperform other state-of-the-art methods.

22 citations


Proceedings ArticleDOI
27 Aug 2019
TL;DR: This work proposes to employ the self-attention mechanism in neural text modeling to achieve the semantic structure induction for rumor detection, and introduces a novel method to preserve the important information of the main news/posts in the final representations of the entire threads to further improve the performance for rumors detection.
Abstract: Fake news and rumors constitute a major problem in social networks recently. Due to the fast information propagation in social networks, it is inefficient to use human labor to detect suspicious news. Automatic rumor detection is thus necessary to prevent devastating effects of rumors on the individuals and society. Previous work has shown that in addition to the content of the news/posts and their contexts (i.e., replies), the relations or connections among those components are important to boost the rumor detection performance. In order to induce such relations between posts and contexts, the prior work has mainly relied on the inherent structures of the social networks (e.g., direct replies), ignoring the potential semantic connections between those objects. In this work, we demonstrate that such semantic relations are also helpful as they can reveal the implicit structures to better capture the patterns in the contexts for rumor detection. We propose to employ the self-attention mechanism in neural text modeling to achieve the semantic structure induction for this problem. In addition, we introduce a novel method to preserve the important information of the main news/posts in the final representations of the entire threads to further improve the performance for rumor detection. Our method matches the main post representations and the thread representations by ensuring that they predict the same latent labels in a multitask learning framework. The extensive experiments demonstrate the effectiveness of the proposed model for rumor detection, yielding the state-of-the-art performance on recent datasets for this problem.

22 citations


Posted Content
TL;DR: This work introduces the c-Eval metric and the corresponding framework to quantify the explainer's quality on feature-based explainers of machine learning image classifiers and conducts extensive experiments of explainers on three different datasets in order to support the adoption of c- Eval in evaluating explainers' performance.
Abstract: Due to high complexity of many modern machine learning models such as deep convolutional networks, understanding the cause of model's prediction is critical. Many explainers have been designed to give us more insights on the decision of complex classifiers. However, there is no common ground on evaluating the quality of different classification methods. Motivated by the needs for comprehensive evaluation, we introduce the c-Eval metric and the corresponding framework to quantify the explainer's quality on feature-based explainers of machine learning image classifiers. Given a prediction and the corresponding explanation on that prediction, c-Eval is the minimum-power perturbation that successfully alters the prediction while keeping the explanation's features unchanged. We also provide theoretical analysis linking the proposed parameter with the portion of predicted object covered by the explanation. Using a heuristic approach, we introduce the c-Eval plot, which not only displays a strong connection between c-Eval and explainers' quality, but also serves as a low-complexity approach of assessing explainers. We finally conduct extensive experiments of explainers on three different datasets in order to support the adoption of c-Eval in evaluating explainers' performance.

17 citations


Journal ArticleDOI
TL;DR: A general case of CIM problem, named Budgeted Competitive Influence Maximization ( BCIM ) problem, is studied, which considers CIM with budget and time constraints under condition of fair competition, and it is found that the objective function is neither submodular nor suppermodular.
Abstract: Competitive Influence Maximization ( CIM ) problem, which seeks a seed set nodes of a player or a company to propagate their product’s information while at the same time their competitors are conducting similar strategies, has been paid much attention recently due to its application in viral marketing. However, existing works neglect the fact that the limited budget and time constraints can play an important role in competitive influence strategy of each company. In addition, based on the the assumption that one of the competitors dominates in the competitive influence process, the majority of prior studies indicate that the competitive influence function (objective function) is monotone and submodular.This led to the fact that CIM can be approximated within a factor of 1 − 1 / e − ϵ by a Greedy algorithm combined with Monte Carlo simulation method. Unfortunately, in a more realistic scenario where there is fair competition among competitors, the objective function is no longer submodular. In this paper, we study a general case of CIM problem, named Budgeted Competitive Influence Maximization ( BCIM ) problem, which considers CIM with budget and time constraints under condition of fair competition. We found that the objective function is neither submodular nor suppermodular. Therefore, it cannot admit Greedy algorithm with approximation ratio of 1 − 1 / e . We propose Sandwich Approximation based on Polling-Based Approximation ( SPBA ), an approximation algorithm based on Sandwich framework and polling-based method. Our experiments on real social network datasets showed the effectiveness and scalability of our algorithm that outperformed other state-of-the-art methods. Specifically, our algorithm is scalable with million-scale networks in only 1.5 min.

14 citations


Journal ArticleDOI
TL;DR: A scenario-based linear MIP model is developed and an exact Markov chain-based algorithm is proposed to solve optimization problems of identifying critical nodes in coupled interdependent networks in the presence of deterministic or probabilistic cascading failure propagations.
Abstract: We consider optimization problems of identifying critical nodes in coupled interdependent networks, that is, choosing a subset of nodes whose deletion causes the maximum network fragmentation (quantified by an appropriate metric) in the presence of deterministic or probabilistic cascading failure propagations. We use two commonly considered network fragmentation metrics: total number of disabled nodes and total number of disabled pair-wise connectivities. First, we discuss computational complexity issues and develop linear mixed integer programming (MIP) formulations for the corresponding optimization problems in the deterministic case. We then extend these problems to the case with probabilistic failure propagations using Conditional Value-at-Risk measure. We develop a scenario-based linear MIP model and propose an exact Markov chain-based algorithm to solve these problems. Finally, we perform a series of computational experiments on synthetic and semi-synthetic networks and discuss some interesting insights that illustrate the properties of the proposed models.

12 citations


Posted Content
TL;DR: In this article, the authors introduce the c-Eval metric and its corresponding framework to quantify the feature-based local explanation's quality, which is the minimum-distortion perturbation that successfully alters the prediction while keeping the explanation's features unchanged.
Abstract: In many modern image-classification applications, understanding the cause of model's prediction can be as critical as the prediction's accuracy itself. Various feature-based local explanations generation methods have been designed to give us more insights on the decision of complex classifiers. Nevertheless, there is no consensus on evaluating the quality of different explanations. In response to this lack of comprehensive evaluation, we introduce the c-Eval metric and its corresponding framework to quantify the feature-based local explanation's quality. Given a classifier's prediction and the corresponding explanation on that prediction, c-Eval is the minimum-distortion perturbation that successfully alters the prediction while keeping the explanation's features unchanged. We then demonstrate how c-Eval can be computed using some modifications on existing adversarial generation libraries. To show that c-Eval captures the importance of input's features, we establish the connection between c-Eval and the features returned by explainers in affine and nearly-affine classifiers. We then introduce the c-Eval plot, which not only displays a strong connection between c-Eval and explainers' quality, but also helps automatically determine explainer's parameters. Since the generation of c-Eval relies on adversarial generation, we provide a demo of c-Eval on adversarial-robust models and show that the metric is applicable in those models. Finally, extensive experiments of explainers on different datasets are conducted to support the adoption of c-Eval in evaluating explainers' performance.

11 citations



Proceedings ArticleDOI
07 Jul 2019
TL;DR: Different approximation algorithms for IMC are proposed using Sandwich approach with a tight submodular function to bound the IMC objective function, and an algorithm with performance guarantee tight to the inapproximability of IMC assuming the exponential time hypothesis.
Abstract: Motivated by various settings, we study a new Influence Maximization problem at the Community level (IMC) which aims at finding k users to maximize the benefit of influenced communities where a community is influenced iff the number of influenced users belong to this community exceeds its predefined threshold. In general, IMC objective function is not submodular nor supermodular, thereby making it very challenging to apply existing greedy solutions of the classic influence maximization (IM) where submodular function is required. Furthermore, the major challenge in the traditional methods for any related IM problem is the inefficiency in estimating the influence spread. IMC brings this difficulty to a higher level when considering influenced communities instead of influencing each individual user. In this paper, we propose different approximation algorithms for IMC: (1) Using Sandwich approach with a tight submodular function to bound the IMC objective function, (2) Activating the top-k influencing nodes found from network sampling. Furthermore, when the activated thresholds of communities are bounded by a constant, we propose an algorithm with performance guarantee tight to the inapproximability of IMC assuming the exponential time hypothesis. Each algorithm has its own strengths in a trade-off between effectiveness and running time, which are illustrated both in theory and comprehensive experimental evaluation.

Journal ArticleDOI
TL;DR: In this paper, a novel economic approach, based on the framework of contract theory, is proposed for providing incentives for LTE over unlicensed channels (LTE-U) in cellular networks.
Abstract: In this paper, a novel economic approach, based on the framework of contract theory, is proposed for providing incentives for LTE over unlicensed channels (LTE-U) in cellular networks. In this model, a mobile network operator (MNO) designs and offers a set of contracts to the users to motivate them to accept being served over the unlicensed bands. A practical model in which the information about the quality-of-service (QoS) required by every user is not known to the MNO and other users is considered. For this contractual model, the closed-form expression of the price charged by the MNO for every user is derived and the problem of spectrum allocation is formulated as a matching game with incomplete information. For the matching problem, a distributed algorithm is proposed to assign the users to the licensed and unlicensed spectra. The simulation results show that the proposed pricing mechanism can increase the fraction of users that achieve their QoS requirements by up to 45% compared to classical algorithms that do not account for users requirements. Moreover, the performance of the proposed algorithm in the case of incomplete information is shown to approach the performance of the same mechanism with complete information.

Journal ArticleDOI
TL;DR: This paper develops the first adaptive solution framework to the dynamic resource problem for maximizing spatial reuse and presents the two algorithms for the adaptive set multicover problem with approximation ratio.
Abstract: Device-to-device (D2D) communication has recently gained much attention for its potential to boost the capacity of cellular systems. D2D enables direct communication between devices while bypassing a base station (BS), hence decreasing the load of BSs and increasing the network throughput via spatial reuse of radio resources. However, the cellular system is highly dynamic, an optimal allocation plan of radio resource to D2D links at one time point can easily become suboptimal when devices move. Thus, to maximize spatial reuse in cellular systems, it is crucial to update the resource allocation adaptively to reflect the current system status. In this paper, we develop the first adaptive solution framework to the dynamic resource problem for maximizing spatial reuse. At the core of the framework, we present the two algorithms for the adaptive set multicover problem with approximation ratio $f$ and $\log n$ , respectively, where $f$ is the frequency of the most frequent element and $n$ is the total number of elements. The experimental results not only show that our solutions have a significant improvement in running time, compared with optimal or approximated offline methods, but also demonstrate their good performance through the resource usage, network throughput, and other metrics.

Book ChapterDOI
23 Oct 2019
TL;DR: This paper proposes a novel decentralized and trustless framework for iterative double auction based on blockchain that adopts the smart contract and state channel technologies to enable a double auction process among parties that do not trust each other, while minimizing the blockchain transactions.
Abstract: One of the major problems in current implementations of iterative double auction is that they rely on a trusted third party to handle the auction process. This imposes the risk of single point of failures and monopoly. In this paper, we aim to tackle this problem by proposing a novel decentralized and trustless framework for iterative double auction based on blockchain. Our design adopts the smart contract and state channel technologies to enable a double auction process among parties that do not trust each other, while minimizing the blockchain transactions. We provide a formal development of the framework and highlight the security of our design against adversaries.

Posted Content
TL;DR: In this article, a scalable algorithm to preserve differential privacy in adversarial learning for deep neural networks (DNNs), with certified robustness to adversarial examples, is proposed.
Abstract: In this paper, we aim to develop a scalable algorithm to preserve differential privacy (DP) in adversarial learning for deep neural networks (DNNs), with certified robustness to adversarial examples. By leveraging the sequential composition theory in DP, we randomize both input and latent spaces to strengthen our certified robustness bounds. To address the trade-off among model utility, privacy loss, and robustness, we design an original adversarial objective function, based on the post-processing property in DP, to tighten the sensitivity of our model. A new stochastic batch training is proposed to apply our mechanism on large DNNs and datasets, by bypassing the vanilla iterative batch-by-batch training in DP DNNs. An end-to-end theoretical analysis and evaluations show that our mechanism notably improves the robustness and scalability of DP DNNs.

Proceedings ArticleDOI
01 Jan 2019
TL;DR: This paper proposes an alternative scheme - predictive variance reduction search (PVRS) – to find a point that maximally reduces the uncertainty at the perceived optima locations and demonstrates the effectiveness of the PVRS on both benchmark functions and real-world applications in standard and batch BO settings.
Abstract: Bayesian optimization (BO) is concerned with efficient optimization using probabilistic methods. Predictive entropy search (PES) is a popular and successful BO strategy to find a point that maximizes the information gained about the optima location of an unknown function. Since the PES analytical form is intractable, it requires approximations and is computationally expensive. These approximations may degrade PES performance in terms of accuracy and efficiency. In this paper, we propose an alternative scheme - predictive variance reduction search (PVRS) – to find a point that maximally reduces the uncertainty at the perceived optima locations. The optimization converges to the true optimum when the uncertainty at all perceived optima locations is vanished. Our novel modification is beneficial in two ways. First, PVRS can be computed in closed-form, unlike the approximations made in PES. Second, PVRS is simple and easy to implement. As a result, the proposed PVRS gains huge speed up for scalable BO whilst showing favorable optimization efficiency. Furthermore, we extend our PVRS framework for batch setting where we select multiple experiments for parallel evaluations at each iteration. Empirically, we demonstrate the effectiveness of the PVRS on both benchmark functions and real-world applications in standard and batch BO settings.

Posted Content
TL;DR: In this article, the authors study the Submodular Cost SubModular Cover problem, which is to minimize the submodular cost required to ensure that the sub-modular benefit function exceeds a given threshold.
Abstract: In this work, we study the Submodular Cost Submodular Cover problem, which is to minimize the submodular cost required to ensure that the submodular benefit function exceeds a given threshold. Existing approximation ratios for the greedy algorithm assume a value oracle to the benefit function. However, access to a value oracle is not a realistic assumption for many applications of this problem, where the benefit function is difficult to compute. We present two incomparable approximation ratios for this problem with an approximate value oracle and demonstrate that the ratios take on empirically relevant values through a case study with the Influence Threshold problem in online social networks.

Book ChapterDOI
18 Nov 2019
TL;DR: Experiments show that the proposed method outperforms the state-of-the-art algorithm which provides the same approximation ratio in terms of the number of required samples and running time.
Abstract: Cost-aware Targeted Viral Marketing (CTVM), a generalization of Influence Maximization (IM), has received a lot of attentions recently due to its commercial values. Previous approximation algorithms for this problem required a large number of samples to ensure approximate guarantee. In this paper, we propose an efficient approximation algorithm which uses fewer samples but provides the same theoretical guarantees based on generating and using important samples in its operation. Experiments on real social networks show that our proposed method outperforms the state-of-the-art algorithm which provides the same approximation ratio in terms of the number of required samples and running time.

Posted Content
TL;DR: BC is proposed, a family of algorithms for approximate $k$-cover that can find $(1-\frac{1}{e} -\epsilon)$-approximation solutions within an \emph{$O(\ep silon^{-2}n \log n)$ space}.
Abstract: Given a weighted hypergraph $\mathcal{H}(V, \mathcal{E} \subseteq 2^V, w)$, the approximate $k$-cover problem seeks for a size-$k$ subset of $V$ that has the maximum weighted coverage by \emph{sampling only a few hyperedges} in $\mathcal{E}$. The problem has emerged from several network analysis applications including viral marketing, centrality maximization, and landmark selection. Despite many efforts, even the best approaches require $O(k n \log n)$ space complexities, thus, cannot scale to, nowadays, humongous networks without sacrificing formal guarantees. In this paper, we propose BCA, a family of algorithms for approximate $k$-cover that can find $(1-\frac{1}{e} -\epsilon)$-approximation solutions within an \emph{$O(\epsilon^{-2}n \log n)$ space}. That is a factor $k$ reduction on space comparing to the state-of-the-art approaches with the same guarantee. We further make BCA more efficient and robust on real-world instances by introducing a novel adaptive sampling scheme, termed DTA.

Proceedings Article
24 May 2019
TL;DR: This work presents two incomparable approximation ratios for this problem with an approximate value oracle and demonstrates that the ratios take on empirically relevant values through a case study with the Influence Threshold problem in online social networks.
Abstract: In this work, we study the Submodular Cost Submodular Cover problem, which is to minimize the submodular cost required to ensure that the submodular benefit function exceeds a given threshold. Existing approximation ratios for the greedy algorithm assume a value oracle to the benefit function. However, access to a value oracle is not a realistic assumption for many applications of this problem, where the benefit function is difficult to compute. We present two incomparable approximation ratios for this problem with an approximate value oracle and demonstrate that the ratios take on empirically relevant values through a case study with the Influence Threshold problem in online social networks.

Proceedings ArticleDOI
01 Dec 2019
TL;DR: It is concluded that misinformation attacks must be considered when developing the security model for Socially-enabled Smart Grid technology and planning mitigation techniques.
Abstract: Utility providers are gradually deploying social networks as a useful addition to the Smart Grid in order to help engage consumers in energy management and efficient usage. Besides its benefits, is there any negative impact to the Smart Grid? In this paper, we investigate the vulnerability of Smart Grid when integrating into social networks, where attackers utilize misinformation propagation in social network to alter electricity customer's behavior with the goal of causing degradation to power infrastructure. Stand in both perspectives of power facility administrator and adversary, we model the vulnerability assessment of the system under an optimization problem, which enables us to provide theoretical analysis and behavior investigation of the system based on the complexity theory. As solving the problem is challenging, we propose heuristic solutions and show their efficiency on assessing the system's vulnerability in the presence of misinformation attacks. Therefore, we conclude that misinformation attacks must be considered when developing the security model for Socially-enabled Smart Grid technology and planning mitigation techniques.

Posted Content
TL;DR: In this paper, the authors proposed a novel Heterogeneous Gaussian Mechanism (HGM) to preserve differential privacy in deep neural networks, with provable robustness against adversarial examples.
Abstract: In this paper, we propose a novel Heterogeneous Gaussian Mechanism (HGM) to preserve differential privacy in deep neural networks, with provable robustness against adversarial examples. We first relax the constraint of the privacy budget in the traditional Gaussian Mechanism from (0, 1] to (0, \infty), with a new bound of the noise scale to preserve differential privacy. The noise in our mechanism can be arbitrarily redistributed, offering a distinctive ability to address the trade-off between model utility and privacy loss. To derive provable robustness, our HGM is applied to inject Gaussian noise into the first hidden layer. Then, a tighter robustness bound is proposed. Theoretical analysis and thorough evaluations show that our mechanism notably improves the robustness of differentially private deep neural networks, compared with baseline approaches, under a variety of model attacks.

Proceedings ArticleDOI
26 Mar 2019
TL;DR: Four algorithms with theoretical performance guarantees for network resilience to perturbation of edge weight are introduced, each of them has its own strength in trade-off between effectiveness and running time.
Abstract: This paper focuses on network resilience to perturbation of edge weight. Other than connectivity, many network applications nowadays rely upon some measure of network distance between a pair of connected nodes. In these systems, a metric related to network functionality is associated to each edge. A pair of nodes only being functional if the weighted, shortest-path distance between the pair is below a given threshold T. Consequently, a natural question is on which degree the change of edge weights can damage the network functionality? With this motivation, we study a new problem, Quality of Service Degradation : given a set of pairs, find a minimum budget to increase the edge weights which ensures the distance between each pair exceeds T . We introduce four algorithms with theoretical performance guarantees for this problem. Each of them has its own strength in trade-off between effectiveness and running time, which are illustrated both in theory and comprehensive experimental evaluation.

Posted Content
TL;DR: In this article, the authors proposed an efficient approximation algorithm which uses fewer samples but provides the same theoretical guarantees based on generating and using important samples in its operation, which outperforms the state-of-the-art algorithm which provided the same approximation ratio in terms of the number of required samples and running time.
Abstract: Cost-aware Targeted Viral Marketing (CTVM), a generalization of Influence Maximization (IM), has received a lot of attentions recently due to its commercial values. Previous approximation algorithms for this problem required a large number of samples to ensure approximate guarantee. In this paper, we propose an efficient approximation algorithm which uses fewer samples but provides the same theoretical guarantees based on generating and using important samples in its operation. Experiments on real social networks show that our proposed method outperforms the state-of-the-art algorithm which provides the same approximation ratio in terms of the number of required samples and running time.

25 Sep 2019
TL;DR: A novel mechanism to preserve differential privacy in adversarial learning for deep neural networks, with provable robustness to adversarial examples, is developed using the sequential composition theory in differential privacy.
Abstract: In this paper, we aim to develop a novel mechanism to preserve differential privacy (DP) in adversarial learning for deep neural networks, with provable robustness to adversarial examples. We leverage the sequential composition theory in differential privacy, to establish a new connection between differential privacy preservation and provable robustness. To address the trade-off among model utility, privacy loss, and robustness, we design an original, differentially private, adversarial objective function, based on the post-processing property in differential privacy, to tighten the sensitivity of our model. Theoretical analysis and thorough evaluations show that our mechanism notably improves the robustness of DP deep neural networks.

Journal ArticleDOI
TL;DR: In this paper, scalable approximation algorithms for the k-cycle transversal problem were proposed to find a minimum-size set of edges that intersects all simple cycles of length k in a network.
Abstract: We study scalable approximation algorithms for the k-cycle transversal problem, which is to find a minimum-size set of edges that intersects all simple cycles of length k in a network. This problem is relevant to network reliability through the important metric of network clustering coefficient of order k. We formulate two algorithms to be both scalable and have good solution quality in practice: CARL and DARC. DARC is able to efficiently update its solution under dynamic node and edge insertion and removal to the network. In our experimental evaluation, we demonstrate that DARC is able to run on networks with billions of 3-cycles within 2 h and is able to dynamically update its solution in microseconds.

Posted Content
TL;DR: In this paper, the authors focus on network resilience to perturbation of edge weight and propose a new problem, ''Quality of Service Degradation'' where given a set of pairs, find a minimum budget to increase the edge weights which ensures the distance between each pair exceeds a given threshold.
Abstract: This paper focuses on network resilience to perturbation of edge weight. Other than connectivity, many network applications nowadays rely upon some measure of network distance between a pair of connected nodes. In these systems, a metric related to network functionality is associated to each edge. A pair of nodes only being functional if the weighted, shortest-path distance between the pair is below a given threshold \texttt{T}. Consequently, a natural question is on which degree the change of edge weights can damage the network functionality? With this motivation, we study a new problem, \textit{Quality of Service Degradation}: given a set of pairs, find a minimum budget to increase the edge weights which ensures the distance between each pair exceeds $\mathtt{T}$. We introduce four algorithms with theoretical performance guarantees for this problem. Each of them has its own strength in trade-off between effectiveness and running time, which are illustrated both in theory and comprehensive experimental evaluation.

Posted Content
TL;DR: Experiments show that the proposed method outperforms the state-of-the-art algorithm which provides the same approximation ratio in terms of the number of required samples and running time.
Abstract: Cost-aware Targeted Viral Marketing (CTVM), a generalization of Influence Maximization (IM), has received a lot of attentions recently due to its commercial values. Previous approximation algorithms for this problem required a large number of samples to ensure approximate guarantee. In this paper, we propose an efficient approximation algorithm which uses fewer samples but provides the same theoretical guarantees based on generating and using important samples in its operation. Experiments on real social networks show that our proposed method outperforms the state-of-the-art algorithm which provides the same approximation ratio in terms of the number of required samples and running time.

Journal ArticleDOI
TL;DR: Results highlight that, as the statistical significance of motifs increases, the system throughput will initially increase and then subsequently decreases, and network operators can obtain statistical significance regions for chain and star motifs that map to the optimal content dissemination performance.
Abstract: Device-to-device (D2D) communication is a promising approach to efficiently disseminate critical or viral information. Reaping the benefits of D2D-enabled networks is contingent upon choosing the optimal content dissemination policy subject to resource and user distribution constraints. In this paper, a novel D2D network analysis framework is proposed to study the impacts of frequently occurring subgraphs, known as motifs, on D2D network performance and to determine an effective content dissemination strategy. In the proposed framework, the distribution of devices in the D2D network is modeled as a Thomas cluster process (TCP), and two graph structures, the star, and chain motifs, are studied in the communication graph. Based on the properties of the TCP, the closed-form analytical expressions for the statistical significance, the outage probability, as well as the average throughput per device, are derived. The simulation results corroborate the analytical derivations and show the influence of different system topologies on the occurrence of motifs and the D2D system throughput. More importantly, the results highlight that, as the statistical significance of motifs increases, the system throughput will initially increase and then subsequently decreases. Hence, the network operators can obtain statistical significance regions for chain and star motifs that map to the optimal content dissemination performance. Furthermore, using the obtained regions and the analytical expressions for statistical significance, network operators can effectively identify which clusters of devices can be leveraged for D2D communications while determining the number of serving devices in each identified cluster.