scispace - formally typeset
Search or ask a question

Showing papers in "Applied Network Science in 2018"


Journal ArticleDOI
TL;DR: This review evaluates how a connectivity-based approach has generated new understanding of structural-functional relationships that characterise complex systems and proposes a ‘common toolbox’ underpinned by network-based approaches that can advance connectivity studies by overcoming existing constraints.
Abstract: In recent years, parallel developments in disparate disciplines have focused on what has come to be termed connectivity; a concept used in understanding and describing complex systems. Conceptualisations and operationalisations of connectivity have evolved largely within their disciplinary boundaries, yet similarities in this concept and its application among disciplines are evident. However, any implementation of the concept of connectivity carries with it both ontological and epistemological constraints, which leads us to ask if there is one type or set of approach(es) to connectivity that might be applied to all disciplines. In this review we explore four ontological and epistemological challenges in using connectivity to understand complex systems from the standpoint of widely different disciplines. These are: (i) defining the fundamental unit for the study of connectivity; (ii) separating structural connectivity from functional connectivity; (iii) understanding emergent behaviour; and (iv) measuring connectivity. We draw upon discipline-specific insights from Computational Neuroscience, Ecology, Geomorphology, Neuroscience, Social Network Science and Systems Biology to explore the use of connectivity among these disciplines. We evaluate how a connectivity-based approach has generated new understanding of structural-functional relationships that characterise complex systems and propose a ‘common toolbox’ underpinned by network-based approaches that can advance connectivity studies by overcoming existing constraints.

108 citations


Journal ArticleDOI
TL;DR: A test for similarity between the inferred and the true network is developed and tested on the consecutive snapshots of a network, and against the ground truth.
Abstract: Most real networks are too large or they are not available for real time analysis Therefore, in practice, decisions are made based on partial information about the ground truth network It is of great interest to have metrics to determine if an inferred network (the partial information network) is similar to the ground truth In this paper we develop a test for similarity between the inferred and the true network Our research utilizes a network visualization tool, which systematically discovers a network, producing a sequence of snapshots of the network We introduce and test our metric on the consecutive snapshots of a network, and against the ground truth To test the scalability of our metric we use a random matrix theory approach while discovering Erdos-Renyi graphs This scaling analysis allows us to make predictions about the performance of the discovery process

52 citations


Journal ArticleDOI
TL;DR: This paper assesses the vulnerability of power grids to targeted attacks based on network science by employing two graph models for power grids as simple and weighted graphs, and formulate different node-attack strategies based on those centrality metrics.
Abstract: Due to the open data policies, nowadays, some countries have their power grid data available online. This may bring a new concern to the power grid operators in terms of malicious threats. In this paper, we assess the vulnerability of power grids to targeted attacks based on network science. By employing two graph models for power grids as simple and weighted graphs, we first calculate the centrality metrics of each node in a power grid. Subsequently, we formulate different node-attack strategies based on those centrality metrics, and empirically analyse the impact of targeted attacks on the structural and the operational performance of power grids. We demonstrate our methodology in the high-voltage transmission networks of 5 European countries and in commonly used IEEE test power grids.

43 citations


Journal ArticleDOI
TL;DR: It is shown that 20% of driver nodes can control dynamic variables acting on the whole network, suggesting that non-repressive strategies such as access to basic education or sanitation can be effective in reducing criminality by changing the perception of driver individuals to norm compliance.
Abstract: Law enforcement and intelligence agencies worldwide struggle to find effective ways to fight organized crime and reduce criminality. However, illegal networks operate outside the law and much of the data collected is classified. Therefore, little is known about the structure, topological weaknesses, and control of criminal networks. We fill this gap by presenting a unique criminal intelligence network built directly by the Brazilian Federal Police for intelligence and investigative purposes. We study its structure, its response to different attack strategies, and its structural controllability. Surprisingly, the network composed of individuals involved in multiple crimes of federal jurisdiction in Brazil has a giant component enclosing more than half of all its edges. We focus on the largest connected cluster of this network and show it has many social network features, such as small-worldness and heavy-tail degree distribution. However, it is less dense and less efficient than typical social networks. The giant component also shows a high degree cutoff that is associated with the lack of trust among individuals belonging to clandestine networks. The giant component of the network is also highly modular (Q=0.96) and thence fragile to module-based attacks. The targets in such attacks, i.e. the nodes connecting distinct communities, may be interpreted as individuals with bridging clandestine activities such as accountants, lawyers, or money changers. The network can be disrupted by the removal of approximately 2% of either its nodes or edges, the negligible difference between both approaches being due to low graph density. Finally, we show that 20% of driver nodes can control dynamic variables acting on the whole network, suggesting that non-repressive strategies such as access to basic education or sanitation can be effective in reducing criminality by changing the perception of driver individuals to norm compliance.

37 citations


Journal ArticleDOI
TL;DR: This work proposes a novel multiple-path asynchronous threshold (MAT) model, in which the model captures not only direct influence from neighboring influencers but also indirect influence passed along by messengers, and develops an effective and efficient heuristic to tackle the influence-maximization problem.
Abstract: Modeling influence diffusion in social networks is an important challenge. We investigate influence-diffusion modeling and maximization in the setting of viral marketing, in which a node’s influence is measured by the number of nodes it can activate to adopt a new technology or purchase a new product. One of the fundamental problems in viral marketing is to find a small set of initial adopters who can trigger the most further adoptions through word-of-mouth-based influence propagation in the network. We propose a novel multiple-path asynchronous threshold (MAT) model, in which we quantify influence and track its diffusion and aggregation. Our MAT model captures not only direct influence from neighboring influencers but also indirect influence passed along by messengers. Moreover, our MAT framework models influence attenuation along diffusion paths, temporal influence decay, and individual diffusion dynamics. Our work is an important step toward a more realistic diffusion model. Further, we develop an effective and efficient heuristic to tackle the influence-maximization problem. Our experiments on four real-life networks demonstrate its excellent performance in terms of both influence spread and time efficiency. Our work provides preliminary but significant insights and implications for diffusion research and marketing practice.

35 citations


Journal ArticleDOI
TL;DR: It is shown that communicability is a simple and reliable way to identify the key concepts and examine their epistemic justification within the collated network.
Abstract: Concept maps, which are network-like visualisations of the inter-linkages between concepts, are used in teaching and learning as representations of students’ understanding of conceptual knowledge and its relational structure. In science education, research on the uses of concept maps has focused much attention on finding methods to identify key concepts that are of the most importance either in supporting or being supported by other concepts in the network. Here we propose a method based on network analysis to examine students’ representations of the relational structure of physics concepts in the form of concept maps. We suggest how the key concepts and their epistemic support can be identified through focusing on the pathways along which the information is passed from one node to another. Towards this end, concept maps are analysed as directed and weighted networks, where nodes are concepts and links represent different types of connections between concepts, and where each link is assumed to provide epistemic support to the node it is connected to. The notion of key concept can then be operationalised through the directed flow of information from one node to another in terms of communicability between the nodes, separately for out-going and in-coming weighted links. Here we analyse a collated concept network based on a sample of 12 original concept maps constructed by university students. We show that communicability is a simple and reliable way to identify the key concepts and examine their epistemic justification within the collated network. The communicabilities of the key nodes in the collated network are compared with communicabilities averaged over the set of 12 individual concept maps. The comparison shows the collated network contains an extensive set of key concepts with good epistemic support. Every individual networks contain a sub-set of these key concepts but with a limited overlap of the sub-sets with other individual networks. The epistemically well substantiated knowledge is thus sparsely distributed over the 12 individual networks.

31 citations


Journal ArticleDOI
TL;DR: This paper proposes a null model that is able to preserve the interlayer connectedness, while taking into account that one of the link types is actually the result of a projection of an underlying bipartite network.
Abstract: In corporate networks, firms are connected through links of corporate ownership and shared directors, connecting the control over major economic actors in our economies in meaningful and consequential ways. Most research thus far focused on the connectedness of firms as a result of one particular link type, analyzing node-specific metrics or global network-based methods to gain insights in the modelled corporate system. In this paper, we aim to understand multiplex corporate networks with multiple types of connections, specifically investigating the network’s essential building blocks: multiplex network motifs. Motifs, which are small subgraph patterns occurring at significantly higher frequencies than in similar random networks, have demonstrated their usefulness in understanding the structure of many types of real-world networks. However, detecting motifs in multiplex networks is nontrivial for two reasons. First of all, there are no out-of-the-box subgraph enumeration algorithms for multiplex networks. Second, existing null models to test network motif significance, are unable to incorporate the interlayer dependencies in the multiplex network. We solve these two issues by introducing a layer encoding algorithm that incorporates the multiplex aspect in the subgraph enumeration phase. In addition, we propose a null model that is able to preserve the interlayer connectedness, while taking into account that one of the link types is actually the result of a projection of an underlying bipartite network. The experimental section considers the corporate network of Germany, in which tens of thousands of firms are connected through several hundred thousand links. We demonstrate how incorporating the multiplex aspect in motif detection is able to reveal new insights that could not be obtained by studying only one type of relationship. In a general sense, the motifs reflect known corporate governance practices related to the monitoring of investments and the concentration of ownership. A substantial fraction of the discovered motifs is typical for an industrialized country such as Germany, whereas others seem specific for certain economic sectors. Interestingly, we find that motifs involving financial firms are over-represented amongst the larger and more complex motifs. This demonstrates the prominent role of the financial sector in Germany’s largely industry-oriented corporate network.

28 citations


Journal ArticleDOI
TL;DR: A new, temporal walk based dynamic centrality measure that models temporal information propagation by considering the order of edge creation and outperforms graph snapshot based static and other recently proposed dynamicCentrality measures in assigning the highest time-aware centrality to the actually relevant nodes of the network.
Abstract: A plethora of centrality measures or rankings have been proposed to account for the importance of the nodes of a network. In the seminal study of Boldi and Vigna (2014), the comparative evaluation of centrality measures was termed a difficult, arduous task. In networks with fast dynamics, such as the Twitter mention or retweet graphs, predicting emerging centrality is even more challenging. Our main result is a new, temporal walk based dynamic centrality measure that models temporal information propagation by considering the order of edge creation. Dynamic centrality measures have already started to emerge in publications; however, their empirical evaluation is limited. One of our main contributions is creating a quantitative experiment to assess temporal centrality metrics. In this experiment, our new measure outperforms graph snapshot based static and other recently proposed dynamic centrality measures in assigning the highest time-aware centrality to the actually relevant nodes of the network. Additional experiments over different data sets show that our method perform well for detecting concept drift in the process that generates the graphs.

27 citations


Journal ArticleDOI
TL;DR: The present findings show that the structure of the mental lexicon influences lexical access in visual word recognition.
Abstract: Network science has been applied to study the structure of the mental lexicon, the part of long-term memory where all the words a person knows are stored. Here the tools of network science are used to study the organization of orthographic word-forms in the mental lexicon and how that might influence visual word recognition. An orthographic similarity network of the English language was constructed such that each node represented an English word, and undirected, unweighted edges were placed between words that differed by an edit distance of 1, a commonly used operationalization of orthographic similarity in psycholinguistics. The largest connected component of the orthographic language network had a small-world structure and a long-tailed degree distribution. Additional analyses were conducted using behavioral data obtained from a psycholinguistic database to determine if network science measures obtained from the orthographic language network could be used to predict how quickly and accurately people process written words. The present findings show that the structure of the mental lexicon influences lexical access in visual word recognition.

26 citations


Journal ArticleDOI
TL;DR: A hydraulically informed surrogate measure of pipe criticality for the resilience analysis of WDNs, called Water Flow Edge Betweenness Centrality (WFEBC), which combines the random walk betweenness centrality with hydraulic (energy) loss principles in pipes is introduced.
Abstract: Water Distribution Networks (WDN) are complex and highly interconnected systems. To maintain operation under failure conditions, WDNs should have built-in resilience based on topological and energy redundancy. There are various methods for analysing the resilience of WDNs based on either hydraulic models or surrogate network measures; however, not a single universally accepted method exists. Hydraulic modeling of disruptive operational scenarios suffer from combinatorial restrictions and uncertainties. Methods that rely on surrogate network measures do not take into account the complex interactions between topological and energy redundancy. To bridge this gap, the presented work introduces a hydraulically informed surrogate measure of pipe criticality for the resilience analysis of WDNs, called Water Flow Edge Betweenness Centrality (WFEBC). The WFEBC combines the random walk betweenness centrality with hydraulic (energy) loss principles in pipes. The proposed network resilience estimation method is applied to a case study network and an operational network. Furthermore, a network decomposition approach is proposed to complement the network estimation method and facilitate its scalability to large operational networks. The described resilience analysis method is benchmarked against a hydraulic model-based analysis of WDN reserve capacity. WFEBC is also applied to assess the improvement in resilience allowed by the implementation of a dynamically adaptive topology in an operational network. The benefits and limitations of the proposed method are discussed.

25 citations


Journal ArticleDOI
TL;DR: The proposed approach builds on learning and discovery and therefore the series of interventions for controlling the network are determined but are not fixed, so system control rules evolve in such a way that they mirror both the structure and dynamics of the system, without having ‘direct’ access to either.
Abstract: In this paper we describe the application of a learning classifier system (LCS) variant known as the eXtended classifier system (XCS) to evolve a set of ‘control rules’ for a number of Boolean network instances. We show that (1) it is possible to take the system to an attractor, from any given state, by applying a set of ‘control rules’ consisting of ternary conditions strings (i.e. each condition component in the rule has three possible states; 0, 1 or #) with associated bit-flip actions, and (2) that it is possible to discover such rules using an evolutionary approach via the application of a learning classifier system. The proposed approach builds on learning (reinforcement learning) and discovery (a genetic algorithm) and therefore the series of interventions for controlling the network are determined but are not fixed. System control rules evolve in such a way that they mirror both the structure and dynamics of the system, without having ‘direct’ access to either.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed an entropy-based method to predict a given percentage of missing links, by identifying them with the most probable non-observed ones, and the probability coefficients are computed by solving opportunely defined null-models over the accessible network structure.
Abstract: Link-prediction is an active research field within network theory, aiming at uncovering missing connections or predicting the emergence of future relationships from the observed network structure. This paper represents our contribution to the stream of research concerning missing links prediction. Here, we propose an entropy-based method to predict a given percentage of missing links, by identifying them with the most probable non-observed ones. The probability coefficients are computed by solving opportunely defined null-models over the accessible network structure. Upon comparing our likelihood-based, local method with the most popular algorithms over a set of economic, financial and food networks, we find ours to perform best, as pointed out by a number of statistical indicators (e.g. the precision, the area under the ROC curve, etc.). Moreover, the entropy-based formalism adopted in the present paper allows us to straightforwardly extend the link-prediction exercise to directed networks as well, thus overcoming one of the main limitations of current algorithms. The higher accuracy achievable by employing these methods - together with their larger flexibility - makes them strong competitors of available link-prediction algorithms.

Journal ArticleDOI
TL;DR: Results analyzing the resilience of networks subjected to node disruptions show that networks synthesized using the proposed model can generally outperform its real-world counterpart.
Abstract: The ubiquity of supply chains along with their increasingly interconnected structure has ignited interest in studying supply chain networks through the lens of complex adaptive systems. A particularly important characteristic of supply chains is the desirable goal of sustaining their operation when exposed to unexpected perturbations. Applied network science methods can be used to analyze topological properties of supply chains and propose models for their growth. Network models focusing on the critical aspect of supply chain resilience may provide insights into the design of supply networks that may quickly recover from disruptions. This is vital for understanding both static and dynamic structures of complex supply networks, and enabling management to make informed decisions and prioritizing particular operations. This paper proposes an action-based perspective for creating a compact probabilistic model for a given real-world supply network. The action-based model consists of a set of rules (actions) that a firm may use to connect with other firms, such that the synthesized networks are topologically resilient. Additionally, it captures the heterogeneous roles of different firms by incorporating domain specific constraints. Results analyzing the resilience of networks subjected to node disruptions show that networks synthesized using the proposed model can generally outperform its real-world counterpart.

Journal ArticleDOI
TL;DR: In this review, rapid advancements in the field of network science in combination with spectral graph theory that enables us to uncover the complexities of various diseases are illustrated.
Abstract: The fundamental understanding of altered complex molecular interactions in a diseased condition is the key to its cure. The overall functioning of these molecules is kind of jugglers play in the cell orchestra and to anticipate these relationships among the molecules is one of the greatest challenges in modern biology and medicine. Network science turned out to be providing a successful and simple platform to understand complex interactions among healthy and diseased tissues. Furthermore, much information about the structure and dynamics of a network is concealed in the eigenvalues of its adjacency matrix. In this review, we illustrate rapid advancements in the field of network science in combination with spectral graph theory that enables us to uncover the complexities of various diseases. Interpretations laid by network science approach have solicited insights into molecular relationships and have reported novel drug targets and biomarkers in various complex diseases.

Journal ArticleDOI
TL;DR: In this article, the authors present a simple, expressive and extensible model for temporal text networks, that can be used as a common ground across different types of networks and analysis tasks, and show how simple procedures to produce views of the model allow the direct application of analysis methods already developed in other domains, from traditional data mining to multilayer network mining.
Abstract: Three fundamental elements to understand human information networks are the individuals (actors) in the network, the information they exchange, that is often observable online as text content (emails, social media posts, etc.), and the time when these exchanges happen. An extremely large amount of research has addressed some of these aspects either in isolation or as combinations of two of them. There are also more and more works studying systems where all three elements are present, but typically using ad hoc models and algorithms that cannot be easily transfered to other contexts. To address this heterogeneity, in this article we present a simple, expressive and extensible model for temporal text networks, that we claim can be used as a common ground across different types of networks and analysis tasks, and we show how simple procedures to produce views of the model allow the direct application of analysis methods already developed in other domains, from traditional data mining to multilayer network mining.

Journal ArticleDOI
TL;DR: In this paper, the authors provide an overview of several existing statistical approaches in network science applicable to weighted directed networks, and apply these methods to the inpatient data of approximately one million people, spanning approximately 17 years.
Abstract: Tools from network science can be utilized to study relations between diseases. Different studies focus on different types of inter-disease linkages. One of them is the comorbidity patterns derived from large-scale longitudinal data of hospital discharge records. Researchers seek to describe comorbidity relations as a network to characterize pathways of disease progressions and to predict future risks. The first step in such studies is the construction of the network itself, which subsequent analyses rest upon. There are different ways to build such a network. In this paper, we provide an overview of several existing statistical approaches in network science applicable to weighted directed networks. We discuss the differences between the null models that these models assume and their applications. We apply these methods to the inpatient data of approximately one million people, spanning approximately 17 years, pertaining to the Montreal Census Metropolitan Area. We discuss the differences in the structure of the networks built by different methods, and different features of the comorbidity relations that they extract. We also present several example applications of these methods.

Journal ArticleDOI
TL;DR: A new view of autonomous control in industrial production is formulated and a set of suggestions with the potential to enhance performance under realistic conditions of scheduling heuristics of jobs in a production process are derived.
Abstract: The scheduling of processes in a network is a core logistic challenge with a multitude of applications in our complex industrialized world Often, scheduling decisions are based on incomplete and unreliable information Here, a simple rule of ’more information, better decisions’ may no longer hold and heuristics balancing global and local information, or centralized and autonomous control, may yield better performance So far, only anecdotal evidence for the potential benefit of autonomous control in scheduling exists Here, we explore this hypothesis within a minimal model derived from scheduling principles and the phenomenology of dynamical processes on graphs In this model, centralized and autonomous control can be represented and quantitatively assessed, performance is well defined and problem complexity can be varied Our model shows that a balance of centralized and autonomous control can enhance the performance in networks of decision-making entities The mechanistic insight gained from the model also reveals the limitations of hybrid control setups: We find that communication at a high hierarchy level can give an advantage to centralized control Counter-intuitively, it arises not from a higher degree of coordination and quicker convergence towards a common solution, but rather from an accelerated sampling of candidate choices leading to a measurable increase in information flow from higher to lower hierarchical levels Our study allows us to formulate a new view of autonomous control in industrial production and derive a set of suggestions with the potential to enhance performance under realistic conditions of scheduling heuristics of jobs in a production process

Journal ArticleDOI
TL;DR: This paper adopts the stabilized Louvain method for network community detection to improve consistency and stability and develops a new method to identify the central nodes based on the temporal evolution of the network structure and track the changes of communities over time.
Abstract: When studying patent data as a way to understand innovation and technological change, the conventional indicators might fall short, and categorizing technologies based on the existing classification systems used by patent authorities could cause inaccuracy and misclassification, as shown in literature. Gao et al. (International Workshop on Complex Networks and their Applications, 2017) have established a method to analyze patent classes of similar technologies as network communities. In this paper, we adopt the stabilized Louvain method for network community detection to improve consistency and stability. Incorporating the overlapping community mapping algorithm, we also develop a new method to identify the central nodes based on the temporal evolution of the network structure and track the changes of communities over time. A case study of Germany’s patent data is used to demonstrate and verify the application of the method and the results. Compared to the non-network metrics and conventional network measures, we offer a heuristic approach with a dynamic view and more stable results.

Journal ArticleDOI
TL;DR: In this paper, the authors explore the conceptual landscape of mathematical research, focusing on homological holes, regions with low connectivity in the simplicial structure, and find that these regions capture some essential feature of research practice in mathematics.
Abstract: In the last years complex networks tools contributed to provide insights on the structure of research, through the study of collaboration, citation and co-occurrence networks. The network approach focuses on pairwise relationships, often compressing multidimensional data structures and inevitably losing information. In this paper we propose for the first time a simplicial complex approach to word co-occurrences, providing a natural framework for the study of higher-order relations in the space of scientific knowledge. Using topological methods we explore the conceptual landscape of mathematical research, focusing on homological holes, regions with low connectivity in the simplicial structure. We find that homological holes are ubiquitous, which suggests that they capture some essential feature of research practice in mathematics. k-dimensional holes die when every concept in the hole appears in an article together with other k+1 concepts in the hole, hence their death may be a sign of the creation of new knowledge, as we show with some examples. We find a positive relation between the size of a hole and the time it takes to be closed: larger holes may represent potential for important advances in the field because they separate conceptually distant areas. We provide further description of the conceptual space by looking for the simplicial analogs of stars and explore the likelihood of edges in a star to be also part of a homological cycle. We also show that authors’ conceptual entropy is positively related with their contribution to homological holes, suggesting that polymaths tend to be on the frontier of research.

Journal ArticleDOI
TL;DR: Users whose centralities and physical activities both change with time and whose evolving centralities are significantly correlated with each other over time are more likely to be introverted as well as anxious compared to those users who are time-stable and do not have a co-evolution relationship.
Abstract: Understanding the relationship between individuals’ social networks and health could help devise public health interventions for reducing incidence of unhealthy behaviors or increasing prevalence of healthy ones. In this context, we explore the co-evolution of individuals’ social network positions and physical activities. We are able to do so because the NetHealth study at the University of Notre Dame has generated both high-resolution longitudinal social network (e.g., SMS) data and high-resolution longitudinal health-related behavioral (e.g., Fitbit physical activity) data. We examine trait differences between (i) users whose social network positions (i.e., centralities) change over time versus those whose centralities remain stable, (ii) users whose Fitbit physical activities change over time versus those whose physical activities remain stable, and (iii) users whose centralities and their physical activities co-evolve, i.e., correlate with each other over time. We find that centralities of a majority of all nodes change with time. These users do not show any trait difference compared to time-stable users. However, if out of all users whose centralities change with time we focus on those whose physical activities also change with time, then the resulting users are more likely to be introverted than time-stable users. Moreover, users whose centralities and physical activities both change with time and whose evolving centralities are significantly correlated (i.e., co-evolve) with evolving physical activities are more likely to be introverted as well as anxious compared to those users who are time-stable and do not have a co-evolution relationship. Our network analysis framework reveals several links between individuals’ social network structure, health-related behaviors, and the other (e.g., personality) traits. In the future, our study could lead to development of a predictive model of social network structure from behavioral/trait information and vice versa.

Journal ArticleDOI
TL;DR: Predicting the onset of complications of diabetes using a network-based approach by identifying fast and slow progressors can aid the physician in developing better diabetes management program for a given patient.
Abstract: Diabetes is a significant health concern with more than 30 million Americans living with diabetes. Onset of diabetes increases the risk for various complications, including kidney disease, myocardial infractions, heart failure, stroke, retinopathy, and liver disease. In this paper, we study and predict the onset of these complications using a network-based approach by identifying fast and slow progressors. That is, given a patient’s diagnosis of diabetes, we predict the likelihood of developing one or more of the possible complications, and which patients will develop complications quickly. This combination of "if a complication will be developed” with ”how fast it will be developed” can aid the physician in developing better diabetes management program for a given patient.

Journal ArticleDOI
TL;DR: Comparing social signatures built from call and text message data, and developing a way of constructing mixed social signatures using both channels, demonstrates individuals vary in how they allocate their communication effort across their personal networks and this variation is persistent over time and across different channels of communication.
Abstract: The structure of egocentric networks reflects the way people balance their need for strong, emotionally intense relationships and a diversity of weaker ties. Egocentric network structure can be quantified with ’social signatures’, which describe how people distribute their communication effort across the members (alters) of their personal networks. Social signatures based on call data have indicated that people mostly communicate with a few close alters; they also have persistent, distinct signatures. To examine if these results hold for other channels of communication, here we compare social signatures built from call and text message data, and develop a way of constructing mixed social signatures using both channels. We observe that all types of signatures display persistent individual differences that remain stable despite the turnover in individual alters. We also show that call, text, and mixed signatures resemble one another both at the population level and at the level of individuals. The consistency of social signatures across individuals for different channels of communication is surprising because the choice of channel appears to be alter-specific with no clear overall pattern, and ego networks constructed from calls and texts overlap only partially in terms of alters. These results demonstrate individuals vary in how they allocate their communication effort across their personal networks and this variation is persistent over time and across different channels of communication.

Journal ArticleDOI
TL;DR: Comparing standard methods for constructing physician networks from patient-physician encounter data with a new method based on clinical episodes of care found the episode-based networks had ties that were more likely to reappear in the future, and appeared to have more fluid community structure.
Abstract: To compare standard methods for constructing physician networks from patient-physician encounter data with a new method based on clinical episodes of care. We used data on 100% of traditional Medicare beneficiaries from 51 nationally representative geographical regions for the years 2005–2010. We constructed networks of physicians based on their shared patients. In the fixed-threshold networks and adaptive-threshold networks, we included data on all patient-physician encounters to form the physician-physician ties, and then subsequently thresholded some proportion of the strongest ties. In contrast, in the episode-based approach, only those patient-physician encounters that occurred within shared clinical episodes treating specific conditions contributed towards physician-physician ties. We extracted clinical episodes in the Medicare data and investigated structural properties of the patient-sharing networks of physicians, temporal dynamics of their ties, and temporal stability of network communities across the two approaches. The episode-based networks accentuated ties between primary care physicians (PCPs) and medical specialists, had ties that were more likely to reappear in the future, and appeared to have more fluid community structure. Constructing physician networks around shared episodes of care is a clinically sound alternative to previous approaches to network construction that does not require arbitrary decisions about thresholding. The resulting networks capture somewhat different aspects of patient-physician encounters.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed a method that identifies spatially contiguous and possibly overlapping components referred to as domains, and identifies the lagged functional relationships between them, which can be used to identify the epicenters of activity of grid cells of the same domain.
Abstract: In real physical systems the underlying spatial components might not have crisp boundaries and their interactions might not be instantaneous. To this end, we propose δ-MAPS; a method that identifies spatially contiguous and possibly overlapping components referred to as domains, and identifies the lagged functional relationships between them. Informally, a domain is a spatially contiguous region that somehow participates in the same dynamic effect or function. The latter will result in highly correlated temporal activity between grid cells of the same domain. δ-MAPS first identifies the epicenters of activity of a domain. Next, it identifies a domain as the maximum possible set of spatially contiguous grid cells that include the detected epicenters and satisfy a homogeneity constraint. After identifying the domains, δ-MAPS infers a functional network between them. The proposed network inference method examines the statistical significance of each lagged correlation between two domains, applies a multiple-testing process to control the rate of false positives, infers a range of potential lag values for each edge, and assigns a weight to each edge reflecting the magnitude of interaction between two domains. δ-MAPS is related to clustering, multivariate statistical techniques and network community detection. However, as we discuss and also show with synthetic data, it is also significantly different, avoiding many of the known limitations of these methods. We illustrate the application of δ-MAPS on data from two domains: climate science and neuroscience. First, the sea-surface temperature climate network identifies some well-known teleconnections (such as the lagged connection between the El Nin $\tilde {o}$ Southern Oscillation and the Indian Ocean). Second, the analysis of resting state fMRI cortical data confirms the presence of known functional resting state networks (default mode, occipital, motor/somatosensory and auditory), and shows that the cortical network includes a backbone of relatively few regions that are densely interconnected.

Journal ArticleDOI
TL;DR: A novel method is proposed for designing a bike network in urban areas based on the number of taxi trips within an urban area, a weighted network abstracted as link weights using the Taxi smart card system of a real city.
Abstract: In this paper, a novel method is proposed for designing a bike network in urban areas. Based on the number of taxi trips within an urban area, a weighted network is abstracted. In this network, nodes are the origins and destinations of taxi trips and the number of trips among them is abstracted as link weights. Data is extracted from the Taxi smart card system of a real city. Then, Communities i.e. clusters of this network are detected using a modularity maximization method. Each community contains the nodes with highest number of trips within the cluster and lowest number of trips with other clusters. Within each community, the nodes close enough to each other for being traveled by bicycle are detected as key points and some non-dominated bike network connecting these nodes are enumerated using a bi-objective optimization model. The total travel cost (distance or time) on the network and the path length are considered as objectives. The method is applied to Isfahan city in Iran and a total of seven regions with some non-dominated bike networks are proposed.

Journal ArticleDOI
TL;DR: Ecological corridors, seen as connecting elements for geographically distant areas dedicated to the preservation of endangered species, can be analyzed for the identification of critical land patches, by ranking cut nodes according to a score that encompasses various criteria for prioritized intervention.
Abstract: Ecological landscape networks represent the current paradigm for the protection of biodiversity. In the analysis of land features that precedes the establishment of land management plans, graph-theoretic approaches become increasingly popular due to their aptness for the representation of connectivity. Ecological corridors, seen as connecting elements for geographically distant areas dedicated to the preservation of endangered species, can be analyzed for the identification of critical land patches, by ranking cut nodes according to a score that encompasses various criteria for prioritized intervention.

Journal ArticleDOI
TL;DR: This paper investigates the application of a novel graph theoretic clustering method, Node-Based Resilience clustering (NBR-Clust), to address the heterogeneity of Autism Spectrum Disorder (ASD) and identify meaningful subgroups and conducts feature extraction analysis to characterize relevant biomarkers that delineate the resulting subgroups.
Abstract: With the growing ubiquity of data in network form, clustering in the context of a network, represented as a graph, has become increasingly important Clustering is a very useful data exploratory machine learning tool that allows us to make better sense of heterogeneous data by grouping data with similar attributes based on some criteria This paper investigates the application of a novel graph theoretic clustering method, Node-Based Resilience clustering (NBR-Clust), to address the heterogeneity of Autism Spectrum Disorder (ASD) and identify meaningful subgroups The hypothesis is that analysis of these subgroups would reveal relevant biomarkers that would provide a better understanding of ASD phenotypic heterogeneity useful for further ASD studies We address appropriate graph constructions suited for representing the ASD phenotype data The sample population is drawn from a very large rigorous dataset: Simons Simplex Collection (SSC) Analysis of the results performed using graph quality measures, internal cluster validation measures, and clinical analysis outcome demonstrate the potential usefulness of resilience measure clustering for biomedical datasets We also conduct feature extraction analysis to characterize relevant biomarkers that delineate the resulting subgroups The optimal results obtained favored predominantly a 5-cluster configuration

Journal ArticleDOI
TL;DR: A methodology is developed to map the multilayer macro-network of financial exposures among institutional sectors across financial instruments (i.e., loans and deposits, debt securities, and equity), and it is illustrated on recently available data to test the effect of the implementation of ECB’s QE on the time evolution of the financial linkages in the multi-layer macro- network of the euro area.
Abstract: Over the last decades, both advanced and emerging economies have experienced a striking increase in the intra-financial activity across different asset classes and increasingly complex contract types, leading to a far more complex financial system. Until the 2007-2008 crisis, the increased financial intensity and complexity was believed beneficial in making the financial system more resilient and less vulnerable to shocks. However, in 2007-2008, the advanced economies suffered the biggest financial crisis since the 1930s, followed by a severe post-crisis recession, questioning the adequacy of traditional tools in predicting, explaining, and responding to periods of financial distress. In particular, the effect of complex interconnections among financial actors on financial stability has been widely acknowledged. A recent debate focused on the effects of unconventional policies aimed at achieving both price and financial stability. Among these unconventional policies, Quantitative Easing (QE, i.e., the large-scale asset purchase programme conducted by a central bank upon the creation of new money) has been recently implemented by the European Central Bank (ECB). In this context, two questions deserve more attention in the literature. First, to what extent, the resources provided to the banking system through QE are transmitted to the real economy. Second, to what extent, the QE may also alter the pattern of intra-financial exposures and what are the implications in terms of financial stability. Here, we address these two questions by developing a methodology to map the multilayer macro-network of financial exposures among institutional sectors across financial instruments (i.e., loans and deposits, debt securities, and equity), and we illustrate our approach on recently available data. We then test the effect of the implementation of ECB’s QE on the time evolution of the financial linkages in the multilayer macro-network of the euro area, as well as the effect on macroeconomic variables, such as consumption, investment, unemployment, growth, and inflation.

Journal ArticleDOI
TL;DR: This paper considers the problem of identifying and tracking communities in graphs that change over time – dynamic community detection – and presents a framework based on Riemannian geometry to aid in this task.
Abstract: A community is a subset of a wider network where the members of that subset are more strongly connected to each other than they are to the rest of the network. In this paper, we consider the problem of identifying and tracking communities in graphs that change over time – dynamic community detection – and present a framework based on Riemannian geometry to aid in this task. Our framework currently supports several important operations such as interpolating between and averaging over graph snapshots. We compare these Riemannian methods with entry-wise linear interpolation and find that the Riemannian methods are generally better suited to dynamic community detection. Next steps with the Riemannian framework include producing a Riemannian least-squares regression method for working with noisy data and developing support methods, such as spectral sparsification, to improve the scalability of our current methods.

Journal ArticleDOI
TL;DR: This study shows that Twitter users exhibit favourable leaning (predominantly neutral or positive) towards impact investing, and uses Twitter as a proxy of the impact investing market, and analyze relevant tweets posted over a period of ten months.
Abstract: The 2008 financial crisis unveiled the intrinsic failures of the financial system as we know it. As a consequence, impact investing started to receive increasing attention, as evidenced by the high market growth rates. The goal of impact investment is to generate social and environmental impact alongside a financial return. In this paper we identify the main players in the sector and how they interact and communicate with each other. We use Twitter as a proxy of the impact investing market, and analyze relevant tweets posted over a period of ten months. We apply network, contents and sentiment analysis on the acquired dataset. Our study shows that Twitter users exhibit favourable leaning (predominantly neutral or positive) towards impact investing. Retweet communities are decentralised and include users from a variety of sectors. Despite some basic common vocabulary used by all retweet communities identified, the vocabulary and the topics discussed by each community vary largely. We note that an additional effort should be made in raising awareness about the sector, especially by policymakers and media outlets. The role of investors and the academia is also discussed, as well as the emergence of hybrid business models within the sector and its connections to the tech industry. This paper extends our previous study, one of the first analyses of Twitter activities in the impact investing market.