scholarly journals Betweenness centrality profiles in trees

2017 ◽  
Vol 5 (5) ◽  
pp. 776-794
Author(s):  
Benjamin Fish ◽  
Rahul Kushwaha ◽  
György Turán

Abstract Betweenness centrality of a vertex in a graph measures the fraction of shortest paths going through the vertex. This is a basic notion for determining the importance of a vertex in a network. The $k$-betweenness centrality of a vertex is defined similarly, but only considers shortest paths of length at most $k$. The sequence of $k$-betweenness centralities for all possible values of $k$ forms the betweenness centrality profile of a vertex. We study properties of betweenness centrality profiles in trees. We show that for scale-free random trees, for fixed $k$, the expectation of $k$-betweenness centrality strictly decreases as the index of the vertex increases. We also analyse worst-case properties of profiles in terms of the distance of profiles from being monotone, and the number of times pairs of profiles can cross. This is related to whether $k$-betweenness centrality, for small values of $k$, may be used instead of having to consider all shortest paths. Bounds are given that are optimal in order of magnitude. We also present some experimental results for scale-free random trees.

2021 ◽  
Vol 7 ◽  
pp. e699
Author(s):  
Martin Mirakyan

Betweenness-centrality is a popular measure in network analysis that aims to describe the importance of nodes in a graph. It accounts for the fraction of shortest paths passing through that node and is a key measure in many applications including community detection and network dismantling. The computation of betweenness-centrality for each node in a graph requires an excessive amount of computing power, especially for large graphs. On the other hand, in many applications, the main interest lies in finding the top-k most important nodes in the graph. Therefore, several approximation algorithms were proposed to solve the problem faster. Some recent approaches propose to use shallow graph convolutional networks to approximate the top-k nodes with the highest betweenness-centrality scores. This work presents a deep graph convolutional neural network that outputs a rank score for each node in a given graph. With careful optimization and regularization tricks, including an extended version of DropEdge which is named Progressive-DropEdge, the system achieves better results than the current approaches. Experiments on both real-world and synthetic datasets show that the presented algorithm is an order of magnitude faster in inference and requires several times fewer resources and time to train.


Complexity ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-16
Author(s):  
Sebastian Wandelt ◽  
Xing Shi ◽  
Xiaoqian Sun

The analysis of real-world systems through the lens of complex networks often requires a node importance function. While many such views on importance exist, a frequently used global node importance measure is betweenness centrality, quantifying the number of times a node occurs on all shortest paths in a network. This centrality of nodes often significantly depends on the presence of nodes in the network; once a node is missing, e.g., due to a failure, other nodes’ centrality values can change dramatically. This observation is, for instance, important when dismantling a network: instead of removing the nodes in decreasing order of their static betweenness, recomputing the betweenness after a removal creates tremendously stronger attacks, as has been shown in recent research. This process is referred to as interactive betweenness centrality. Nevertheless, very few studies compute the interactive betweenness centrality, given its high computational costs, a worst-case runtime complexity of O(N∗∗4) in the number of nodes in the network. In this study, we address the research questions, whether approximations of interactive betweenness centrality can be obtained with reduction of computational costs and how much quality/accuracy needs to be traded in order to obtain a significant reduction. At the heart of our interactive betweenness approximation framework, we use a set of established betweenness approximation techniques, which come with a wide range of parameter settings. Given that we are interested in the top-ranked node(s) for interactive dismantling, we tune these methods accordingly. Moreover, we explore the idea of batch removal, where groups of top-k ranked nodes are removed before recomputation of betweenness centrality values. Our experiments on real-world and random networks show that specific variants of the approximate interactive betweenness framework allow for a speedup of two orders of magnitude, compared to the exact computation, while obtaining near-optimal results. This work contributes to the analysis of complex network phenomena, with a particular focus on obtaining scalable techniques.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Silvia Zaoli ◽  
Piero Mazzarisi ◽  
Fabrizio Lillo

AbstractBetweenness centrality quantifies the importance of a vertex for the information flow in a network. The standard betweenness centrality applies to static single-layer networks, but many real world networks are both dynamic and made of several layers. We propose a definition of betweenness centrality for temporal multiplexes. This definition accounts for the topological and temporal structure and for the duration of paths in the determination of the shortest paths. We propose an algorithm to compute the new metric using a mapping to a static graph. We apply the metric to a dataset of $$\sim 20$$ ∼ 20 k European flights and compare the results with those obtained with static or single-layer metrics. The differences in the airports rankings highlight the importance of considering the temporal multiplex structure and an appropriate distance metric.


2017 ◽  
Vol 4 (3) ◽  
pp. 187-200
Author(s):  
Dianne S. V. de Medeiros ◽  
Miguel Elias M. Campista ◽  
Nathalie Mitton ◽  
Marcelo Dias de Amorim ◽  
Guy Pujolle

Robotica ◽  
2022 ◽  
pp. 1-17
Author(s):  
Jie Liu ◽  
Chaoqun Wang ◽  
Wenzheng Chi ◽  
Guodong Chen ◽  
Lining Sun

Abstract At present, the frontier-based exploration has been one of the mainstream methods in autonomous robot exploration. Among the frontier-based algorithms, the method of searching frontiers based on rapidly exploring random trees consumes less computing resources with higher efficiency and performs well in full-perceptual scenarios. However, in the partially perceptual cases, namely when the environmental structure is beyond the perception range of robot sensors, the robot often lingers in a restricted area, and the exploration efficiency is reduced. In this article, we propose a decision-making method for robot exploration by integrating the estimated path information gain and the frontier information. The proposed method includes the topological structure information of the environment on the path to the candidate frontier in the frontier selection process, guiding the robot to select a frontier with rich environmental information to reduce perceptual uncertainty. Experiments are carried out in different environments with the state-of-the-art RRT-exploration method as a reference. Experimental results show that with the proposed strategy, the efficiency of robot exploration has been improved obviously.


2022 ◽  
Vol 14 (1) ◽  
pp. 0-0

In the domain of cyber security, the defence mechanisms of networks has traditionally been placed in a reactionary role. Cyber security professionals are therefore disadvantaged in a cyber-attack situation due to the fact that it is vital that they maneuver such attacks before the network is totally compromised. In this paper, we utilize the Betweenness Centrality network measure (social property) to discover possible cyber-attack paths and then employ computation of similar personality of nodes/users to generate predictions about possible attacks within the network. Our method proposes a social recommender algorithm called socially-aware recommendation of cyber-attack paths (SARCP), as an attack predictor in the cyber security defence domain. In a social network, SARCP exploits and delivers all possible paths which can result in cyber-attacks. Using a real-world dataset and relevant evaluation metrics, experimental results in the paper show that our proposed method is favorable and effective.


2006 ◽  
Vol DMTCS Proceedings vol. AG,... (Proceedings) ◽  
Author(s):  
Michael Drmota

International audience The purpose of this survey is to present recent results concerning concentration properties of extremal parameters of random discrete structures. A main emphasis is placed on the height and maximum degree of several kinds of random trees. We also provide exponential tail estimates for the height distribution of scale-free trees.


Author(s):  
Stefan Schmid ◽  
Nicolas Schnepf ◽  
Jiří Srba

AbstractTo ensure a high availability, communication networks provide resilient routing mechanisms that quickly change routes upon failures. However, a fundamental algorithmic question underlying such mechanisms is hardly understood: how to verify whether a given network reroutes flows along feasible paths, without violating capacity constraints, for up to k link failures? We chart the algorithmic complexity landscape of resilient routing under link failures, considering shortest path routing based on link weights as e.g. deployed in the ECMP protocol. We study two models: a pessimistic model where flows interfere in a worst-case manner along equal-cost shortest paths, and an optimistic model where flows are routed in a best-case manner, and we present a complete picture of the algorithmic complexities. We further propose a strategic search algorithm that checks only the critical failure scenarios while still providing correctness guarantees. Our experimental evaluation on a benchmark of Internet and datacenter topologies confirms an improved performance of our strategic search by several orders of magnitude.


2021 ◽  
Author(s):  
Yuhu Qiu ◽  
Tianyang Lyu ◽  
Xizhe Zhang ◽  
Ruozhou Wang

Network decrease caused by the removal of nodes is an important evolution process that is paralleled with network growth. However, many complex network models usually lacked a sound decrease mechanism. Thus, they failed to capture how to cope with decreases in real life. The paper proposed decrease mechanisms for three typical types of networks, including the ER networks, the WS small-world networks and the BA scale-free networks. The proposed mechanisms maintained their key features in continuous and independent decrease processes, such as the random connections of ER networks, the long-range connections based on nearest-coupled network of WS networks and the tendency connections and the scale-free feature of BA networks. Experimental results showed that these mechanisms also maintained other topology characteristics including the degree distribution, clustering coefficient, average length of shortest-paths and diameter during decreases. Our studies also showed that it was quite difficult to find an efficient decrease mechanism for BA networks to withstand the continuous attacks at the high-degree nodes, because of the unequal status of nodes.


2017 ◽  
Vol 2017 ◽  
pp. 1-9 ◽  
Author(s):  
Marek Šimon ◽  
Iveta Dirgová Luptáková ◽  
Ladislav Huraj ◽  
Marián Hosťovecký ◽  
Jiří Pospíchal

Usually, the existence of a complex network is considered an advantage feature and efforts are made to increase its robustness against an attack. However, there exist also harmful and/or malicious networks, from social ones like spreading hoax, corruption, phishing, extremist ideology, and terrorist support up to computer networks spreading computer viruses or DDoS attack software or even biological networks of carriers or transport centers spreading disease among the population. New attack strategy can be therefore used against malicious networks, as well as in a worst-case scenario test for robustness of a useful network. A common measure of robustness of networks is their disintegration level after removal of a fraction of nodes. This robustness can be calculated as a ratio of the number of nodes of the greatest remaining network component against the number of nodes in the original network. Our paper presents a combination of heuristics optimized for an attack on a complex network to achieve its greatest disintegration. Nodes are deleted sequentially based on a heuristic criterion. Efficiency of classical attack approaches is compared to the proposed approach on Barabási-Albert, scale-free with tunable power-law exponent, and Erdős-Rényi models of complex networks and on real-world networks. Our attack strategy results in a faster disintegration, which is counterbalanced by its slightly increased computational demands.


Sign in / Sign up

Export Citation Format

Share Document