Network Science (2005)
Chapter: 8 conclusions and recommendations, 8 conclusions and recommendations.
In this chapter the committee combines its findings into conclusions and offers recommendations. First, it collects the factual findings presented in Chapters 2 - 7 into three overarching conclusions concerning the importance of networks and the current state of knowledge about them. Next, it articulates specific conclusions that are directly responsive to Items 1 through 3 of the statement of task. Finally, in response to Item 4, the committee provides its recommendations, including for research initiatives. Box 8-1 summarizes how the report responds to the statement of task.
OVERARCHING CONCLUSIONS
Conclusion 1. Networks are pervasive in all aspects of life: biological, physical, and social. They are indispensable to the workings of a global economy and to the defense of the United States against both conventional military threats and the threat of terrorism.
Conclusion 1 was developed in Chapters 2 and 3 and summarized in Tables 2-1 , 2-2 , and 3-1 and the discussions surrounding them. It sets the stage for the committee’s inquiry into the state of knowledge about these networks.
Conclusion 2. Fundamental knowledge about the prediction of the properties of complex networks is primitive.
Given the pervasiveness and vital importance of networks, one might assume that a lot is known about them. As documented in Chapters 5 and 6 , however, this is not the case. Although the technology for constructing and operating engineered physical networks is sophisticated, critical questions about their robustness, stability, scaling, and performance cannot be answered with confidence without extensive simulation and testing. For large global networks, even simulations are often inadequate. The design and operation of network components (such things as computers, routers, or radios) are based on fundamental knowledge gleaned from physics, chemistry, and materials science. However, there is no comparable fundamental knowledge that allows the a priori prediction of the properties of complex assemblies of these components into networks. Indeed, such networks are expected to exhibit emergent behaviors—that is, behaviors that cannot be predicted or anticipated from the known behaviors of their components. In the case of social and biological networks, even the properties of the components are poorly known. A huge gap exists between the demand for knowledge about the networks on which our lives depend and the availability of that knowledge.
The committee learned that developing predictive models of the behavior of large, complex networks is difficult. There are relatively few rigorous results to describe the scaling of their behaviors with increasing size. Surprisingly, this is true for common engineered networks like the Internet as well as for social and biological networks.
Simulation rather than analysis is the research tool of choice. In the case of social networks, even simulation is vastly complicated by the diversity and complexity of the agents that are the nodes of the networks—humans or groups of humans “in the wild.” Which of their many properties are relevant for developing mathematical models of a particular phenomenon? Existing models of social networks, moreover, represent highly simplified situations and not necessarily ones that are relevant to the Army or network-centric warfare.
Finally, the notion of using network models in biology is relatively new. Controversy swirls around their utility, indeed around that of systems biology itself. In spite of a burgeoning literature on the structure of simple networks, the advancement of the field to allow relating basic scientific results to applications of societal and military interest still lies mostly in the future.
Conclusion 3. Current funding policies and priorities are unlikely to provide adequate fundamental knowledge about large complex networks.
Fundamental knowledge is created and stockpiled in disciplinary environments, mostly at universities, and then used as required by (vertically integrated) industries to provide the products and services required by customers, including the military. This fundamental knowledge is different in kind from empirical knowledge gleaned during the development of technology and products. You get what you measure. Suppliers of fundamental knowledge measure publications, presentations, students supervised, awards received, and other metrics associated with individual investigators. The knowledge accumulates along traditional disciplinary lines because this is where the rewards are found. Large team activities are
relatively rare (except in medicine and large-scale physics experiments) and are mostly left to the consumers of the fundamental knowledge, who must supplement the fundamental knowledge generously with empirical knowledge to convert it into the goods and services desired by the paying customer.
This scheme worked marvelously for more than a half a century, when the United States dominated the world and industries were vertically integrated. With the onset of the global economy in the 1990s, however, the situation began to change dramatically, for a number of reasons. First, knowledge, investment capital, technology, and technical labor are becoming globally available commodities. Second, economic activity, including R&D, is becoming global in scale. Third, these two trends are making the networks on which we depend ever larger and more complex and their susceptibility to disruption ever greater.
This traditional scheme does not work well for generating knowledge about global networks, because focused, coordinated efforts are needed. Thus, there is a huge difference between the social and financial arrangements needed to gain fundamental knowledge about large, complex networks in a global environment and the arrangements that worked so well to provide such knowledge for the design and production of smaller, less complex entities in a national environment. Any successful effort to create the knowledge necessary to secure robust, reliable scalable global networks must come to grips with this reality.
Overall, the committee is led to a view of networks as pervasive in and vital to modern society, yet understood only as well as the solar system was understood in Ptolemy’s time. The military has made networks the centerpiece of its transformation effort without a methodology to design networks in the physical and information domains in a predictive way for network-centric operations (NCO). Further, according to the DOD Office of Force Transformation, research in the cognitive and social domains has yet to yield advances comparable to the technological developments in the information domain. At the same time, current efforts by academia to describe networks are fragmented and disjointed. Relatively little of the current research on networks promises to create a science of networks that will generate knowledge adequate to meet the demand.
In short, there is a massive disconnect between the importance of networks in modern society and military affairs on the one hand and, on the other, the support of coherent R&D activities that would raise current network technologies and capabilities to the next level. The Army alone cannot transform this situation, but it can make a beginning.
SPECIFIC CONCLUSIONS
Items 1 and 2 in the statement of task inquire into the appropriateness of a field of investigation called network science and its definition, content, and the research challenges that would characterize it. Elements of a field of network science have begun to emerge in different disciplines spanning engineering, biological, and social networks. The emerging field is concerned with the development and analysis of network representations to create predictive models of observed physical, biological, and social phenomena.
The remarkable diversity and pervasiveness of network ideas renders the study of network science a highly leveraged topic for both civilian and military investment. The provisional consensus around its core content clearly defines the notion of network science. By making an investment in network science, the Army could forge a single approach to a diverse collection of applications.
Conclusion 4. Network science is an emerging field of investigation whose support will address important societal problems, including the Army’s pursuit of network-centric operations capabilities.
Although the boundaries of network science are fuzzy, there is broad agreement on key topics that should be included within the field, the types of tools that must be developed, and the research challenges that should be investigated. These were documented in Chapters 3 and 4 .
Conclusion 5. There is a consensus among the practitioners of research on networks for physical, biological, social, and information applications on the topics that make up network science.
Responses to its questionnaire greatly assisted the committee in determining “the key research challenges to enable progress in network science.” These responses establish that there is a fair degree of consensus on these challenges across practitioners in diverse applications areas.
Conclusion 6. There are seven major research challenges the surmounting of which will enable progress in network science:
Dynamics, spatial location, and information propagation in networks. Better understanding of the relationship between the architecture of a network and its function is needed.
Modeling and analysis of very large networks. Tools, abstractions, and approximations are needed that allow reasoning about large-scale networks, as well as techniques for modeling networks characterized by noisy and incomplete data.
Design and synthesis of networks. Techniques are needed to design or modify a network to obtain desired properties.
Increasing the level of rigor and mathematical structure. Many of the respondents to the questionnaire felt that the current state of the art in network
science did not have an appropriately rigorous mathematical basis.
Abstracting common concepts across fields. The disparate disciplines need common concepts defined across network science.
Better experiments and measurements of network structure. Current data sets on large-scale networks tend to be sparse, and tools for investigating their structure and function are limited.
Robustness and security of networks. Finally, there is a clear need to better understand and design networked systems that are both robust to variations in the components (including localized failures) and secure against hostile intent.
These challenges are elaborated in terms of specific research issues and their theoretical, experimental, and practical difficulties in Chapter 7 and Appendix E within the framework of exploring various investment scenarios. The scenarios respond to Item 3 in the statement of task.
Although all the military services have a vision of the future in which engineered communications networks play a fundamental role, there is no methodology for ensuring that these networks are scalable, reliable, robust, and secure. Of particular importance is the ability to design networks whose behaviors are predictable in their intended domains of applications. This also is true in the commercial sphere. Creation of such a methodology is an especially pressing task because global commercial networks can also be exploited by criminal and terrorist social networks.
Conclusion 7. The high value attached to the efficient and failure-free operation of global engineered networks makes their design, scaling, and operation a national priority.
RECOMMENDATIONS
The statement of task requests investment recommendations from the committee. Options for these recommendations are explored in Chapter 7 and Appendix E . The committee documents in Chapters 2 and 3 that the impact of networks on society transcends their impact on military applications, although both are vital aspects of the total picture. Chapters 3 and 4 explain that the current state of knowledge about networks does not support the design and operation of complex global networks for current military, political, and economic applications. Advances in network science are essential to developing adequate knowledge for these applications.
Recommendation 1. The federal government should initiate a focused program of research and development to close the gap between currently available knowledge about networks and the knowledge required to characterize and sustain the complex global networks on which the well-being of the United States has come to depend.
This recommendation is buttressed by centuries of evidence that disruptive social networks (e.g., terrorists, criminals) learn to exploit evolving infrastructure networks (e.g., communications or transportation) in ways that the creators of these networks did not anticipate. The global war on terrorism, which is a main driver of military transformation, is only one recent manifestation of this general pattern. Society has the same need in other areas, such as control of criminal activities perpetrated using the global airline and information infrastructures. Addressing problems resulting from the interaction of social and engineered networks is an example of a compelling national issue that transcends the transformation of the military and that is largely untouched by current research on networks.
Within this broad context, Recommendations 1a, 1b, and 1c provide the Army with three options:
Recommendation 1a. The Army, in coordination with other federal agencies, should underwrite a broad network research initiative that includes substantial resources for both military and nonmilitary applications that would address military, economic, criminal, and terrorist threats.
The Army can lead the country in creating a base of network knowledge that can support applications for both the Army and the country at large. Maximum impact could be obtained by a coordinated effort across a variety of federal agencies, including the DOD and the Department of Homeland Security, to create a focused national program of network research that would develop applications to support not only NCO but also countermeasures against international terrorist and criminal threats.
Alternatively, if the Army is restricted to working just with the DOD, it should initiate a focused program to create an achievable vision of NCO capabilities across all the services.
Recommendation 1b. If the Army wants to exploit fully applications in the information domain for military operations in a reasonable time frame and at an affordable cost, it should champion the initiation of a high-priority, focused DOD effort to create a realizable vision of the associated capabilities and to lay out a trajectory for its realization.
Finally, if the Army elects to apply the insight from the committee primarily to its own operations, it can still provide leadership in network science research.
TABLE 8-1 Network Research Areas
Recommendation 1c. The Army should support an aggressive program of both basic and applied research to improve its NCO capabilities.
Specific areas of research of interest to the Army are shown in Table 8-1 . This table expresses the committee’s assessment of the relative priorities for these areas, the time frames in which one might reasonably expect them to be consummated as actionable technology investment options, and the degree of commercial interest in exploiting promising options. Specific research problems and sample projects are given in Appendix E . The committee notes that both trained personnel and promising research problems exist in many of these areas, so the Army should be able to create a productive program readily.
By selecting from among Recommendations 1a through 1c an option that is ambitious yet achievable, the Army can lead the country in creating a base of knowledge emanating from network science that is adequate to support applications on which both the Army and the country at large depend. Regardless of which option (or options) are adopted, Army initiatives in network science should be grounded in basic research.
Recommendation 1d. The initiatives recommended in 1, 1a, 1b, and 1c should include not only theoretical studies but also the experimental testing of new ideas in settings sufficiently realistic to verify or disprove their use for intended applications.
Recommendations 1, 1a, 1b, and 1c span only part of the investment opportunity space—namely, those segments of the space described in Scenarios 2 and 3 in Chapter 7 and Appendix E . They will involve substantial changes in how the Army invests its R&D dollars and in how it plans and manages these investments.
The Army also has the opportunity associated with Scenario 1 in Chapter 7 , which involves funding a small program of basic research in network science. This investment of relatively small amounts of Army risk capital funds would create a base of knowledge and personnel from which the Army could launch an attack on practical problems that arise as it tries to provide NCO capabilities.
Investments in basic (6.1) research in network science can generate significant value; however, the committee wants to be crystal clear that such investments have no immediate prospects of impacting the design, testing, evaluation, or sourcing of NCO capabilities. They would create additional knowledge that builds the core content of network science, and they would train researchers who could also be recruited by the Army for later efforts. While the knowledge generated would probably be less valuable than in the case
of Scenarios 2 and 3, the cost is less and implementation can be immediate.
If the Army elects to exploit Scenario 1, the committee offers the following two further recommendations:
Recommendation 2. The Army should make a modest investment of at least $10 million per year to support a diverse portfolio of basic (6.1) network research that promises high leverage for the dollars invested and is clearly different from existing investments by other federal agencies like the National Science Foundation (NSF), the Department of Energy (DOE), and the National Institutes of Health (NIH).
This modest level of investment is compatible with the Army’s current R&D portfolio. There is an adequate supply of promising research topics and talented researchers to make this investment productive. Additionally, it can be implemented within the Army’s current R&D management work processes, although some enhancements along the lines noted in Chapter 7 and Appendix E would improve the return on this investment.
To identify the topics in basic network science research that would bring the most value to NCO, the committee recalls that the open system architectures for computer networks consist of layers, each of which performs a special function regarded as a “service” by the layers above. It is useful to distinguish among the lower (physical and transport) layers of this architecture, the higher (applications) layers that are built on top of them to offer services to the people, and the cognitive and social networks that are built higher still, on top of the services-to-humans layers.
Research on the lower layers of the network architecture is relatively mature. Improving the services offered at these levels is more of an engineering problem than one requiring basic research. The most immediate payoffs from network science are likely to result from research associated with the upper levels of the network architecture and the social networks that are built at an even higher level upon their outputs. This is where the committee thinks that Army investments are most likely to create the greatest value.
An area of particular promise that has little or no current investment is the social implications of NCO for the organizational structure and command and control. Basic research could provide valuable insight into how military personnel use advanced information exchange capabilities to improve combat effectiveness. For example, one might study how troops in combat could use these capabilities to make better decisions. Additional basic research in the core content of network science might help to determine how the Army can most productively utilize the capabilities of its advanced information infrastructure.
Recommendation 3. The Army should fund a basic research program to explore the interaction between information networks and the social networks that utilize them.
The Army can implement Recommendations 2 and 3 within the confines of its present policies and procedures. They require neither substantial replanning nor the orchestration of joint Army/university/industry research projects. They create significant value and are actionable immediately.
The committee’s Recommendations 1, 1a through 1d, 2, and 3 give the Army an actionable menu of options that span the opportunity space available. By selecting and implementing appropriate items from this menu, the Army can develop a robust network science to “enable progress toward achieving Network-Centric Warfare capabilities,” as requested in the statement of task.
This page intentionally left blank.
The military is currently attempting to develop itself into a force capable of networkcentric operations. While this effort has highlighted the military's dependence on interacting networks, it has also shown that there is a huge gap between what we need to know about networks and our fundamental knowledge about network behavior. This gap makes the military vision of NCO problematic. To help address this problem, the Army asked the NRC to find out whether identifying and funding a new field of "network science" could help close this gap. This report presents an examination of networks andthe military, an analysis of the promise, content, and challenges of network science, and an assessment of approaches to creating value from network science.
READ FREE ONLINE
Welcome to OpenBook!
You're looking at OpenBook, NAP.edu's online reading room since 1999. Based on feedback from you, our users, we've made some improvements that make it easier than ever to read thousands of publications on our website.
Do you want to take a quick tour of the OpenBook's features?
Show this book's table of contents , where you can jump to any chapter by name.
...or use these buttons to go back to the previous chapter or skip to the next one.
Jump up to the previous page or down to the next one. Also, you can type in a page number and press Enter to go directly to that page in the book.
Switch between the Original Pages , where you can read the report as it appeared in print, and Text Pages for the web version, where you can highlight and search the text.
To search the entire text of this book, type in your search term here and press Enter .
Share a link to this book page on your preferred social network or via email.
View our suggested citation for this chapter.
Ready to take your reading offline? Click here to buy this book in print or download it as a free PDF, if available.
Get Email Updates
Do you enjoy reading reports from the Academies online for free ? Sign up for email notifications and we'll let you know about new publications in your areas of interest when they're released.
Networking is central to modern computing, from WANs connecting cell phones to massive data stores, to the data-center interconnects that deliver seamless storage and fine-grained distributed computing. Because our distributed computing infrastructure is a key differentiator for the company, Google has long focused on building network infrastructure to support our scale, availability, and performance needs, and to apply our expertise and infrastructure to solve similar problems for Cloud customers. Our research combines building and deploying novel networking systems at unprecedented scale, with recent work focusing on fundamental questions around data center architecture, cloud virtual networking, and wide-area network interconnects. We helped pioneer the use of Software Defined Networking, the application of ML to networking, and the development of large-scale management infrastructure including telemetry systems. We are also addressing congestion control and bandwidth management, capacity planning, and designing networks to meet traffic demands. We build cross-layer systems to ensure high network availability and reliability. By publishing our findings at premier research venues, we continue to engage both academic and industrial partners to further the state of the art in networked systems.
Recent Publications
Some of our teams.
Cloud networking
Global networking
Network infrastructure
We're always looking for more talented, passionate people.
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
- View all journals
- My Account Login
- Explore content
- About the journal
- Publish with us
- Sign up for alerts
- Open access
- Published: 26 November 2019
Comparing methods for comparing networks
- Mattia Tantardini 1 ,
- Francesca Ieva 2 , 3 ,
- Lucia Tajoli 4 &
- Carlo Piccardi 5
Scientific Reports volume 9 , Article number: 17557 ( 2019 ) Cite this article
64k Accesses
142 Citations
9 Altmetric
Metrics details
- Complex networks
- Computational science
With the impressive growth of available data and the flexibility of network modelling, the problem of devising effective quantitative methods for the comparison of networks arises. Plenty of such methods have been designed to accomplish this task: most of them deal with undirected and unweighted networks only, but a few are capable of handling directed and/or weighted networks too, thus properly exploiting richer information. In this work, we contribute to the effort of comparing the different methods for comparing networks and providing a guide for the selection of an appropriate one. First, we review and classify a collection of network comparison methods, highlighting the criteria they are based on and their advantages and drawbacks. The set includes methods requiring known node-correspondence, such as DeltaCon and Cut Distance, as well as methods not requiring a priori known node-correspondence, such as alignment-based, graphlet-based, and spectral methods, and the recently proposed Portrait Divergence and NetLSD. We test the above methods on synthetic networks and we assess their usability and the meaningfulness of the results they provide. Finally, we apply the methods to two real-world datasets, the European Air Transportation Network and the FAO Trade Network, in order to discuss the results that can be drawn from this type of analysis.
Similar content being viewed by others
Metrics for network comparison using egonet feature distributions
Degree difference: a simple measure to characterize structural heterogeneity in complex networks
Network mutual information measures for graph similarity
Introduction.
The research on complex networks has exploded in the last two decades, thanks to the great power and flexibility of network models in describing real-world applications coming from very diverse scientific and technological areas, including social sciences, economics, biology, computer science, telecommunications, transportation, and many others 1 , 2 , 3 . A typical data-driven study begins with modelling or interpreting real-world data with a network, and then continues by exploiting a broad set of algorithmic tools, in order to highlight a number of network features from the local scale (nodes and edges) to the global scale (the overall network), passing through an intermediate (meso-)scale where the important structure and role of suitable subnetworks is possibly evidenced 4 , 5 .
With the dramatic growth of available data, and the corresponding growth of network models, researchers often face the problem of comparing networks, i.e., finding and quantifying similarities and differences between networks. In whatever application area, it is straightforward to find relevant examples where this problem is truly of interest: in international economics, one may want to compare the trade structure of different product categories; in transportation, the flight networks of different airlines; in biology, the interaction structure of different protein complexes; in social media, the propagation cascade of news; etc. Or, when temporally-stamped data are available, one may want to spot anomalous instants in a temporal series of graphs–e.g., an event that abruptly modifies the connection pattern among the users of a social media. Defining metrics for quantifying the similarity/difference between networks opens the door to clustering, i.e., grouping networks according to their similarity: for example, given a set of brain networks derived from NMR data, find whether and how ill-subjects’ networks are different from healthy-subjects’ ones.
To perform network comparison, a measure of the distance between graphs needs to be defined. This is a highly non trivial task, which requires a trade off among effectiveness of the results, interpretability, and computational efficiency, all features that, in addition, are often sensitive to the specific domain of application. Not surprisingly, the literature on this topic is extremely abundant, and plenty of different methods have been proposed. Yet much work has still to be done in terms of their comparative analysis, namely classifying which methods are the best and in which situations, and understanding the insights they can provide when used to compare a set of real-world networks. A few critical reviews of the literature on this subject have already been compiled (e.g., Soundarajan et al . 6 , Emmert-Streib et al . 7 , and Donnat and Holmes 8 ). Our aim is to complement these works by systematically testing many of the available methods in a comparative fashion.
This paper is organised as follows. In the next section, we discuss the general problem of network comparison and we recall the definition and main features of many of the methods available in the literature, from naїve approaches to a few of the most recent ones. This gives an overview of the state-of-the-art tools and of the strategies for network comparison. Then we restrict our attention to a subset of such methods, and we analyse them with a suitably designed battery of tests on synthetic networks, in order to highlight pros and cons of each method. We subsequently illustrate two examples of the use of network comparison methods on real-world data, specifically the European Air Transportation Network and the FAO Trade Network. We conclude by summarizing the results of our work and by highlighting which methods reveal to be the most effective and in which situations. All the analyses are carried out using the codes made available by the authors of each method (see the Supplementary Information file, Sec. S2 , for details).
Measuring the Distance Between Networks
The network comparison problem derives from the graph isomorphism problem. Two (undirected, unweighted) graphs G 1 ( V 1 , E 1 ) and G 2 ( V 2 , E 2 ) are isomorphic if there exists a one-to-one correspondence Φ mapping the node set V 1 onto V 2 such that the edge ( u , v ) ∈ E 1 if and only if the edge (Φ( u ), Φ( v )) ∈ E 2 9 . The complexity of the graph isomorphism problem , i.e., checking whether two finite graphs are isomorphic, is unknown in rigorous terms 10 , 11 , 12 : nonetheless, efficient algorithms exist for many classes of graphs 13 . In any case, isomorphism is an exact graph matching : if used as a distance for comparison, it gives a binary outcome: the graphs are either isomorphic, i.e., identical, or not. This information is poor, however, because networks are almost never identical in applications, and one is interested in assessing to what extent they are similar. To effectively compare networks, we need to move to inexact graph matching , i.e., define a real-valued distance which, as a minimal requirement, has the property of converging to zero as the networks approach isomorphism.
Searching for accurate and effective tools to compare networks has pushed the research in many different directions, leading to a wide variety of methods and algorithms. We present a short review of several of the most used approaches to network comparison, acknowledging that the literature is very abundant and we cannot cover it exhaustively neither want to repeat already established results. Following previous approaches 6 , we partition the comparison methods based on whether the induced distances are dependent from the correspondence of nodes. In the former case– Known Node-Correspondence ( KNC )–the two networks have the same node set (or at least a common subset), and the pairwise correspondence between nodes is known. Thus, typically, only graphs of the same size and coming from the same application domain can be compared. In the latter case– Unknown Node-Correspondence ( UNC )–ideally any pair of graphs (even with different sizes, densities, or coming from different application fields) can be compared: typically these methods summarize the global structure into one or more statistics, which are then elaborated to define a distance. This latter notion of distance thus reflects the difference in the global structure of the networks.
For example, imagine that one wants to compare the European air transportation networks of the airlines A and B. The node sets (i.e., the European airports) are the same, thus a KNC method can be applied to quantify to what extent the two sets of edges are similar, i.e., to what extent the two airlines offer the same set of flights. If the exercise is (pairwise) extended to all airlines, the overall results will allow one to cluster airlines supplying similar sets of connections. But the same dataset could alternatively be analysed, with a different aim, by a UNC method, to highlight pairs of airlines whose network is globally structurally similar. Then, for example, one might discover that airlines A and B have both a markedly star-like flight network, but the first is based in Amsterdam (the centre of the star) and the second in Berlin. Here, extending the analysis to all airlines might cluster sets of airlines with similar strategies or business models.
A screening of the literature reveals that the latter problem is far more studied–and that biology is the most popular application field–so that the number of available UNC methods is much larger than that of the KNC ones. Below we present several methods used for network comparison, briefly explaining the details of their approach. In doing that, we will mostly privilege methods of general use, that is, applicable to directed and weighted networks too–this significantly narrows the set of candidates. In the next section, we will numerically compare the performance of a subset of these methods.
Known node-correspondence (KNC) methods
Difference of the adjacency matrices.
The simplest and näive measures are obtained by directly computing the difference of the adjacency matrices of the two networks. Then any matrix norm can be used, e.g., Euclidean, Manhattan, Canberra, or Jaccard 14 . All of them are suitable to compare all kinds of graphs (directed or not, weighted or not), with the exception of the Jaccard distance which needs to be extended to the Weighted Jaccard distance 15 (the definitions of the four norms are recalled in the Supplementary Information file, Sec. S1 ). Although this straightforward approach is rarely used in network comparison, we include it in the pool and consider it as a baseline approach.
It is based on the comparison of the similarities between all node pairs in the two graphs 16 , 17 . The similarity matrix of a graph is defined by S = [ s ij ] = [ I + ε 2 D − εA ] −1 , where A is the adjacency matrix, D = diag( k i ) is the degree matrix, k i is the degree of node i , and ε > 0 is a small constant. The rationale of the method is that just measuring the overlap of the two edge sets might not work in practice, because not all edges have the same importance. Instead, the difference between r -step paths, r = 2, 3, …, provides a much more sensitive measure. As a matter of fact, it can be shown 16 that s ij depends on all the r -paths connecting ( i , j ). The DeltaCon distance between the N × N similarity matrices S 1 = [ s ij 1 ] and S 2 = [ s ij 2 ] is finally defined using the Matusita distance:
Equation ( 1 ) assures that DeltaCon satisfies the usual axioms of distances. Moreover, it can be shown 16 , 17 that it also satisfies a few desirable properties regarding the impact of specific changes. Such properties are: changes leading to disconnected graphs are more penalized; in weighted graphs, the bigger the weight of the removed edge is, the greater the impact on the distance; a change has more impact in low density graphs than in denser graphs with equal size; random changes produce smaller impacts than targeted ones.
The computational complexity of the DeltaCon algorithm is quadratic in the number of nodes. To improve execution speed, an approximated version was proposed, which restricts the computation of the similarity matrices to groups of randomly chosen nodes 16 : this version has linear complexity in the number of edges and groups. Finally, DeltaCon was extended 17 to make it able to find which nodes or edges are most responsible for the differences between the two graphs.
Cut distance
This method 12 is based on the notion of cut weight, which is standard in graph theory and also used in community detection 4 . Given a (possibly directed, weighted) graph G = ( V , E ) with edge weights w ij , i , j ∈ V , and two disjoint node sets S , T ⊂ V , the cut weight is defined as \({e}_{G}(S,T)={\sum }_{i\in S,j\in T}\,{w}_{ij}\) , i.e., the total weight of the edges crossing the cut from S to T . The cut distance between two graphs G 1 ( V , E 1 ) and G 2 ( V , E 2 ) with the same node set is then defined as
where S C = V \ S . Thus two networks are similar if they have similar cut weight for all possible network bipartitions. The maximization is performed through genetic algorithms, which makes the comparison of large networks (thousands of nodes and larger) unfeasible. On the other hand, this is one of the few methods able to compare directed, weighted graphs.
Unknown node-correspondence (UNC) methods
Global statistics.
Simple metrics can be obtained by comparing the value of network statistics, such as the clustering coefficient 18 , 19 , 20 , 21 , the diameter 19 , 21 , or the average distance 18 . Although being intuitive and typically computationally efficient, often this approach does not yield robust results. As a matter of fact, similar values of network statistics do not necessarily imply similar network structures (e.g., see the discussion in ref. 22 ) and indeed, the comparison often fails in catching important local features. On the other hand, these simple metrics offer a computationally low-cost alternative that can be useful for a first analysis.
Mesoscopic response functions (MRFs)
This method exploits the information carried by the mesoscopic properties of the networks, i.e., their modular structure 23 . Three functions–called MRFs–are defined, the Hamiltonian H ( λ ), the partition entropy S ( λ ), and the number of communities η ( λ ), which describe the properties of a given network at different mesoscopic scales: the parameter λ tunes the fragmentation of the network into communities. The network distance is defined for a given set of graphs: for each network pair, the distances between corresponding MRFs are defined by standard function metrics, then the first principal component obtained from PCA is taken as distance. This entails non-comparability among different datasets, since the distance between two networks depends on the dataset they are part of. On the other hand, it is the only available method based on mesoscale properties, and allows one to consider both weighted and unweighted undirected networks. The computational efficiency of the method mostly depends on the efficiency of the community detection algorithm used.
Graphlet-based methods
Graphlets are small, connected, non-isomorphic subgraphs of large networks. Originally proposed for undirected (unweighted) networks 18 , their use has been subsequently extended to directed networks 24 , 25 . They encode important information about the structure of the network and provide a valuable tool for comparison. The different types of graphlets need to be enumerated, and this can be done in two ways, i.e., by taking into account or not their automorphism orbits 22 , which differentiate the roles of the nodes in each graphlet (see Fig. 1 , where graphlets are enumerated from G 0 to G 29 and orbits from O 0 to O 72 ). Usually graphlets with more than five nodes are not considered, both for computational reasons and due to repetition of smaller graphlets within their structure.
Graphlets (2- to 5-node) in undirected, unweighted networks (from ref. 22 ). The 30 graphlets defined by Pržulj et al . 18 are labeled G 0 to G 29 . In each graphlet, nodes with the same shading belong to the same automorphism orbit O 0 to O 72 , i.e., they have the same characteristics and are indistinguishable to each other 22 .
Counting all graphlets of a network is, in principle, a very demanding task: given a graph with N nodes and L edges, the worst-case running time for counting 2- to k -node graphlets (for both the undirected and directed case) with a complete enumeration strategy is O ( N k ): a tighter upper bound gives O ( Nk max k −1 ), where k max is the maximum node degree of the graph 20 , 24 . In practice, these pessimistic bounds are never reached: thanks to the sparsity of real-world networks, and exploiting wiser counting strategies, large improvements are possible. Hočevar and Demšar 26 proposed the ORCA algorithm, based on a particular counting strategy. Its complexity is O ( Lk max + T 4 ) for the enumeration of 2- to 4-node graphlets, and O ( Lk max 2 + T 5 ) for 2- to 5-node graphlets, where T 4 and T 5 are terms which are negligible in most situations. Aparicio, Ribeiro and Silva 25 proposed another approach based on a particular data structure, the G-Trie 27 : it demonstrated higher performances with respect to ORCA, but its theoretical upper bound is not provided.
Graphlet-based network distances are based on graphlet counts, which can be organized in several ways:
Relative Graphlets Frequency Distance ( RGFD ) 18 . The 29 graphlets from 3 to 5 nodes are counted in each network. Then the distance is defined as \(d({G}_{1},{G}_{2})={\sum }_{i=1}^{29}\,|{F}_{i}({G}_{1})-{F}_{i}({G}_{2})|\) , where F i ( ⋅ ) denotes the count of graphlet i normalized with respect to the total number of graphlets in the graph.
Graphlet Degree Distribution Agreement ( GDDA ) 22 . The 73 automorphism orbits of graphlets from 2 to 5 nodes are counted in each network G . For each orbit j = 0, 1, …, 72, the graphlet degree distribution (GDD) d G j ( k ), which is the number of nodes in G touching k times that orbit, is computed. This quantity is first scaled as d G j ( k )/ k , and then normalized by the total area T G j under the j -th GDD, obtaining N G j ( k ) = ( d G j ( k )/ k )/ T G j . Then, the agreement of the j -th GDD between networks G 1 and G 2 is defined as
and the final GDDA distance is taken as the geometric or arithmetic mean of all the 73 agreements A j ( G 1 , G 2 ).
Graphlets Correlation Distance ( GCD ): Yaveroglu et al . 19 investigated the dependencies among graphlets and found that some orbits are redundant, i.e., their count can actually be obtained from the counts of other orbits. Discarding the redundant orbits led to the definition of a more sensible and efficient measure. For instance, graphlets up to 4 nodes have 4 redundant orbits, and discarding them reduces the number of orbits to 11 from the original 15. For a given N -node network G , the N graphlets degree vectors 28 , i.e., the count of the considered orbits that each node touches, are appended row by row to form a N × 11 matrix. Then, the Spearman’s correlation coefficient is computed between all column pairs, obtaining the 11 × 11 Graphlet Correlation Matrix GCM G . The GCD distance between graphs G 1 and G 2 is finally defined as the Euclidean distance between the upper triangular parts of the matrices GCM G 1 and GCM G 2 . Yaveroglu et al . showed that GCD-11 (the distance with non-redundant orbits) outperformed GCD-15 (with redundant orbits), but also GCD-73 and GCD-56, the distances based on 5-node graphlets with or without reduntant orbits, respectively. Also, GCD-11 performed better than other graphlet-based distances in recognizing different network models.
NetDis 29 : It compares graphlet counts in overlapping neighbourhoods of nodes, rather than in the entire network: more specifically, it considers the 2-step ego-network of each node. The rationale comes from the observation that network size and density strongly influence global graphlet counts, and that such effect can be attenuated by restricting to local subnetworks. Graphlet counts (from G 0 to G 29 ) are computed for each ego-network and then normalized with respect to the expected counts from a null model. Denote by S w ( G ) the sum over all ego-networks of the normalized count of graphlet w in graph G . Then, for a given size k ∈ {3, 4, 5} of the graphlets:
where M ( k ) is a normalizing constant forcing netD 2 S ( k ) ∈ [−1, 1]. Finally, the NetDis distance for k -node graphlets is defined as
Note that the NetDis measure actually depends on k , which is therefore a parameter to be selected. Yaveroglu et al . 20 pointed out a few critical aspects of the method, such as the choice of a null model, the computational efficiency, and the performances, which are overall inferior to those of other graphlet-based distances.
GRAFENE 21 . In this method, the graphlet degree vectors for graphlets G 0 to G 29 are first computed for each node, and then scaled in [0, 1] dividing each component by the total counts of the corresponding graphlet in the whole network. Principal Component Analysis is then performed over the rescaled graphlet degree vectors, and the first r components that account for at least 90% of the total variability are kept. The distance between the two networks is defined as 1 − d cos ( R 1 , R 2 ), where d cos is the cosine similarity and R 1 , R 2 are the first r principal components for the two graphs. The use of PCA is a novel idea within the graphlet-based methods, that improves the quality of results and the computational performances. Tests on synthetic networks showed that GRAFENE performs at least as well as the other alignment-free methods, and outperforms all other methods on real networks 21 .
Multiple approaches 24 , 25 have extended the applicability of graphlet-based methods introducing directed graphlets , with the aim of comparing directed (yet unweighted) networks. This allowed to define directed versions of a few of the existing distances 24 : the Directed Relative Frequency Distance ( DRGFD ), the Directed Graphlet Degree Distribution Agreement ( DGDDA ) and the Directed Graphlets Correlation Distance ( DGCD ).
Alignment-based methods
These methods create a mapping between the nodes of two graphs ( alignment procedure) trying to maximize an objective function that captures the quality of matching. They originated in computational biology, where a number of algorithms have been proposed (e.g., refs. 30 , 31 , 32 ): most of them, however, rely on the biological information associated to the network and, as such, have not general applicability. The GRAAL (Graph Aligner) family (GRAAL 33 , H- 34 , MI- 11 , C- 35 , and L-GRAAL 36 ), on the contrary, performs network alignment based on the network topology (except the most recent L-GRAAL, which can exploit both biology and topology). Given the graphs G 1 = ( V 1 , E 1 ) and G 2 = ( V 2 , E 2 ) with | V 1 | ≤ | V 2 |, a node-mapping f : V 1 → V 2 is defined, which induces an edge-mapping g : V 1 × V 1 → V 2 × V 2 such that g ( E 1 ) = {( f ( u ), f ( v )): ( u , v ) ∈ E 1 }. The objective function that measures the quality of the node alignment f is then expressed by the edge correctness
i.e., the fraction of edges in E 1 aligned to edges in E 2 . Solving the alignment problem boils out to finding f that maximizes EC . The GRAAL family proposes a number of different approaches: GRAAL and H-GRAAL assign a similarity score to each pair ( u 1 , u 2 ) ∈ V 1 × V 2 based on the graphlet degrees (see above) of the two nodes, thus accounting for the similarity of their neighbourhoods. Then for the alignment GRAAL uses a greedy “seed-and-extend” algorithm trying to maximize the aggregate similarity, while H-GRAAL solves the same problem as an optimal assignment (Hungarian algorithm). In MI-GRAAL a confidence score, computed by taking into account various node statistics (e.g., degree, clustering coefficient, etc.) is assigned to each pair ( u 1 , u 2 ) ∈ V 1 × V 2 , then nodes are aligned starting from the pairs with highest score. The method is customizable, since any node statistics can be used to compute the confidence score–this allows one the comparison of directed and weighted networks. C-GRAAL does not require an explicit node similarity measure (which however can be incorporated, if available) but works on networks topology with an iterative common-neighbors-based algorithm. Finally, L-GRAAL optimizes a novel objective function that takes into account both sequence-based protein conservation and graphlet-based interaction conservation, by using an alignment heuristic based on integer programming and Lagrangian relaxation. The main drawback of these methods is their computational efficiency, which scales at least quadratically with the number of nodes.
Spectral methods
Here the rationale is that, since the spectrum of the representation matrix of a network (adjacency or Laplacian matrix) carries information about its structure, comparing spectra provides metrics for comparing networks. Different approaches are used: Wilson and Zhu 37 proposed to simply take the Euclidean distance between the two spectra, while Gera et al . 38 proposed to take as distance the p -value of a nonparametric test assessing whether the two spectra come from the same distribution. Despite the ease of use, spectral methods prove to suffer from many drawbacks, including cospectrality between different graphs, dependence on the matrix representation, and abnormal sensitivity (small changes in the graph’s structure can produce large changes in the spectrum).
NetLSD (Network laplacian spectral descriptor)
This method 39 summarizes the features of the (undirected, unweighted) graph G by a vector derived from the solution of the “heat equation” ∂ u t /∂ t = − Lu t , where u t is an N -dimensional vector and L = I − D −1/2 AD −1/2 is the normalized Laplacian matrix. This resembles the dynamics of a continuous-time random walker (e.g., ref. 40 ) so that the solution is a pagerank-like centrality 41 . L is symmetric and can be written as L = ΦΛΦ T via spectral decomposition, hence the closed-form solution is given by the N × N “heat kernel” matrix
whose entry ( H t ) ij is the heat transferred from node i to j at time t . NetLSD condenses the graph representation in the heat trace signature
The continuous-time function h t is finally transformed into a finite-dimensional vector by sampling over a suitable time interval, and the distance between two networks G 1 , G 2 is taken as the norm of the vector difference between h ( G 1 ) and h ( G 2 ). The competitive performance of NetLSD is demonstrated in a few machine learning classification tasks. The time complexity is O ( N 3 ), if the full eigendecomposition of the Laplacian is carried out. This would limit the scope of applicability to a few thousands nodes: for that, approximation schemes are devised for reconstructing the spectrum after computing only a limited number of eigenvalues. In this way, networks up to 10 6 nodes can be treated. Although not discussed in the paper, the method is readily applicable to weighted graphs, whereas the extension to directed networks appears non trivial due to the different spectral characteristics.
Portrait divergence
It is a recent method 42 based on a graph invariant which encodes the distribution of the shortest-path lengths in a graph: the network portrait 43 is a matrix B whose entry B lk , l = 0, 1, …, d ( d is the graph diameter), k = 0, 1, …, N − 1, is the number of nodes having k nodes at shortest-path distance l . The definition also extends to directed and weighted networks–in the weighted case, a binning strategy is needed to manage real-valued path lengths. The network portrait is a powerful summary of the topological features of the graph–e.g., the number of nodes and edges, the degree distribution, the distribution of the next-nearest neighbours, and the number of shortest paths of length l can straightforwardly be recovered from B . The Portrait Divergence distance between graphs G 1 and G 2 is then defined as follows. First, the probability P ( k , l ) (and similarly Q ( k , l ) for the second graph) of randomly choosing two nodes at distance l and, for one of the two nodes, to have k nodes at distance l , is computed:
where n c is the number of nodes in the connected component c . Then, the Portrait Divergence distance is defined using the Jensen-Shannon divergence:
where M = ( P + Q )/2 is the mixture distribution of P and Q , and KL ( ⋅ || ⋅ ) is the Kullback-Liebler divergence. The method is computationally efficient for small and medium size graphs, since it is quadratic in the number of nodes, and can naturally handle disconnected networks.
Graph kernels
A graph kernel k ( G 1 , G 2 ) is a non-negative function of two feature vectors f 1 , f 2 , respectively representative of the two graphs G 1 , G 2 . In abstract form, a graph kernel implements a (generalized) inner product of the two graphs, which is taken as a measure of their similarity. The proposal of using kernel methods for graph comparison is attributed to refs. 44 , 45 - see also refs. 46 , 47 for a unified description and survey.
The approach is very general, as feature vectors can be defined in very different forms. A recent paper 48 , introducing an R/Python implementation, summarizes 14 different kernel types among the most popular ones: the majority of them are based, in different forms, on statistics on node/edge labels (thus they fall out of the scope of our work, as we do not assume labels on nodes/edges). Two of them are based on graphlet count, and the remaining on the comparison of random walks on the two graphs. But this list is by no means exhaustive, as many other proposals are found in the literature (e.g., ref. 49 defines a Laplacian graph kernel). Therefore, graph kernel methods can be considered as a general framework where diversified network features can be included. It follows that selecting the proper graph kernel for the problem at hand can be critical, also considering the non trivial computational requirements of this class of methods: we refer the reader to the discussion in ref. 47 .
Bayes’ modeling of a network population
The network comparison problem can be addressed using a Bayesian nonparametric approach. Durante et al . 50 proposed a mixture model to describe a population of networks, interpreted as realizations of a network-valued random variable, and in particular to infer the parameters of the probability mass function of such variable. This is not a distance-based method, since no explicit distance is defined to compare networks. Instead, the use of a Dirichlet process prior naturally yields a clustering of the input networks thanks to the discreteness of the resulting measure.
Persistent homology
Homology is an algebraic-topological measurement of the structure of an undirected unweighted network which, based on the number and dimension of cliques and cycles, exploits the information carried by the mesoscale structure of the network. The generalization to weighted graphs is possible by considering persistent homology , which tracks the evolution of cycles when a sequence of unweighted graphs is obtained by thresholding the network at distinct edge weights. This technique was used by Sizemore et al . 51 , where suitable quantities derived from persistent homology are jointly used as features to classify networks via hierarchical clustering, showing a superior performance with respect to using standard graph statistics. In the paper, however, no network distance is explicitly defined.
In this section, we select some of the previously described methods and we define and carry out tests on synthetic networks to assess the performance of each method. Finally we illustrate the application to real-world network data. As a general rule, we only consider methods for which a source/executable code is freely made available by the authors of the method (details of the used codes are in the Supplementary Information file, Sec. S2 ), in order to fairly test each method in terms of correctness (exceptions are the baseline and spectral methods, requiring the computations of simple distances, which were directly coded by ourselves).
More specifically, among the KNC methods we included:
The difference of the adjacency matrices (with Euclidean, Manhattan, Canberra and Jaccard norms), as a baseline approach;
DeltaCon, due to its desirable properties above described, and with the aim of testing a non-trivial KNC method. We also tested the implementation for directed networks.
Among the UNC methods, we selected:
The (absolute) difference of the clustering coefficient and of the diameter, as baseline approaches;
For alignment-based methods, MI-GRAAL, which allows to extract additional information from the node mapping;
For graphlet-based methods, GCD-11 and DCGD-129. The first one is for undirected networks and it was proved to be very effective in discriminating synthetic networks of different topologies 19 . The second one is the directed version of GCD-11, except that, differently from GCD-11, DGCD-129 does not exclude redundant orbits;
For spectral methods, the Wilson and Zhu 37 approach: we define three distances by computing the Euclidean distance between the spectra of the adjacency matrices, Laplacians, and Symmetric Normalized Laplacians 52 (SNLs) (the approach by Gera et al . 38 was discarded since no code was available).
NetLSD, based on the solution of a dynamical (“heat”) equation;
Finally, Portrait Divergence, which is naturally able to deal with undirected and directed networks.
No source/executable code was available for Cut Distance and Persistent Homology. The Bayesian method was excluded since no direct comparison is possible with the other approaches, due to its inherently different nature. Finally, GRAFENE was excluded because, despite it is defined for general networks, the executable provided by the authors is strongly domain-dependent, i.e., it requires a biologically-based network definition.
We summarize in Table 1 the methods selected for testing, classified with respect to the type of network they can manage and to the nature of the method itself. In the UNC column, we generically insert “Global statistics” in all rows, although the statistics to be concretely used might depend on the network type. We point out that the tests we carried out on synthetic networks are restricted to the unweighted case: the definition of suitable weighted benchmark networks and of adequate testing strategies for them is a complex and delicate task that goes beyond the scope of the present work. Note that, in the unweighted (binary) case, the MAN and CAN distances actually yield the same result as the (square of the) EUC distance; thus the three distances are actually the same and only the EUC distance will be considered in the analysis.
Perturbation tests
We perform successive perturbations starting from an original graph and we measure, after each perturbation, the distance of the obtained graph from the original one. This is aimed at checking to what extent the considered network distance has a “regular” behaviour. As a matter of fact, we expect that whatever distance we use, it should tend to zero when the perturbations tend to zero, i.e., the distance between very similar graphs is small; moreover, the distance should increase monotonically with the number of perturbations, meaning that fluctuations, if any, should be very limited. In addition, for perturbations that tend to randomize the graph (see below), the distance should saturate to some asymptotic value after a large number of perturbations, because when the graph has become fully randomized it remains such for any further perturbation (see ref. 16 for a formal discussion on the properties of similarity measures).
The following types of perturbations are considered:
Removal ( pREM-test ): a connected node pair is picked uniformly at random and the edge is removed (the graph density decreases).
Addition ( pADD-test ): a non-connected node pair is picked uniformly at random and an edge is added (the graph density increases).
Random switching ( pRSW-test ): combines the previous two: a connected node pair is picked uniformly at random and the edge is removed; then a non-connected node pair is picked uniformly at random and an edge is added (the graph density does not change).
Degree-preserving switching ( pDSW-test ): two edges are picked uniformly at random and swapped: if we pick ( i , j ) and ( u , v ), we delete them and insert the new edges ( i , v ) and ( j , u ), provided they are not already existing (the graph density and the degree distribution do not change).
In addition, for directed networks:
Change of direction ( pDIR-test ): a connected node pair is picked uniformly at random and the edge direction is reversed (the graph density does not change).
We consider two families of networks, i.e., undirected and directed, and three models for each family, i.e., Erdös-Rényi (ER) 53 , Barabási-Albert (BA) 54 and Lancichinetti-Fortunato-Radicchi (LFR) 55 , 56 (see ref. 16 for alternative proposals of synthetic test networks for comparison methods). The three models, which are well-known and widely used as test-beds in a variety of network science problems 2 , have increasing structural complexity. In the undirected case, an ER network 53 is built by setting the number of nodes and links, and then by randomly assigning links to node pairs selected uniformly at random. This yields a network with almost homogenous degree distribution, i.e., all nodes are statistically equivalent and have small fluctuations around the average degree. In a BA network 54 , on the other hand, a few nodes with very high degree coexist with a majority of medium/small degree nodes, giving rise to a strongly inhomogeneous (power-law) degree distribution, as encountered in a number of real-world datasets. LFR networks 55 , 56 add a further level of complexity, namely modular structure: the network is partitioned into communities, which are subgraphs with large internal density yet loosely connected to the other communities, and not only the nodes degree, but also the communities size are power-law distributed, to mimic features typically found in real data. The three models, originally developed for undirected networks, have been generalized to the directed case (see Supplementary Information file, Sec. S3 , for details on the creation of undirected and directed networks).
For each family/model pair, we create two graphs with 1000 nodes, respectively with density 0.01 and 0.05. On each network, we perform 10 different replications of 1000 successive perturbations of the types above defined. For the undirected and directed cases, respectively, we test the methods in the first and second row of Table 1 . For “Global statistics”, we use the clustering coefficient and the diameter for undirected networks, and the diameter only for directed networks. For density 0.05 and for all directed networks, we exclude MI-GRAAL from the analysis because it turns out to be computationally too heavy.
Figures 2 and 3 summarize the results obtained on undirected networks with density 0.01 (see Supplementary Information file, Figs. S1 and S2 , for density 0.05). We notice that essentially all distances have a “regular” behaviour, in the sense above discussed. The only remarkable exception is the diameter distance, which remains zero on a broad range of perturbations for most network models, thus proving inadequate as a network distance. Confidence bands are generally tight, except for MI-GRAAL and GCD-11 which fluctuate more. In a few cases, the distances tend to saturate–this is especially evident in switching tests ( pRSW-test and pDSW-test ). Overall, these results denote that all distances are well defined and can be properly used for comparison–except the diameter, which therefore will not be considered in the remainder.
Perturbations tests: results of the Removal ( pREM-test ) and Addition ( pADD-test ) tests for the 12 distances and the 3 undirected models ER, BA, LFR (density 0.01). Solid (dashed) lines are the mean (±3 std) values obtained over the 10 replications of the perturbation histories.
Perturbations tests: results of the Random switching ( pRSW-test ) and Degree-preserving switching ( pDSW-test ) tests for the 12 distances and the 3 undirected models ER, BA, LFR (density 0.01). Solid (dashed) lines are the mean (±3 std) values obtained over the 10 replications of the perturbation histories.
At the same time, the tests reveal a different behaviour for the two classes of distances KNC and UNC (Table 1 ). KNC distances turn out to be practically insensitive to the network model and even to the type of perturbation: for these distances, a perturbation acts in the same way in all the considered models. This is not unexpected: for example, EUC simply measures the (square root of the) number of edges that have been added, removed or switched, disregarding the topology in which these changes happen: a very weak characterization of the effects of the perturbations.
UNC distances, on the contrary, show quite different values and patterns for different models and perturbations. The three spectral distances behave quite differently from one test to another. The trend is linear in pADD-test and pREM-test and independent on the network model, while for pRSW-test and pDSW-test the distance is definitely larger for LFR networks, in almost all cases. This behaviour captures the higher sensitivity of LFR networks to random rewiring, due to the progressive destruction of the built-in community structure via removal of intra-communities edges and creation of inter-community connections. Indeed, the same behaviour is shown by the clustering coefficient and the NetLSD distance. On the other hand, pRSW-test affects BA networks too, at least for a few of the distances, because hubs will easily lose connections. Noticeably, ER networks are weakly affected by random rewiring of both types, if measured by spectral distances: random rewiring yields of course different graphs, but structurally equivalent to the original one.
Coming to the last UNC distances, we notice the comparatively large variability of GCD-11. PDIV, on the other hand, has a peculiar behaviour: the distance has a steep increase in the very first perturbations, then either it increases further but at a slower rate, or it immediately saturates. The latter is the case, e.g., of ER and BA models under pDSW-test : under random rewiring, the ER topology yields equivalent graphs that are almost equally distant from the original one, but interestingly the same happens for the BA structure: we argue that after some pDSW-test steps the shortest paths distribution is almost the same for all the perturbed graphs, so that the PDIV distance becomes constant.
The MI-GRAAL distance shows a behaviour different from all the other distances. The confidence bands reveal a large variability, as above pointed out. In a few extreme cases (e.g., pREM-test and pDSW-test on LFR graphs, or pADD-test and pDSW-test on ER graphs), the confidence band spans almost the entire [0, 1] range, namely, a few perturbations may lead to a graph apparently equal or totally different from the original one, depending on which edges are perturbed. This unpleasant behaviour is probably due to the large sensitivity to perturbations of the local (node) features. Recall that MI-GRAAL distance builds a map between the most similar nodes of the compared networks. Perturbations induce differences in the individual node features, i.e., degree, clustering coefficient, and betweenness centrality, that can obscure and shuffle the correct node correspondence. Incidentally, ER networks seem the most suffering ones, presumably because nodes are weakly characterized.
Finally, Fig. 4 summarizes the results of the change of direction ( pDIR-test ) test on directed networks with density 0.01 (see Supplementary Information file, Figs. S3 to S5 , for density 0.05). Again, for the same reasons as above, the simplest distances (EUC and JAC) show a behaviour which is very regular and insensitive to the network model. Again the diameter distance has a pathological behaviour and should be discarded in any case. What emerges in the remaining two UNC distances, on the other hand, is their strong sensitivity in the case of BA networks, where (see DGCD-129 and PDIV panels) even the first few perturbations yield large distances. To explain this effect, it should be emphasized that in our directed BA networks, by construction, hubs are nodes having systematically large in-degree while the out-degree is homogeneously distributed (see Supplementary Information file, Sec. S3 , for details on the construction of directed networks). It turns out that few changes in the edge directions are sufficient to dramatically change the number and type of graphlets (DGCD-129) and the structure of the shortest paths (PDIV). Such a strong sensitivity could be an issue in some applications, although it is certainly amplified by the specific structure of our benchmark networks.
Perturbations tests: results of the Change of direction ( pDIR-test ) test for the 6 distances (see Table 1 , MI-GRAAL excluded for computational reasons) and the 3 directed models ER, BA, LFR (density 0.01). Solid (dashed) lines are the mean (±3 std) values obtained over the 10 replications of the perturbation histories.
Clustering tests
In this section, several tests are proposed aimed at assessing the effectiveness of each method in recognizing and grouping together networks with the same structural features, i.e., originated from the same model. In other words, to yield a good performance, a method should be able to assign small distance to network pairs coming from the same model but large distance to pairs coming from different models. We adopt a typical approach for testing performance of methods operating in a not-supervised setting, i.e., simulating controlled scenarios where groups (or, as in this case, network models) are known, then running the algorithms as in the unsupervised case, and finally checking their ability in properly identifying groups coming from different models. Note that only UNC methods are suitable to this task, since we assume that no node correspondence is available. Thus, with reference to Table 1 , we test the methods of the UNC column, in the first and second row, respectively, for undirected and directed networks. In both cases we exclude MI-GRAAL, which proved to be computationally too heavy in most of the experimental settings below described. We also restrict the analysis of “Global statistics” to the clustering coefficient (only for undirected networks), given the bad performance of the diameter in the perturbation tests above described.
We choose again ER, BA, and LFR network models. For each model, we consider two sizes (1000 and 2000 nodes), two densities (0.01 and 0.05), and the two families undirected/directed (see Supplementary Information file, Sec. S3 , for details on network construction), and we generate 5 networks for each one of the 3 × 2 × 2 model/parameter set, for a total of 60 undirected and 60 directed networks. We then compute all the pairwise distances for each method, ending up with two (i.e., undirected/directed) 60 × 60 distance matrices.
In Figs. 5 and 6 we visualize the results by means of dendrograms (built with ward.D2 linkage 57 and Optimal Leaf Ordering 58 ) where, to aid interpretation, we label each leaf (=network) with two colours, one denoting the network model and the other the size/density. This is in line with the aforementioned purpose: since we can only control the network generating mechanisms, and not the realizations obtained by each of them, we simulate a setting as controlled as possible, mimicking a supervised context, and then we assess the performances of the different methods when adopted for building clusters, as in the unsupervised setting. Thus the supervised context is used only to set the benchmarking object, then the algorithms run in a completely unsupervised way. For this reason, as above emphasized, it is reasonable to expect that an effective method should ideally be able to cluster all networks coming from the same model family together, without being disturbed from the different sizes and densities.
Clustering tests: dendrograms for the undirected case (the dendrograms EIG-LAP, EIG-SNL, and PDIV are similar to EIG-ADJ and are reported in the Supplementary Information file, Fig. S6 ).
Clustering tests: dendrograms for the directed case. Left: Portrait Divergence distance PDIV. Right: DGCD-129 distance.
In the undirected case (Fig. 5 ), we observe that all methods are able, more or less precisely, to correctly group networks of the same class, but only if they have the same size/density. This is the case for clustering coefficient distance, EIG-ADJ (as well as EIG-LAP, EIG-SNL and PDIV, not reported in the figure), and NetLSD. GCD-11 achieves a definitely better performance, as it proves able to group all LFR networks, along with the BA graphs with low densities, and by identifying two other clear clusters: one containing ER graphs with low density, and the other BA and ER graphs with high densities. Overall, the results reveal a quite strong dependence of all methods, except GCD-11, on size and density of the graphs, a result not fully satisfactory.
In the directed case (Fig. 6 ), PDIV behaves better than in the undirected case, since it is able to group together all the BA graphs, while ER and LFR graphs are grouped together in another large cluster within which density becomes the discriminating factor. On the other hand, DGCD-129 is able to achieve a perfect grouping of the network models.
A systematic approach to comparatively quantify the performance of different methods is the Precision-Recall framework 19 , 20 , 24 : for a given network distance, one defines a threshold ε > 0 and classifies two networks as belonging to the same model class if their distance is less than ε . Given that the correct classes are known, the accuracy of classifying all network pairs can be quantified by Precision and Recall. The procedure is then repeated by varying ε , obtaining the Precision-Recall curve (Fig. 7 ) which, ideally, should have Precision equal to 1 for any Recall value. Overall, the curves highlight the superiority of the graphlet-based approaches (GCD-11 and DGCD-129, respectively) in comparison to all other methods we have tested–their curve is above all the others in almost the entire curve range.
Clustering tests. Precision-Recall curves for the methods used for undirected (left) and directed (right) networks.
A quantification of the performance of each method can be obtained by the Area Under the Precision Recall curve (AUPR), which should be 1 in the ideal case. For the left panel of Fig. 7 , AUPR ranges from 0.386 (NetLSD) to 0.688 (GCD-11), for the right panel AUPR is 0.685 for PDIV and 0.928 for DGCD-129. We note that all methods perform better than a random classifier (AUPR equal to 0.322), but the improvement is very moderate for spectral methods and NetLSD. The performance of the clustering coefficient distance (undirected networks) is surprisingly good, with AUPR equal to 0.620. At least in our testing environment, this simple statistics appears to be a viable low-cost alternative to more sophisticated distances.
In addition, the AUPR measure can be used to confirm that, if only networks of the same size/density are considered, all the analysed methods yield an essentially correct classification–a feature already highlighted above when discussing the dendrograms. Indeed, we report that in this case almost all methods of Fig. 7 obtain an AUPR larger than 0.8 in the undirected case, and practically equal to 1 in the directed case (see Supplementary Information file, Tables S3 and S4 ).
Tests on real-world networks
In this section we assess the performance of the above methods in grouping sets of real-world networks: our aim is to understand more deeply how these methods work and which type of results one should expect from their use. We analyse multiplex networks, since their layers can be seen as individual (ordinary) networks defined over the same node set. This is a desirable property for our analysis, because it enables the use of both KNC and UNC methods. One can therefore ascertain whether the two classes of methods provide different insights of the same problem.
European air transportation network
This network has 448 nodes, representing European airports, and 37 layers, corresponding to the airline companies (see Supplementary Information file, Table S5 , for the list), for a total of 3588 undirected and unweighted edges 59 . We test all the distances of the first row of Table 1 , namely three KNC distances (recall that MAN and CAN are equivalent to EUC in this setting) and eight UNC distances (with the clustering coefficient as “Global statistics”).
Figure 8 displays the eleven 37 × 37 (symmetric) distance matrices (the corresponding dendrograms are in the Supplementary Information file, Fig. S7 ). We firstly observe that the two classes of distances yield qualitatively different output: KNC distances are almost uniform over the entire matrix (JAC above all), while UNC methods obtain different degrees of differentiation, remarkably for EIG-SNL and GCD-11. EUC and DCON essentially identify one single (trivial) cluster including all the networks except Ryanair (#2). This is not unexpected, since the Ryanair layer differs from the others in having by far the largest number of edges and (non isolated) nodes. In turn, this demonstrates that KNC distances are strongly affected by size and density: two graphs with different sizes or densities are classified as distant just because of edge counting, regardless of their topological properties.
European Air Transportation Network: Distance matrices for three KNC distances (first row) and eight UNC distances (second to last row). The colour in entry ( i , j ) is the distance between layers i and j of the multiplex air transportation network. Distances are coded by colours from blue (large distance) to red (small distance).
The three spectral distances have similar behaviour. Dendrograms identify two main clusters, a small one containing airlines #1, #2, #3, and, according to the method, #5, #6 or #26, and a large one containing all the rest (actually EIG-SNL reveals a third, small cluster, which however has no clear interpretation). The small cluster groups together the airlines with the highest number of nodes and connections (>75 nodes, >110 edges). This confirms the strong dependence of the three spectral distances on size and density, as highlighted in the synthetic tests.
To avoid ambiguities in the interpretation of the heatmaps of Fig. 8 , we remark that the presence of a group of airlines with comparatively smaller internal distance (i.e., a diagonal block with hotter colour) does not imply the existence of a well defined group, unless the group itself displays comparatively higher distance (colder colour) to the remaining airlines. That is why, for example, in the EUC case airlines #34, #19, #12, etc., do not form a well separated group, as well as the hotter coloured diagonal blocks in the EIG-ADJ case are not significant since there is no sharp colour separation (i.e., cold colour outside the block) but a continuum of distances among airlines.
The patterns of the clustering coefficient and of the NetLSD distance matrices are qualitatively similar to that of the spectral distances. NetLSD matrix does not lend to a clear interpretation. The clustering coefficient matrix is easy to read, on the contrary, thanks to the straightforward definition of the specific network distance: airline #35 (Wideore) has by far the largest clustering coefficient. In addition, the heatmap reveals two weakly separated groups, which indeed correspond to the groups of airlines with, respectively, larger (>0.1) and smaller clustering coefficient.
MI-GRAAL and PDIV distances do not show any meaningful grouping of networks, except that the former groups together three pure-star networks (#9, #31, #33) but misses the fourth one existing in the database (#18), and the latter groups a few networks having the same maximum degree (#17, #19, #23, and #13, #20, #31, #34) but in fact misses a few others. Finally, the distance matrix and the dendrogram for GCD-11 show three main clusters. One of them contains all the four pure-star layers in the dataset (#9, #18, #31, #33), while the second groups seven of the eight airlines with comparatively large (>0.15) clustering coefficient (#2, #3, #6, #15, #16, #21, #26, #35). The third cluster contains all the remaining layers. In addition, airlines #4 and #28 are also paired to form a small but well identified cluster: it turns out that they both have zero clustering coefficient without being a star graph. Overall, these results denote that GCD-11 is quite effective in classifying networks based on purely topological features.
We pointed out above that KNC distances are unable to extract meaningful clusters of networks. Nonetheless they provide some useful result, as they pair networks on a regional basis, i.e., they put at a comparatively small distance airlines based in the same nation/region. For example, EUC dendrogram pairs #1 with #6 (Germany), #8 with #13 (Scandinavia), #30 with #37 (Greece), and #9 with #27 (Netherlands). JAC pairs together the aforementioned airlines and, in addition, #3 with #4 (UK), #12 with #22 (Spain), #2 with #23 (Ireland), and #16 with #33 (Hungary). DCON (see Fig. 9 ) groups #8 with #13 (Scandinavia), #12, #21, #22 (Spain), #30 with #37 (Greece), #1 with #6 (Germany), and #24 with #29 (Germany again). This kind of grouping is not recovered by UNC distances (see, e.g., DCON vs GCD-11 in Fig. 9 ). As a matter of fact, KNC distances are all based on some sort of node similarity. Airlines based in the same nation share the same national airports; moreover, in many instances the two airlines paired by KNC methods are the leading national company and a low-cost company, offering different journey and price conditions on the same routes. Consequently the nodes of two airlines based in the same nation are typically similar and this reduces the distance between the corresponding networks.
European Air Transportation Network: dendrograms for DCON and GCD-11, with airlines coloured according to the nations they are based in. Contrarily to UNC distances (e.g. GCD-11), KNC distances (e.g., DCON) are able in many cases to group airlines which are based in the same nation.
FAO trade network
We now consider the FAO Trade Network of food, agricultural and animal products 60 . It is composed of 364 layers, each one representing a different product, sharing 214 nodes, corresponding to countries, for a total number of 318346 connections. If layers/products are matched with the WTO Harmonized System classification 61 , we find that FAO products belong to seven out of the fifteen major HS categories: Animals and Animal Products ; Vegetable Products ; Foodstuffs ; Chemicals ; Plastic and Rubbers ; Raw Hides, Skins and Leathers ; Textiles . Each layer is directed and weighted, the weights representing the export value (thousands of USD) of a product from one country to another. Here we discuss a few results derived on the binarized version of the network, which is obtained, for each product p , by retaining only the links departing from countries c having Revealed Comparative Advantage RCA cp > 1, i.e., countries which are significant exporters of that product (see refs. 62 , 63 . and Supplementary Information file, eq. ( S1) , for details). We test the distances of the second row of Table 1 , namely three KNC distances (since MAN and CAN are equivalent to EUC) and two of the UNC distances (as usual, we exclude MI-GRAAL due to the high computational cost, and we do not consider the clustering coefficient since we analysed it as “Global statistics” for undirected networks only).
The cluster analysis summarized by the dendrograms of Fig. 10 clearly shows that the naive expectation that products organize in groups according to their category is not fulfilled: HS categories turn out to be too broad and diversified to induce similar trade topologies. Vegetable Products , for example, is the most represented category: it includes such a diversity of products that the corresponding networks are also extremely diverse, not only in terms of number and location of producing/consuming countries, but also for the possible existence of intermediate countries where products are imported, transformed, and eventually re-exported.
FAO Trade Network: cluster analysis for the methods under test. In the upper colour bar, products are classified according to WTO Harmonized System. In the bottom bar, a few specific categories are evidenced.
Nonetheless, cluster analysis is able to spot interesting similarities. For example, some agricultural products (such as tropical fruits or citrus fruits) can be produced only in specific regions of the world and thus only by few countries. Thus, we expect that the distance between the corresponding layers will be small–this obviously holds true if KNC distances are used. This expectation is indeed confirmed. All KNC distances group together (see the bottom colour bars in Fig. 10 ) all the citrus fruits (namely Oranges ; Lemons and limes ; Grapefruit ; Tangerines, mandarins, clementines, satsumas ) present in the dataset, and EUC also puts Olives preserved and Oil, olive, virgin close to them. This is not surprising, since the latter are mostly produced in the same areas as citrus, for climatic reasons. Most of the tropical fruits are also gathered together: EUC mainly finds two groups, one which contains Cocoa, beans ; Cocoa, powder and cake ; Cocoa, paste ; Cocoa, butter ; Bananas ; Pineapples ; Mangoes, mangosteens, guavas ; and the other one which contains Vanilla ; Coconuts ; Coconuts, desiccated ; Cinnamon ; Nutmeg ; Pepper ; Coffee, green . Similarly, DCON also identifies two major groups of tropical fruits. In addition, another group of specific products, namely those related to juices, are grouped together by KNC distances. Notably, the groups of citrus fruits and citrus juices are not adjacent to each other, revealing that production and processing of fresh fruit mostly take place in different countries thus yielding different trade networks.
Finally, UNC distances, applied to the FAO dataset, identify those few products with a very peculiar trade pattern. For example, the leftmost leaves of the DGCD-129 dendrogram (Fig. 10 ), which have large distance from all other leaves, correspond to trade networks with a pronounced star-like structure. They relate to very peculiar products, such as Bulgur ; Maple sugar and syrups ; Margarine, liquid ; Beehives ; Camels ; and others. Some of them are produced and exported by only one country (e.g., Maple sugar and syrup ) or are imported significantly by very few countries (e.g., Beehives ) and thus their trade patterns are highly centralized.
Discussion and Concluding Remarks
We reviewed a broad set of methods available in the literature for network comparison, assessing their performances both on synthetic benchmarks and on real-world multilayer networks. We relied on a well-accepted dichotomy of the methods based on their functioning. The first class (Known Node-Correspondence) gathers all the methods, such as Euclidean, Jaccard or DeltaCon distances, which require a priori to know the correspondence between the nodes of the compared networks: they are suitable to quantify the differences between graphs representing various connection patterns among the same node set, as in the EU Air Transportation case, or to quantify changes or spotting anomalies occurring in a graph subject to temporal evolution. The second class (Unknown Node-Correspondence) collects all the methods which do not require any a priori knowledge of the correspondence between nodes. Methods of this class, such as spectral distances, graphlet-based measures, and Portrait Divergence, are specifically suited for structural comparison, namely to provide information about how much, and in what sense, the structures of graphs differ.
To evaluate the performance of each method, we carried out two classes of tests–perturbation and clustering–with the use of synthetic networks. Virtually all methods demonstrated a fairly good behaviour under perturbation tests (the diameter distance being the only exception), in the sense that all distances tend to zero as the similarity of the networks increases, tend to a plateau after a large number of perturbations, and do not fluctuate too much if perturbations are repeated. On the contrary, the clustering tests highlighted different behaviours and performances. When networks of the same size and density are considered, in both the undirected and directed case most methods are reasonably able to discriminate between different structures, often achieving a perfect classification. However, when considering networks of different sizes and densities, the results change considerably. In the undirected case, the graphlet-based measure GCD-11 is the best performing distance in discriminating between different network topologies and clearly outperforms all the other methods whereas, on the other hand, a few methods only perform slightly better than a random classifier. In the directed case, the graphlet-based measure DGCD-129 is able to achieve an almost perfect classification, whereas, for example, Portrait Divergence distance performs much better than in the undirected case. Therefore, since in many real-world applications density and size of the graphs may vary considerably, graphlet-based measures prove to be the most reliable tools to investigate the differences between networks structures. As a matter of fact, while in the perturbation tests graphlet methods display behaviours comparable to the other methods in terms of regularity, in clustering tests GCD-11 and DGCD-129 demonstrate performances overall superior to the other methods. This emerges from the dendrograms (Figs. 5 and 6 ) and from the Precision/Recall diagrams (Fig. 7 ), the latter synthesized by the AUPR values (see Supplementary Information file, Tables S3 and S4 ). Needless to say, these conclusions are related to our specific test-bed, and we cannot exclude that different evidences could emerge from the analysis, e.g., of different network topologies, where some other method could prove to be more effective.
The methods analysed were able to recover important features in the real-world case studies analysed. For example, in the European Air Transportation Network they paired at small distances airlines based in the same nation, which are expected to have high nodes similarities since they share the same airports and routes. In the FAO Trade Network, they were able to group specific products whose countries of production are in common due to climate reasons. In this latter case, results were obtained after the original weighted network was reduced to a binary (unweighted) version. In this respect, considerable work is still needed to develop and systematically test methods able to fully exploit the available information on weights, an effort that could potentially yield a significant improvement in network classification. A common feature of methods with Known Node-Correspondence, that emerged from the analysis of real-world case studies, is their strong dependence on density and size of the graphs–a feature that could be a strength or a weakness according to the application. More in general, the analysis of real-world datasets has shown that, even when a network measure is unable to induce a clear partition of the entire set of networks, it can nonetheless highlight important subsets. In this sense, using different methods–based on different notions of network similarity–can provide rich and diversified information.
Newman, M. E. J. Networks: An Introduction . (Oxford University Press, 2010).
Barabási, A. L. Network Science . (Cambridge University Press, 2016).
Latora, V., Nicosia, V. & Russo, G. Complex Networks: Principles, Methods and Applications . (Cambridge University Press, 2017).
Fortunato, S. Community detection in graphs. Phys. Rep. 486 , 75–174 (2010).
Article ADS MathSciNet Google Scholar
Rombach, P., Porter, M. A., Fowler, J. H. & Mucha, P. J. Core-periphery structure in networks (revisited). SIAM Rev. 59 , 619–646 (2017).
Article MathSciNet MATH Google Scholar
Soundarajan, S., Eliassi-Rad, T. & Gallagher, B. A guide to selecting a network similarity method. In Proc . 2014 SIAM Int. Conf. on Data Mining , 1037–1045 (2014).
Emmert-Streib, F., Dehmer, M. & Shi, Y. Fifty years of graph matching, network alignment and network comparison. Inf. Sci. 346-347 , 180–197 (2016).
Donnat, C. & Holmes, S. M. Tracking network dynamics: A survey of distances and similarity metrics. Ann. Appl. Stat. 12 , 971–1012 (2018).
Chartrand, G. & Zhang, P. A first course in graph theory . (Courier Corporation, 2012).
Toda, S. Graph isomorphism: Its complexity and algorithms. In Rangan, C., Raman, V. & Ramanujam, R. (eds) Foundations of Software Technology and Theoretical Computer Science , vol. 1738 of Lecture Notes in Computer Science , 341–341 (Springer Berlin Heidelberg, 1999).
Kuchaiev, O. & Pržulj, N. Integrative network alignment reveals large regions of global network similarity in yeast and human. Bioinformatics 27 , 1390–1396 (2011).
Article CAS PubMed Google Scholar
Liu, Q., Dong, Z. & Wang, E. Cut based method for comparing complex networks. Sci. Rep. 8 , 5134 (2018).
Article ADS CAS PubMed PubMed Central Google Scholar
Foggia, P., Sansone, C. & Vento, M. A performance comparison of five algorithms for graph isomorphism. In Proc. 3rd IAPR-TC15 Workshop Graph-Based Representations in Pattern Recognition , 188–199 (2001).
Hand, D., Mannila, H. & Smyth, P. Principles of Data Mining . (The MIT Press, 2001).
Ioffe, S. Improved consistent sampling, weighted minhash and L1 sketching. In Proc . 2010 IEEE Int. Conf. Data Mining , 246–255 (2010).
Koutra, D., Vogelstein, J. T. & Faloutsos, C. Deltacon: A principled massive-graph similarity function. Proceedings of the 2013 SIAM International Conference on Data Mining 162–170 (2013).
Koutra, D., Shah, N., Vogelstein, J. T., Gallagher, B. & Faloutsos, C. Deltacon: A principled massive-graph similarity function with attribution. ACM Trans. Knowl. Discov . Data 10 , 1–43 (2016).
Google Scholar
Pržulj, N., Corneil, D. G. & Jurisica, I. Modeling interactome: scale-free or geometric? Bioinformatics 20 , 3508–3515 (2004).
Yaveroglu, O. N. et al . Revealing the hidden language of complex networks. Sci. Rep. 4 , 4547 (2014).
Article CAS PubMed PubMed Central Google Scholar
Yaveroglu, O. N., Milenkovic, T. & Przulj, N. Proper evaluation of alignment-free network comparison methods. Bioinformatics 31 , 2697–2704 (2015).
Faisal, F. E. et al . Grafene: Graphlet-based alignment-free network approach integrates 3d structural and sequence (residue order) data to improve protein structural comparison. Sci. Rep. 7 , 14890 (2017).
Przulj, N. Biological network comparison using graphlet degree distribution. Bioinformatics 23 , 177–183 (2007).
Article CAS Google Scholar
Onnela, J.-P. et al . Taxonomies of networks from community structure. Phys. Rev. E 86 , 036104 (2012).
Article ADS CAS Google Scholar
Sarajlić, A., Malod-Dognin, N., Yaveroğlu, O. N. & Pržulj, N. Graphlet-based characterization of directed networks. Sci. Rep. 6 , 35098 (2016).
Aparício, D., Ribeiro, P. & Silva, F. Network comparison using directed graphlets. ArXiv: 1511.01964 (2015).
Hočevar, T. & Demšar, J. A combinatorial approach to graphlet counting. Bioinformatics 30 , 559–565 (2014).
Ribeiro, P. & Silva, F. G-tries: A data structure for storing and finding subgraphs. Data Min. Knowl. Discov. 28 , 337–377 (2014).
Milenković, T. & Przulj, N. Uncovering biological network function via graphlet degree signatures. Cancer Inform. 6 , 257–273 (2008).
Article PubMed Google Scholar
Ali, W., Rito, T., Reinert, G., Sun, F. & Deane, C. M. Alignment-free protein interaction network comparison. European Conference on Computational Biology 30 , i430–i437 (2014).
CAS Google Scholar
Kelley, B. P. et al . PathBLAST: a tool for alignment of protein interaction networks. Nucleic Acids Res. 32 , W83–W88 (2004).
Flannick, J., Novak, A., Balaji, S., Harley, H. & Batzglou, S. GRAEMLIN: general and robust alignment of multiple large interaction networks. Genome Res. 16 , 1169–1181 (2006).
Liao, C.-S., Lu, K., Baym, M., Singh, R. & Berger, B. IsoRankN: spectral methods for global alignment of multiple protein networks. Bioinformatics 25 , i253–258 (2009).
Kuchaiev, O., Milenkovic, T., Memisevic, V., Hayes, W. & Przulj, N. Topological network alignment uncovers biological function and phylogeny. J. R. Soc. Interface 7 , 1341–1354 (2010).
Article PubMed PubMed Central Google Scholar
Milenkovic, T., Ng, W. L., Hayes, W. & Przulj, N. Optimal network alignment with graphlet degree vectors. Cancer Inform. 9 , CIN.S4744 (2010).
Article Google Scholar
Memisevic, V. & Przulj, N. C-GRAAL: Common-neighbors-based global graph alignment of biological networks. Integr. Biol. 4 , 734–743 (2012).
Malod-Dognin, N. & Przulj, N. L-GRAAL: Lagrangian graphlet-based network aligner. Bioinformatics 31 , 2182–2189 (2015).
Wilson, R. C. & Zhu, P. A study of graph spectra for comparing graphs and trees. Pattern Recognit. 41 , 2833–2841 (2008).
Article MATH Google Scholar
Gera, R. et al . Identifying network structure similarity using spectral graph theory. Applied Network . Science 3 , 2 (2018).
Tsitsulin, A., Mottin, D., Karras, P., Bronstein, A. & Müller, E. NetLSD: Hearing the shape of a graph. In Proc. 24th ACM SIGKDD Int. Conf. Knowledge Discovery & Data Mining , KDD ‘18, 2347–2356 (ACM, New York, NY, USA, 2018).
Masuda, N., Porter, M. A. & Lambiotte, R. Random walks and diffusion on networks. Phys. Rep. 716-717 , 1–58 (2017).
Article ADS MathSciNet MATH Google Scholar
Chung, F. The heat kernel as the pagerank of a graph. Proc. Natl. Acad. Sci. USA 104 , 19735–19740 (2007).
Article ADS Google Scholar
Bagrow, J. P. & Bollt, E. M. An information-theoretic, all-scales approach to comparing networks. arXiv:1804.03665 (2018).
Bagrow, J. P., Bollt, E. M., Skufca, J. D. & ben Avraham, D. Portraits of complex networks. Europhysics Letters 81 , 68004 (2008).
Gartner, T., Flach, P. & Wrobel, S. On graph kernels: Hardness results and efficient alternatives. In Scholkopf, B. and Warmuth, M. K. (ed.) Learning Theory and Kernel Machines , vol. 2777 of Lecture Notes in Artificial Intelligence , 129–143 (Springer-Verlag Berlin, 2003).
Kashima, H., Tsuda, K. & Inokuchi, A. Marginalized kernels between labeled graphs. In Proc. 20th Int. Conf. on Machine Learning ( ICML-2003 ), 321–328 (2003).
Vishwanathan, S. V. N., Schraudolph, N. N., Kondor, R. & Borgwardt, K. M. Graph kernels. J. Mach. Learn. Res. 11 , 1201–1242 (2010).
MathSciNet MATH Google Scholar
Ghosh, S., Das, N., Goncalves, T., Quaresma, P. & Kundu, M. The journey of graph kernels through two decades. Comput. Sci. Rev. 27 , 88–111 (2018).
Sugiyama, M., Ghisu, M. E., Llinares-Lopez, F. & Borgwardt, K. graphkernels: R and Python packages for graph comparison. Bioinformatics 34 , 530–532 (2018).
Risi, K. & Horace, P. The multiscale Laplacian graph kernel. In Proc. 30th Conf. on Neural Information Processing Systems ( NIPS 2016) (2016).
Durante, D., Dunson, D. B. & Vogelstein, J. T. Nonparametric Bayes modeling of populations of networks. Journal of the American Statistical Association 112 , 1516–1530 (2017).
Article MathSciNet CAS Google Scholar
Sizemore, A., Giusti, C. & Bassett, D. S. Classification of weighted networks through mesoscale homological features. J. Complex Netw. 5 , 245–273 (2016).
MathSciNet Google Scholar
Chung, F. R. K. Spectral Graph Theory , vol. 92 of CBMS Regional Conference Series in Mathematics . (American Mathematical Society, 1997).
Erdös, P. & Rényi, A. On random graphs, I. Publ. Math.-Debr. 6 , 290–297 (1959).
MATH Google Scholar
Barabasi, A.-L. & Albert, R. Emergence of scaling in random networks. Science 286 , 509–512 (1999).
Article ADS MathSciNet CAS PubMed MATH Google Scholar
Lancichinetti, A., Fortunato, S. & Radicchi, F. Benchmark graphs for testing community detection algorithms. Physical Review E 78 , 046110 (2008).
Lancichinetti, A. & Fortunato, S. Benchmarks for testing community detection algorithms on directed and weighted graphs with overlapping communities. Physical Review E 80 , 016118 (2009).
Murtagh, F. & Legendre, P. Ward’s hierarchical agglomerative clustering method: Which algorithms implement Ward’s criterion? Journal of Classification 31 , 274–295 (2014).
Bar-Joshep, Z., Gifford, D. K. & Jaakkola, T. S. Fast optimal leaf ordering for hierarchical clustering. Bioinformatics 17 , 22–29 (2001).
Cardillo, A. et al . Emergence of network features from multiplexity. Sci. Rep. 3 , 1344 (2013).
Domenico, M. D., Nicosia, V., Arenas, A. & Latora, V. Structural reducibility of multilayer networks. Nature Communications 6 , 6864 (2015).
Article ADS CAS PubMed Google Scholar
United Nations International Trade Statistics. Harmonized Commodity Description and Coding Systems (HS). https://unstats.un.org/unsd/tradekb/Knowledgebase/50018/Harmonized-Commodity-Description-and-Coding-Systems-HS (2017).
Balassa, B. Trade liberalisation and revealed comparative advantage. The Manchester School 33 , 99–123 (1965).
Piccardi, C. & Tajoli, L. Complexity, centralization and fragility in economic networks. PLoS ONE 13 , e0208265 (2018).
Download references
Acknowledgements
The authors are grateful to James Bagrow for providing an updated version of the code for Portrait Divergence.
Author information
Authors and affiliations.
Moxoff SpA, via Schiaffino 11/A, 20158, Milano, Italy
Mattia Tantardini
MOX - Modelling and Scientific Computing Lab, Department of Mathematics, Politecnico di Milano, Via Bonardi 9, 20133, Milano, Italy
Francesca Ieva
CADS - Center for Analysis, Decisions and Society, Human Technopole, 20157, Milano, Italy
Department of Management, Economics and Industrial Engineering, Politecnico di Milano, Via Lambruschini 4/b, 20156, Milano, Italy
Lucia Tajoli
Department of Electronics, Information and Bioengineering, Politecnico di Milano, Piazza Leonardo da Vinci 32, 20133, Milano, Italy
Carlo Piccardi
You can also search for this author in PubMed Google Scholar
Contributions
M.T., F.I., L.T., and C.P. conceived the research, conducted the experiments, wrote and reviewed the manuscript.
Corresponding author
Correspondence to Carlo Piccardi .
Ethics declarations
Competing interests.
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Supplementary information, rights and permissions.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ .
Reprints and permissions
About this article
Cite this article.
Tantardini, M., Ieva, F., Tajoli, L. et al. Comparing methods for comparing networks. Sci Rep 9 , 17557 (2019). https://doi.org/10.1038/s41598-019-53708-y
Download citation
Received : 16 July 2019
Accepted : 25 October 2019
Published : 26 November 2019
DOI : https://doi.org/10.1038/s41598-019-53708-y
Share this article
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
This article is cited by
Altered resting-state functional connectivity in newborns with hypoxic ischemic encephalopathy assessed using high-density functional near-infrared spectroscopy.
- Lingkai Tang
- Lilian M. N. Kebaya
- Emma G. Duerden
Scientific Reports (2024)
Intrinsic dimension as a multi-scale summary statistics in network modeling
- Iuri Macocco
- Antonietta Mira
- Alessandro Laio
Altered functional connectivity in preterm neonates with intraventricular hemorrhage assessed using functional near-infrared spectroscopy
An embedding-based distance for temporal graphs.
- Lorenzo Dall’Amico
- Alain Barrat
- Ciro Cattuto
Nature Communications (2024)
- Helcio Felippe
- Federico Battiston
- Alec Kirkley
Communications Physics (2024)
Quick links
- Explore articles by subject
- Guide to authors
- Editorial policies
Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.
Advancements and Challenges in Networking Technologies: A Comprehensive Survey
Ieee account.
- Change Username/Password
- Update Address
Purchase Details
- Payment Options
- Order History
- View Purchased Documents
Profile Information
- Communications Preferences
- Profession and Education
- Technical Interests
- US & Canada: +1 800 678 4333
- Worldwide: +1 732 981 0060
- Contact & Support
- About IEEE Xplore
- Accessibility
- Terms of Use
- Nondiscrimination Policy
- Privacy & Opting Out of Cookies
A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.
IMAGES
VIDEO
COMMENTS
There are seven major research challenges the surmounting of which will enable progress in network science: Dynamics, spatial location, and information propagation in networks. Better understanding of the relationship between the architecture of a network and its function is needed.
Remote access and data communication render networks vulnerable to various threats. Thus, network security is essential for data sharing and communication. This article emphasises network...
Computer networking is an emerging and fascinating field of study. A computer network allows communication and exchange of information between two or more independent computers and other...
Our research combines building and deploying novel networking systems at unprecedented scale, with recent work focusing on fundamental questions around data center architecture, cloud virtual networking, and wide-area network interconnects.
Motivated by these successes, researchers in the field of networking apply deep learning models for Network Traffic Monitoring and Analysis (NTMA) applications, e.g., traffic classification and prediction. This paper provides a comprehensive review on applications of deep learning in NTMA.
First, we review and classify a collection of network comparison methods, highlighting the criteria they are based on and their advantages and drawbacks.
This survey article discusses techniques, difficulties, and future research by summarizing each technology. To help researchers and practitioners make educated judgments and enhance networking technologies.
A research goal that the social and physical sciences have shared has been to explain the formation of network ties and, more generally, to predict a host of network properties, such as the clusteredness of networks or the distributions of node centrality.
In order to reveal the deep social structure, this paper analyzes the structure of social networks from three levels: microlevel, mesolevel, and macrolevel and reveals the origin, development, perfection, and latest achievements of complex network models.
We reviewed and coded 89 articles and described the main social theories used in the network effectiveness literature and the associated mechanisms that translate network structures to network effects. We also explain how scholars operationalize those theoretical mechanisms through network measures.