No Arabic abstract
Much research has been done on studying the diffusion of ideas or technologies on social networks including the textit{Influence Maximization} problem and many of its variations. Here, we investigate a type of inverse problem. Given a snapshot of the diffusion process, we seek to understand if the snapshot is feasible for a given dynamic, i.e., whether there is a limited number of nodes whose initial adoption can result in the snapshot in finite time. While similar questions have been considered for epidemic dynamics, here, we consider this problem for variations of the deterministic Linear Threshold Model, which is more appropriate for modeling strategic agents. Specifically, we consider both sequential and simultaneous dynamics when deactivations are allowed and when they are not. Even though we show hardness results for all variations we consider, we show that the case of sequential dynamics with deactivations allowed is significantly harder than all others. In contrast, sequential dynamics make the problem trivial on cliques even though its complexity for simultaneous dynamics is unknown. We complement our hardness results with structural insights that can help better understand diffusions of social networks under various dynamics.
Graphs have been utilized as a powerful tool to model pairwise relationships between people or objects. Such structure is a special type of a broader concept referred to as hypergraph, in which each hyperedge may consist of an arbitrary number of nodes, rather than just two. A large number of real-world datasets are of this form - for example, list of recipients of emails sent from an organization, users participating in a discussion thread or subject labels tagged in an online question. However, due to complex representations and lack of adequate tools, little attention has been paid to exploring the underlying patterns in these interactions. In this work, we empirically study a number of real-world hypergraph datasets across various domains. In order to enable thorough investigations, we introduce the multi-level decomposition method, which represents each hypergraph by a set of pairwise graphs. Each pairwise graph, which we refer to as a k-level decomposed graph, captures the interactions between pairs of subsets of k nodes. We empirically find that at each decomposition level, the investigated hypergraphs obey five structural properties. These properties serve as criteria for evaluating how realistic a hypergraph is, and establish a foundation for the hypergraph generation problem. We also propose a hypergraph generator that is remarkably simple but capable of fulfilling these evaluation metrics, which are hardly achieved by other baseline generator models.
As individuals communicate, their exchanges form a dynamic network. We demonstrate, using time series analysis of communication in three online settings, that network structure alone can be highly revealing of the diversity and novelty of the information being communicated. Our approach uses both standard and novel network metrics to characterize how unexpected a network configuration is, and to capture a networks ability to conduct information. We find that networks with a higher conductance in link structure exhibit higher information entropy, while unexpected network configurations can be tied to information novelty. We use a simulation model to explain the observed correspondence between the evolution of a networks structure and the information it carries.
We study the structure of loops in networks using the notion of modulus of loop families. We introduce a new measure of network clustering by quantifying the richness of families of (simple) loops. Modulus tries to minimize the expected overlap among loops by spreading the expected link-usage optimally. We propose weighting networks using these expected link-usages to improve classical community detection algorithms. We show that the proposed method enhances the performance of certain algorithms, such as spectral partitioning and modularity maximization heuristics, on standard benchmarks.
We study social networks and focus on covert (also known as hidden) networks, such as terrorist or criminal networks. Their structures, memberships and activities are illegal. Thus, data about covert networks is often incomplete and partially incorrect, making interpreting structures and activities of such networks challenging. For legal reasons, real data about active covert networks is inaccessible to researchers. To address these challenges, we introduce here a network generator for synthetic networks that are statistically similar to a real network but void of personal information about its members. The generator uses statistical data about a real or imagined covert organization network. It generates randomized instances of the Stochastic Block model of the network groups but preserves this network organizational structure. The direct use of such anonymized networks is for training on them the research and analytical tools for finding structure and dynamics of covert networks. Since these synthetic networks differ in their sets of edges and communities, they can be used as a new source for network analytics. First, they provide alternative interpretations of the data about the original network. The distribution of probabilities for these alternative interpretations enables new network analytics. The analysts can find community structures which are frequent, therefore stable under perturbations. They may also analyze how the stability changes with the strength of perturbation. For covert networks, the analysts can quantify statistically expected outcomes of interdiction. This kind of analytics applies to all complex network in which the data are incomplete or partially incorrect.
Network structures, consisting of nodes and edges, have applications in almost all subjects. A set of nodes is called a community if the nodes have strong interrelations. Industries (including cell phone carriers and online social media companies) need community structures to allocate network resources and provide proper and accurate services. However, all the current detection algorithms are motivated by the practical problems, whose applicabilities in other fields are open to question. Thence, for a new community problem, researchers need to derive algorithms ad hoc, which is arduous and even unnecessary. In this paper, we represent a general procedure to find community structures in practice. We mainly focus on two typical types of networks: transmission networks and similarity networks. We reduce them to a unified graph model, based on which we propose a general method to define and detect communities. Readers can specialize our general algorithm to accommodate their problems. In the end, we also give a demonstration to show how the algorithm works.