text stringlengths 17 3.36M | source stringlengths 3 333 | __index_level_0__ int64 0 518k |
|---|---|---|
The rapid growth of content distribution on the Internet has brought with it proportional increases in the costs of distributing content. Adding to distribution costs is the fact that digital content is easily duplicable, and hence can be shared in an illicit peer-to-peer (P2P) manner that generates no revenue for the content provider. In this paper, we study whether the content provider can recover lost revenue through a more innovative approach to distribution. In particular, we evaluate the benefits of a hybrid revenue-sharing system that combines a legitimate P2P swarm and a centralized client-server approach. We show how the revenue recovered by the content provider using a server-supported legitimate P2P swarm can exceed that of the monopolistic scheme by an order of magnitude. Our analytical results are obtained in a fluid model, and supported by stochastic simulations. | Incentives for P2P-Assisted Content Distribution: If You Can't Beat 'Em,
Join 'Em | 8,900 |
An online user joins multiple social networks in order to enjoy different services. On each joined social network, she creates an identity and constitutes its three major dimensions namely profile, content and connection network. She largely governs her identity formulation on any social network and therefore can manipulate multiple aspects of it. With no global identifier to mark her presence uniquely in the online domain, her online identities remain unlinked, isolated and difficult to search. Earlier research has explored the above mentioned dimensions, to search and link her multiple identities with an assumption that the considered dimensions have been least disturbed across her identities. However, majority of the approaches are restricted to exploitation of one or two dimensions. We make a first attempt to deploy an integrated system (Finding Nemo) which uses all the three dimensions of an identity to search for a user on multiple social networks. The system exploits a known identity on one social network to search for her identities on other social networks. We test our system on two most popular and distinct social networks - Twitter and Facebook. We show that the integrated system gives better accuracy than the individual algorithms. We report experimental findings in the report. | Finding Nemo: Searching and Resolving Identities of Users Across Online
Social Networks | 8,901 |
With the growing popularity and usage of online social media services, people now have accounts (some times several) on multiple and diverse services like Facebook, LinkedIn, Twitter and YouTube. Publicly available information can be used to create a digital footprint of any user using these social media services. Generating such digital footprints can be very useful for personalization, profile management, detecting malicious behavior of users. A very important application of analyzing users' online digital footprints is to protect users from potential privacy and security risks arising from the huge publicly available user information. We extracted information about user identities on different social networks through Social Graph API, FriendFeed, and Profilactic; we collated our own dataset to create the digital footprints of the users. We used username, display name, description, location, profile image, and number of connections to generate the digital footprints of the user. We applied context specific techniques (e.g. Jaro Winkler similarity, Wordnet based ontologies) to measure the similarity of the user profiles on different social networks. We specifically focused on Twitter and LinkedIn. In this paper, we present the analysis and results from applying automated classifiers for disambiguating profiles belonging to the same user from different social networks. UserID and Name were found to be the most discriminative features for disambiguating user profiles. Using the most promising set of features and similarity metrics, we achieved accuracy, precision and recall of 98%, 99%, and 96%, respectively. | Studying User Footprints in Different Online Social Networks | 8,902 |
Este art\'iculo presenta un an\'alisis de los tweets recolectados el 28 de Octubre de 2012, en el contexto de las elecciones municipales de 2012 en Chile. Dicho an\'alisis se realiza mediante una metodolog\'ia basada en literatura previa, en particular en t\'ecnicas de recuperaci\'on de la informaci\'on y de an\'alisis de espacios de informaci\'on. Como resultado, se determinan: 1) caracter\'isticas demogr\'aficas b\'asicas de la poblaci\'on virtual chilena, incluyendo su distribuci\'on geogr\'afica, 2) el contenido que caracteriza a cada regi\'on, y c\'omo fluye informaci\'on entre regiones, y 3) el grado de representatividad de la poblaci\'on virtual participante en el evento con respecto a la poblaci\'on f\'isica. Se determina que la muestra obtenida es representativa de la poblaci\'on en t\'erminos de distribuci\'on geogr\'afica, que el centralismo que afecta al pa\'is se ve reflejado en Twitter, y que, a pesar de los sesgos poblacionales, es posible identificar el contenido que caracteriza a cada regi\'on. Se finaliza con una discusi\'on de las implicaciones y conclusiones pr\'acticas de este trabajo, as\'i como futuras aplicaciones. | Ornitología Virtual: Caracterizando a #Chile en Twitter | 8,903 |
We make a formal sociologic study of the concept of free will. By using the language of mathematics and logic, we define what we call everlasting societies. Everlasting societies never age: persons never age, and the goods of the society are indestructible. The infinite history of an everlasting society unfolds by following deterministic and probabilistic laws that do their best to satisfy the free will of all the persons of the society. We define three possible kinds of histories for everlasting societies: primitive histories, good histories, and golden histories. In primitive histories, persons are inherently selfish, and they use their free will to obtain the personal ownerships of all the goods of the society. In good histories, persons are inherently good, and they use their free will to distribute the goods of the society. In good histories, a person is not only able to desire the personal ownership of goods, but is also able to desire that a good be owned by another person. In golden histories, free will is bound by the ethic of reciprocity, which states that "you should wish upon others as you would like others to wish upon yourself". In golden societies, the ethic of reciprocity becomes a law that partially binds free will, and that must be abided at all times. In other words, the verb "should" becomes the verb "must". | A Formal Sociologic Study of Free Will | 8,904 |
With the popularity of Social Networking Services (SNS), more and more sensitive information are stored online and associated with SNS accounts. The obvious value of SNS accounts motivates the usage stealing problem -- unauthorized, stealthy use of SNS accounts on the devices owned/used by account owners without any technology hacks. For example, anxious parents may use their kids' SNS accounts to inspect the kids' social status; husbands/wives may use their spouses' SNS accounts to spot possible affairs. Usage stealing could happen anywhere in any form, and seriously invades the privacy of account owners. However, there is no any currently known defense against such usage stealing. To an SNS operator (e.g., Facebook Inc.), usage stealing is hard to detect using traditional methods because such attackers come from the same IP addresses/devices, use the same credentials, and share the same accounts as the owners do. In this paper, we propose a novel continuous authentication approach that analyzes user browsing behavior to detect SNS usage stealing incidents. We use Facebook as a case study and show that it is possible to detect such incidents by analyzing SNS browsing behavior. Our experiment results show that our proposal can achieve higher than 80% detection accuracy within 2 minutes, and higher than 90% detection accuracy after 7 minutes of observation time. | Is Somebody Watching Your Facebook Newsfeed? | 8,905 |
This article presents a characterization of the web space from Chile in 2007. The characterization shows distributions of sites and domains, analysis of document content and server configuration. In addition, the network structure of the chilean Web is analyzed, determining components based on hyperlink structure at the document and site levels. Original Abstract: En este art\'iculo se muestra una caracterizaci\'on del espacio web de Chile para el a\~no 2007. Se muestran distribuciones de sitios y dominios, caracterizaci\'on del contenido en base a tipos de documento, asi como configuraci\'on de los servidores. Se estudia la estructura de la red creada mediante hiperv\'inculos en los documentos y c\'omo las diferentes componentes de esta estructura var\'ian cuando los hiperv\'inculos son agregados a nivel de sitios. | Caracterizando la Web Chilena | 8,906 |
In this paper, we consider the problem of locating the information source with sparse observations. We assume that a piece of information spreads in a network following a heterogeneous susceptible-infected-recovered (SIR) model and that a small subset of infected nodes are reported, from which we need to find the source of the information. We adopt the sample path based estimator developed in [1], and prove that on infinite trees, the sample path based estimator is a Jordan infection center with respect to the set of observed infected nodes. In other words, the sample path based estimator minimizes the maximum distance to observed infected nodes. We further prove that the distance between the estimator and the actual source is upper bounded by a constant independent of the number of infected nodes with a high probability on infinite trees. Our simulations on tree networks and real world networks show that the sample path based estimator is closer to the actual source than several other algorithms. | A Robust Information Source Estimator with Sparse Observations | 8,907 |
Many real-world processes evolve in cascades over complex networks, whose topologies are often unobservable and change over time. However, the so-termed adoption times when blogs mention popular news items, individuals in a community catch an infectious disease, or consumers adopt a trendy electronics product are typically known, and are implicitly dependent on the underlying network. To infer the network topology, a \textit{dynamic} structural equation model is adopted to capture the relationship between observed adoption times and the unknown edge weights. Assuming a slowly time-varying topology and leveraging the sparse connectivity inherent to social networks, edge weights are estimated by minimizing a sparsity-regularized exponentially-weighted least-squares criterion. To this end, solvers with complementary strengths are developed by leveraging (pseudo) real-time sparsity-promoting proximal gradient iterations, the improved convergence rate of accelerated variants, or reduced computational complexity of stochastic gradient descent. Numerical tests with both synthetic and real data demonstrate the effectiveness of the novel algorithms in unveiling sparse dynamically-evolving topologies, while accounting for external influences in the adoption times. Key events in the recent succession of political leadership in North Korea, explain connectivity changes observed in the associated network inferred from global cascades of online media. | Dynamic Structural Equation Models for Social Network Topology Inference | 8,908 |
Role analysis in online communities allows us to understand and predict users behavior. Though several approaches have been followed, there is still lack of generalization of their methods and their results. In this paper, we discuss about the ground theory of roles and search for a consistent and computable definition that allows the automatic detection of roles played by users in forum threads on the internet. We analyze the web site IMDb to illustrate the discussion. | Analyse des rôles dans les communautés virtuelles : définitions et
premières expérimentations sur IMDb | 8,909 |
Traditional network generation models attempt to replicate global structural properties (degree distribution, average distance, clustering coefficient, communities, etc.) through synthetic link formation mechanisms such as triadic closure or preferential attachment. In this work, we study the evolution of a very big communication network coming from mobile telephony and we analyse the link formation process. A first study conducted on the standard mechanisms allows observing that several mechanisms are responsible for the properties observed in this network. In a second study, we characterize more precisely the link formation process by searching for correlations between the probability of creating a new link and some individual properties such as the degree, the clustering coefficient and the age of the nodes. | Dynamiques globales et locales dans un réseau de
télécommunications | 8,910 |
The modern age has seen an exponential growth of social network data available on the web. Analysis of these networks reveal important structural information about these networks in particular and about our societies in general. More often than not, analysis of these networks is concerned in identifying similarities among social networks and how they are different from other networks such as protein interaction networks, computer networks and food web. In this paper, our objective is to perform a critical analysis of different social networks using structural metrics in an effort to highlight their similarities and differences. We use five different social network datasets which are contextually and semantically different from each other. We then analyze these networks using a number of different network statistics and metrics. Our results show that although these social networks have been constructed from different contexts, they are structurally similar. We also review the snowball sampling method and show its vulnerability against different network metrics. | Are all Social Networks Structurally Similar? A Comparative Study using
Network Statistics and Metrics | 8,911 |
Decelle et al.\cite{Decelle11} conjectured the existence of a sharp threshold for community detection in sparse random graphs drawn from the stochastic block model. Mossel et al.\cite{Mossel12} established the negative part of the conjecture, proving impossibility of meaningful detection below the threshold. However the positive part of the conjecture remained elusive so far. Here we solve the positive part of the conjecture. We introduce a modified adjacency matrix $B$ that counts self-avoiding paths of a given length $\ell$ between pairs of nodes and prove that for logarithmic $\ell$, the leading eigenvectors of this modified matrix provide non-trivial detection, thereby settling the conjecture. A key step in the proof consists in establishing a {\em weak Ramanujan property} of matrix $B$. Namely, the spectrum of $B$ consists in two leading eigenvalues $\rho(B)$, $\lambda_2$ and $n-2$ eigenvalues of a lower order $O(n^{\epsilon}\sqrt{\rho(B)})$ for all $\epsilon>0$, $\rho(B)$ denoting $B$'s spectral radius. $d$-regular graphs are Ramanujan when their second eigenvalue verifies $|\lambda|\le 2 \sqrt{d-1}$. Random $d$-regular graphs have a second largest eigenvalue $\lambda$ of $2\sqrt{d-1}+o(1)$ (see Friedman\cite{friedman08}), thus being {\em almost} Ramanujan. Erd\H{o}s-R\'enyi graphs with average degree $d$ at least logarithmic ($d=\Omega(\log n)$) have a second eigenvalue of $O(\sqrt{d})$ (see Feige and Ofek\cite{Feige05}), a slightly weaker version of the Ramanujan property. However this spectrum separation property fails for sparse ($d=O(1)$) Erd\H{o}s-R\'enyi graphs. Our result thus shows that by constructing matrix $B$ through neighborhood expansion, we regularize the original adjacency matrix to eventually recover a weak form of the Ramanujan property. | Community detection thresholds and the weak Ramanujan property | 8,912 |
Recent years have seen tremendous growth of many online social networks such as Facebook, LinkedIn and MySpace. People connect to each other through these networks forming large social communities providing researchers rich datasets to understand, model and predict social interactions and behaviors. New contacts in these networks can be formed either due to an individual's demographic profile such as age group, gender, geographic location or due to network's structural dynamics such as triadic closure and preferential attachment, or a combination of both demographic and structural characteristics. A number of network generation models have been proposed in the last decade to explain the structure, evolution and processes taking place in different types of networks, and notably social networks. Network generation models studied in the literature primarily consider structural properties, and in some cases an individual's demographic profile in the formation of new social contacts. These models do not present a mechanism to combine both structural and demographic characteristics for the formation of new links. In this paper, we propose a new network generation algorithm which incorporates both these characteristics to model growth of a network.We use different publicly available Facebook datasets as benchmarks to demonstrate the correctness of the proposed network generation model. | Demographic and Structural Characteristics to Rationalize Link Formation
in Online Social Networks | 8,913 |
In this paper we present a web-based information system which is a portfolio social network (PSN) that provides solutions to recruiters and job seekers. The proposed system enables users to create portfolios so that he/she can add his specializations with piece of code, if any, specifically for software engineers, which is accessible online. The unique feature of the system is to enable the recruiters to quickly view the prominent skills of the users. A comparative analysis of the proposed system with the state of the art systems is presented. The comparative study reveals that the proposed system has advanced functionalities. | PSN: Portfolio Social Network | 8,914 |
Online Social Network (OSN) is one of the most hottest services in the past years. It preserves the life of users and provides great potential for journalists, sociologists and business analysts. Crawling data from social network is a basic step for social network information analysis and processing. As the network becomes huge and information on the network updates faster than web pages, crawling is more difficult because of the limitations of band-width, politeness etiquette and computation power. To extract fresh information from social network efficiently and effectively, this paper presents a novel crawling method and discusses parallelization architecture of social network. To discover the feature of social network, we gather data from real social network, analyze them and build a model to describe the discipline of users' behavior. With the modeled behavior, we propose methods to predict users' behavior. According to the prediction, we schedule our crawler more reasonably and extract more fresh information with parallelization technologies. Experimental results demonstrate that our strategies could obtain information from OSN efficiently and effectively. | Parallelization in Extracting Fresh Information from Online Social
Network | 8,915 |
We study user behavior in the courses offered by a major Massive Online Open Course (MOOC) provider during the summer of 2013. Since social learning is a key element of scalable education in MOOCs and is done via online discussion forums, our main focus is in understanding forum activities. Two salient features of MOOC forum activities drive our research: 1. High decline rate: for all courses studied, the volume of discussions in the forum declines continuously throughout the duration of the course. 2. High-volume, noisy discussions: at least 30% of the courses produce new discussion threads at rates that are infeasible for students or teaching staff to read through. Furthermore, a substantial portion of the discussions are not directly course-related. We investigate factors that correlate with the decline of activity in the online discussion forums and find effective strategies to classify threads and rank their relevance. Specifically, we use linear regression models to analyze the time series of the count data for the forum activities and make a number of observations, e.g., the teaching staff's active participation in the discussion increases the discussion volume but does not slow down the decline rate. We then propose a unified generative model for the discussion threads, which allows us both to choose efficient thread classifiers and design an effective algorithm for ranking thread relevance. Our ranking algorithm is further compared against two baseline algorithms, using human evaluation from Amazon Mechanical Turk. The authors on this paper are listed in alphabetical order. For media and press coverage, please refer to us collectively, as "researchers from the EDGE Lab at Princeton University, together with collaborators at Boston University and Microsoft Corporation." | Learning about social learning in MOOCs: From statistical analysis to
generative model | 8,916 |
The notion of community structure is particularly useful when analyzing complex networks, because it provides an intermediate level, compared to the more classic global (whole network) and local (node neighborhood) approaches. The concept of community role of a node was derived from this base, in order to describe the position of a node in a network depending on its connectivity at the community level. However, the existing approaches are restricted to undirected networks, use topological measures which do not consider all aspects of community-related connectivity, and their role identification methods are not generalizable to all networks. We tackle these limitations by generalizing and extending the measures, and using an unsupervised approach to determine the roles. We then illustrate the applicability of our method by analyzing a Twitter network.We show how our modifications allow discovering the fact some particular users called social capitalists occupy very specific roles in this system. --- La notion de structure de communaut\'es est particuli\`erement utile pour \'etudier les r\'eseaux complexes, car elle am\`ene un niveau d'analyse interm\'ediaire, par opposition aux plus classiques niveaux local (voisinage des noeuds) et global (r\'eseau entier). Le concept de r\^ole communautaire permet de d\'ecrire le positionnement d'un noeud en fonction de sa connectivit\'e communautaire. Cependant, les approches existantes sont restreintes aux r\'eseaux non-orient\'es, utilisent des mesures topologiques ne consid\'erant pas tous les aspects de la connectivit\'e communautaire, et des m\'ethodes d'identification des r\^oles non-g\'en\'eralisables \`a tous les r\'eseaux. Nous proposons de r\'esoudre ces probl\`emes en g\'en\'eralisant les mesures existantes, et en utilisant une m\'ethode non-supervis\'ee pour d\'eterminer les r\^oles. Nous illustrons l'int\'er\^et de notre m\'ethode en l'appliquant au r\'eseau de Twitter. Nous montrons que nos modifications mettent en \'evidence les r\^oles sp\'ecifiques d'utilisateurs particuliers du r\'eseau, nomm\'es capitalistes sociaux. | Identification de rôles communautaires dans des réseaux orientés
appliquée à Twitter | 8,917 |
Many complex systems are modeled through complex networks whose analysis reveals typical topological properties. Amongst those, the community structure is one of the most studied. Many methods are proposed to detect communities, not only in plain, but also in attributed, directed or even dynamic networks. A community structure takes the form of a partition of the node set, which must then be characterized relatively to the properties of the studied system. We propose a method to support such a characterization task. We define a sequence-based representation of networks, combining temporal information, topological measures, and nodal attributes. We then characterize communities using the most representative emerging sequential patterns of its nodes. This also allows detecting unusual behavior in a community. We describe an empirical study of a network of scientific collaborations.---De nombreux syst\`emes complexes sont \'etudi\'es via l'analyse de r\'eseaux dits complexes ayant des propri\'et\'es topologiques typiques. Parmi cellesci, les structures de communaut\'es sont particuli\`erement \'etudi\'ees. De nombreuses m\'ethodes permettent de les d\'etecter, y compris dans des r\'eseaux contenant des attributs nodaux, des liens orient\'es ou \'evoluant dans le temps. La d\'etection prend la forme d'une partition de l'ensemble des noeuds, qu'il faut ensuite caract\'eriser relativement au syst\`eme mod\'elis\'e. Nous travaillons sur l'assistance \`a cette t\^ache de caract\'erisation. Nous proposons une repr\'esentation des r\'eseaux sous la forme de s\'equences de descripteurs de noeuds, qui combinent les informations temporelles, les mesures topologiques, et les valeurs des attributs nodaux. Les communaut\'es sont caract\'eris\'ees au moyen des motifs s\'equentiels \'emergents les plus repr\'esentatifs issus de leurs noeuds. Ceci permet notamment la d\'etection de comportements inhabituels au sein d'une communaut\'e. Nous d\'ecrivons une \'etude empirique sur un r\'eseau de collaboration scientifique. | Une méthode pour caractériser les communautés des réseaux
dynamiques à attributs | 8,918 |
The position of the nodes within a network topology largely determines the level of their involvement in various networking functions. Yet numerous node centrality indices, proposed to quantify how central individual nodes are in this respect, yield very different views of their relative significance. Our first contribution in this paper is then an exhaustive survey and categorization of centrality indices along several attributes including the type of information (local vs. global) and processing complexity required for their computation. We next study the seven most popular of those indices in the context of Internet vulnerability to address issues that remain under-explored in literature so far. First, we carry out a correlation study to assess the consistency of the node rankings those indices generate over ISP router-level topologies. For each pair of indices, we compute the full ranking correlation, which is the standard choice in literature, and the percentage overlap between the k top nodes. Then, we let these rankings guide the removal of highly central nodes and assess the impact on both the connectivity properties and traffic-carrying capacity of the network. Our results confirm that the top-k overlap predicts the comparative impact of indices on the network vulnerability better than the full-ranking correlation. Importantly, the locally computed degree centrality index approximates closely the global indices with the most dramatic impact on the traffic-carrying capacity; whereas, its approximative power in terms of connectivity is more topology-dependent. | The Multiple Instances of Node Centrality and their Implications on the
Vulnerability of ISP Networks | 8,919 |
In this paper, we consider the problem of latent sentiment detection in Online Social Networks such as Twitter. We demonstrate the benefits of using the underlying social network as an Ising prior to perform network aided sentiment detection. We show that the use of the underlying network results in substantially lower detection error rates compared to strictly features-based detection. In doing so, we introduce a novel communications-oriented framework for characterizing the probability of error, based on information-theoretic analysis. We study the variation of the calculated error exponent for several stylized network topologies such as the complete network, the star network and the closed-chain network, and show the importance of the network structure in determining detection performance. | Latent Sentiment Detection in Online Social Networks: A
Communications-oriented View | 8,920 |
With an expansive and ubiquitously available gold mine of educational data, Massive Open Online courses (MOOCs) have become the an important foci of learning analytics research. In this paper, we investigate potential reasons as to why are these digitalized learning repositories being plagued with huge attrition rates. We analyze an ongoing online course offered in Coursera using a social network perspective, with an objective to identify students who are actively participating in course discussions and those who are potentially at a risk of dropping off. We additionally perform extensive forum analysis to visualize student's posting patterns longitudinally. Our results provide insights that can assist educational designers in establishing a pedagogical basis for decision-making while designing MOOCs. We infer prominent characteristics about the participation patterns of distinct groups of students in the networked learning community, and effectively discover important discussion threads. These methods can, despite the otherwise prohibitive number of students involved, allow an instructor to leverage forum behavior to identify opportunities for support. | Supporting MOOC Instruction with Social Network Analysis | 8,921 |
The spread of non-indigenous species (NIS) through the global shipping network (GSN) has enormous ecological and economic cost throughout the world. Previous attempts at quantifying NIS invasions have mostly taken "bottom-up" approaches that eventually require the use of multiple simplifying assumptions due to insufficiency and/or uncertainty of available data. By modeling implicit species exchanges via a graph abstraction that we refer to as the Species Flow Network (SFN), a different approach that exploits the power of network science methods in extracting knowledge from largely incomplete data is presented. Here, coarse-grained species flow dynamics are studied via a graph clustering approach that decomposes the SFN to clusters of ports and inter-cluster connections. With this decomposition of ports in place, NIS flow among clusters can be very efficiently reduced by enforcing NIS management on a few chosen inter-cluster connections. Furthermore, efficient NIS management strategy for species exchanges within a cluster (often difficult due higher rate of travel and pathways) are then derived in conjunction with ecological and environmental aspects that govern the species establishment. The benefits of the presented approach include robustness to data uncertainties, implicit incorporation of "stepping-stone" spread of invasive species, and decoupling of species spread and establishment risk estimation. Our analysis of a multi-year (1997--2006) GSN dataset using the presented approach shows the existence of a few large clusters of ports with higher intra-cluster species flow that are fairly stable over time. Furthermore, detailed investigations were carried out on vessel types, ports, and inter-cluster connections. Finally, our observations are discussed in the context of known NIS invasions and future research directions are also presented. | Patterns of Ship-borne Species Spread: A Clustering Approach for Risk
Assessment and Management of Non-indigenous Species Spread | 8,922 |
Despite the several advantages commonly attributed to social networks such as easiness and immediacy to communicate with acquaintances and friends, significant privacy threats provoked by unexperienced or even irresponsible users recklessly publishing sensitive material are also noticeable. Yet, a different, but equally hazardous privacy risk might arise from social networks profiling the online activity of their users based on the timestamp of the interactions between the former and the latter. In order to thwart this last type of commonly neglected attacks, this paper presents a novel, smart deferral mechanism for messages in online social networks. Such solution suggests intelligently delaying certain messages posted by end users in social networks in a way that the observed online-activity profile generated by the attacker does not reveal any time-based sensitive information. Conducted experiments as well as a proposed architecture implementing this approach demonstrate the suitability and feasibility of our mechanism. | Smart Deferral of Messages for Privacy Protection in Online Social
Networks | 8,923 |
This research explains the importance of transculturality in social networking in a wide variety of activities of our daily life. We focus our analysis to online activities that use social richness, analyzing societies in Yakutia (A Russian Republic), Macau in China, Uberl\^andia in Brazil and Juarez City in Mexico, all with people descending from Japanese people. To this end, we performed surveys to gathering information about salient aspects of upgrade and combined them using social data mining techniques to profile a number of behavioural patterns and choices that describe social networking behaviours in these societies. | Never forget, whom was my ancestors: A cross-cultural analysis from
Yonsei (fourth-generation Nikkei) in four societies using Data Mining | 8,924 |
Conventional online social networks (OSNs) are implemented in a centralized manner. Although centralization is a convenient way for implementing OSNs, it has several well known drawbacks. Chief among them are the risks they pose to the security and privacy of the information maintained by the OSN; and the loss of control over the information contributed by individual members. These concerns prompted several attempts to create decentralized OSNs, or DOSNs. The basic idea underlying these attempts, is that each member of a social network keeps its data under its own control, instead of surrendering it to a central host; providing access to it to other members of the OSN according to its own access-control policy. Unfortunately all existing DOSN projects have a very serious limitation. Namely, they are unable to subject the membership of a DOSN, and the interaction between its members, to any global policy. We adopt the decentralization idea underlying DOSNs, complementing it with a means for specifying and enforcing a wide range of policies over the membership of a social community, and over the interaction between its disparate distributed members. And we do so in a scalable fashion. | Establishing Global Policies over Decentralized Online Social Networks | 8,925 |
Despite the robustness of complex networks has been extensively studied in the last decade, there still lacks a unifying framework able to embrace all the proposed metrics. In the literature there are two open issues related to this gap: (a) how to dimension several metrics to allow their summation and (b) how to weight each of the metrics. In this work we propose a solution for the two aforementioned problems by defining the $R^*$-value and introducing the concept of \emph{robustness surface} ($\Omega$). The rationale of our proposal is to make use of Principal Component Analysis (PCA). We firstly adjust to 1 the initial robustness of a network. Secondly, we find the most informative robustness metric under a specific failure scenario. Then, we repeat the process for several percentage of failures and different realizations of the failure process. Lastly, we join these values to form the robustness surface, which allows the visual assessment of network robustness variability. Results show that a network presents different robustness surfaces (i.e., dissimilar shapes) depending on the failure scenario and the set of metrics. In addition, the robustness surface allows the robustness of different networks to be compared. | Robustness surfaces of complex networks | 8,926 |
The proliferation of social media such as real time microblogging and online reputation systems facilitate real time sensing of social patterns and behavior. In the last decade, sensing and decision making in social networks have witnessed significant progress in the electrical engineering, computer science, economics, finance, and sociology research communities. Research in this area involves the interaction of dynamic random graphs, socio-economic analysis, and statistical inference algorithms. This monograph provides a survey, tutorial development, and discussion of four highly stylized examples: social learning for interactive sensing; tracking the degree distribution of social networks; sensing and information diffusion; and coordination of decision making via game-theoretic learning. Each of the four examples is motivated by practical examples, and comprises of a literature survey together with careful problem formulation and mathematical analysis. Despite being highly stylized, these examples provide a rich variety of models, algorithms and analysis tools that are readily accessible to a signal processing, control/systems theory, and applied mathematics audience. | Interactive Sensing and Decision Making in Social Networks | 8,927 |
The unprecedented use of social media through smartphones and other web-enabled mobile devices has enabled the rapid adoption of platforms like Twitter. Event detection has found many applications on the web, including breaking news identification and summarization. The recent increase in the usage of Twitter during crises has attracted researchers to focus on detecting events in tweets. However, current solutions have focused on static Twitter data. The necessity to detect events in a streaming environment during fast paced events such as a crisis presents new opportunities and challenges. In this paper, we investigate event detection in the context of real-time Twitter streams as observed in real-world crises. We highlight the key challenges in this problem: the informal nature of text, and the high volume and high velocity characteristics of Twitter streams. We present a novel approach to address these challenges using single-pass clustering and the compression distance to efficiently detect events in Twitter streams. Through experiments on large Twitter datasets, we demonstrate that the proposed framework is able to detect events in near real-time and can scale to large and noisy Twitter streams. | From Tweets to Events: Exploring a Scalable Solution for Twitter Streams | 8,928 |
Exponential-family random graph models (ERGMs) are probabilistic network models that are parametrized by sufficient statistics based on structural (i.e., graph-theoretic) properties. The ergm package for the R statistical computing system is a collection of tools for the analysis of network data within an ERGM framework. Many different network properties can be employed as sufficient statistics for ERGMs by using the model terms defined in the ergm package; this functionality can be expanded by the creation of packages that code for additional network statistics. Here, our focus is on the addition of statistics based on graphlets. Graphlets are small, connected, and non-isomorphic induced subgraphs that describe the topological structure of a network. We introduce an R package called ergm.graphlets that enables the use of graphlet properties of a network within the ergm package of R. The ergm.graphlets package provides a complete list of model terms that allows to incorporate statistics of any 2-, 3-, 4- and 5-node graphlet into ERGMs. The new model terms of ergm.graphlets package enable both ERG modelling of global structural properties and investigation of relationships between nodal attributes (i.e., covariates) and local topologies around nodes. | ergm.graphlets: A Package for ERG Modeling Based on Graphlet Statistics | 8,929 |
There are many indexes (measures or metrics) in Social Network Analysis (SNA), like density, cohesion, etc. We have defined a new SNA index called "comfortability". In this paper, core comfortable team of a social network is defined based on graph theoretic concepts and some of their structural properties are analyzed. Comfortability is one of the important attributes (characteristics) for a successful team work. So, it is necessary to find a comfortable and successful team in any given social network. It is proved that forming core comfortable team in any network is NP-Complete using the concepts of domination in graph theory. Next, we give two polynomial-time approximation algorithms for finding such a core comfortable team in any given network with performance ratio O(ln \Delta), where \Delta is the maximum degree of a given network (graph). The time complexity of the algorithm is proved to be O(n^{3}), where n is the number of persons (vertices) in the network (graph). It is also proved that the algorithms give good results in scale-free networks. | Comfortability of a Team in Social Networks | 8,930 |
Recent work [Hazy 2012] has demonstrated computationally that collectives that are organized into networks which govern the flow of resources can learn to recognize newly emerging opportunities distributed in the environment. This paper argues that the system does this through a process analogous to neural network learning with relative status playing the role of synaptic weights. Hazy showed computationally that learning of this type can occur even when resource allocation decision makers have no direct visibility into the environment, have no direct understanding of the opportunity, and are not involved in their exploitation except to the extent that they evaluate the success or failure of funded projects. Effectively, the system of interactions learns which individuals have the best access to information and other resources within the ecosystem. Hazy [2012] calls this previously unidentified emergence phenomenon: Influence Process Structural Learning (IPSL). In the prior model of IPSL, a three-tiered organizational structure was predetermined in the model design [Hazy 2012]. These initial conditions delimit the extent to which the emergence of collective intelligence can be posited because the model itself assumes a defined structure. This work contributes to the field by extending the IPSL argument for collective intelligence to a holistic emergence argument. It begins by briefly reviewing previously published work. It continues the conversation by adding two additional steps: Firstly, it shows how a three-tier organizing structure might emerge through known complexity mechanisms. In this case the mechanism identified is preferential attachment [Barabasi 2002]. Secondly, the paper shows how collective intelligence can emerge within a system of agents when the influence structure among these agents is treated as a the genetic algorithm. | Influence Process Structural Learning and the Emergence of Collective
Intelligence | 8,931 |
Given the rapid proliferation of advanced information technologies, including the Internet, modern humans can easily access vast amount of socially transmitted information. Intuitively, this situation is isomorphic to some eusocial insects that are known to solve the exploration-exploitation dilemma collectively through information transfer (e.g., honeybees [Seeley et al., 1991]; and ants [Shaffer, Sasaki & Pratt, 2013]). Yet, in contrast from the eusocial insects, whose colonies are composed of kin, human collective performance may be affected by an inherent free-rider problem [Bolton & Harris, 1999; Kameda, Tsukasaki, Hastie & Berg, 2011]. Specifically, in groups involving non-kin members, it is expected that free-riders, who allow others to search for better alternatives and then exploit their findings through social learning ("information scroungers"), will frequently appear, and consequently undermine the advantage of collective intelligence [Rogers, 1998; Kameda & Nakanishi, 2003]. | Less-is-more in a 5-star rating system: an experimental study of human
combined decisions in a multi-armed bandit problem | 8,932 |
This paper analyzes correlations and causalities between Bitcoin market indicators and Twitter posts containing emotional signals on Bitcoin. Within a timeframe of 104 days (November 23rd 2013 - March 7th 2014), about 160,000 Twitter posts containing "bitcoin" and a positive, negative or uncertainty related term were collected and further analyzed. For instance, the terms "happy", "love", "fun", "good", "bad", "sad" and "unhappy" represent positive and negative emotional signals, while "hope", "fear" and "worry" are considered as indicators of uncertainty. The static (daily) Pearson correlation results show a significant positive correlation between emotional tweets and the close price, trading volume and intraday price spread of Bitcoin. However, a dynamic Granger causality analysis does not confirm a statistically significant effect of emotional Tweets on Bitcoin market values. To the contrary, the analyzed data shows that a higher Bitcoin trading volume Granger causes more signals of uncertainty within a 24 to 72-hour timeframe. This result leads to the interpretation that emotional sentiments rather mirror the market than that they make it predictable. Finally, the conclusion of this paper is that the microblogging platform Twitter is Bitcoin's virtual trading floor, emotionally reflecting its trading dynamics. | Nowcasting the Bitcoin Market with Twitter Signals | 8,933 |
The only acceptable form of polling in the multi-billion dollar survey research field utilizes representative samples. We argue that with proper statistical adjustment, non-representative polling can provide accurate predictions, and often in a much more timely and cost-effective fashion. We demonstrate this by applying multilevel regression and post-stratification (MRP) to a 2012 election survey on the Xbox gaming platform. Not only do the transformed top-line projections from this data closely trend standard indicators, but we use the unique nature of the data's size and panel to answer a meaningful political puzzle. We find that reported swings in public opinion polls are generally not due to actual shifts in vote intention, but rather are the result of temporary periods of relatively low response rates among supporters of the reportedly slumping candidate. This work shows great promise for using non-representative polling to measure public opinion and the first product of this new polling technique raises the possibility that decades of large, reported swings in public opinion-including the perennial "convention bounce"-are mostly artifacts of sampling bias. | The Mythical Swing Voter | 8,934 |
In an online prediction market, forecasters who could not see the current state of the market until they made their own separate estimates moved their estimates closer to the market forecast when the current state of the market became known. Their first edits to the market forecast were very similar to the first edits of forecasters who could always see the current state of the market, and forecasters in both conditions had similar accuracy. These results suggest that our more elaborate forecast elicitation method might not improve forecasts and that any anchoring on the state of the market does not constitute an error in judgment. | Interval Elicitation of Forecasts in a Prediction Market Reveals Lack of
Anchoring "Bias" | 8,935 |
In this paper we address the problem of inferring social structure and dominance relationships in a group of rhesus macaques (a species of monkey) using only position data captured using RFID tags. Automatic inference of the social structure in an animal group enables a number of important capabilities, including: 1) A verifiable measure of how the social structure is affected by an intervention such as a change in the environment, or the introduction of another animal, and 2) A potentially significant reduction in person hours normally used for assessing these changes. Social structure in a group is an important indicator of its members' relative level of access to resources and has interesting implications for an individual's health and learning in groups. There are two main quantitative criteria assessed in order to infer the social structure; Time spent close to conspecifics, and displacements. An interaction matrix is used to represent the total duration of events detected as grooming behavior between any two monkeys. This forms an undirected tie-strength (closeness of relationships) graph. A directed graph of hierarchy is constructed by using the well cited assumption of a linear hierarchy for rhesus macaques. Events that contribute to the adjacency matrix for this graph are withdrawals or displacements where a lower ranked monkey moves away from a higher ranked monkey. Displacements are one of the observable behaviors that can act as a strong indication of tie-strength and dominance. To quantify the directedness of interaction during these events we construct histograms of the dot products of motion orientation and relative position. This gives us a measure of how much time a monkey spends in moving towards or away from other group members. | Inferring Social Structure and Dominance Relationships Between Rhesus
macaques using RFID Tracking Data | 8,936 |
Background: Suicide issue is of great concern in China. Social media provides an active approach to understanding suicide individuals in terms of their behavior and language use. Aims: This study investigates how suicide Microblog users in China act and speak differently on social media from others. Methods: Hypothesis testing in behavioral and linguistic features was performed between a target group of 33 Chinese Microblog users who have committed suicide and a control group of 30 active users without suicidal ideation. Results: Suicide group significantly outnumbered control group in the extent of openly published posts and self-reference, and the intensity of using 7 word categories: negative words/social process words/cognitive process words/emotion process words/negative emotion words/exclusive words/physiological process words. Limitations: Information collection and confirmation of suicide users remain difficult. Conclusions: It is revealed that suicide people vary from others in certain behavioral and linguistic features in social media. This study fills the niche of suicide studies by noting specified indicators of suicide ideation for Chinese individuals online, providing insights of constructing an online alarm system for early detection and intervention of suicidal individuals. | How did the Suicide Act and Speak Differently Online? Behavioral and
Linguistic Features of China's Suicide Microblog Users | 8,937 |
In this paper, we explore the task of gender classification using limited network data with an application to Fotolog. We take a heuristic approach to automating gender inference based on username, followers and network structure. We test our approach on a subset of 100,000 nodes and analyze our results to find that there is a lot of value in these limited information and that there is great promise in further pursuing this approach to classification. | Gender Prediction in Social Media | 8,938 |
We provide a dynamic policy for the rapid containment of a contagion process modeled as an SIS epidemic on a bounded degree undirected graph with n nodes. We show that if the budget $r$ of curing resources available at each time is ${\Omega}(W)$, where $W$ is the CutWidth of the graph, and also of order ${\Omega}(\log n)$, then the expected time until the extinction of the epidemic is of order $O(n/r)$, which is within a constant factor from optimal, as well as sublinear in the number of nodes. Furthermore, if the CutWidth increases only sublinearly with n, a sublinear expected time to extinction is possible with a sublinearly increasing budget $r$. | An efficient curing policy for epidemics on graphs | 8,939 |
Understanding dynamics of evolution in large social networks is an important problem. In this paper, we characterize evolution in large multi-relational social networks. The proliferation of online media such as Twitter, Facebook, Orkut and MMORPGs\footnote{Massively Multi-player Online Role Playing Games} have created social networking data at an unprecedented scale. Sony's Everquest 2 is one such example. We used game multi-relational networks to reveal the dynamics of evolution in a multi-relational setting by macroscopic study of the game network. Macroscopic analysis involves fragmenting the network into smaller portions for studying the dynamics within these sub-networks, referred to as `communities'. From an evolutionary perspective of multi-relational network analysis, we have made the following contributions. Specifically, we formulated and analyzed various metrics to capture evolutionary properties of networks. We find that co-evolution rates in trust based `communities' are approximately $60\%$ higher than the trade based `communities'. We also find that the trust and trade connections within the `communities' reduce as their size increases. Finally, we study the interrelation between the dynamics of trade and trust within `communities' and find interesting results about the precursor relationship between the trade and the trust dynamics within the `communities'. | Understanding Co-evolution in Large Multi-relational Social Networks | 8,940 |
The degree distribution is an important characteristic of complex networks. In many data analysis applications, the networks should be represented as fixed-length feature vectors and therefore the feature extraction from the degree distribution is a necessary step. Moreover, many applications need a similarity function for comparison of complex networks based on their degree distributions. Such a similarity measure has many applications including classification and clustering of network instances, evaluation of network sampling methods, anomaly detection, and study of epidemic dynamics. The existing methods are unable to effectively capture the similarity of degree distributions, particularly when the corresponding networks have different sizes. Based on our observations about the structure of the degree distributions in networks over time, we propose a feature extraction and a similarity function for the degree distributions in complex networks. We propose to calculate the feature values based on the mean and standard deviation of the node degrees in order to decrease the effect of the network size on the extracted features. The proposed method is evaluated using different artificial and real network datasets, and it outperforms the state of the art methods with respect to the accuracy of the distance function and the effectiveness of the extracted features. | Feature Extraction from Degree Distribution for Comparison and Analysis
of Complex Networks | 8,941 |
Comfortability is one of the important attributes (characteristics) for a successful team work in any organization. It is necessary to find a comfortable and successful team in any given social network. We have introduced "comfortability" as a new SNA index. Comfortable team exists only in some social networks. In this paper, we analyze the existence of comfortable team in product graphs, such as strong product and Lexicographic product of two given graphs. | Existence of Comfortable Team in some Special Social Networks | 8,942 |
The online communities available on the Web have shown to be significantly interactive and capable of collectively solving difficult tasks. Nevertheless, it is still a challenge to decide how a task should be dispatched through the network due to the high diversity of the communities and the dynamically changing expertise and social availability of their members. We introduce CrowdSTAR, a framework designed to route tasks across and within online crowds. CrowdSTAR indexes the topic-specific expertise and social features of the crowd contributors and then uses a routing algorithm, which suggests the best sources to ask based on the knowledge vs. availability trade-offs. We experimented with the proposed framework for question and answering scenarios by using two popular social networks as crowd candidates: Twitter and Quora. | CrowdSTAR: A Social Task Routing Framework for Online Communities | 8,943 |
In this paper, we apply social network analytic methods to unveil the structural dynamics of a popular open source goal oriented IRC community, Ubuntu. The primary objective is to track the development of this ever growing community over time using a social network lens and examine the dynamically changing participation patterns of people. Specifically, our research seeks out to investigate answers to the following question: How can the communication dynamics help us in delineating important substructures in the IRC network? This gives an insight into how open source learning communities function internally and what drives the exhibited IRC behavior. By application of a consistent set of social network metrics, we discern factors that affect people's embeddedness in the overall IRC network, their structural influence and importance as discussion initiators or responders. Deciphering these informal connections are crucial for the development of novel strategies to improve communication and foster collaboration between people conversing in the IRC channel, there by stimulating knowledge flow in the network. Our approach reveals a novel network skeleton, that more closely resembles the behavior of participants interacting online. We highlight bottlenecks to effective knowledge dissemination in the IRC, so that focused attention could be provided to communities with peculiar behavioral patterns. Additionally, we explore interesting research directions in augmenting the study of communication dynamics in the IRC. | Towards Investigating Substructures and Role Recognition in Goal
Oriented Online Communities | 8,944 |
In this paper, we characterise the notion of preferential attachment in networks as action at a distance, and argue that it can only be an emergent phenomenon -- the actual mechanism by which networks grow always being the closing of triangles. After a review of the concepts of triangle closing and preferential attachment, we present our argument, as well as a simplified model in which preferential attachment can be derived mathematically from triangle closing. Additionally, we perform experiments on synthetic graphs to demonstrate the emergence of preferential attachment in graph growth models based only on triangle closing. | The Problem of Action at a Distance in Networks and the Emergence of
Preferential Attachment from Triadic Closure | 8,945 |
Internet is fast becoming critically important to commerce, industry and individuals. Search Engine (SE) is the most vital component for communication network and also used for discover information for users or people. Search engine optimization (SEO) is the process that is mostly used to increasing traffic from free, organic or natural listings on search engines and also helps to increase website ranking. It includes techniques like link building, directory submission, classified submission etc. but SMO, on the other hand, is the process of promoting your website on social media platforms. It includes techniques like RSS feeds, social news and bookmarking sites, video and blogging sites, as well as social networking sites, such as Facebook, Twitter, Google+, Tumblr, Pinterest, Instagram etc.Social media optimization is becoming increasingly important for search engine optimization, as search engines are increasingly utilizing the recommendations of users of social networks to rank pages in the search engine result pages. Since it is more difficult to tip the influence the search engines in this way. Social Media Optimization (SMO) may also use to generate traffic on a website, promote your business at the center of social marketing place and increase ranking. | Social Media Impact on Website Ranking | 8,946 |
Events in an online social network can be categorized roughly into endogenous events, where users just respond to the actions of their neighbors within the network, or exogenous events, where users take actions due to drives external to the network. How much external drive should be provided to each user, such that the network activity can be steered towards a target state? In this paper, we model social events using multivariate Hawkes processes, which can capture both endogenous and exogenous event intensities, and derive a time dependent linear relation between the intensity of exogenous events and the overall network activity. Exploiting this connection, we develop a convex optimization framework for determining the required level of external drive in order for the network to reach a desired activity level. We experimented with event data gathered from Twitter, and show that our method can steer the activity of the network more accurately than alternatives. | Shaping Social Activity by Incentivizing Users | 8,947 |
Rather than anonymizing social graphs by generalizing them to super nodes/edges or adding/removing nodes and edges to satisfy given privacy parameters, recent methods exploit the semantics of uncertain graphs to achieve privacy protection of participating entities and their relationship. These techniques anonymize a deterministic graph by converting it into an uncertain form. In this paper, we propose a generalized obfuscation model based on uncertain adjacency matrices that keep expected node degrees equal to those in the unanonymized graph. We analyze two recently proposed schemes and show their fitting into the model. We also point out disadvantages in each method and present several elegant techniques to fill the gap between them. Finally, to support fair comparisons, we develop a new tradeoff quantifying framework by leveraging the concept of incorrectness in location privacy research. Experiments on large social graphs demonstrate the effectiveness of our schemes. | Anonymizing Social Graphs via Uncertainty Semantics | 8,948 |
Online Social Media (OSM) is extensively used by contemporary Internet users to communicate, socialize and disseminate information. This has led to the creation of a distinct online social identity which in turn has created the need of online social reputation management techniques. A significant percentage of OSM users utilize various methods to drive and manage their reputation on OSM. This has given rise to underground markets which buy/sell fraudulent accounts, `likes', `comments' (Facebook, Instagram) and `followers' (Twitter) to artificially boost their social reputation. In this study, we present an anatomy of purchased followers on Twitter and their behaviour. We illustrate in detail the profile characteristics, content sharing and behavioural patterns of purchased follower accounts. Previous studies have analyzed the purchased follower markets and customers. Ours is the first study which analyzes the anatomy of purchased followers accounts. Some of the key insights of our study show that purchased followers have a very high unfollow entropy rate and low social engagement with their friends. In addition, we noticed that purchased follower accounts have significant difference in their interaction and content sharing patterns in comparison to random Twitter users. We also found that underground markets do not follow their service policies and guarantees they provide to customer. Our study highlights the key identifiers for suspicious follow behaviour. We then built a supervised learning mechanism to predict suspicious follower behaviour with 88.2% accuracy. We believe that understanding the anatomy and characteristics of purchased followers can help detect suspicious follower behaviour and fraudulent accounts to a larger extent. | Followers or Phantoms? An Anatomy of Purchased Twitter Followers | 8,949 |
This article reviews the problem of degree of closeness and interaction level in a social network by ranking users based on similarity score. This similarity is measured on the basis of social, geographic, educational, professional, shared interests, pages liked, mutual interested groups or communities and mutual friends. The technique addresses the problem of matching user profiles in its globality by providing a suitable matching framework able to consider all profiles' attributes and finding the similarity by new ways of string metrics. It is able to discover the biggest possible number of profiles that are similar to the target user profile, which the existing techniques are unable to detect. Attributes were assigned weights manually; string and semantic similarity metrics were used to compare attributes values thus predicting the most similar profiles. Profile based similarity show the exact relationship between users and this similarity between user profiles reflects closeness and interaction between users. | User Profile Relationships using String Similarity Metrics in Social
Networks | 8,950 |
In this work, we study the problem of clearing contamination spreading through a large network where we model the problem as a graph searching game. The problem can be summarized as constructing a search strategy that will leave the graph clear of any contamination at the end of the searching process in as few steps as possible. We show that this problem is NP-hard even on directed acyclic graphs and provide an efficient approximation algorithm. We experimentally observe the performance of our approximation algorithm in relation to the lower bound on several large online networks including Slashdot, Epinions and Twitter. The experiments reveal that in most cases our algorithm performs near optimally. | Clearing Contamination in Large Networks | 8,951 |
This paper presents an experimental study to investigate the learning and decision making behavior of individuals in a human society. Social learning is used as the mathematical basis for modelling interaction of individuals that aim to perform a perceptual task interactively. A psychology experiment was conducted on a group of undergraduate students at the University of British Columbia to examine whether the decision (action) of one individual affects the decision of the subsequent individuals. The major experimental observation that stands out here is that the participants of the experiment (agents) were affected by decisions of their partners in a relatively large fraction (60%) of trials. We fit a social learning model that mimics the interactions between participants of the psychology experiment. Misinformation propagation (also known as data incest) within the society under study is further investigated in this paper. | Social Learning in a Human Society: An Experimental Study | 8,952 |
This paper models the temporal adoption dynamics of an abstracted Internet technology or service, where the instantaneous net value of the service perceived by each (current or potential) user / customer incorporates three key features: i) user service affinity heterogeneity, ii) a network externality, and iii) a subscription cost. Internet technologies and services with network externalities face a "chicken and egg" adoption problem in that the service requires an established customer base in order to attract new customers. In this paper we study cost subsidization as a means to "reach the knee", at which point the externality drives rapid service adoption, and thereby change the equilibrium service fractional adoption level from an initial near-zero level to a final near-one level (full adoption). We present three simple subsidy models and evaluate them under two natural performance metrics: i) the duration required for the subsidized service to reach a given target adoption level, and ii) the aggregate cost of the subsidy born by the service provide. First, we present a "two target adoption subsidy" (TTAS) that subsidizes the cost so as to keep the fraction of users with positive net utility at a (constant) target level until the actual adoption target is reached. Second, we study a special case of the above where the target ensures all users have positive net utility, corresponding to a "quickest adoption" subsidy (QAS). Third, we introduce an approximation of QAS, called AQAS, that only requires the service provider adjust the subsidy level a prescribed number of times. Fourth, we study equilibria and their stability under uniformly and normally distributed user service affinities, highlighting the unstable equilibrium in each case as the natural target adoption level for the provider. Finally, we provide a fictional case study to illustrate the results. | Facilitating adoption of Internet technologies and services with
externalities via cost subsidization | 8,953 |
One of the major sources of trending news, events and opinion in the current age is micro blogging. Twitter, being one of them, is extensively used to mine data about public responses and event updates. This paper intends to propose methods to filter tweets to obtain the most accurately descriptive tweets, which communicates the content of the trend. It also potentially ranks the tweets according to relevance. The principle behind the ranking mechanism would be the assumed tendencies in the natural language used by the users. The mapping frequencies of occurrence of words and related hash tags is used to create a weighted score for each tweet in the sample space obtained from twitter on a particular trend. | Mining and Analyzing Twitter trends: Frequency based ranking of
descriptive Tweets | 8,954 |
Unlike popular videos, which would have already achieved high viewership numbers by the time they are declared popular, YouTube trending videos represent content that targets viewers attention over a relatively short time, and has the potential of becoming popular. Despite their importance and visibility, YouTube trending videos have not been studied or analyzed thoroughly. In this paper, we present our findings for measuring, analyzing, and comparing key aspects of YouTube trending videos. Our study is based on collecting and monitoring high-resolution time-series of the viewership and related statistics of more than 8,000 YouTube videos over an aggregate period of nine months. Since trending videos are declared as such just several hours after they are uploaded, we are able to analyze trending videos time-series across critical and sufficiently-long durations of their lifecycle. In addition, we analyze the profile of users who upload trending videos, to potentially identify the role that these users profile plays in getting their uploaded videos trending. Furthermore, we conduct a directional-relationship analysis among all pairs of trending videos time-series that we have monitored. We employ Granger Causality (GC) with significance testing to conduct this analysis. Unlike traditional correlation measures, our directional-relationship analysis provides a deeper insight onto the viewership pattern of different categories of trending videos. Trending videos and their channels have clear distinct statistical attributes when compared to other YouTube content that has not been labeled as trending. Our results also reveal a highly asymmetric directional-relationship among different categories of trending videos. Our directionality analysis also shows a clear pattern of viewership toward popular categories, whereas some categories tend to be isolated. | Trending Videos: Measurement and Analysis | 8,955 |
The problem of cascading failures in cyber-physical systems is drawing much attention in lieu of different network models for a diverse range of applications. While many analytic results have been reported for the case of large networks, very few of them are readily applicable to finite-size networks. This paper studies cascading failures in finite-size geometric networks where the number of nodes is on the order of tens or hundreds as in many real-life networks. First, the impact of the tolerance parameter on network resiliency is investigated. We quantify the network reaction to initial disturbances of different sizes by measuring the damage imposed on the network. Lower and upper bounds on the number of failures are derived to characterize such damages. Such finite-size analysis reveals the decisiveness and criticality of taking action within the first few stages of failure propagation in preventing a cascade. By studying the trend of the bounds as the number of nodes increases, we observe a phase transition phenomenon in terms of the tolerance parameter. The critical value of the tolerance parameter, known as the threshold, is further derived. The findings of this paper, in particular, shed light on how to choose the tolerance parameter appropriately such that a cascade of failures could be avoided. | Cascading Failures in Finite-Size Random Geometric Networks | 8,956 |
Social networks and their applications have become extremely popular during the last years, mostly targeting users via the web. However, it has been recently observed an interest to offer social network services to mobile users. Telecom operators attempt to integrate existing social networks to their systems or develop new ones, in order to offer new services to their subscribers. Subsequently, emphasis is given to the user-context modeling, as well as to the integration of sources that leads to the summarized collection of information anchored to the user; such as its location or its mobile device type, etc. In this paper we discuss the most important factors and challenges encountered during the design of such a system on architectural, technological and tool level. | Design issues for distributed mobile social networks | 8,957 |
Online Social Networks (OSNs), such as Facebook and Twitter, have become an integral part of our daily lives. There are hundreds of OSNs, each with its own focus in that each offers particular services and functionalities. Recent studies show that many OSN users create several accounts on multiple OSNs using the same or different personal information. Collecting all the available data of an individual from several OSNs and fusing it into a single profile can be useful for many purposes. In this paper, we introduce novel machine learning based methods for solving Entity Resolution (ER), a problem for matching user profiles across multiple OSNs. The presented methods are able to match between two user profiles from two different OSNs based on supervised learning techniques, which use features extracted from each one of the user profiles. By using the extracted features and supervised learning techniques, we developed classifiers which can perform entity matching between two profiles for the following scenarios: (a) matching entities across two OSNs; (b) searching for a user by similar name; and (c) de-anonymizing a user's identity. The constructed classifiers were tested by using data collected from two popular OSNs, Facebook and Xing. We then evaluated the classifiers' performances using various evaluation measures, such as true and false positive rates, accuracy, and the Area Under the receiver operator Curve (AUC). The constructed classifiers were evaluated and their classification performance measured by AUC was quite remarkable, with an AUC of up to 0.982 and an accuracy of up to 95.9% in identifying user profiles across two OSNs. | Matching Entities Across Online Social Networks | 8,958 |
Graphs are a powerful way to model interactions and relationships in data from a wide variety of application domains. In this setting, entities represented by vertices at the "center" of the graph are often more important than those associated with vertices on the "fringes". For example, central nodes tend to be more critical in the spread of information or disease and play an important role in clustering/community formation. Identifying such "core" vertices has recently received additional attention in the context of {\em network experiments}, which analyze the response when a random subset of vertices are exposed to a treatment (e.g. inoculation, free product samples, etc). Specifically, the likelihood of having many central vertices in any exposure subset can have a significant impact on the experiment. We focus on using $k$-cores and core numbers to measure the extent to which a vertex is central in a graph. Existing algorithms for computing the core number of a vertex require the entire graph as input, an unrealistic scenario in many real world applications. Moreover, in the context of network experiments, the subgraph induced by the treated vertices is only known in a probabilistic sense. We introduce a new method for estimating the core number based only on the properties of the graph within a region of radius $\delta$ around the vertex, and prove an asymptotic error bound of our estimator on random graphs. Further, we empirically validate the accuracy of our estimator for small values of $\delta$ on a representative corpus of real data sets. Finally, we evaluate the impact of improved local estimation on an open problem in network experimentation posed by Ugander et al. | Locally Estimating Core Numbers | 8,959 |
A premise at a heart of network analysis is that entities in a network derive utilities from their connections. The {\em influence} of a seed set $S$ of nodes is defined as the sum over nodes $u$ of the {\em utility} of $S$ to $u$. {\em Distance-based} utility, which is a decreasing function of the distance from $S$ to $u$, was explored in several successful research threads from social network analysis and economics: Network formation games [Bloch andJackson 2007], Reachability-based influence [Richardson and Domingos 2002, Kempe et al. 2003], "threshold" influence [Gomez-Rodriguez et al. 2011], and {\em closeness centrality} [Bavelas 1948]. We formulate a model that unifies and extends this previous work and address the two fundamental computational problems in this domain: {\em Influence oracles} and {\em influence maximization} (IM). An oracle performs some preprocessing, after which influence queries for arbitrary seed sets can be efficiently computed. With IM, we seek a set of nodes of a given size with maximum influence. Since the IM problem is computationally hard, we instead seek a {\em greedy sequence} of nodes, with each prefix having influence that is at least $1-1/e$ of that of the optimal seed set of the same size. We present the first highly scalable algorithms for both problems, providing statistical guarantees on approximation quality and near-linear worst-case bounds on the computation. We perform an experimental evaluation which demonstrates the effectiveness of our designs on networks with hundreds of millions of edges. | Distance-Based Influence in Networks: Computation and Maximization | 8,960 |
Network visualization allows a quick glance at how nodes (or actors) are connected by edges (or ties). A conventional network diagram of "contact tree" maps out a root and branches that represent the structure of nodes and edges, often without further specifying leaves or fruits that would have grown from small branches. By furnishing such a network structure with leaves and fruits, we reveal details about "contacts" in our ContactTrees that underline ties and relationships. Our elegant design employs a bottom-up approach that resembles a recent attempt to understand subjective well-being by means of a series of emotions. Such a bottom-up approach to social-network studies decomposes each tie into a series of interactions or contacts, which help deepen our understanding of the complexity embedded in a network structure. Unlike previous network visualizations, ContactTrees can highlight how relationships form and change based upon interactions among actors, and how relationships and networks vary by contact attributes. Based on a botanical tree metaphor, the design is easy to construct and the resulting tree-like visualization can display many properties at both tie and contact levels, a key ingredient missing from conventional techniques of network visualization. We first demonstrate ContactTrees using a dataset consisting of three waves of 3-month contact diaries over the 2004-2012 period, then compare ContactTrees with alternative tools and discuss how this tool can be applied to other types of datasets. | ContactTrees: A Technique for Studying Personal Network Data | 8,961 |
Scalable probabilistic modeling and prediction in high dimensional multivariate time-series is a challenging problem, particularly for systems with hidden sources of dependence and/or homogeneity. Examples of such problems include dynamic social networks with co-evolving nodes and edges and dynamic student learning in online courses. Here, we address these problems through the discovery of hierarchical latent groups. We introduce a family of Conditional Latent Tree Models (CLTM), in which tree-structured latent variables incorporate the unknown groups. The latent tree itself is conditioned on observed covariates such as seasonality, historical activity, and node attributes. We propose a statistically efficient framework for learning both the hierarchical tree structure and the parameters of the CLTM. We demonstrate competitive performance in multiple real world datasets from different domains. These include a dataset on students' attempts at answering questions in a psychology MOOC, Twitter users participating in an emergency management discussion and interacting with one another, and windsurfers interacting on a beach in Southern California. In addition, our modeling framework provides valuable and interpretable information about the hidden group structures and their effect on the evolution of the time series. | Are you going to the party: depends, who else is coming? [Learning
hidden group dynamics via conditional latent tree models] | 8,962 |
Finding dense substructures in a graph is a fundamental graph mining operation, with applications in bioinformatics, social networks, and visualization to name a few. Yet most standard formulations of this problem (like clique, quasiclique, k-densest subgraph) are NP-hard. Furthermore, the goal is rarely to find the "true optimum", but to identify many (if not all) dense substructures, understand their distribution in the graph, and ideally determine relationships among them. Current dense subgraph finding algorithms usually optimize some objective, and only find a few such subgraphs without providing any structural relations. We define the nucleus decomposition of a graph, which represents the graph as a forest of nuclei. Each nucleus is a subgraph where smaller cliques are present in many larger cliques. The forest of nuclei is a hierarchy by containment, where the edge density increases as we proceed towards leaf nuclei. Sibling nuclei can have limited intersections, which enables discovering overlapping dense subgraphs. With the right parameters, the nucleus decomposition generalizes the classic notions of k-cores and k-truss decompositions. We give provably efficient algorithms for nucleus decompositions, and empirically evaluate their behavior in a variety of real graphs. The tree of nuclei consistently gives a global, hierarchical snapshot of dense substructures, and outputs dense subgraphs of higher quality than other state-of-the-art solutions. Our algorithm can process graphs with tens of millions of edges in less than an hour. | Finding the Hierarchy of Dense Subgraphs using Nucleus Decompositions | 8,963 |
Social media have become part of modern news reporting, used by journalists to spread information and find sources, or as a news source by individuals. The quest for prominence and recognition on social media sites like Twitter can sometimes eclipse accuracy and lead to the spread of false information. As a way to study and react to this trend, we introduce {\sc TwitterTrails}, an interactive, web-based tool ({\tt twittertrails.com}) that allows users to investigate the origin and propagation characteristics of a rumor and its refutation, if any, on Twitter. Visualizations of burst activity, propagation timeline, retweet and co-retweeted networks help its users trace the spread of a story. Within minutes {\sc TwitterTrails} will collect relevant tweets and automatically answer several important questions regarding a rumor: its originator, burst characteristics, propagators and main actors according to the audience. In addition, it will compute and report the rumor's level of visibility and, as an example of the power of crowdsourcing, the audience's skepticism towards it which correlates with the rumor's credibility. We envision {\sc TwitterTrails} as valuable tool for individual use, but we especially for amateur and professional journalists investigating recent and breaking stories. Further, its expanding collection of investigated rumors can be used to answer questions regarding the amount and success of misinformation on Twitter. | Investigating Rumor Propagation with TwitterTrails | 8,964 |
Arguably one of the most important features of Twitter is the support for "retweets" or messages re-posted verbatim by a user that were originated by someone else. (This does not include modified tweets that sometimes are referred to as retweets.) Despite the fact that retweets are routinely studied and reported, many important questions remain about user motivation for their use and their significance. In this paper we answer the question of what users indicate when they retweet. We do so in a comprehensive fashion, by employing a user survey, a study of user profiles, and a meta-analysis of over 100 research publications from three related major conferences. Our findings indicate that retweeting indicates not only interest in a message, but also trust in the message and the originator, and agreement with the message contents. However, the findings are significantly weaker for journalists, some of whom beg to differ declaring so in their own user profiles. On the other hand, the inclusion of hashtags strengthens the signal of agreement, especially when the hashtags are related to politics. While in the past there have been additional claims in the literature about possible reasons for retweeting, many of them are not supported, especially given the technical changes introduced recently by Twitter. | Do Retweets indicate Interest, Trust, Agreement? (Extended Abstract) | 8,965 |
Internet users and businesses are increasingly using online social networks (OSN) to drive audience traffic and increase their popularity. In order to boost social presence, OSN users need to increase the visibility and reach of their online profile, like - Facebook likes, Twitter followers, Instagram comments and Yelp reviews. For example, an increase in Twitter followers not only improves the audience reach of the user but also boosts the perceived social reputation and popularity. This has led to a scope for an underground market that provides followers, likes, comments, etc. via a network of fraudulent and compromised accounts and various collusion techniques. In this paper, we landscape the underground markets that provide Twitter followers by studying their basic building blocks - merchants, customers and phony followers. We charecterize the services provided by merchants to understand their operational structure and market hierarchy. Twitter underground markets can operationalize using a premium monetary scheme or other incentivized freemium schemes. We find out that freemium market has an oligopoly structure with few merchants being the market leaders. We also show that merchant popularity does not have any correlation with the quality of service provided by the merchant to its customers. Our findings also shed light on the characteristics and quality of market customers and the phony followers provided. We draw comparison between legitimate users and phony followers, and find out key identifiers to separate such users. With the help of these differentiating features, we build a supervised learning model to predict suspicious following behaviour with an accuracy of 89.2%. | What They Do in Shadows: Twitter Underground Follower Market | 8,966 |
Measuring socioeconomic deprivation of cities in an accurate and timely fashion has become a priority for governments around the world, as the massive urbanization process we are witnessing is causing high levels of inequalities which require intervention. Traditionally, deprivation indexes have been derived from census data, which is however very expensive to obtain, and thus acquired only every few years. Alternative computational methods have been proposed in recent years to automatically extract proxies of deprivation at a fine spatio-temporal level of granularity; however, they usually require access to datasets (e.g., call details records) that are not publicly available to governments and agencies. To remedy this, we propose a new method to automatically mine deprivation at a fine level of spatio-temporal granularity that only requires access to freely available user-generated content. More precisely, the method needs access to datasets describing what urban elements are present in the physical environment; examples of such datasets are Foursquare and OpenStreetMap. Using these datasets, we quantitatively describe neighborhoods by means of a metric, called {\em Offering Advantage}, that reflects which urban elements are distinctive features of each neighborhood. We then use that metric to {\em (i)} build accurate classifiers of urban deprivation and {\em (ii)} interpret the outcomes through thematic analysis. We apply the method to three UK urban areas of different scale and elaborate on the results in terms of precision and recall. | Measuring Urban Deprivation from User Generated Content | 8,967 |
Due to popularity surge social networks became lucrative targets for spammers and guerilla marketers, who are trying to game ranking systems and broadcast their messages at little to none cost. Ranking systems, for example Twitter's Trends, can be gamed by scripted users also called bots, who are automatically or semi-automatically twitting essentially the same message. Judging by the prices and abundance of supply from PR firms this is an easy to implement and widely used tactic, at least in Russian blogosphere. Aggregative analysis of social networks should at best mark those messages as spam or at least correctly downplay their importance as they represent opinions only of a few, if dedicated, users. Hence bot detection plays a crucial role in social network mining and analysis. In this paper we propose technique called RepRank which could be viewed as Markov chain based model for reputation propagation on graphs utilizing simultaneous trust and anti-trust propagation and provide effective numerical approach for its computation. Comparison with another models such as TrustRank and some of its modifications on sample of 320000 Russian speaking Twitter users is presented. The dataset is presented as well. | Algebraic reputation model RepRank and its application to spambot
detection | 8,968 |
We introduce the diffusion and superposition distances as two metrics to compare signals supported in the nodes of a network. Both metrics consider the given vectors as initial temperature distributions and diffuse heat trough the edges of the graph. The similarity between the given vectors is determined by the similarity of the respective diffusion profiles. The superposition distance computes the instantaneous difference between the diffused signals and integrates the difference over time. The diffusion distance determines a distance between the integrals of the diffused signals. We prove that both distances define valid metrics and that they are stable to perturbations in the underlying network. We utilize numerical experiments to illustrate their utility in classifying signals in a synthetic network as well as in classifying ovarian cancer histologies using gene mutation profiles of different patients. We also reinterpret diffusion as a transformation of interrelated feature spaces and use it as preprocessing tool for learning. We use diffusion to increase the accuracy of handwritten digit classification. | Diffusion and Superposition Distances for Signals Supported on Networks | 8,969 |
Web Service is one of the most important information sharing technologies on the web and one of the example of service oriented processing. To guarantee accurate execution of web services operations, they must be accountable with regulations of the social networks in which they sign up. This operations implement using controls called 'Commitment'. This paper studies commitments, then has an overview on existing researches, web service execution method using commitments and information sharing methods between web services based on commitments and social networks. A key challenge in this technique is consistency ensuring in execution time. The aim of this study is presenting an algorithm for consistency ensuring between commitments. An application designed for proving correctness of algorithm. | Consistency of Commitments in Social Web Services | 8,970 |
In this paper, we study the problem of allocating ads to users through the viral-marketing lens. Advertisers approach the host with a budget in return for the marketing campaign service provided by the host. We show that allocation that takes into account the propensity of ads for viral propagation can achieve significantly better performance. However, uncontrolled virality could be undesirable for the host as it creates room for exploitation by the advertisers: hoping to tap uncontrolled virality, an advertiser might declare a lower budget for its marketing campaign, aiming at the same large outcome with a smaller cost. This creates a challenging trade-off: on the one hand, the host aims at leveraging virality and the network effect to improve advertising efficacy, while on the other hand the host wants to avoid giving away free service due to uncontrolled virality. We formalize this as the problem of ad allocation with minimum regret, which we show is NP-hard and inapproximable w.r.t. any factor. However, we devise an algorithm that provides approximation guarantees w.r.t. the total budget of all advertisers. We develop a scalable version of our approximation algorithm, which we extensively test on four real-world data sets, confirming that our algorithm delivers high quality solutions, is scalable, and significantly outperforms several natural baselines. | Viral Marketing Meets Social Advertising: Ad Allocation with Minimum
Regret | 8,971 |
We study asymptotic dynamical patterns that emerge among a set of nodes interacting in a dynamically evolving signed random network, where positive links carry out standard consensus and negative links induce relative-state flipping. A sequence of deterministic signed graphs define potential node interactions that take place independently. Each node receives a positive recommendation consistent with the standard consensus algorithm from its positive neighbors, and a negative recommendation defined by relative-state flipping from its negative neighbors. After receiving these recommendations, each node puts a deterministic weight to each recommendation, and then encodes these weighted recommendations in its state update through stochastic attentions defined by two Bernoulli random variables. We establish a number of conditions regarding almost sure convergence and divergence of the node states. We also propose a condition for almost sure state clustering for essentially weakly balanced graphs, with the help of several martingale convergence lemmas. Some fundamental differences on the impact of the deterministic weights and stochastic attentions to the node state evolution are highlighted between the current relative-state-flipping model and the state-flipping model considered in Altafini 2013 and Shi et al. 2014. | Emergent Behaviors over Signed Random Dynamical Networks:
Relative-State-Flipping Model | 8,972 |
Social Media involve many shared items, such as photos, which may concern more than one user. The first challenge we address in this paper is to develop a way for users of such items to take a decision on to whom to share these items. This is not an easy problem, as users' privacy preferences for the same item may conflict, so an approach that just merges in some way the users' privacy preferences may provide unsatisfactory results. We propose a negotiation mechanism for users to agree on a compromise for the conflicts found. The second challenge we address in this paper relates to the exponential complexity of such a negotiation mechanism, which could make it too slow to be used in practice in a Social Media infrastructure. To address this, we propose heuristics that reduce the complexity of the negotiation mechanism and show how substantial benefits can be derived from the use of these heuristics through extensive experimental evaluation that compares the performance of the negotiation mechanism with and without these heuristics. Moreover, we show that one such heuristic makes the negotiation mechanism produce results fast enough to be used in actual Social Media infrastructures with near-optimal results. | Privacy Policy Negotiation in Social Media | 8,973 |
The friendship paradox is revisited by considering both local and global averages of friends. How the economics of attention affects the recruitment of friends is examined. Statistical implications of varying individual attentions are investigated and it is argued that this is one reason why the mean of friends is higher than the median in social networks. The distribution of friends skews to the right for two other reasons: (i) the presence of institutional nodes that increase the mean; and (ii) the dormancy of many of the nodes. The difference between friends and friends of friends is a measure of the structural information about the network. | Friendship Paradox and Attention Economics | 8,974 |
The shortest path problem is among the most fundamental combinatorial optimization problems to answer reachability queries. It is hard to deter-mine which vertices or edges are visited during shortest path traversals. In this paper, we provide an empirical analysis on how traversal algorithms behave on social networks. First, we compute the shortest paths between set of vertices. Each shortest path is considered as one transaction. Second, we utilize the pat-tern mining approach to identify the frequency of occurrence of the vertices. We evaluate the results in terms of network properties, i.e. degree distribution, clustering coefficient. | Shortest Path Analysis in Social Graphs | 8,975 |
This paper presents an extensive data-based analysis of the non-profit democratic hospitality exchange service bewelcome.org. We hereby pursuit the goal of determining the factors influencing its growth. It also provides general insights on internet-based hospitality exchange services. The other investigated services are hospitalityclub.org and couchsurfing.org. Communities using the three services are interconnected -- comparing their data provides additional information. | Bewelcome.org -- a non-profit democratic hospex service set up for
growth | 8,976 |
Nowadays, the exponentially growing of the Web renders the problem of correlation among different topics of paramount importance. The proposed model can be used to study the evolution of network depicted by different topics on the web correlated by a dynamic "fluid" of tags among them. The fluid-dynamic model depicted is completely evolutive, thus it is able to describe the dynamic situation of a network at every instant of time. This overcomes the difficulties encountered by many static models. The theory permits the development of efficient numerical schemes also for very large networks. This is possible since dynamic flow at junctions is modeled in a simple and computationally convenient way (resorting to a linear programming problem). The obtained model consists of a single conservation law and is on one side simple enough to permit a complete understanding, on the other side reach enough to detect the evolution of the dynamic network. | Liquidity on Web Dynamic Network | 8,977 |
The prospect of finding love may be scary but the prospect of committing to a relationship for the rest of your life is almost certainly scary. The secretary problem is a parallel to romantic decision making where an individual decides when to be satisfied with a selection choice in the face of uncertain future options. However, the secretary problem and its variations still do not provide a practical solution in a world where individual preference, goals, and societal context create a highly complex space of values that factor into decision making. In light of these complexities, we offer a general process that can determine the value of romantic options in a highly personal context. This algorithm is currently being developed into a service that will be available in 2015 for the general public. | "Should I break up with my girlfriend? Will I find another?" Or: An
Algorithm for the Forecasting of Romantic Options | 8,978 |
This paper considers online reputation and polling systems where individuals make recommendations based on their private observations and recommendations of friends. Such interaction of individuals and their social influence is modelled as social learning on a directed acyclic graph. Data incest (misinformation propagation) occurs due to unintentional re-use of identical actions in the for- mation of public belief in social learning; the information gathered by each agent is mistakenly considered to be independent. This results in overconfidence and bias in estimates of the state. Necessary and sufficient conditions are given on the structure of information exchange graph to mitigate data incest. Incest removal algorithms are presented. Experimental results on human subjects are presented to illustrate the effect of social influence and data incest on decision making. These experimental results indicate that social learning protocols require careful design to handle and mitigate data incest. The incest removal algorithms are illustrated in an expectation polling system where participants in a poll respond with a summary of their friends' beliefs. Finally, the principle of revealed preferences arising in micro-economics theory is used to parse Twitter datasets to determine if social sensors are utility maximizers and then determine their utility functions. | Online Reputation and Polling Systems: Data Incest, Social Learning and
Revealed Preferences | 8,979 |
The abundance of online user data has led to a surge of interests in understanding the dynamics of social relationships using computational methods. Utilizing users' items adoption data, we develop a new method to compute the Granger-causal (GC) relationships among users. In order to handle the high dimensional and sparse nature of the adoption data, we propose to model the relationships among users in latent space instead of the original data space. We devise a Linear Dynamical Topic Model (LDTM) that can capture the dynamics of the users' items adoption behaviors in latent (topic) space. Using the time series of temporal topic distributions learned by LDTM, we conduct Granger causality tests to measure the social correlation relationships between pairs of users. We call the combination of our LDTM and Granger causality tests as Temporal Social Correlation. By conducting extensive experiments on bibliographic data, where authors are analogous to users, we show that the ordering of authors' name on their publications plays a statistically significant role in the interaction of research topics among the authors. We also present a case study to illustrate the correlational relationships between pairs of authors. | Using Linear Dynamical Topic Model for Inferring Temporal Social
Correlation in Latent Space | 8,980 |
Social network based trust relationships present a critical foundation for designing trustworthy systems, such as Sybil defenses, secure routing, and anonymous/censorshipresilient communications. A key issue in the design of such systems, is the revelation of users' trusted social contacts to an adversary-information that is considered sensitive in today's society. In this work, we focus on the challenge of preserving the privacy of users' social contacts, while still enabling the design of social trust based applications. First, we propose LinkMirage, a community detection based algorithm for anonymizing links in social network topologies; LinkMirage preserves community structures in the social topology while anonymizing links within the communities. LinkMirage considers the evolution of the social network topologies, and minimizes privacy leakage due to temporal dynamics of the system. Second, we define metrics for quantifying the privacy and utility of a time series of social topologies with anonymized links. We analyze the privacy and utility provided by LinkMirage both theoretically, as well as using real world social network topologies: a Facebook dataset with 870K links and a large-scale Google+ dataset with 940M links. We find that our approach significantly outperforms the existing state-of-art. Finally, we demonstrate the applicability of LinkMirage in real-world applications such as Sybil defenses, reputation systems, anonymity systems and vertex anonymity. We also prototype LinkMirage as a Facebook application such that real world systems can bootstrap privacy-preserving trust relationships without the cooperation of the OSN operators. | LinkMirage: How to Anonymize Links in Dynamic Social Systems | 8,981 |
In epistemic community, people are said to be selected on their knowledge contribution to the project (articles, codes, etc.) However, the socialization process is an important factor for inclusion, sustainability as a contributor, and promotion. Finally, what does matter to be promoted? being a good contributor? being a good animator? knowing the boss? We explore this question looking at the process of election for administrator in the English Wikipedia community. We modeled the candidates according to their revisions and/or social attributes. These attributes are used to construct a predictive model of promotion success, based on the candidates's past behavior, computed thanks to a random forest algorithm. Our model combining knowledge contribution variables and social networking variables successfully explain 78% of the results which is better than the former models. It also helps to refine the criterion for election. If the number of knowledge contributions is the most important element, social interactions come close second to explain the election. But being connected with the future peers (the admins) can make the difference between success and failure, making this epistemic community a very social community too. | Social Interactions vs Revisions, What is important for Promotion in
Wikipedia? | 8,982 |
We constructed a product network based on the sales data collected and provided by a Fortune 500 speciality retailer. The structure of the network is dominated by small isolated components, dense clique-based communities, and sparse stars and linear chains and pendants. We used the identified structural elements (tiles) to organize products into mini-categories -- compact collections of potentially complementary and substitute items. The mini-categories extend the traditional hierarchy of retail products (group - class - subcategory) and may serve as building blocks towards exploration of consumer projects and long-term customer behavior. | Building Mini-Categories in Product Networks | 8,983 |
Most sampling techniques for online social networks (OSNs) are based on a particular sampling method on a single graph, which is referred to as a statistics. However, various realizing methods on different graphs could possibly be used in the same OSN, and they may lead to different sampling efficiencies, i.e., asymptotic variances. To utilize multiple statistics for accurate measurements, we formulate a mixture sampling problem, through which we construct a mixture unbiased estimator which minimizes asymptotic variance. Given fixed sampling budgets for different statistics, we derive the optimal weights to combine the individual estimators; given fixed total budget, we show that a greedy allocation towards the most efficient statistics is optimal. In practice, the sampling efficiencies of statistics can be quite different for various targets and are unknown before sampling. To solve this problem, we design a two-stage framework which adaptively spends a partial budget to test different statistics and allocates the remaining budget to the inferred best statistics. We show that our two-stage framework is a generalization of 1) randomly choosing a statistics and 2) evenly allocating the total budget among all available statistics, and our adaptive algorithm achieves higher efficiency than these benchmark strategies in theory and experiment. | Sampling Online Social Networks via Heterogeneous Statistics | 8,984 |
The present article serves as an erratum to our paper of the same title, which was presented and published in the KDD 2014 conference. In that article, we claimed falsely that the objective function defined in Section 1.4 is non-monotone submodular. We are deeply indebted to Debmalya Mandal, Jean Pouget-Abadie and Yaron Singer for bringing to our attention a counter-example to that claim. Subsequent to becoming aware of the counter-example, we have shown that the objective function is in fact NP-hard to approximate to within a factor of $O(n^{1-\epsilon})$ for any $\epsilon > 0$. In an attempt to fix the record, the present article combines the problem motivation, models, and experimental results sections from the original incorrect article with the new hardness result. We would like readers to only cite and use this version (which will remain an unpublished note) instead of the incorrect conference version. | Stability of Influence Maximization | 8,985 |
Online dating sites have become popular platforms for people to look for potential romantic partners. Different from traditional user-item recommendations where the goal is to match items (e.g., books, videos, etc) with a user's interests, a recommendation system for online dating aims to match people who are mutually interested in and likely to communicate with each other. We introduce similarity measures that capture the unique features and characteristics of the online dating network, for example, the interest similarity between two users if they send messages to same users, and attractiveness similarity if they receive messages from same users. A reciprocal score that measures the compatibility between a user and each potential dating candidate is computed and the recommendation list is generated to include users with top scores. The performance of our proposed recommendation system is evaluated on a real-world dataset from a major online dating site in China. The results show that our recommendation algorithms significantly outperform previously proposed approaches, and the collaborative filtering-based algorithms achieve much better performance than content-based algorithms in both precision and recall. Our results also reveal interesting behavioral difference between male and female users when it comes to looking for potential dates. In particular, males tend to be focused on their own interest and oblivious towards their attractiveness to potential dates, while females are more conscientious to their own attractiveness to the other side of the line. | Reciprocal Recommendation System for Online Dating | 8,986 |
In this paper, we present the first results of our ongoing early-stage research on a realtime disaster detection and monitoring tool. Based on Wikipedia, it is language-agnostic and leverages user-generated multimedia content shared on online social networking sites to help disaster responders prioritize their efforts. We make the tool and its source code publicly available as we make progress on it. Furthermore, we strive to publish detected disasters and accompanying multimedia content following the Linked Data principles to facilitate its wide consumption, redistribution, and evaluation of its usefulness. | Disaster Monitoring with Wikipedia and Online Social Networking Sites:
Structured Data and Linked Data Fragments to the Rescue? | 8,987 |
The largest eigenvalue of the adjacency matrix of a network (referred to as the spectral radius) is an important metric in its own right. Further, for several models of epidemic spread on networks (e.g., the `flu-like' SIS model), it has been shown that an epidemic dies out quickly if the spectral radius of the graph is below a certain threshold that depends on the model parameters. This motivates a strategy to control epidemic spread by reducing the spectral radius of the underlying network. In this paper, we develop a suite of provable approximation algorithms for reducing the spectral radius by removing the minimum cost set of edges (modeling quarantining) or nodes (modeling vaccinations), with different time and quality tradeoffs. Our main algorithm, \textsc{GreedyWalk}, is based on the idea of hitting closed walks of a given length, and gives an $O(\log^2{n})$-approximation, where $n$ denotes the number of nodes; it also performs much better in practice compared to all prior heuristics proposed for this problem. We further present a novel sparsification method to improve its running time. In addition, we give a new primal-dual based algorithm with an even better approximation guarantee ($O(\log n)$), albeit with slower running time. We also give lower bounds on the worst-case performance of some of the popular heuristics. Finally we demonstrate the applicability of our algorithms and the properties of our solutions via extensive experiments on multiple synthetic and real networks. | Approximation Algorithms for Reducing the Spectral Radius to Control
Epidemic Spread | 8,988 |
Information spread in social media depends on a number of factors, including how the site displays information, how users navigate it to find items of interest, users' tastes, and the `virality' of information, i.e., its propensity to be adopted, or retweeted, upon exposure. Probabilistic models can learn users' tastes from the history of their item adoptions and recommend new items to users. However, current models ignore cognitive biases that are known to affect behavior. Specifically, people pay more attention to items at the top of a list than those in lower positions. As a consequence, items near the top of a user's social media stream have higher visibility, and are more likely to be seen and adopted, than those appearing below. Another bias is due to the item's fitness: some items have a high propensity to spread upon exposure regardless of the interests of adopting users. We propose a probabilistic model that incorporates human cognitive biases and personal relevance in the generative model of information spread. We use the model to predict how messages containing URLs spread on Twitter. Our work shows that models of user behavior that account for cognitive factors can better describe and predict user behavior in social media. | VIP: Incorporating Human Cognitive Biases in a Probabilistic Model of
Retweeting | 8,989 |
The 5th annual international conference on Collaborative Innovation Networks Conference (COINS) takes place at Keio University from March 12 to 14, 2015. COINS15 brings together practitioners, researchers and students of the emerging science of collaboration to share their work, learn from each other, and get inspired through creative new ideas. Where science, design, business and art meet, COINS15 looks at the emerging forces behind the phenomena of open-source, creative, entrepreneurial and social movements. Through interactive workshops, professional presentations, and fascinating keynotes, COINS15 combines a wide range of interdisciplinary fields such as social network analysis, group dynamics, design and visualization, information systems, collective action and the psychology and sociality of collaboration. | Proceedings of the 5th International Conference on Collaborative
Innovation Networks COINs15, Tokyo, Japan March 12-14, 2015 | 8,990 |
Contributing to history has never been as easy as it is today. Anyone with access to the Web is able to play a part on Wikipedia, an open and free encyclopedia. Wikipedia, available in many languages, is one of the most visited websites in the world and arguably one of the primary sources of knowledge on the Web. However, not everyone is contributing to Wikipedia from a diversity point of view; several groups are severely underrepresented. One of those groups is women, who make up approximately 16% of the current contributor community, meaning that most of the content is written by men. In addition, although there are specific guidelines of verifiability, notability, and neutral point of view that must be adhered by Wikipedia content, these guidelines are supervised and enforced by men. In this paper, we propose that gender bias is not about participation and representation only, but also about characterization of women. We approach the analysis of gender bias by defining a methodology for comparing the characterizations of men and women in biographies in three aspects: meta-data, language, and network structure. Our results show that, indeed, there are differences in characterization and structure. Some of these differences are reflected from the off-line world documented by Wikipedia, but other differences can be attributed to gender bias in Wikipedia content. We contextualize these differences in feminist theory and discuss their implications for Wikipedia policy. | First Women, Second Sex: Gender Bias in Wikipedia | 8,991 |
Wikipedia introduced a new social function "wiki-thanks". "Wiki-thanks" enable editors to send thanks to other editors' contributions. In this paper, we aim to investigate this new social tool from different cultural perspectives. To achieve this goal, we analyze "wiki-thanks" log events and compared the English, German, Spanish, Chinese, Japanese, Korean, and Finish language Wikipedias. | WIKI THANKS: Cultural Differences in Thanks Network of
Different-Language Wikipedias | 8,992 |
We investigate the sensitivity of epidemic behavior to a bounded susceptibility constraint -- susceptible nodes are infected by their neighbors via the regular SI/SIS dynamics, but subject to a cap on the infection rate. Such a constraint is motivated by modern social networks, wherein messages are broadcast to all neighbors, but attention spans are limited. Bounded susceptibility also arises in distributed computing applications with download bandwidth constraints, and in human epidemics under quarantine policies. Network epidemics have been extensively studied in literature; prior work characterizes the graph structures required to ensure fast spreading under the SI dynamics, and long lifetime under the SIS dynamics. In particular, these conditions turn out to be meaningful for two classes of networks of practical relevance -- dense, uniform (i.e., clique-like) graphs, and sparse, structured (i.e., star-like) graphs. We show that bounded susceptibility has a surprising impact on epidemic behavior in these graph families. For the SI dynamics, bounded susceptibility has no effect on star-like networks, but dramatically alters the spreading time in clique-like networks. In contrast, for the SIS dynamics, clique-like networks are unaffected, but star-like networks exhibit a sharp change in extinction times under bounded susceptibility. Our findings are useful for the design of disease-resistant networks and infrastructure networks. More generally, they show that results for existing epidemic models are sensitive to modeling assumptions in non-intuitive ways, and suggest caution in directly using these as guidelines for real systems. | The Behavior of Epidemics under Bounded Susceptibility | 8,993 |
In this paper we study the differences in historical worldview between Western and Eastern cultures, represented through the English, Chinese, Japanese, and German Wikipedia. In particular, we analyze the historical networks of the World's leaders since the beginning of written history, comparing them in the four different Wikipedias. | Cultural Anthropology Through the Lens of Wikipedia - A Comparison of
Historical Leadership Networks in the English, Chinese, Japanese and German
Wikipedia | 8,994 |
For maximizing influence spread in a social network, given a certain budget on the number of seed nodes, we investigate the effects of selecting and activating the seed nodes in multiple phases. In particular, we formulate an appropriate objective function for two-phase influence maximization under the independent cascade model, investigate its properties, and propose algorithms for determining the seed nodes in the two phases. We also study the problem of determining an optimal budget-split and delay between the two phases. | A Multi-phase Approach for Improving Information Diffusion in Social
Networks | 8,995 |
The rise of social media provides a great opportunity for people to reach out to their social connections to satisfy their information needs. However, generic social media platforms are not explicitly designed to assist information seeking of users. In this paper, we propose a novel framework to identify the social connections of a user able to satisfy his information needs. The information need of a social media user is subjective and personal, and we investigate the utility of his social context to identify people able to satisfy it. We present questions users post on Twitter as instances of information seeking activities in social media. We infer soft community memberships of the asker and his social connections by integrating network and content information. Drawing concepts from the social foci theory, we identify answerers who share communities with the asker w.r.t. the question. Our experiments demonstrate that the framework is effective in identifying answerers to social media questions. | Leveraging Social Foci for Information Seeking in Social Media | 8,996 |
The social presence theory in social psychology suggests that computer-mediated online interactions are inferior to face-to-face, in-person interactions. In this paper, we consider the scenarios of organizing in person friend-making social activities via online social networks (OSNs) and formulate a new research problem, namely, Hop-bounded Maximum Group Friending (HMGF), by modeling both existing friendships and the likelihood of new friend making. To find a set of attendees for socialization activities, HMGF is unique and challenging due to the interplay of the group size, the constraint on existing friendships and the objective function on the likelihood of friend making. We prove that HMGF is NP-Hard, and no approximation algorithm exists unless P = NP. We then propose an error-bounded approximation algorithm to efficiently obtain the solutions very close to the optimal solutions. We conduct a user study to validate our problem formulation and per- form extensive experiments on real datasets to demonstrate the efficiency and effectiveness of our proposed algorithm. | Maximizing Friend-Making Likelihood for Social Activity Organization | 8,997 |
Studies have shown that each person is more inclined to enjoy a group activity when 1) she is interested in the activity, and 2) many friends with the same interest join it as well. Nevertheless, even with the interest and social tightness information available in online social networks, nowadays many social group activities still need to be coordinated manually. In this paper, therefore, we first formulate a new problem, named Participant Selection for Group Activity (PSGA), to decide the group size and select proper participants so that the sum of personal interests and social tightness of the participants in the group is maximized, while the activity cost is also carefully examined. To solve the problem, we design a new randomized algorithm, named Budget-Aware Randomized Group Selection (BARGS), to optimally allocate the computation budgets for effective selection of the group size and participants, and we prove that BARGS can acquire the solution with a guaranteed performance bound. The proposed algorithm was implemented in Facebook, and experimental results demonstrate that social groups generated by the proposed algorithm significantly outperform the baseline solutions. | Scale-Adaptive Group Optimization for Social Activity Planning | 8,998 |
Research issues and data mining techniques for product recommendation and viral marketing have been widely studied. Existing works on seed selection in social networks do not take into account the effect of product recommendations in e-commerce stores. In this paper, we investigate the seed selection problem for viral marketing that considers both effects of social influence and item inference (for product recommendation). We develop a new model, Social Item Graph (SIG), that captures both effects in form of hyperedges. Accordingly, we formulate a seed selection problem, called Social Item Maximization Problem (SIMP), and prove the hardness of SIMP. We design an efficient algorithm with performance guarantee, called Hyperedge-Aware Greedy (HAG), for SIMP and develop a new index structure, called SIG-index, to accelerate the computation of diffusion process in HAG. Moreover, to construct realistic SIG models for SIMP, we develop a statistical inference based framework to learn the weights of hyperedges from data. Finally, we perform a comprehensive evaluation on our proposals with various baselines. Experimental result validates our ideas and demonstrates the effectiveness and efficiency of the proposed model and algorithms over baselines. | When Social Influence Meets Item Inference | 8,999 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.