text stringlengths 17 3.36M | source stringlengths 3 333 | __index_level_0__ int64 0 518k |
|---|---|---|
While the polls have been the most trusted source for election predictions for decades, in the recent presidential election they were called inaccurate and biased. How inaccurate were the polls in this election and can social media beat the polls as an accurate election predictor? Polls from several news outlets and sentiment analysis on Twitter data were used, in conjunction with the results of the election, to answer this question and outline further research on the best method for predicting the outcome of future elections. | Election Bias: Comparing Polls and Twitter in the 2016 U.S. Election | 9,300 |
The 2016 U.S. presidential election has witnessed the major role of Twitter in the year's most important political event. Candidates used this social media platform extensively for online campaigns. Meanwhile, social media has been filled with rumors, which might have had huge impacts on voters' decisions. In this paper, we present a thorough analysis of rumor tweets from the followers of two presidential candidates: Hillary Clinton and Donald Trump. To overcome the difficulty of labeling a large amount of tweets as training data, we detect rumor tweets by matching them with verified rumor articles. We analyze over 8 million tweets collected from the followers of the two candidates. Our results provide answers to several primary concerns about rumors in this election, including: which side of the followers posted the most rumors, who posted these rumors, what rumors they posted, and when they posted these rumors. The insights of this paper can help us understand the online rumor behaviors in American politics. | Detection and Analysis of 2016 US Presidential Election Related Rumors
on Twitter | 9,301 |
This paper describes a Research Prototype, WebSelect, designed to assist online media planners in deciding which websites to target for a media campaign. The salient feature of the prototype is its ability to capture and utilize the overlap information in website traffic and use it for media planning exercises. In addition, the prototype possesses the capability to include targeting parameters like users age group and income along with the varying advertising costs across different websites. The prototype uses a genetic algorithm at the backend to select the final subset to target from the possible website set. | WebSelect: A Research Prototype for Optimizing Ad Exposures based on
Network Structure | 9,302 |
Network sparsification aims to reduce the number of edges of a network while maintaining its structural properties; such properties include shortest paths, cuts, spectral measures, or network modularity. Sparsification has multiple applications, such as, speeding up graph-mining algorithms, graph visualization, as well as identifying the important network edges. In this paper we consider a novel formulation of the network-sparsification problem. In addition to the network, we also consider as input a set of communities. The goal is to sparsify the network so as to preserve the network structure with respect to the given communities. We introduce two variants of the community-aware sparsification problem, leading to sparsifiers that satisfy different connectedness community properties. From the technical point of view, we prove hardness results and devise effective approximation algorithms. Our experimental results on a large collection of datasets demonstrate the effectiveness of our algorithms. | Community-aware network sparsification | 9,303 |
With mobile shopping surging in popularity, people are spending ever more money on digital purchases through their mobile devices and phones. However, few large-scale studies of mobile shopping exist. In this paper we analyze a large data set consisting of more than 776M digital purchases made on Apple mobile devices that include songs, apps, and in-app purchases. We find that 61% of all the spending is on in-app purchases and that the top 1% of users are responsible for 59% of all the spending. These big spenders are more likely to be male and older, and less likely to be from the US. We study how they adopt and abandon individual app, and find that, after an initial phase of increased daily spending, users gradually lose interest: the delay between their purchases increases and the spending decreases with a sharp drop toward the end. Finally, we model the in-app purchasing behavior in multiple steps: 1) we model the time between purchases; 2) we train a classifier to predict whether the user will make a purchase from a new app or continue purchasing from the existing app; and 3) based on the outcome of the previous step, we attempt to predict the exact app, new or existing, from which the next purchase will come. The results yield new insights into spending habits in the mobile digital marketplace. | iPhone's Digital Marketplace: Characterizing the Big Spenders | 9,304 |
In this paper, we study the Cost-aware Target Viral Marketing (CTVM) problem, a generalization of Influence Maximization (IM). CTVM asks for the most cost-effective users to influence the most relevant users. In contrast to the vast literature, we attempt to offer exact solutions. As the problem is NP-hard, thus, exact solutions are intractable, we propose TipTop, a $(1-\epsilon)$-optimal solution for arbitrary $\epsilon>0$ that scales to very large networks such as Twitter. At the heart of TipTop lies an innovative technique that reduces the number of samples as much as possible. This allows us to exactly solve CTVM on a much smaller space of generated samples using Integer Programming. Furthermore, TipTop lends a tool for researchers to benchmark their solutions against the optimal one in large-scale networks, which is currently not available. | TipTop: (Almost) Exact Solutions for Influence Maximization in
Billion-scale Networks | 9,305 |
Understanding the dynamics by which urban areas attract visitors is significant for urban development in cities. In addition, identifying services that relate to highly attractive districts is useful to make policies regarding the placement of such places. Thus, we present a framework for classifying districts in cities by their attractiveness to visitors, and relating Points of Interests (POIs) types to districts' attraction patterns. We used Origin-Destination matrices (ODs) mined from cell phone data that capture the flow of trips between each pair of places in Riyadh, Saudi Arabia. We define the attraction profile for a place based on three main statistical features: The amount of visitors a place received, the distribution of distance traveled by visitors on the road network, and the spatial spread of where visitors come from. We use a hierarchical clustering algorithm to classify all places in the city by their features of attraction. We detect three types of Urban Attractors in Riyadh during the morning period: Global which are significant places in the city, Downtown which the central business district and Residential attractors. In addition, we uncover what makes these places different in terms of attraction patterns. We used a statistical significance testing approach to rigorously quantify the relationship between Points of Interests (POIs) types (services) and the 3 patterns of Urban Attractors we detected. The proposed framework can be used for detecting the attraction patterns given by type of services related to each pattern. This is a critical piece of information to inform trip distribution models. | Urban Attractors: Discovering Patterns in Regions of Attraction in
Cities | 9,306 |
Robustness in response to unexpected events is always desirable for real-world networks. To improve the robustness of any networked system, it is important to analyze vulnerability to external perturbation such as random failures or adversarial attacks occurring to elements of the network. In this paper, we study an emerging problem in assessing the robustness of complex networks: the vulnerability of the clustering of the network to the failure of network elements. Specifically, we identify vertices whose failures will critically damage the network by degrading its clustering, evaluated through the average clustering coefficient. This problem is important because any significant change made to the clustering, resulting from element-wise failures, could degrade network performance such as the ability for information to propagate in a social network. We formulate this vulnerability analysis as an optimization problem, prove its NP-completeness and non-monotonicity, and we offer two algorithms to identify the vertices most important to clustering. Finally, we conduct comprehensive experiments in synthesized social networks generated by various well-known models as well as traces of real social networks. The empirical results over other competitive strategies show the efficacy of our proposed algorithms. | Vulnerability of Clustering under Node Failure in Complex Networks | 9,307 |
We consider the Threshold Activation Problem (TAP): given social network $G$ and positive threshold $T$, find a minimum-size seed set $A$ that can trigger expected activation of at least $T$. We introduce the first scalable, parallelizable algorithm with performance guarantee for TAP suitable for datasets with millions of nodes and edges; we exploit the bicriteria nature of solutions to TAP to allow the user to control the running time versus accuracy of our algorithm through a parameter $\alpha \in (0,1)$: given $\eta > 0$, with probability $1 - \eta$ our algorithm returns a solution $A$ with expected activation greater than $T - 2 \alpha T$, and the size of the solution $A$ is within factor $1 + 4 \alpha T + \log ( T )$ of the optimal size. The algorithm runs in time $O \left( \alpha^{-2}\log \left( n / \eta \right) (n + m) |A| \right)$, where $n$, $m$, refer to the number of nodes, edges in the network. The performance guarantee holds for the general triggering model of internal influence and also incorporates external influence, provided a certain condition is met on the cost-effectivity of seed selection. | Scalable Bicriteria Algorithms for the Threshold Activation Problem in
Online Social Networks | 9,308 |
Motivated by the two paradoxical facts that the marginal cost of following one extra candidate is close to zero and that the majority of Twitter users choose to follow only one or two candidates, we study the Twitter follow behaviors observed in the 2016 U.S. presidential election. Specifically, we complete the following tasks: (1) analyze Twitter follow patterns of the presidential election on Twitter, (2) use negative binomial regression to study the effects of gender and occupation on the number of candidates that one follows, and (3) use multinomial logistic regression to investigate the effects of gender, occupation and celebrities on the choice of candidates to follow. | When Follow is Just One Click Away: Understanding Twitter Follow
Behavior in the 2016 U.S. Presidential Election | 9,309 |
Source localization, the act of finding the originator of a disease or rumor in a network, has become an important problem in sociology and epidemiology. The localization is done using the infection state and time of infection of a few designated sensor nodes; however, maintaining sensors can be very costly in practice. We propose the first online approach to source localization: We deploy a priori only a small number of sensors (which reveal if they are reached by an infection) and then iteratively choose the best location to place new sensors in order to localize the source. This approach allows for source localization with a very small number of sensors; moreover, the source can be found while the epidemic is still ongoing. Our method applies to a general network topology and performs well even with random transmission delays. | Back to the Source: an Online Approach for Sensor Placement and Source
Localization | 9,310 |
We present in this paper a multicultural approach to social media marketing analytics, applied in two Facebook brand pages: French (individualistic culture, the country home of the brand) versus Saudi Arabian (collectivistic culture, one of its country hosts), which are published by an internationalbeauty \& cosmetics firm. Using social network analysis and content analysis, we identify the most popular posts and the most influential users within these two brand pages and highlight the different communities emerging from brand and users interactions. These communities seem to be culture oriented when they are constructed around socialization branded posts and product-line oriented when advertising branded posts are concerned. | Social media marketing analytics : a multicultural approach applied to
the beauty \& cosmetic sector | 9,311 |
Information can propagate among Online Social Network (OSN) users at a high speed, which makes the OSNs become important platforms for viral marketing. Although the viral marketing related problems in OSNs have been extensively studied in the past decade, the existing works all assume known propagation rates and are not able to solve the scenario when the rates may dynamically increase for popular topics. In this paper, we propose a novel model, Dynamic Influence Propagation (DIP), which allows propagation rates to change during the diffusion and can be used for describing information propagation in OSNs more realistically. Based on DIP, we define a new research problem: Threshold Activation Problem under DIP (TAP-DIP). TAP-DIP is more generalized than TAP and can be used for studying the DIP model. However, it adds another layer of complexity over the already \#P-hard TAP problem. Despite it hardness, we are able to approximate TAP-DIP with $O(\log|V|)$ ratio. Our solution consists of two major parts: 1) the Lipschitz optimization technique and 2) a novel solution to the general version of TAP, the Multi-TAP problem. We experimentally test our solution Using various real OSN datasets, and demonstrate that our solution not only generates high-quality yet much smaller seed sets when being aware of the rate increase, but also is scalable. In addition, considering DIP or not has a significant difference in seed set selection. | Popular Topics Spread Faster: New Dimension for Influence Propagation in
Online Social Networks | 9,312 |
In this paper, we study how rumors in Online Social Networks (OSNs) may impact the performance of device-to-device (D2D) communication. As D2D is a new technology, people may choose not to use it when believed in rumors of its negative impacts. Thus, the cellular network with underlaying D2D is vulnerable to OSNs as rumors in OSNs may decrement the throughput of the cellular network in popular content delivery scenarios. To analyze the vulnerability, we introduce the problem of finding the most critical nodes in the OSN such that the throughput of a content delivery scenario is minimized when a rumor starts from those nodes. We then propose an efficient solution to the critical nodes detection problem. The severity of such vulnerability is supported by extensive experiments in various simulation settings, from which we observe up to $40\%$ reduction in network throughput. | Vulnerability of D2D Communications from Interconnected Social Networks | 9,313 |
We consider discrete-time distributed averaging algorithms over multi-agent networks with measurement noises and time-varying random graph flows. Each agent updates its state by relative states between neighbours with both additive and multiplicative measurement noises. The network structure is modeled by time-varying random digraphs, which may be spatially and temporally dependent. By developing difference inequalities of proper stochastic Lyapunov function, the algebraic graph theory and martingale convergence theory, we obtain sufficient conditions for stochastic approximation type algorithms to achieve mean square and almost sure average consensus. We prove that all states of agents converge to a common variable in mean square and almost surely if the graph flow is conditionally balanced and uniformly conditionally jointly connected. The mathematical expectation of the common variable is right the average of initial values, and the upper bound of the mean square steady-state error is given quantitatively related to the weights, the algorithm gain and the energy level of the noises. | Distributed Averaging With Random Network Graphs and Noises | 9,314 |
In a former paper we simplified the proof of a theorem on personalized random walk that is fundamental to graph nodes clustering and generalized it to bipartite graphs for a specific case where the proobability of random jump was proprtional to the number of links of "personally prefereed" nodes. In this paper we turn to the more complex issue of graphs in which the random jump follows uniform distribution. | Traditional PageRank versus Network Capacity Bound | 9,315 |
An increasing number of social network mental disorders (SNMDs), such as Cyber-Relationship Addiction, Information Overload, and Net Compulsion, have been recently noted. Symptoms of these mental disorders are usually observed passively today, resulting in delayed clinical intervention. In this paper, we argue that mining online social behavior provides an opportunity to actively identify SNMDs at an early stage. It is challenging to detect SNMDs because the mental factors considered in standard diagnostic criteria (questionnaire) cannot be observed from online social activity logs. Our approach, new and innovative to the practice of SNMD detection, does not rely on self-revealing of those mental factors via questionnaires. Instead, we propose a machine learning framework, namely, Social Network Mental Disorder Detection (SNMDD), that exploits features extracted from social network data to accurately identify potential cases of SNMDs. We also exploit multi-source learning in SNMDD and propose a new SNMD-based Tensor Model (STM) to improve the performance. Our framework is evaluated via a user study with 3126 online social network users. We conduct a feature analysis, and also apply SNMDD on large-scale datasets and analyze the characteristics of the three SNMD types. The results show that SNMDD is promising for identifying online social network users with potential SNMDs. | Mining Online Social Data for Detecting Social Network Mental Disorders | 9,316 |
Network centrality plays an important role in many applications. Central nodes in social networks can be influential, driving opinions and spreading news or rumors.In hyperlinked environments, such as the Web, where users navigate via clicks, central content receives high traffic, becoming targets for advertising campaigns. While there is an extensive amount of work on centrality measures and their efficient computation, controlling nodes' centrality via network updates is a more recent and challenging problem. Performing minimal modifications to a network to achieve a desired property falls under the umbrella of network design problems. This paper is focused on improving the coverage centrality of a set of nodes, which is the number of pairs of nodes that have a shortest path passing through the set, by adding edges to the network. We prove strong inapproximability results and propose a greedy algorithm for maximizing coverage centrality. To ensure scalability to large networks, we also design an efficient sampling algorithm for the problem. In addition to providing an extensive empirical evaluation of our algorithms, we also show that, under some realistic constraints, the proposed solutions achieve almost-optimal approximation for coverage centrality maximization. | Maximizing Coverage Centrality via Network Design: Extended Version | 9,317 |
Users on Twitter are commonly identified by their profile names. These names are used when directly addressing users on Twitter, are part of their profile page URLs, and can become a trademark for popular accounts, with people referring to celebrities by their real name and their profile name, interchangeably. Twitter, however, has chosen to not permanently link profile names to their corresponding user accounts. In fact, Twitter allows users to change their profile name, and afterwards makes the old profile names available for other users to take. In this paper, we provide a large-scale study of the phenomenon of profile name reuse on Twitter. We show that this phenomenon is not uncommon, investigate the dynamics of profile name reuse, and characterize the accounts that are involved in it. We find that many of these accounts adopt abandoned profile names for questionable purposes, such as spreading malicious content, and using the profile name's popularity for search engine optimization. Finally, we show that this problem is not unique to Twitter (as other popular online social networks also release profile names) and argue that the risks involved with profile-name reuse outnumber the advantages provided by this feature. | What's in a Name? Understanding Profile Name Reuse on Twitter | 9,318 |
Influence maximization is the problem of finding influential users, or nodes, in a graph so as to maximize the spread of information. It has many applications in advertising and marketing on social networks. In this paper, we study a highly generic version of influence maximization, one of optimizing influence campaigns by sequentially selecting "spread seeds" from a set of influencers, a small subset of the node population, under the hypothesis that, in a given campaign, previously activated nodes remain "persistently" active throughout and thus do not yield further rewards. This problem is in particular relevant for an important form of online marketing, known as influencer marketing, in which the marketers target a sub-population of influential people, instead of the entire base of potential buyers. Importantly, we make no assumptions on the underlying diffusion model and we work in a setting where neither a diffusion network nor historical activation data are available. We call this problem online influencer marketing with persistence (in short, OIMP). We first discuss motivating scenarios and present our general approach. We introduce an estimator on the influencers' remaining potential -- the expected number of nodes that can still be reached from a given influencer -- and justify its strength to rapidly estimate the desired value, relying on real data gathered from Twitter. We then describe a novel algorithm, GT-UCB, relying on upper confidence bounds on the remaining potential. We show that our approach leads to high-quality spreads on both simulated and real datasets, even though it makes almost no assumptions on the diffusion medium. Importantly, it is orders of magnitude faster than state-of-the-art influence maximization methods, making it possible to deal with large-scale online scenarios. | Algorithms for Online Influencer Marketing | 9,319 |
What is the best way to describe a user in a social network with just a few numbers? Mathematically, this is equivalent to assigning a vector representation to each node in a graph, a process called graph embedding. We propose a novel framework, GEM-D that unifies most of the past algorithms such as LapEigs, DeepWalk and node2vec. GEM-D achieves its goal by decomposing any graph embedding algorithm into three building blocks: node proximity function, warping function and loss function. Based on thorough analysis of GEM-D, we propose a novel algorithm, called UltimateWalk, which outperforms the most-recently proposed state-of-the-art DeepWalk and node2vec. The contributions of this work are: (1) The proposed framework, GEM-D unifies the past graph embedding algorithms and provides a general recipe of how to design a graph embedding; (2) the nonlinearlity in the warping function contributes significantly to the quality of embedding and the exponential function is empirically optimal; (3) the proposed algorithm, UltimateWalk is one-click (no user-defined parameters), scalable and has a closed-form solution. | Fast, Warped Graph Embedding: Unifying Framework and One-Click Algorithm | 9,320 |
Cyber-epidemics, the widespread of fake news or propaganda through social media, can cause devastating economic and political consequences. A common countermeasure against cyber-epidemics is to disable a small subset of suspected social connections or accounts to effectively contain the epidemics. An example is the recent shutdown of 125,000 ISIS-related Twitter accounts. Despite many proposed methods to identify such subset, none are scalable enough to provide high-quality solutions in nowadays billion-size networks. To this end, we investigate the Spread Interdiction problems that seek most effective links (or nodes) for removal under the well-known Linear Threshold model. We propose novel CPU-GPU methods that scale to networks with billions of edges, yet, possess rigorous theoretical guarantee on the solution quality. At the core of our methods is an $O(1)$-space out-of-core algorithm to generate a new type of random walks, called Hitting Self-avoiding Walks (HSAWs). Such a low memory requirement enables handling of big networks and, more importantly, hiding latency via scheduling of millions of threads on GPUs. Comprehensive experiments on real-world networks show that our algorithms provides much higher quality solutions and are several order of magnitude faster than the state-of-the art. Comparing to the (single-core) CPU counterpart, our GPU implementations achieve significant speedup factors up to 177x on a single GPU and 338x on a GPU pair. | Blocking Self-avoiding Walks Stops Cyber-epidemics: A Scalable GPU-based
Approach | 9,321 |
Strong and supportive social relationships are fundamental to our well-being. However, there are costs to their maintenance, resulting in a trade-off between quality and quantity, a typical strategy being to put a lot of effort on a few high-intensity relationships while maintaining larger numbers of less close relationships. It has also been shown that there are persistent individual differences in this pattern; some individuals allocate their efforts more uniformly across their networks, while others strongly focus on their closest relationships. Furthermore, some individuals maintain more stable networks than others. Here, we focus on how personality traits of individuals affect this picture, using mobile phone calls records and survey data from the Mobile Territorial Lab (MTL) study. In particular, we look at the relationship between personality traits and the (i) persistence of social signatures, namely the similarity of the social signature shape of an individual measured in different time intervals; (ii) the turnover in egocentric networks, that is, differences in the set of alters present at two consecutive temporal intervals; and (iii) the rank dynamics defined as the variation of alter rankings in egocentric networks in consecutive intervals. We observe that some traits have effects on the stability of the social signatures as well as network turnover and rank dynamics. As an example, individuals who score highly in the Openness to Experience trait tend to have higher levels of network turnover and larger alter rank variations. On broader terms, our study shows that personality traits clearly affect the ways in which individuals maintain their personal networks. | Personality Traits and Ego-network Dynamics | 9,322 |
Cascades on online networks have been a popular subject of study in the past decade, and there is a considerable literature on phenomena such as diffusion mechanisms, virality, cascade prediction, and peer network effects. However, a basic question has received comparatively little attention: how desirable are cascades on a social media platform from the point of view of users? While versions of this question have been considered from the perspective of the producers of cascades, any answer to this question must also take into account the effect of cascades on their audience. In this work, we seek to fill this gap by providing a consumer perspective of cascade. Users on online networks play the dual role of producers and consumers. First, we perform an empirical study of the interaction of Twitter users with retweet cascades. We measure how often users observe retweets in their home timeline, and observe a phenomenon that we term the "Impressions Paradox": the share of impressions for cascades of size k decays much slower than frequency of cascades of size k. Thus, the audience for cascades can be quite large even for rare large cascades. We also measure audience engagement with retweet cascades in comparison to non-retweeted content. Our results show that cascades often rival or exceed organic content in engagement received per impression. This result is perhaps surprising in that consumers didn't opt in to see tweets from these authors. Furthermore, although cascading content is widely popular, one would expect it to eventually reach parts of the audience that may not be interested in the content. Motivated by our findings, we posit a theoretical model that focuses on the effect of cascades on the audience. Our results on this model highlight the balance between retweeting as a high-quality content selection mechanism and the role of network users in filtering irrelevant content. | Cascades: A view from Audience | 9,323 |
Social Networking accounts for a significant chunk of interest among various online activities~\cite{smith2009social}. The proclivity of being social, online, has been ingrained in us so much that we are actively producing content for the rest of the world to see or take interest in our whereabouts, our meals, our opinions, photographs etc. Yelp (https://www.yelp.com/), seamlessly, integrates this very aspect of people in its portal. It engages people to write reviews about the businesses they have availed the services of, rate them, add photographs, tags, follow other people and their activities, etc. In this paper we examine and present the co-relation between a user's rating and the influence of the people, that the user follows, on the user for a particular business. The group of people that the user follows is commonly referred as friends of the user. We also analyze if a user can get influenced, if a business has a certain number of reviews already present or if the reviews have been written by elite reviewers (a reviewer who, according to Yelp, has contributed exceptionally in engaging the community in the form of consistency in writing reviews, as well as the quality of the reviews). Our analysis, through correlation and regression techniques, is able to prove that the user's rating remains unaffected by the number of people a user was friends with nor does the existing number of reviews and presence of elite reviewers helps in influencing a user. What shapes a user's rating is the overall experience, that the user had at the restaurant. | Social Data Analysis: A Study on Friend Rating Influence | 9,324 |
Selfies have become increasingly fashionable in the social media era. People are willing to share their selfies in various social media platforms such as Facebook, Instagram and Flicker. The popularity of selfie have caught researchers' attention, especially psychologists. In computer vision and machine learning areas, little attention has been paid to this phenomenon as a valuable data source. In this paper, we focus on exploring the deeper personal patterns behind people's different kinds of selfie-posting behaviours. We develop this work based on a dataset of WeChat, one of the most extensively used instant messaging platform in China. In particular, we first propose an unsupervised approach to classify the images posted by users. Based on the classification result, we construct three types of user-level features that reflect user preference, activity and posting habit. Based on these features, for a series of selfie related tasks, we build classifiers that can accurately predict two sets of users with opposite selfie-posting behaviours. We have found that people's interest, activity and posting habit have a great influence on their selfie-posting behaviours. For example, the classification accuracy between selfie-posting addict and nonaddict reaches 89.36%. We also prove that using user's image information to predict these behaviours achieve better performance than using text information. More importantly, for each set of users with a specific selfie-posting behaviour, we extract and visualize significant personal patterns about them. In addition, we cluster users and extract their high-level attributes, revealing the correlation between these attributes and users' selfie-posting behaviours. In the end, we demonstrate that users' selfie-posting behaviour, as a good predictor, could predict their different preferences toward these high-level attributes accurately. | A Selfie is Worth a Thousand Words: Mining Personal Patterns behind User
Selfie-posting Behaviours | 9,325 |
Online communities are changing the ways that creative professionals such as artists and designers share ideas, receive feedback, and find inspiration. While they became increasingly popular, there have been few studies so far. In this paper, we investigate Behance, an online community site for creatives to maintain relationships with others and showcase their works from various fields such as graphic design, illustration, photography, and fashion. We take a quantitative approach to study three research questions about the site. What attract followers and appreciation of artworks on Behance? what patterns of activity exist around topics? And, lastly, does color play a role in attracting appreciation? In summary, being male suggests more followers and appreciations, most users focus on a few topics, and grayscale colors mean fewer appreciations. This work serves as a preliminary overview of a creative community that later studies can build on. | Creative Community Demystified: A Statistical Overview of Behance | 9,326 |
This paper analyzes the impact of peer effects on electricity consumption of a network of rational, utility-maximizing users. Users derive utility from consuming electricity as well as consuming less energy than their neighbors. However, a disutility is incurred for consuming more than their neighbors. To maximize the profit of the load-serving entity that provides electricity to such users, we develop a two-stage game-theoretic model, where the entity sets the prices in the first stage. In the second stage, consumers decide on their demand in response to the observed price set in the first stage so as to maximize their utility. To this end, we derive theoretical statements under which such peer effects reduce aggregate user consumption. Further, we obtain expressions for the resulting electricity consumption and profit of the load serving entity for the case of perfect price discrimination and a single price under complete information, and approximations under incomplete information. Simulations suggest that exposing only a selected subset of all users to peer effects maximizes the entity's profit. | How Peer Effects Influence Energy Consumption | 9,327 |
Modeling the popularity dynamics of an online item is an important open problem in computational social science. This paper presents an in-depth study of popularity dynamics under external promotions, especially in predicting popularity jumps of online videos, and determining effective and efficient schedules to promote online content. The recently proposed Hawkes Intensity Process (HIP) models popularity as a non-linear interplay between exogenous stimuli and the endogenous reactions. Here, we propose two novel metrics based on HIP: to describe popularity gain per unit of promotion, and to quantify the time it takes for such effects to unfold. We make increasingly accurate forecasts of future popularity by including information about the intrinsic properties of the video, promotions it receives, and the non-linear effects of popularity ranking. We illustrate by simulation the interplay between the unfolding of popularity over time, and the time-sensitive value of resources. Lastly, our model lends a novel explanation of the commonly adopted periodic and constant promotion strategy in advertising, as increasing the perceived viral potential. This study provides quantitative guidelines about setting promotion schedules considering content virality, timing, and economics. | Online Popularity under Promotion: Viral Potential, Forecasting, and the
Economics of Time | 9,328 |
Reciprocity in directed networks points to user's willingness to return favors in building mutual interactions. High reciprocity has been widely observed in many directed social media networks such as following relations in Twitter and Tumblr. Therefore, reciprocal relations between users are often regarded as a basic mechanism to create stable social ties and play a crucial role in the formation and evolution of networks. Each reciprocity relation is formed by two parasocial links in a back-and-forth manner with a time delay. Hence, understanding the delay can help us gain better insights into the underlying mechanisms of network dynamics. Meanwhile, the accurate prediction of delay has practical implications in advancing a variety of real-world applications such as friend recommendation and marketing campaign. For example, by knowing when will users follow back, service providers can focus on the users with a potential long reciprocal delay for effective targeted marketing. This paper presents the initial investigation of the time delay in reciprocal relations. Our study is based on a large-scale directed network from Tumblr that consists of 62.8 million users and 3.1 billion user following relations with a timespan of multiple years (from 31 Oct 2007 to 24 Jul 2013). We reveal a number of interesting patterns about the delay that motivate the development of a principled learning model to predict the delay in reciprocal relations. Experimental results on the above mentioned dynamic networks corroborate the effectiveness of the proposed delay prediction model. | Understanding and Predicting Delay in Reciprocal Relations | 9,329 |
With the growing popularity of online social media, identifying influential users in these social networks has become very popular. Existing works have studied user attributes, network structure and user interactions when measuring user influence. In contrast to these works, we focus on user behavioural characteristics. We investigate the temporal dynamics of user activity patterns and how these patterns affect user interactions. We assimilate such characteristics into a PageRank based temporal influence ranking model (TIR) to identify influential users. The transition probability in TIR is predicted by a logistic regression model and the random walk, biased according to users' temporal activity patterns. Experiments demonstrate that TIR has better performance and is more stable than the existing models in global influence ranking and friend recommendation. | Finding Influentials in Twitter: A Temporal Influence Ranking Model | 9,330 |
Social media has played an important role in shaping political discourse over the last decade. At the same time, it is often perceived to have increased political polarization, thanks to the scale of discussions and their public nature. In this paper, we try to answer the question of whether political polarization in the US on Twitter has increased over the last eight years. We analyze a large longitudinal Twitter dataset of 679,000 users and look at signs of polarization in their (i) network - how people follow political and media accounts, (ii) tweeting behavior - whether they retweet content from both sides, and (iii) content - how partisan the hashtags they use are. Our analysis shows that online polarization has indeed increased over the past eight years and that, depending on the measure, the relative change is 10%-20%. Our study is one of very few with such a long-term perspective, encompassing two US presidential elections and two mid-term elections, providing a rare longitudinal analysis. | A Long-Term Analysis of Polarization on Twitter | 9,331 |
Increasing evidence suggests that a growing amount of social media content is generated by autonomous entities known as social bots. In this work we present a framework to detect such entities on Twitter. We leverage more than a thousand features extracted from public data and meta-data about users: friends, tweet content and sentiment, network patterns, and activity time series. We benchmark the classification framework by using a publicly available dataset of Twitter bots. This training data is enriched by a manually annotated collection of active Twitter users that include both humans and bots of varying sophistication. Our models yield high accuracy and agreement with each other and can detect bots of different nature. Our estimates suggest that between 9% and 15% of active Twitter accounts are bots. Characterizing ties among accounts, we observe that simple bots tend to interact with bots that exhibit more human-like behaviors. Analysis of content flows reveals retweet and mention strategies adopted by bots to interact with different target groups. Using clustering analysis, we characterize several subclasses of accounts, including spammers, self promoters, and accounts that post content from connected applications. | Online Human-Bot Interactions: Detection, Estimation, and
Characterization | 9,332 |
The goal of cluster analysis in survival data is to identify clusters that are decidedly associated with the survival outcome. Previous research has explored this problem primarily in the medical domain with relatively small datasets, but the need for such a clustering methodology could arise in other domains with large datasets, such as social networks. Concretely, we wish to identify different survival classes in a social network by clustering the users based on their lifespan in the network. In this paper, we propose a decision tree based algorithm that uses a global normalization of $p$-values to identify clusters with significantly different survival distributions. We evaluate the clusters from our model with the help of a simple survival prediction task and show that our model outperforms other competing methods. | Identifying User Survival Types via Clustering of Censored Social
Network Data | 9,333 |
In this paper, we study the implications of the commonplace assumption that most social media studies make with respect to the nature of message shares (such as retweets) as a predominantly positive interaction. By analyzing two large longitudinal Brazilian Twitter datasets containing 5 years of conversations on two polarizing topics - Politics and Sports - we empirically demonstrate that groups holding antagonistic views can actually retweet each other more often than they retweet other groups. We show that assuming retweets as endorsement interactions can lead to misleading conclusions with respect to the level of antagonism among social communities, and that this apparent paradox is explained in part by the use of retweets to quote the original content creator out of the message's original temporal context, for humor and criticism purposes. As a consequence, messages diffused on online media can have their polarity reversed over time, what poses challenges for social and computer scientists aiming to classify and track opinion groups on online media. On the other hand, we found that the time users take to retweet a message after it has been originally posted can be a useful signal to infer antagonism in social platforms, and that surges of out-of-context retweets correlate with sentiment drifts triggered by real-world events. We also discuss how such evidences can be embedded in sentiment analysis models. | Antagonism also Flows through Retweets: The Impact of Out-of-Context
Quotes in Opinion Polarization Analysis | 9,334 |
Social media platforms provide continuous access to user generated content that enables real-time monitoring of user behavior and of events. The geographical dimension of such user behavior and events has recently caught a lot of attention in several domains: mobility, humanitarian, or infrastructural. While resolving the location of a user can be straightforward, depending on the affordances of their device and/or of the application they are using, in most cases, locating a user demands a larger effort, such as exploiting textual features. On Twitter for instance, only 2% of all tweets are geo-referenced. In this paper, we present a system for zoomed-in grounding (below city level) for short messages (e.g., tweets). The system combines different natural language processing and machine learning techniques to increase the number of geo-grounded tweets, which is essential to many applications such as disaster response and real-time traffic monitoring. | QT2S: A System for Monitoring Road Traffic via Fine Grounding of Tweets | 9,335 |
Many people dream to become famous, YouTube video makers also wish their videos to have a large audience, and product retailers always hope to expose their products to customers as many as possible. Do these seemingly different phenomena share a common structure? We find that fame, popularity, or exposure, could be modeled as a node's discoverability on some properly defined network, and all of the previously mentioned phenomena can be commonly stated as a target node wants to be discovered easily by the other nodes in the network. In this work, we explicitly define a node's discoverability in a network, and formulate a general node discoverability optimization problem, where the goal is to create a budgeted set of incoming edges to the target node so as to optimize the target node's discoverability in the network. Although the optimization problem is proven to be NP-hard, we find that the defined discoverability measures have good properties that enable us to use a greedy algorithm to find provably near-optimal solutions. The computational complexity of a greedy algorithm is dominated by the time cost of an oracle call, i.e., calculating the marginal gain of a node. To scale up the oracle call over large networks, we propose an estimation-and-refinement approach, that provides a good trade-off between estimation accuracy and computational efficiency. Experiments conducted on real-world networks demonstrate that our method is thousands of times faster than an exact method using dynamic programming, thereby allowing us to solve the node discoverability optimization problem on large networks. | Optimizing Node Discovery on Networks: Problem Definitions, Fast
Algorithms, and Observations | 9,336 |
Several network embedding models have been developed for unsigned networks. However, these models based on skip-gram cannot be applied to signed networks because they can only deal with one type of link. In this paper, we present our signed network embedding model called SNE. Our SNE adopts the log-bilinear model, uses node representations of all nodes along a given path, and further incorporates two signed-type vectors to capture the positive or negative relationship of each edge along the path. We conduct two experiments, node classification and link prediction, on both directed and undirected signed networks and compare with four baselines including a matrix factorization method and three state-of-the-art unsigned network embedding models. The experimental results demonstrate the effectiveness of our signed network embedding. | SNE: Signed Network Embedding | 9,337 |
The risks of publishing privacy-sensitive data have received considerable attention recently. Several de-anonymization attacks have been proposed to re-identify individuals even if data anonymization techniques were applied. However, there is no theoretical quantification for relating the data utility that is preserved by the anonymization techniques and the data vulnerability against de-anonymization attacks. In this paper, we theoretically analyze the de-anonymization attacks and provide conditions on the utility of the anonymized data (denoted by anonymized utility) to achieve successful de-anonymization. To the best of our knowledge, this is the first work on quantifying the relationships between anonymized utility and de-anonymization capability. Unlike previous work, our quantification analysis requires no assumptions about the graph model, thus providing a general theoretical guide for developing practical de-anonymization/anonymization techniques. Furthermore, we evaluate state-of-the-art de-anonymization attacks on a real-world Facebook dataset to show the limitations of previous work. By comparing these experimental results and the theoretically achievable de-anonymization capability derived in our analysis, we further demonstrate the ineffectiveness of previous de-anonymization attacks and the potential of more powerful de-anonymization attacks in the future. | Quantification of De-anonymization Risks in Social Networks | 9,338 |
Among the topics discussed in Social Media, some lead to controversy. A number of recent studies have focused on the problem of identifying controversy in social media mostly based on the analysis of textual content or rely on global network structure. Such approaches have strong limitations due to the difficulty of understanding natural language, and of investigating the global network structure. In this work we show that it is possible to detect controversy in social media by exploiting network motifs, i.e., local patterns of user interaction. The proposed approach allows for a language-independent and fine- grained and efficient-to-compute analysis of user discussions and their evolution over time. The supervised model exploiting motif patterns can achieve 85% accuracy, with an improvement of 7% compared to baseline structural, propagation-based and temporal network features. | A Motif-based Approach for Identifying Controversy | 9,339 |
Social media has become an indispensable part of the everyday lives of millions of people around the world. It provides a platform for expressing opinions and beliefs, communicated to a massive audience. However, this ease with which people can express themselves has also allowed for the large scale spread of propaganda and hate speech. To prevent violating the abuse policies of social media platforms and also to avoid detection by automatic systems like Google's Conversation AI, racists have begun to use a code (a movement termed Operation Google). This involves substituting references to communities by benign words that seem out of context, in hate filled posts or Tweets. For example, users have used the words Googles and Bings to represent the African-American and Asian communities, respectively. By generating the list of users who post such content, we move a step forward from classifying tweets by allowing us to study the usage pattern of these concentrated set of users. | Detecting the Hate Code on Social Media | 9,340 |
In this paper, we are interested in understanding the interrelationships between mainstream and social media in forming public opinion during mass crises, specifically in regards to how events are framed in the mainstream news and on social networks and to how the language used in those frames may allow to infer political slant and partisanship. We study the lingual choices for political agenda setting in mainstream and social media by analyzing a dataset of more than 40M tweets and more than 4M news articles from the mass protests in Ukraine during 2013-2014 - known as "Euromaidan" - and the post-Euromaidan conflict between Russian, pro-Russian and Ukrainian forces in eastern Ukraine and Crimea. We design a natural language processing algorithm to analyze at scale the linguistic markers which point to a particular political leaning in online media and show that political slant in news articles and Twitter posts can be inferred with a high level of accuracy. These findings allow us to better understand the dynamics of partisan opinion formation during mass crises and the interplay between main- stream and social media in such circumstances. | Identifying Partisan Slant in News Articles and Twitter during Political
Crises | 9,341 |
This paper presents methods to compare networks where relationships between pairs of nodes in a given network are defined. We define such network distance by searching for the optimal method to embed one network into another network, prove that such distance is a valid metric in the space of networks modulo permutation isomorphisms, and examine its relationship with other network metrics. The network distance defined can be approximated via multi-dimensional scaling, however, the lack of structure in networks results in poor approximations. To alleviate such problem, we consider methods to define the interiors of networks. We show that comparing interiors induced from a pair of networks yields the same result as the actual network distance between the original networks. Practical implications are explored by showing the ability to discriminate networks generated by different models. | Network Comparison: Embeddings and Interiors | 9,342 |
Gender differences is a phenomenon around the world actively researched by social scientists. Traditionally, the data used to support such studies is manually obtained, often through surveys with volunteers. However, due to their inherent high costs because of manual steps, such traditional methods do not quickly scale to large-size studies. We here investigate a particular aspect of gender differences: preferences for venues. To that end we explore the use of check-in data collected from Foursquare to estimate cultural gender preferences for venues in the physical world. For that, we first demonstrate that by analyzing the check-in data in various regions of the world we can find significant differences in preferences for specific venues between gender groups. Some of these significant differences reflect well-known cultural patterns. Moreover, we also gathered evidence that our methodology offers useful information about gender preference for venues in a given region in the real world. This suggests that gender and venue preferences observed may not be independent. Our results suggests that our proposed methodology could be a promising tool to support studies on gender preferences for venues at different spatial granularities around the world, being faster and cheaper than traditional methods, besides quickly capturing changes in the real world. | Gender Matters! Analyzing Global Cultural Gender Preferences for Venues
Using Social Sensing | 9,343 |
Incorporating graphs in the analysis of multivariate signals is becoming a standard way to understand the interdependency of activity recorded at different sites. The new research frontier in this direction includes the important problem of how to assess dynamic changes of signal activity. We address this problem in a novel way by defining the graph-variate signal alongside methods for its analysis. Essentially, graph-variate signal analysis leverages graphs of reliable connectivity information to filter instantaneous bivariate functions of the multivariate signal. This opens up a new and robust approach to analyse joint signal and network dynamics at sample resolution. Furthermore, our method can be formulated as instantaneous networks on which standard network analysis can be implemented. When graph connectivity is estimated from the multivariate signal itself, the appropriate consideration of instantaneous graph signal functions allows for a novel dynamic connectivity measure-- graphvariate dynamic (GVD) connectivity-- which is robust to spurious short-term dependencies. Particularly, we present appropriate functions for three pertinent connectivity metrics-- correlation, coherence and the phase-lag index. We show that our approach can determine signals with a single correlated couple against wholly uncorrelated data of up to 128 nodes in signal size (1 out of 8128 weighted edges). GVD connectivity is also shown to be more robust than i) other GSP approaches at detecting a randomly traveling spheroid on a 3D grid and ii) standard dynamic connectivity in determining differences in EEG restingstate and task-related activity. We also demonstrate its use in revealing hidden depth correlations from geophysical gamma ray data. We expect that the methods and framework presented will provide new approaches to data analysis in a variety of applied settings. | Graph-Variate Signal Analysis | 9,344 |
In 2010, a paper entitled "From Obscurity to Prominence in Minutes: Political Speech and Real-time search" won the Best Paper Prize of the Web Science 2010 Conference. Among its findings were the discovery and documentation of what was termed a "Twitter-bomb", an organized effort to spread misinformation about the democratic candidate Martha Coakley through anonymous Twitter accounts. In this paper, after summarizing the details of that event, we outline the recipe of how social networks are used to spread misinformation. One of the most important steps in such a recipe is the "infiltration" of a community of users who are already engaged in conversations about a topic, to use them as organic spreaders of misinformation in their extended subnetworks. Then, we take this misinformation spreading recipe and indicate how it was successfully used to spread fake news during the 2016 U.S. Presidential Election. The main differences between the scenarios are the use of Facebook instead of Twitter, and the respective motivations (in 2010: political influence; in 2016: financial benefit through online advertising). After situating these events in the broader context of exploiting the Web, we seize this opportunity to address limitations of the reach of research findings and to start a conversation about how communities of researchers can increase their impact on real-world societal issues. | The Fake News Spreading Plague: Was it Preventable? | 9,345 |
Language in social media is extremely dynamic: new words emerge, trend and disappear, while the meaning of existing words can fluctuate over time. Such dynamics are especially notable during a period of crisis. This work addresses several important tasks of measuring, visualizing and predicting short term text representation shift, i.e. the change in a word's contextual semantics, and contrasting such shift with surface level word dynamics, or concept drift, observed in social media streams. Unlike previous approaches on learning word representations from text, we study the relationship between short-term concept drift and representation shift on a large social media corpus - VKontakte posts in Russian collected during the Russia-Ukraine crisis in 2014-2015. Our novel contributions include quantitative and qualitative approaches to (1) measure short-term representation shift and contrast it with surface level concept drift; (2) build predictive models to forecast short-term shifts in meaning from previous meaning as well as from concept drift; and (3) visualize short-term representation shift for example keywords to demonstrate the practical use of our approach to discover and track meaning of newly emerging terms in social media. We show that short-term representation shift can be accurately predicted up to several weeks in advance. Our unique approach to modeling and visualizing word representation shifts in social media can be used to explore and characterize specific aspects of the streaming corpus during crisis events and potentially improve other downstream classification tasks including real-time event detection. | Measuring, Predicting and Visualizing Short-Term Change in Word
Representation and Usage in VKontakte Social Network | 9,346 |
Social media expose millions of users every day to information campaigns --- some emerging organically from grassroots activity, others sustained by advertising or other coordinated efforts. These campaigns contribute to the shaping of collective opinions. While most information campaigns are benign, some may be deployed for nefarious purposes. It is therefore important to be able to detect whether a meme is being artificially promoted at the very moment it becomes wildly popular. This problem has important social implications and poses numerous technical challenges. As a first step, here we focus on discriminating between trending memes that are either organic or promoted by means of advertisement. The classification is not trivial: ads cause bursts of attention that can be easily mistaken for those of organic trends. We designed a machine learning framework to classify memes that have been labeled as trending on Twitter.After trending, we can rely on a large volume of activity data. Early detection, occurring immediately at trending time, is a more challenging problem due to the minimal volume of activity data that is available prior to trending.Our supervised learning framework exploits hundreds of time-varying features to capture changing network and diffusion patterns, content and sentiment information, timing signals, and user meta-data. We explore different methods for encoding feature time series. Using millions of tweets containing trending hashtags, we achieve 75% AUC score for early detection, increasing to above 95% after trending. We evaluate the robustness of the algorithms by introducing random temporal shifts on the trend time series. Feature selection analysis reveals that content cues provide consistently useful signals; user features are more informative for early detection, while network and timing features are more helpful once more data is available. | Early Detection of Promoted Campaigns on Social Media | 9,347 |
Background: Social media public health campaigns have the advantage of tailored messaging at low cost and large reach, but little is known about what would determine their feasibility as tools for inducing attitude and behavior change. Objective: The aim of this study was to test the feasibility of designing, implementing, and evaluating a social media-enabled intervention for skin cancer prevention. Conclusions: Social media-disseminated public health messages reached more than 23% of the Northern Ireland population. A Web-based survey suggested that the campaign might have contributed to improved knowledge and attitudes toward skin cancer among the target population. Findings suggested that shocking and humorous messages generated greatest impressions and engagement, but information-based messages were likely to be shared most. The extent of behavioral change as a result of the campaign remains to be explored, however, the change of attitudes and knowledge is promising. Social media is an inexpensive, effective method for delivering public health messages. However, existing and traditional process evaluation methods may not be suitable for social media. | Tweet for Behavior Change: Using Social Media for the Dissemination of
Public Health Messages | 9,348 |
It has been recently reported that the reciprocity of real-life weighted networks is very pronounced, however its impact on dynamical processes is poorly understood. In this paper, we study random walks in a scale-free directed weighted network with a trap at the central hub node, where the weight of each directed edge is dominated by a parameter controlling the extent of network reciprocity. We derive two expressions for the mean first passage time (MFPT) to the trap, by using two different techniques, the results of which agree well with each other. We also analytically determine all the eigenvalues as well as their multiplicities for the fundamental matrix of the dynamical process, and show that the largest eigenvalue has an identical dominant scaling as that of the MFPT.We find that the weight parameter has a substantial effect on the MFPT, which behaves as a power-law function of the system size with the power exponent dependent on the parameter, signaling the crucial role of reciprocity in random walks occurring in weighted networks. | Effects of reciprocity on random walks in weighted networks | 9,349 |
As a typical Cyber-Physical System (CPS), smart water distribution networks require monitoring of underground water pipes with high sample rates for precise data analysis and water network control. Due to poor underground wireless channel quality and long-range communication requirements, high transmission power is typically adopted to communicate high-speed sensor data streams; posing challenges for long term sustainable monitoring. In this paper, we develop the first sustainable water sensing system, exploiting energy harvesting opportunities from water flows. Our system does this by scheduling the transmission of a subset of the data streams, while other correlated streams are estimated using auto-regressive models based on the sound-velocity propagation of pressure signals inside water networks. To compute the optimal scheduling policy, we formalize a stochastic optimization problem to maximize the estimation reliability, while ensuring the system's sustainable operation under dynamic conditions. We develop Data Transmission Scheduling (DTS), an asymptotically optimal scheme; and FAST-DTS, a lightweight online algorithm that can adapt to arbitrary energy and correlation dynamics. Using over 170 days of real data from our smart water system deployment and conducting in-vitro experiments to our small-scale testbed; our evaluation demonstrates that Fast-DTS significantly outperforms three alternatives, considering data reliability, energy utilization, and sustainable operation. | Reliability or Sustainability: Optimal Data Stream Estimation and
Scheduling in Smart Water Networks | 9,350 |
In this paper, a comparative study was conducted between complex networks representing origin and destination survey data. Similarities were found between the characteristics of the networks of Brazilian cities with networks of foreign cities. Power laws were found in the distributions of edge weights and this scale - free behavior can occur due to the economic characteristics of the cities. | Análise comparativa de pesquisas de origens e destinos: uma abordagem
baseada em Redes Complexas | 9,351 |
A multimodal network encodes relationships between the same set of nodes in multiple settings, and network alignment is a powerful tool for transferring information and insight between a pair of networks. We propose a method for multimodal network alignment that computes a matrix which indicates the alignment, but produces the result as a low-rank factorization directly. We then propose new methods to compute approximate maximum weight matchings of low-rank matrices to produce an alignment. We evaluate our approach by applying it on synthetic networks and use it to de-anonymize a multimodal transportation network. | Multimodal Network Alignment | 9,352 |
Today, users are reading the news through social platforms. These platforms are built to facilitate crowd engagement, but not necessarily disseminate useful news to inform the masses. Hence, the news that is highly engaged with may not be the news that best informs. While predicting news popularity has been well studied, it has not been studied in the context of crowd manipulations. In this paper, we provide some preliminary results to a longer term project on crowd and platform manipulations of news and news popularity. In particular, we choose to study known features for predicting news popularity and how those features may change on reddit.com, a social platform used commonly for news aggregation. Along with this, we explore ways in which users can alter the perception of news through changing the title of an article. We find that news on reddit is predictable using previously studied sentiment and content features and that posts with titles changed by reddit users tend to be more popular than posts with the original article title. | The Impact of Crowds on News Engagement: A Reddit Case Study | 9,353 |
Polarized topics often spark discussion and debate on social media. Recent studies have shown that polarized debates have a specific clustered structure in the endorsement net- work, which indicates that users direct their endorsements mostly to ideas they already agree with. Understanding these polarized discussions and exposing social media users to content that broadens their views is of paramount importance. The contribution of this demonstration is two-fold. (i) A tool to visualize retweet networks about controversial issues on Twitter. By using our visualization, users can understand how polarized discussions are shaped on Twitter, and explore the positions of the various actors. (ii) A solution to reduce polarization of such discussions. We do so by exposing users to information which presents a contrarian point of view. Users can visually inspect our recommendations and understand why and how these would play out in terms of the retweet network. Our demo (https://users.ics.aalto.fi/kiran/reducingControversy/ homepage) provides one of the first steps in developing automated tools that help users explore, and possibly escape, their echo chambers. The ideas in the demo can also help content providers design tools to broaden their reach to people with different political and ideological backgrounds. | Exposing Twitter Users to Contrarian News | 9,354 |
Collaborative crowdsourcing has become a popular approach to organizing work across the globe. Being global also means being vulnerable to shocks -- unforeseen events that disrupt crowds -- that originate from any country. In this study, we examine changes in collaborative behavior of editors of Chinese Wikipedia that arise due to the 2005 government censor- ship in mainland China. Using the exogenous variation in the fraction of editors blocked across different articles due to the censorship, we examine the impact of reduction in group size, which we denote as the shock level, on three collaborative behavior measures: volume of activity, centralization, and conflict. We find that activity and conflict drop on articles that face a shock, whereas centralization increases. The impact of a shock on activity increases with shock level, whereas the impact on centralization and conflict is higher for moderate shock levels than for very small or very high shock levels. These findings provide support for threat rigidity theory -- originally introduced in the organizational theory literature -- in the context of large-scale collaborative crowds. | Shocking the Crowd: The Effect of Censorship Shocks on Chinese Wikipedia | 9,355 |
Recently, two influential PNAS papers have shown how our preferences for 'Hello Kitty' and 'Harley Davidson', obtained through Facebook likes, can accurately predict details about our personality, religiosity, political attitude and sexual orientation (Konsinski et al. 2013; Youyou et al 2015). In this paper, we make the claim that though the wide variety of Facebook likes might predict such personal traits, even more accurate and generalizable results can be reached through applying a contexts-specific, parsimonious data strategy. We built this claim by predicting present day voter intention based solely on likes directed toward posts from political actors. Combining the online and offline, we join a subsample of surveyed respondents to their public Facebook activity and apply machine learning classifiers to explore the link between their political liking behaviour and actual voting intention. Through this work, we show how even a single well-chosen Facebook like, can reveal as much about our political voter intention as hundreds of random likes. Further, by including the entire political like history of the respondents, our model reaches prediction accuracies above previous multiparty studies (60-70%). We conclude the paper by discussing how a parsimonious data strategy applied, with some limitations, allow us to generalize our findings to the 1,4 million Danes with at least one political like and even to other political multiparty systems. | Parsimonious Data: How a single Facebook like predicts voting behaviour
in multiparty systems | 9,356 |
Recent research has shown a substantial active presence of bots in online social networks (OSNs). In this paper we utilise our past work on studying bots (Stweeler) to comparatively analyse the usage and impact of bots and humans on Twitter, one of the largest OSNs in the world. We collect a large-scale Twitter dataset and define various metrics based on tweet metadata. We divide and filter the dataset in four popularity groups in terms of number of followers. Using a human annotation task we assign 'bot' and 'human' ground-truth labels to the dataset, and compare the annotations against an online bot detection tool for evaluation. We then ask a series of questions to discern important behavioural bot and human characteristics using metrics within and among four popularity groups. From the comparative analysis we draw important differences as well as surprising similarities between the two entities, thus paving the way for reliable classification of automated political infiltration, advertisement campaigns, and general bot detection. | An in-depth characterisation of Bots and Humans on Twitter | 9,357 |
Numerus works that study the spread of information in social networks include a spreading mechanism in which a set of nodes is initially infected (i.e. seeded), followed by a viral process, which spontaneously spread the message through the nodes of the network. These models are used to describe the spread of rumors as well as the spread of new products and services. In reality however, it is quite rare that a product or service spreads through a social networks solely by viral forces. It is more common, that the spreader would invests a continuous active effort of its sales representatives in order to enhance the spread. The Latent Viral Marketing Model is a spreading model that fits this reality. Along the description of the model, the paper continues and proposes a simple Scheduling of Seeding heuristics, which recommends the node to seed at each period. A large study of empirical simulations shows that under a wide range of realistic initial conditions, the Scheduling Seeding method improves a product adoption rate by 29%-94% in comparison to existing state-of-the-art seeding methods. | Latent Viral Marketing, Concepts and Control Methods | 9,358 |
We propose a framework to measure, evaluate, and rank campaign effectiveness in the ongoing 2016 U.S. presidential election. Using Twitter data collected from Sept. 2015 to Jan. 2016, we first uncover the tweeting tactics of the candidates and second, using negative binomial regression and exploiting the variations in 'likes,' we evaluate the effectiveness of these tactics. Thirdly, we rank the candidates' campaign tactics by calculating the conditional expectation of their generated 'likes.' We show that while Ted Cruz and Marco Rubio put much weight on President Obama, this tactic is not being well received by their supporters. We demonstrate that Hillary Clinton's tactic of linking herself to President Obama resonates well with her supporters but the same is not true for Bernie Sanders. In addition, we show that Donald Trump is a major topic for all the other candidates and that the women issue is equally emphasized in Sanders' campaign as in Clinton's. Finally, we suggest two ways that politicians can use the feedback mechanism in social media to improve their campaign: (1) use feedback from social media to improve campaign tactics within social media; (2) prototype policies and test the public response from the social media. | Tactics and Tallies: A Study of the 2016 U.S. Presidential Campaign
Using Twitter 'Likes' | 9,359 |
The complex topology of real networks allows its actors to change their functional behavior. Network models provide better understanding of the evolutionary mechanisms being accountable for the growth of such networks by capturing the dynamics in the ways network agents interact and change their behavior. Considerable amount of research efforts is required for developing novel network modeling techniques to understand the structural properties such networks, reproducing similar properties based on empirical evidence, and designing such networks efficiently. First, we demonstrate how to construct social interaction networks using social media data and then present the key findings obtained from the network analytics. We analyze the characteristics and growth of such interaction networks, examine the network properties and derive important insights based on the theories of network science literature. We also discuss the application of such networks as a useful tool to effectively disseminate targeted information during planned special events. We observed that the degree-distributions of such networks follow power-law that is indicative of the existence of fewer nodes in the network with higher levels of interactions, and many other nodes with less interactions. While the network elements and average user degree grow linearly each day, densities of such networks tend to become zero. Largest connected components exhibit higher connectivity (density) when compared with the whole graph. Network radius and diameter become stable over time evidencing the small-world property. We also observe increased transitivity and higher stability of the power-law exponents as the networks grow. Data is specific to the Purdue University community and two large events, namely Purdue Day of Giving and Senator Bernie Sanders' visit to Purdue University as part of Indiana Primary Election 2016. | Analyzing Social Interaction Networks from Twitter for Planned Special
Events | 9,360 |
The minimum conductance problem is an NP-hard graph partitioning problem. Apart from the search for bottlenecks in complex networks, the problem is very closely related to the popular area of network community detection. In this paper, we tackle the minimum conductance problem as a pseudo-Boolean optimisation problem and propose a memetic algorithm to solve it. An efficient local search strategy is established. Our memetic algorithm starts by using this local search strategy with different random strings to sample a set of diverse initial solutions. This is followed by an evolutionary phase based on a steady-state framework and two intensification subroutines. We compare the algorithm to a wide range of multi-start local search approaches and classical genetic algorithms with different crossover operators. The experimental results are presented for a diverse set of real-world networks. These results indicate that the memetic algorithm outperforms the alternative stochastic approaches. | A Memetic Algorithm for the Minimum Conductance Graph Partitioning
Problem | 9,361 |
Recent researches have seen an upsurge in the analysis of consumer reviews. Although, several dimensions have been explored, less is known on the temporal dynamics of events that happen over the lifecycle of online products. What are the dominant sales patterns? How are they affected by review count, rating, helpfulness and sentiment? How is trust characterized and what are its effects on sales and revenue? What happens during a market competition? When does a takeover/recovery happen and by what percentage do sales increase on a takeover? This work aims to answer these fundamental research questions based on a sales time-series analysis of reviews of over 1 million products from Amazon.com. We discover novel temporal patterns of sales and interesting correlations of sales with the ratings. We find that trust and helpfulness are important for higher revenue. Based on the analyses, we propose a model to forecast sales that significantly outperforms other baselines. We then explore the phenomena of market competition. Particularly, we characterize different factors that govern survival/death of a product under competition and a model for competition forecast. Experimental results on large-scale reviews demonstrate the effectiveness of the proposed approaches. | Characterizing Product Lifecycle in Online Marketing: Sales, Trust,
Revenue, and Competition Modeling | 9,362 |
In this paper, we present computational models to predict Twitter users' attitude towards a specific brand through their personal and social characteristics. We also predict their likelihood to take different actions based on their attitudes. In order to operationalize our research on users' attitude and actions, we collected ground-truth data through surveys of Twitter users. We have conducted experiments using two real world datasets to validate the effectiveness of our attitude and action prediction framework. Finally, we show how our models can be integrated with a visual analytics system for customer intervention. | Computational Models for Attitude and Actions Prediction | 9,363 |
Spatially-situated opinions that can be held with different degrees of conviction lead to spatio-temporal patterns such as clustering (homophily), polarization, and deadlock. Our goal is to understand how sensitive these patterns are to changes in the local nature of interactions. We introduce two different mixing mechanisms, spatial relocation and non-local interaction ("telephoning"), to an earlier fully spatial model (no mixing). Interestingly, the mechanisms that create deadlock in the fully spatial model have the opposite effect when there is a sufficient amount of mixing. With telephoning, not only is polarization and deadlock broken up, but consensus is hastened. The effects of mixing by relocation are even more pronounced. Further insight into these dynamics is obtained for selected parameter regimes via comparison to the mean-field differential equations. | Spatial Opinion Dynamics and the Effects of Two Types of Mixing | 9,364 |
In many Twitter studies, it is important to know where a tweet came from in order to use the tweet content to study regional user behavior. However, researchers using Twitter to understand user behavior often lack sufficient geo-tagged data. Given the huge volume of Twitter data there is a need for accurate automated geolocating solutions. Herein, we present a new method to predict a Twitter user's location based on the information in a single tweet. We integrate text and user profile meta-data into a single model using a convolutional neural network. Our experiments demonstrate that our neural model substantially outperforms baseline methods, achieving 52.8% accuracy and 92.1% accuracy on city-level and country-level prediction respectively. | On Predicting Geolocation of Tweets using Convolutional Neural Networks | 9,365 |
In many systems privacy of users depends on the number of participants applying collectively some method to protect their security. Indeed, there are numerous already classic results about revealing aggregated data from a set of users. The conclusion is usually as follows: if you have enough friends to "aggregate" the private data, you can safely reveal your private information. Apart from data aggregation, it has been noticed that in a wider context privacy can be often reduced to being hidden in a crowd. Generally, the problems is how to create such crowd. This task may be not easy in some distributed systems, wherein gathering enough "individuals" is hard for practical reasons. Such example are social networks (or similar systems), where users have only a limited number of semi trusted contacts and their aim is to reveal some aggregated data in a privacy preserving manner. This may be particularly problematic in the presence of a strong adversary that can additionally corrupt some users. We show two methods that allow to significantly amplify privacy with only limited number of local operations and very moderate communication overhead. Except theoretical analysis we show experimental results on topologies of real-life social networks to demonstrate that our methods can significantly amplify privacy of chosen aggregation protocols even facing a massive attack of a powerful adversary. We believe however that our results can have much wider applications for improving security of systems based on locally trusted relations. | How to Cooperate Locally to Improve Global Privacy in Social Networks?
On Amplification of Privacy Preserving Data Aggregation | 9,366 |
Many network analysis tasks in social sciences rely on pre-existing data sources that were created with explicit relations or interactions between entities under consideration. Examples include email logs, friends and followers networks on social media, communication networks, etc. In these data, it is relatively easy to identify who is connected to whom and how they are connected. However, most of the data that we encounter on a daily basis are unstructured free-text data, e.g., forums, online marketplaces, etc. It is considerably more difficult to extract network data from unstructured text. In this work, we present an end-to-end system for analyzing unstructured text data and transforming the data into structured graphs that are directly applicable to a downstream application. Specifically, we look at social media data and attempt to predict the most indicative words from users' posts. The resulting keywords can be used to construct a context+content network for downstream processing such as graph-based analysis and learning. With that goal in mind, we apply our methods to the application of cross-domain entity resolution. The performance of the resulting system with automatic keywords shows improvement over the system with user-annotated hashtags. | Making Sense of Unstructured Text Data | 9,367 |
In this paper, we present a novel approach based on the random walk process for finding meaningful representations of a graph model. Our approach leverages the transient behavior of many short random walks with novel initialization mechanisms to generate model discriminative features. These features are able to capture a more comprehensive structural signature of the underlying graph model. The resulting representation is invariant to both node permutation and the size of the graph, allowing direct comparison between large classes of graphs. We test our approach on two challenging model selection problems: the discrimination in the sparse regime of an Erd\"{o}s-Renyi model from a stochastic block model and the planted clique problem. Our representation approach achieves performance that closely matches known theoretical limits in addition to being computationally simple and scalable to large graphs. | Graph Model Selection via Random Walks | 9,368 |
Matrix Factorization (MF) is a very popular method for recommendation systems. It assumes that the underneath rating matrix is low-rank. However, this assumption can be too restrictive to capture complex relationships and interactions among users and items. Recently, Local LOw-Rank Matrix Approximation (LLORMA) has been shown to be very successful in addressing this issue. It just assumes the rating matrix is composed of a number of low-rank submatrices constructed from subsets of similar users and items. Although LLORMA outperforms MF, how to construct such submatrices remains a big problem. Motivated by the availability of rich social connections in today's recommendation systems, we propose a novel framework, i.e., Social LOcal low-rank Matrix Approximation (SLOMA), to address this problem. To the best of our knowledge, SLOMA is the first work to incorporate social connections into the local low-rank framework. Furthermore, we enhance SLOMA by applying social regularization to submatrices factorization, denoted as SLOMA++. Therefore, the proposed model can benefit from both social recommendation and the local low-rank assumption. Experimental results from two real-world datasets, Yelp and Douban, demonstrate the superiority of the proposed models over LLORMA and MF. | Collaborative Filtering with Social Local Models | 9,369 |
Sleep condition is closely related to an individual's health. Poor sleep conditions such as sleep disorder and sleep deprivation affect one's daily performance, and may also cause many chronic diseases. Many efforts have been devoted to monitoring people's sleep conditions. However, traditional methodologies require sophisticated equipment and consume a significant amount of time. In this paper, we attempt to develop a novel way to predict individual's sleep condition via scrutinizing facial cues as doctors would. Rather than measuring the sleep condition directly, we measure the sleep-deprived fatigue which indirectly reflects the sleep condition. Our method can predict a sleep-deprived fatigue rate based on a selfie provided by a subject. This rate is used to indicate the sleep condition. To gain deeper insights of human sleep conditions, we collected around 100,000 faces from selfies posted on Twitter and Instagram, and identified their age, gender, and race using automatic algorithms. Next, we investigated the sleep condition distributions with respect to age, gender, and race. Our study suggests among the age groups, fatigue percentage of the 0-20 youth and adolescent group is the highest, implying that poor sleep condition is more prevalent in this age group. For gender, the fatigue percentage of females is higher than that of males, implying that more females are suffering from sleep issues than males. Among ethnic groups, the fatigue percentage in Caucasian is the highest followed by Asian and African American. | Large-Scale Sleep Condition Analysis Using Selfies from Social Media | 9,370 |
This paper examines the problem of adaptive influence maximization in social networks. As adaptive decision making is a time-critical task, a realistic feedback model has been considered, called myopic. In this direction, we propose the myopic adaptive greedy policy that is guaranteed to provide a (1 - 1/e)-approximation of the optimal policy under a variant of the independent cascade diffusion model. This strategy maximizes an alternative utility function that has been proven to be adaptive monotone and adaptive submodular. The proposed utility function considers the cumulative number of active nodes through the time, instead of the total number of the active nodes at the end of the diffusion. Our empirical analysis on real-world social networks reveals the benefits of the proposed myopic strategy, validating our theoretical results. | Adaptive Submodular Influence Maximization with Myopic Feedback | 9,371 |
This paper addresses the discount pricing in word-of-mouth (WOM) marketing. A new discount strategy known as the Infection-Based Discount (IBD) strategy is proposed. The basic idea of the IBD strategy lies in that each customer enjoys a discount that is linearly proportional to his/her influence in the WOM network. To evaluate the performance of the IBD strategy, the WOM spreading process is modeled as a dynamic model known as the DPA model, and the performance of the IBD strategy is modeled as a function of the basic discount. Next, the influence of different factors, including the basic discount and the WOM network, on the dynamics of the DPA model is revealed experimentally. Finally, the influence of different factors on the performance of the IBD strategy is uncovered experimentally. On this basis, some promotional measures are recommended. | A discount strategy in word-of-mouth marketing and its assessment | 9,372 |
As compared to the traditional advertising, the word-of-mouth (WOM) communications have striking advantages such as significantly lower cost and rapid delivery; this is especially the case with the popularity of online social networks. This paper addresses the issue of maximizing the overall profit of a WOM marketing campaign. A marketing process with both positive and negative WOM is modeled as a dynamical model knwn as the SIPNS model, and the profit maximization problem is modeled as a constrained optimization problem. The influence of different factors on the dynamics of the SIPNS model is revealed experimentally. Also, the impact of different factors on the expected overall profit of a WOM marketing campaign is uncovered experimentally. On this basis, some promotion strategies are suggested. To our knowledge, this is the first time a WOM marketing campaign is treated this way. | Maximizing the overall profit of a word-of-mouth marketing campaign: A
modeling study | 9,373 |
This paper treats the privacy-preserving publication of social graphs in the presence of active adversaries, that is, adversaries with the ability to introduce sybil nodes in the graph prior to publication and leverage them to create unique fingerprints for a set of victim nodes and re-identify them after publication. Stemming from the notion of $(k,\ell)$-anonymity, we introduce $(k,\ell)$-anonymous transformations, characterising graph perturbation methods that ensure protection from active adversaries levaraging up to $\ell$ sybil nodes. Additionally, we introduce a new privacy property: $(k,\ell)$-adjacency anonymity, which relaxes the assumption made by $(k,\ell)$-anonymity that adversaries can control all distances between sybil nodes and the rest of the nodes in the graph. The new privacy property is in turn the basis for a new type of graph perturbation: $(k,\ell)$-adjacency anonymous transformations. We propose algorithms for obtaining $(k,1)$-adjacency anonymous transformations for arbitrary values of $k$, as well as $(2,\ell)$-adjacency anonymous transformations for small values of $\ell$. | Rethinking $(k,\ell)$-anonymity in social graphs: $(k,\ell)$-adjacency
anonymity and $(k,\ell)$-(adjacency) anonymous transformations | 9,374 |
In recent years, there have been efforts to collect human contact traces during social events (e.g., conferences) using Bluetooth devices (e.g., mobile phones, iMotes). The results of these studies have enabled the ability to do the crowd-sourcing task from within the crowd, in order to answer questions, such as: what is the current density of the crowd, or how many people are attending the event? However, in those studies, the sensing devices are usually distributed and configured in a certain manner. For example, the number of devices is fixed, people register for the devices on a volunteering basis. In this paper, we treat the above problem as an optimization problem and draw the connection to the vertex cover problem in graph theory. Since finding the optimal solution for minimum vertex cover problem is NP-complete, approximation algorithms have to be used. However, we will show that the well-known approximation algorithms do not perform well with the crowd-sensing task. In this paper, we propose the notions of node observability and coverage utility score and design a new context-aware approximation algorithm to find vertex cover that is tailored for crowd-sensing task. In addition, we design human-centric bootstrapping strategies to make initial assignment of sensing devices based on meta information about the participants (e.g., interests, friendship). The motivation is to assign the sensing task to a more "socialized" device to obtain better sensing coverage. We perform comprehensive experiments on real-world data traces obtained from previous experimental studies in conference and academic social context. The results show that our proposed approach significantly outperforms the baseline approximation algorithms in terms of sensing coverage. | Crowdsensing in Opportunistic Mobile Social Networks: A Context-aware
and Human-centric Approach | 9,375 |
Researchers have attempted to model information diffusion and topic trends and lifecycle on online social networks. They have investigated the role of content, social connections and communities, familiarity and behavioral similarity in this context. The current article presents a survey of representative models that perform topic analysis, capture information diffusion, and explore the properties of social connections in the context of online social networks. The article concludes with a set of outlines of open problems and possible directions of future research interest. This article is intended for researchers to identify the current literature, and explore possibilities to improve the art. | Literature Survey on Interplay of Topics, Information Diffusion and
Connections on Social Networks | 9,376 |
The fast growth of social networks and their privacy requirements in recent years, has lead to increasing difficulty in obtaining complete topology of these networks. However, diffusion information over these networks is available and many algorithms have been proposed to infer the underlying networks by using this information. The previously proposed algorithms only focus on inferring more links and do not pay attention to the important characteristics of the underlying social networks In this paper, we propose a novel algorithm, called DANI, to infer the underlying network structure while preserving its properties by using the diffusion information. Moreover, the running time of the proposed method is considerably lower than the previous methods. We applied the proposed method to both real and synthetic networks. The experimental results showed that DANI has higher accuracy and lower run time compared to well-known network inference methods. | DANI: A Fast Diffusion Aware Network Inference Algorithm | 9,377 |
Personalized PageRank (PPR) is a measure of the importance of a node from the perspective of another (we call these nodes the $\textit{target}$ and the $\textit{source}$, respectively). PPR has been used in many applications, such as offering a Twitter user (the source) recommendations of who to follow (targets deemed important by PPR); additionally, PPR has been used in graph-theoretic problems such as community detection. However, computing PPR is infeasible for large networks like Twitter, so efficient estimation algorithms are necessary. In this work, we analyze the relationship between PPR estimation complexity and clustering. First, we devise algorithms to estimate PPR for many source/target pairs. In particular, we propose an enhanced version of the existing single pair estimator $\texttt{Bidirectional-PPR}$ that is more useful as a primitive for many pair estimation. We then show that the common underlying graph can be leveraged to efficiently and jointly estimate PPR for many pairs, rather than treating each pair separately using the primitive algorithm. Next, we show the complexity of our joint estimation scheme relates closely to the degree of clustering among the sources and targets at hand, indicating that estimating PPR for many pairs is easier when clustering occurs. Finally, we consider estimating PPR when several machines are available for parallel computation, devising a method that leverages our clustering findings, specifically the quantities computed $\textit{in situ}$, to assign tasks to machines in a manner that reduces computation time. This demonstrates that the relationship between complexity and clustering has important consequences in a practical distributed setting. | On the role of clustering in Personalized PageRank estimation | 9,378 |
Most real world dynamic networks are evolved very fast with time. It is not feasible to collect the entire network at any given time to study its characteristics. This creates the need to propose local algorithms to study various properties of the network. In the present work, we estimate degree rank of a node without having the entire network. The proposed methods are based on the power law degree distribution characteristic or sampling techniques. The proposed methods are simulated on synthetic networks, as well as on real world social networks. The efficiency of the proposed methods is evaluated using absolute and weighted error functions. Results show that the degree rank of a node can be estimated with high accuracy using only $1\%$ samples of the network size. The accuracy of the estimation decreases from high ranked to low ranked nodes. We further extend the proposed methods for random networks and validate their efficiency on synthetic random networks, that are generated using Erd\H{o}s-R\'{e}nyi model. Results show that the proposed methods can be efficiently used for random networks as well. | Degree Ranking Using Local Information | 9,379 |
This work is produced by researchers at the Neuropolitics Research Lab, School of Social and Political Science and the School of Informatics at the University of Edinburgh. In this report we provide an analysis of the social media posts on the British general election 2017 over the month running up to the vote. We find that pro-Labour sentiment dominates the Twitter conversation around GE2017 and that there is also a disproportionate presence of the Scottish National Party (SNP), given the UK-wide nature of a Westminster election. Substantive issues have featured much less prominently and in a less sustained manner in the Twitter debate than pro and anti leader and political party posts. However, the issue of Brexit has provided a consistent backdrop to the GE2017 conversation and has rarely dropped out of the top three most popular hashtags in the last month. Brexit has been the issue of the GE2017 campaign, eclipsing even the NHS. We found the conversation in the GE2017 Twitter debate to be heavily influenced both by external events and by the top-down introduction of hashtags by broadcast media outlets, often associated with specific programmes and the mediatised political debates. Hashtags like these have a significant impact on the shape of the data collected from Twitter and might distort studies with short data-collection windows but are usually short-lived with little long term impact on the Twitter conversation. If the current polling is to be believed Jeremy Corbyn is unlikely to do as badly as was anticipated when the election was first called. Traditional media sources were slow to pick up on this change in public opinion whereas this trend could be seen early on in social media and throughout the month of May. | UK General Election 2017: a Twitter Analysis | 9,380 |
Recent studies suggest that human emotions diffuse in not only real-world communities but also online social media. More and more mechanisms beyond emotion contagion are revealed, including emotion correlations which indicate their influence and the coupling of emotion diffusion and network structure such as tie strength. Besides, different emotions might even compete in shaping the public opinion. However, a comprehensive model that considers up-to-date findings to replicate the patterns of emotion contagion in online social media is still missing. In this paper, to bridge this vital gap, we propose an agent-based emotion contagion model which combines features of emotion influence and tie strength preference in the dissemination process. The simulation results indicate that anger-dominated users have higher vitality than joy-dominated ones, and anger prefers weaker ties than joy in diffusion, which could make it easier to spread between online groups. Moreover, anger's high influence makes it competitive and easily to dominate the community, especially when negative public events occur. It is also surprisingly revealed that as the ratio of anger approaches joy with a gap less than 10%, angry tweets and users will eventually dominate the online social media and arrives the collective outrage in the cyber space. The critical gap disclosed here can be indeed warning signals at early stages for outrage controlling in online social media. All the parameters of the presented model can be easily estimated from the empirical observations and their values from historical data could help reproduce the emotion contagion of different circumstances. Our model would shed lights on the study of multiple issues like forecasting of emotion contagion in terms of computer simulations. | An agent-based model for emotion contagion and competition in online
social media | 9,381 |
Many people use social media to seek information during disasters while lacking access to traditional information sources. In this study, we analyze Twitter data to understand information spreading activities of social media users during hurricane Sandy. We create multiple subgraphs of Twitter users based on activity levels and analyze network properties of the subgraphs. We observe that user information sharing activity follows a power-law distribution suggesting the existence of few highly active nodes in disseminating information and many other nodes being less active. We also observe close enough connected components and isolates at all levels of activity, and networks become less transitive, but more assortative for larger subgraphs. We also analyze the association between user activities and characteristics that may influence user behavior to spread information during a crisis. Users become more active in spreading information if they are centrally placed in the network, less eccentric, and have higher degrees. Our analysis provides insights on how to exploit user characteristics and network properties to spread information or limit the spreading of misinformation during a crisis event. | Understanding Information Spreading in Social Media during Hurricane
Sandy: User Activity and Network Properties | 9,382 |
Nowadays as the world population has become more interconnected and is relying on faster transportation methods, simplified connections and shorter commuting times, we witness a rapid increase in human mobility. In this situation unveiling and understanding human mobility patterns have become a crucial issue to support decisions and prediction activities when managing the complexity of the today social organization. In practice, the mobility pattern of each individual person consists of the sequence of visited places. Those places and their correlations represent the foundation of most modelling and activity researches for understanding human mobility. Even though visited places underpin almost the majority of works in this field, their features remain largely unknown because in previous works, they have been mainly considered as uncharacterized spot points in an area or social gathering places, without considering the roles and importance of places to the behavior of each single user. | Human Mobility Mining through Head/Tail Breaks | 9,383 |
A signed network is a network with each link associated with a positive or negative sign. Models for nodes interacting over such signed networks, where two different types of interactions take place along the positive and negative links, respectively, arise from various biological, social, political, and economic systems. As modifications to the conventional DeGroot dynamics for positive links, two basic types of negative interactions along negative links, namely the opposing rule and the repelling rule, have been proposed and studied in the literature. This paper reviews a few fundamental convergence results for such dynamics over deterministic or random signed networks under a unified algebraic-graphical method. We show that a systematic tool of studying node state evolution over signed networks can be obtained utilizing generalized Perron-Frobenius theory, graph theory, and elementary algebraic recursions. | Dynamics over Signed Networks | 9,384 |
Traditional frameworks for dynamic graphs have relied on processing only the stream of edges added into or deleted from an evolving graph, but not any additional related information such as the degrees or neighbor lists of nodes incident to the edges. In this paper, we propose a new edge sampling framework for big-graph analytics in dynamic graphs which enhances the traditional model by enabling the use of additional related information. To demonstrate the advantages of this framework, we present a new sampling algorithm, called Edge Sample and Discard (ESD). It generates an unbiased estimate of the total number of triangles, which can be continuously updated in response to both edge additions and deletions. We provide a comparative analysis of the performance of ESD against two current state-of-the-art algorithms in terms of accuracy and complexity. The results of the experiments performed on real graphs show that, with the help of the neighborhood information of the sampled edges, the accuracy achieved by our algorithm is substantially better. We also characterize the impact of properties of the graph on the performance of our algorithm by testing on several Barabasi-Albert graphs. | On Counting Triangles through Edge Sampling in Large Dynamic Graphs | 9,385 |
Peatland fires and haze events are disasters with national, regional and international implications. The phenomena lead to direct damage to local assets, as well as broader economic and environmental losses. Satellite imagery is still the main and often the only available source of information for disaster management. In this article, we test the potential of social media to assist disaster management. To this end, we compare insights from two datasets: fire hotspots detected via NASA satellite imagery and almost all GPS-stamped tweets from Sumatra Island, Indonesia, posted during 2014. Sumatra Island is chosen as it regularly experiences a significant number of haze events, which affect citizens in Indonesia as well as in nearby countries including Malaysia and Singapore. We analyse temporal correlations between the datasets and their geo-spatial interdependence. Furthermore, we show how Twitter data reveals changes in users' behavior during severe haze events. Overall, we demonstrate that social media is a valuable source of complementary and supplementary information for haze disaster management. Based on our methodology and findings, an analytics tool to improve peatland fire and haze disaster management by the Indonesian authorities is under development. | Mining Social Media to Inform Peatland Fire and Haze Disaster Management | 9,386 |
What tweet features are associated with higher effectiveness in tweets? Through the mining of 122 million engagements of 2.5 million original tweets, we present a systematic review of tweet time, entities, composition, and user account features. We show that the relationship between various features and tweeting effectiveness is non-linear; for example, tweets that use a few hashtags have higher effectiveness than using no or too many hashtags. This research closely relates to various industrial applications that are based on tweet features, including the analysis of advertising campaigns, the prediction of user engagement, the extraction of signals for automated trading, etc. | Mining Features Associated with Effective Tweets | 9,387 |
Social bots are currently regarded an influential but also somewhat mysterious factor in public discourse and opinion making. They are considered to be capable of massively distributing propaganda in social and online media and their application is even suspected to be partly responsible for recent election results. Astonishingly, the term `Social Bot' is not well defined and different scientific disciplines use divergent definitions. This work starts with a balanced definition attempt, before providing an overview of how social bots actually work (taking the example of Twitter) and what their current technical limitations are. Despite recent research progress in Deep Learning and Big Data, there are many activities bots cannot handle well. We then discuss how bot capabilities can be extended and controlled by integrating humans into the process and reason that this is currently the most promising way to go in order to realize effective interactions with other humans. | Social Bots: Human-Like by Means of Human Control? | 9,388 |
The problem of maximizing information diffusion, given a certain budget expressed in terms of the number of seed nodes, is an important topic in social networks research. Existing literature focuses on single phase diffusion where all seed nodes are selected at the beginning of diffusion and all the selected nodes are activated simultaneously. This paper undertakes a detailed investigation of the effect of selecting and activating seed nodes in multiple phases. Specifically, we study diffusion in two phases assuming the well-studied independent cascade model. First, we formulate an objective function for two-phase diffusion, investigate its properties, and propose efficient algorithms for finding seed nodes in the two phases. Next, we study two associated problems: (1) budget splitting which seeks to optimally split the total budget between the two phases and (2) scheduling which seeks to determine an optimal delay after which to commence the second phase. Our main conclusions include: (a) under strict temporal constraints, use single phase diffusion, (b) under moderate temporal constraints, use two-phase diffusion with a short delay while allocating most of the budget to the first phase, and (c) when there are no temporal constraints, use two-phase diffusion with a long delay while allocating roughly one-third of the budget to the first phase. | Information Diffusion in Social Networks in Two Phases | 9,389 |
We present HARP, a novel method for learning low dimensional embeddings of a graph's nodes which preserves higher-order structural features. Our proposed method achieves this by compressing the input graph prior to embedding it, effectively avoiding troublesome embedding configurations (i.e. local minima) which can pose problems to non-convex optimization. HARP works by finding a smaller graph which approximates the global structure of its input. This simplified graph is used to learn a set of initial representations, which serve as good initializations for learning representations in the original, detailed graph. We inductively extend this idea, by decomposing a graph in a series of levels, and then embed the hierarchy of graphs from the coarsest one to the original graph. HARP is a general meta-strategy to improve all of the state-of-the-art neural algorithms for embedding graphs, including DeepWalk, LINE, and Node2vec. Indeed, we demonstrate that applying HARP's hierarchical paradigm yields improved implementations for all three of these methods, as evaluated on both classification tasks on real-world graphs such as DBLP, BlogCatalog, CiteSeer, and Arxiv, where we achieve a performance gain over the original implementations by up to 14% Macro F1. | HARP: Hierarchical Representation Learning for Networks | 9,390 |
Sociotechnological and geospatial processes exhibit time varying structure that make insight discovery challenging. This paper proposes a new statistical model for such systems, modeled as dynamic networks, to address this challenge. It assumes that vertices fall into one of k types and that the probability of edge formation at a particular time depends on the types of the incident nodes and the current time. The time dependencies are driven by unique seasonal processes, which many systems exhibit (e.g., predictable spikes in geospatial or web traffic each day). The paper defines the model as a generative process and an inference procedure to recover the seasonal processes from data when they are unknown. Evaluation with synthetic dynamic networks show the recovery of the latent seasonal processes that drive its formation. | Seasonality in Dynamic Stochastic Block Models | 9,391 |
Social media is considered a democratic space in which people connect and interact with each other regardless of their gender, race, or any other demographic factor. Despite numerous efforts that explore demographic factors in social media, it is still unclear whether social media perpetuates old inequalities from the offline world. In this paper, we attempt to identify gender and race of Twitter users located in U.S. using advanced image processing algorithms from Face++. Then, we investigate how different demographic groups (i.e. male/female, Asian/Black/White) connect with other. We quantify to what extent one group follow and interact with each other and the extent to which these connections and interactions reflect in inequalities in Twitter. Our analysis shows that users identified as White and male tend to attain higher positions in Twitter, in terms of the number of followers and number of times in user's lists. We hope our effort can stimulate the development of new theories of demographic information in the online space. | White, Man, and Highly Followed: Gender and Race Inequalities in Twitter | 9,392 |
Social network analysis is a prominent approach to investigate interpersonal relationships. Most studies use self-report data to quantify the connections between participants and construct social networks. In recent years smartphones have been used as an alternative to map networks by assessing the proximity between participants based on Bluetooth and GPS data. While most studies have handed out specially programmed smartphones to study participants, we developed an application for iOS and Android to collect Bluetooth data from participants own smartphones. In this study, we compared the networks estimated with the smartphone app to those obtained from sociometric badges and self-report data. Participants (n=21) installed the app on their phone and wore a sociometric badge during office hours. Proximity data was collected for 4 weeks. A contingency table revealed a significant association between proximity data (rho = 0.17, p<0.0001), but the marginal odds were higher for the app (8.6%) than for the badges (1.3%), indicating that dyads were more often detected by the app. We then compared the networks that were estimated using the proximity and self-report data. All three networks were significantly correlated, although the correlation with self-reported data was lower for the app (rho = 0.25) than for badges (rho = 0.67). The scanning rates of the app varied considerably between devices and was lower on iOS than on Android. The association between the app and the badges increased when the network was estimated between participants whose app recorded more regularly. These findings suggest that the accuracy of proximity networks can be further improved by reducing missing data and restricting the interpersonal distance at which interactions are detected. | Validation of a smartphone app to map social networks of proximity | 9,393 |
We study the problem of inferring network topology from information cascades, in which the amount of time taken for information to diffuse across an edge in the network follows an unknown distribution. Unlike previous studies, which assume knowledge of these distributions, we only require that diffusion along different edges in the network be independent together with limited moment information (e.g., the means). We introduce the concept of a separating vertex set for a graph, which is a set of vertices in which for any two given distinct vertices of the graph, there exists a vertex whose distance to them are different. We show that a necessary condition for reconstructing a tree perfectly using distance information between pairs of vertices is given by the size of an observed separating vertex set. We then propose an algorithm to recover the tree structure using infection times, whose differences have means corresponding to the distance between two vertices. To improve the accuracy of our algorithm, we propose the concept of redundant vertices, which allows us to perform averaging to better estimate the distance between two vertices. Though the theory is developed mainly for tree networks, we demonstrate how the algorithm can be extended heuristically to general graphs. Simulations using synthetic and real networks, and experiments using real-world data suggest that our proposed algorithm performs better than some current state-of-the-art network reconstruction methods. | Network Topology Inference Using Information Cascades with Limited
Statistical Knowledge | 9,394 |
Discovering the underlying structures present in large real world graphs is a fundamental scientific problem. Recent work at the intersection of formal language theory and graph theory has found that a Hyperedge Replacement Grammar (HRG) can be extracted from a tree decomposition of any graph. This HRG can be used to generate new graphs that share properties that are similar to the original graph. Because the extracted HRG is directly dependent on the shape and contents of the of tree decomposition, it is unlikely that informative graph-processes are actually being captured with the extraction algorithm. To address this problem, the current work presents a new extraction algorithm called temporal HRG (tHRG) that learns HRG production rules from a temporal tree decomposition of the graph. We observe problems with the assumptions that are made in a temporal HRG model. In experiments on large real world networks, we show and provide reasoning as to why tHRG does not perform as well as HRG and other graph generators. | A Temporal Tree Decomposition for Generating Temporal Graphs | 9,395 |
Social media are pervaded by unsubstantiated or untruthful rumors, that contribute to the alarming phenomenon of misinformation. The widespread presence of a heterogeneous mass of information sources may affect the mechanisms behind the formation of public opinion. Such a scenario is a florid environment for digital wildfires when combined with functional illiteracy, information overload, and confirmation bias. In this essay, we focus on a collection of works aiming at providing quantitative evidence about the cognitive determinants behind misinformation and rumor spreading. We account for users' behavior with respect to two distinct narratives: a) conspiracy and b) scientific information sources. In particular, we analyze Facebook data on a time span of five years in both the Italian and the US context, and measure users' response to i) information consistent with one's narrative, ii) troll contents, and iii) dissenting information e.g., debunking attempts. Our findings suggest that users tend to a) join polarized communities sharing a common narrative (echo chambers), b) acquire information confirming their beliefs (confirmation bias) even if containing false claims, and c) ignore dissenting information. | Misinformation spreading on Facebook | 9,396 |
Recent advancements in complex network analysis are encouraging and may provide useful insights when applied in software engineering domain, revealing properties and structures that cannot be captured by traditional metrics. In this paper, we analyzed the topological properties of Hibernate library, a well-known Java-based software through the extraction of its static call graph. The results reveal a complex network with small-world and scale-free characteristics while displaying a strong propensity on forming communities. | Complex Networks Analysis for Software Architecture: an Hibernate Call
Graph Study | 9,397 |
Many real-world networks such as social networks consist of strategic agents. The topology of these networks often plays a crucial role in determining the ease and speed with which certain information driven tasks can be accomplished. Consequently, growing a stable network having a certain desired topology is of interest. Motivated by this, we study the following important problem: given a certain desired topology, under what conditions would best response link alteration strategies adopted by strategic agents, uniquely lead to formation of a stable network having the given topology. This problem is the inverse of the classical network formation problem where we are concerned with determining stable topologies, given the conditions on the network parameters. We study this interesting inverse problem by proposing (1) a recursive model of network formation and (2) a utility model that captures key determinants of network formation. Building upon these models, we explore relevant topologies such as star graph, complete graph, bipartite Turan graph, and multiple stars with interconnected centers. We derive a set of sufficient conditions under which these topologies uniquely emerge, study their social welfare properties, and investigate the effects of deviating from the derived conditions. | Formation of Stable Strategic Networks with Desired Topologies | 9,398 |
In this paper, we present quantitative and qualitative analysis of the top retweeted tweets (viral tweets) pertaining to the US presidential elections from September 1, 2016 to Election Day on November 8, 2016. For everyday, we tagged the top 50 most retweeted tweets as supporting or attacking either candidate or as neutral/irrelevant. Then we analyzed the tweets in each class for: general trends and statistics; the most frequently used hashtags, terms, and locations; the most retweeted accounts and tweets; and the most shared news and links. In all we analyzed the 3,450 most viral tweets that grabbed the most attention during the US election and were retweeted in total 26.3 million times accounting over 40% of the total tweet volume pertaining to the US election in the aforementioned period. Our analysis of the tweets highlights some of the differences between the social media strategies of both candidates, the penetration of their messages, and the potential effect of attacks on both | Trump vs. Hillary: What went Viral during the 2016 US Presidential
Election | 9,399 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.