Alert button
Picture for Rossano Schifanella

Rossano Schifanella

Alert button

Modeling Teams Performance Using Deep Representational Learning on Graphs

Jun 29, 2022
Francesco Carli, Pietro Foini, Nicolò Gozzi, Nicola Perra, Rossano Schifanella

Figure 1 for Modeling Teams Performance Using Deep Representational Learning on Graphs
Figure 2 for Modeling Teams Performance Using Deep Representational Learning on Graphs
Figure 3 for Modeling Teams Performance Using Deep Representational Learning on Graphs
Figure 4 for Modeling Teams Performance Using Deep Representational Learning on Graphs

The large majority of human activities require collaborations within and across formal or informal teams. Our understanding of how the collaborative efforts spent by teams relate to their performance is still a matter of debate. Teamwork results in a highly interconnected ecosystem of potentially overlapping components where tasks are performed in interaction with team members and across other teams. To tackle this problem, we propose a graph neural network model designed to predict a team's performance while identifying the drivers that determine such an outcome. In particular, the model is based on three architectural channels: topological, centrality, and contextual which capture different factors potentially shaping teams' success. We endow the model with two attention mechanisms to boost model performance and allow interpretability. A first mechanism allows pinpointing key members inside the team. A second mechanism allows us to quantify the contributions of the three driver effects in determining the outcome performance. We test model performance on a wide range of domains outperforming most of the classical and neural baselines considered. Moreover, we include synthetic datasets specifically designed to validate how the model disentangles the intended properties on which our model vastly outperforms baselines.

* This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible 
Viaarxiv icon

Mobile Recognition of Wikipedia Featured Sites using Deep Learning and Crowd-sourced Imagery

Nov 04, 2019
Jimin Tan, Anastasios Noulas, Diego Sáez, Rossano Schifanella

Figure 1 for Mobile Recognition of Wikipedia Featured Sites using Deep Learning and Crowd-sourced Imagery
Figure 2 for Mobile Recognition of Wikipedia Featured Sites using Deep Learning and Crowd-sourced Imagery
Figure 3 for Mobile Recognition of Wikipedia Featured Sites using Deep Learning and Crowd-sourced Imagery
Figure 4 for Mobile Recognition of Wikipedia Featured Sites using Deep Learning and Crowd-sourced Imagery

Rendering Wikipedia content through mobile and augmented reality mediums can enable new forms of interaction in urban-focused user communities facilitating learning, communication and knowledge exchange. With this objective in mind, in this work we develop a mobile application that allows for the recognition of notable sites featured on Wikipedia. The application is powered by a deep neural network that has been trained on crowd-sourced imagery describing sites of interest, such as buildings, statues, museums or other physical entities that are present and visually accessible in an urban environment. We describe an end-to-end pipeline that describes data collection, model training and evaluation of our application considering online and real world scenarios. We identify a number of challenges in the site recognition task which arise due to visual similarities amongst the classified sites as well as due to noise introduce by the surrounding built environment. We demonstrate how using mobile contextual information, such as user location, orientation and attention patterns can significantly alleviate such challenges. Moreover, we present an unsupervised learning technique to de-noise crowd-sourced imagery which improves classification performance further.

Viaarxiv icon

Notable Site Recognition on Mobile Devices using Deep Learning with Crowd-sourced Imagery

Oct 22, 2019
Jimin Tan, Anastasios Noulas, Diego Sáez, Rossano Schifanella

Figure 1 for Notable Site Recognition on Mobile Devices using Deep Learning with Crowd-sourced Imagery
Figure 2 for Notable Site Recognition on Mobile Devices using Deep Learning with Crowd-sourced Imagery
Figure 3 for Notable Site Recognition on Mobile Devices using Deep Learning with Crowd-sourced Imagery
Figure 4 for Notable Site Recognition on Mobile Devices using Deep Learning with Crowd-sourced Imagery

In this work we design a mobile system that is able to automatically recognise sites of interest and project relevant information to a user that navigates the city. First, we build a collection of notable sites to bootstrap our system using Wikipedia. We then exploit online services such as Google Images and Flickr to collect large collections of crowdsourced imagery describing those sites. These images are then used to train minimal deep learning architectures that can effectively be transmitted and deployed to mobile devices becoming accessible to users through a dedicated application. Conducting an evaluation performing a series of online and real world experiments we enlist a number of key challenges that make the successful deployment of site recognition system difficult and highlight the importance of incorporating mobile contextual information to facilitate the visual recognition task. Similarity in the feature maps of objects that undergo identification, the presence of noise in crowdsourced imagery and arbitrary user induced inputs are among the factors the impede correct classification for deep learning models. We show how curating the training data through the application of a class-specific image denoising method and the incorporation of information such as user location, orientation and attention patterns can allow for significant improvement in classification accuracy and the election of a system that can effectively be used to recognise sites in the wild two out of three times.

Viaarxiv icon

Multimodal Social Media Analysis for Gang Violence Prevention

Jul 23, 2018
Philipp Blandfort, Desmond Patton, William R. Frey, Svebor Karaman, Surabhi Bhargava, Fei-Tzin Lee, Siddharth Varia, Chris Kedzie, Michael B. Gaskell, Rossano Schifanella, Kathleen McKeown, Shih-Fu Chang

Figure 1 for Multimodal Social Media Analysis for Gang Violence Prevention
Figure 2 for Multimodal Social Media Analysis for Gang Violence Prevention
Figure 3 for Multimodal Social Media Analysis for Gang Violence Prevention
Figure 4 for Multimodal Social Media Analysis for Gang Violence Prevention

Gang violence is a severe issue in major cities across the U.S. and recent studies [Patton et al. 2017] have found evidence of social media communications that can be linked to such violence in communities with high rates of exposure to gang activity. In this paper we partnered computer scientists with social work researchers, who have domain expertise in gang violence, to analyze how public tweets with images posted by youth who mention gang associations on Twitter can be leveraged to automatically detect psychosocial factors and conditions that could potentially assist social workers and violence outreach workers in prevention and early intervention programs. To this end, we developed a rigorous methodology for collecting and annotating tweets. We gathered 1,851 tweets and accompanying annotations related to visual concepts and the psychosocial codes: aggression, loss, and substance use. These codes are relevant to social work interventions, as they represent possible pathways to violence on social media. We compare various methods for classifying tweets into these three classes, using only the text of the tweet, only the image of the tweet, or both modalities as input to the classifier. In particular, we analyze the usefulness of mid-level visual concepts and the role of different modalities for this tweet classification task. Our experiments show that individually, text information dominates classification performance of the loss class, while image information dominates the aggression and substance use classes. Our multimodal approach provides a very promising improvement (18% relative in mean average precision) over the best single modality approach. Finally, we also illustrate the complexity of understanding social media data and elaborate on open challenges.

Viaarxiv icon

Beautiful and damned. Combined effect of content quality and social ties on user engagement

Nov 01, 2017
Luca M. Aiello, Rossano Schifanella, Miriam Redi, Stacey Svetlichnaya, Frank Liu, Simon Osindero

Figure 1 for Beautiful and damned. Combined effect of content quality and social ties on user engagement
Figure 2 for Beautiful and damned. Combined effect of content quality and social ties on user engagement
Figure 3 for Beautiful and damned. Combined effect of content quality and social ties on user engagement
Figure 4 for Beautiful and damned. Combined effect of content quality and social ties on user engagement

User participation in online communities is driven by the intertwinement of the social network structure with the crowd-generated content that flows along its links. These aspects are rarely explored jointly and at scale. By looking at how users generate and access pictures of varying beauty on Flickr, we investigate how the production of quality impacts the dynamics of online social systems. We develop a deep learning computer vision model to score images according to their aesthetic value and we validate its output through crowdsourcing. By applying it to over 15B Flickr photos, we study for the first time how image beauty is distributed over a large-scale social system. Beautiful images are evenly distributed in the network, although only a small core of people get social recognition for them. To study the impact of exposure to quality on user engagement, we set up matching experiments aimed at detecting causality from observational data. Exposure to beauty is double-edged: following people who produce high-quality content increases one's probability of uploading better photos; however, an excessive imbalance between the quality generated by a user and the user's neighbors leads to a decline in engagement. Our analysis has practical implications for improving link recommender systems.

* 13 pages, 12 figures, final version published in IEEE Transactions on Knowledge and Data Engineering (Volume: PP, Issue: 99) 
Viaarxiv icon

Detecting Sarcasm in Multimodal Social Platforms

Aug 08, 2016
Rossano Schifanella, Paloma de Juan, Joel Tetreault, Liangliang Cao

Figure 1 for Detecting Sarcasm in Multimodal Social Platforms
Figure 2 for Detecting Sarcasm in Multimodal Social Platforms
Figure 3 for Detecting Sarcasm in Multimodal Social Platforms
Figure 4 for Detecting Sarcasm in Multimodal Social Platforms

Sarcasm is a peculiar form of sentiment expression, where the surface sentiment differs from the implied sentiment. The detection of sarcasm in social media platforms has been applied in the past mainly to textual utterances where lexical indicators (such as interjections and intensifiers), linguistic markers, and contextual information (such as user profiles, or past conversations) were used to detect the sarcastic tone. However, modern social media platforms allow to create multimodal messages where audiovisual content is integrated with the text, making the analysis of a mode in isolation partial. In our work, we first study the relationship between the textual and visual aspects in multimodal posts from three major social media platforms, i.e., Instagram, Tumblr and Twitter, and we run a crowdsourcing task to quantify the extent to which images are perceived as necessary by human annotators. Moreover, we propose two different computational frameworks to detect sarcasm that integrate the textual and visual modalities. The first approach exploits visual semantics trained on an external dataset, and concatenates the semantics features with state-of-the-art textual features. The second method adapts a visual neural network initialized with parameters trained on ImageNet to multimodal sarcastic posts. Results show the positive effect of combining modalities for the detection of sarcasm across platforms and methods.

* 10 pages, 3 figures, final version published in the Proceedings of ACM Multimedia 2016 
Viaarxiv icon

An Image is Worth More than a Thousand Favorites: Surfacing the Hidden Beauty of Flickr Pictures

May 15, 2015
Rossano Schifanella, Miriam Redi, Luca Aiello

Figure 1 for An Image is Worth More than a Thousand Favorites: Surfacing the Hidden Beauty of Flickr Pictures
Figure 2 for An Image is Worth More than a Thousand Favorites: Surfacing the Hidden Beauty of Flickr Pictures
Figure 3 for An Image is Worth More than a Thousand Favorites: Surfacing the Hidden Beauty of Flickr Pictures
Figure 4 for An Image is Worth More than a Thousand Favorites: Surfacing the Hidden Beauty of Flickr Pictures

The dynamics of attention in social media tend to obey power laws. Attention concentrates on a relatively small number of popular items and neglecting the vast majority of content produced by the crowd. Although popularity can be an indication of the perceived value of an item within its community, previous research has hinted to the fact that popularity is distinct from intrinsic quality. As a result, content with low visibility but high quality lurks in the tail of the popularity distribution. This phenomenon can be particularly evident in the case of photo-sharing communities, where valuable photographers who are not highly engaged in online social interactions contribute with high-quality pictures that remain unseen. We propose to use a computer vision method to surface beautiful pictures from the immense pool of near-zero-popularity items, and we test it on a large dataset of creative-commons photos on Flickr. By gathering a large crowdsourced ground truth of aesthetics scores for Flickr images, we show that our method retrieves photos whose median perceived beauty score is equal to the most popular ones, and whose average is lower by only 1.5%.

* ICWSM 2015 
Viaarxiv icon

6 Seconds of Sound and Vision: Creativity in Micro-Videos

Nov 14, 2014
Miriam Redi, Neil O Hare, Rossano Schifanella, Michele Trevisiol, Alejandro Jaimes

Figure 1 for 6 Seconds of Sound and Vision: Creativity in Micro-Videos
Figure 2 for 6 Seconds of Sound and Vision: Creativity in Micro-Videos
Figure 3 for 6 Seconds of Sound and Vision: Creativity in Micro-Videos
Figure 4 for 6 Seconds of Sound and Vision: Creativity in Micro-Videos

The notion of creativity, as opposed to related concepts such as beauty or interestingness, has not been studied from the perspective of automatic analysis of multimedia content. Meanwhile, short online videos shared on social media platforms, or micro-videos, have arisen as a new medium for creative expression. In this paper we study creative micro-videos in an effort to understand the features that make a video creative, and to address the problem of automatic detection of creative content. Defining creative videos as those that are novel and have aesthetic value, we conduct a crowdsourcing experiment to create a dataset of over 3,800 micro-videos labelled as creative and non-creative. We propose a set of computational features that we map to the components of our definition of creativity, and conduct an analysis to determine which of these features correlate most with creative video. Finally, we evaluate a supervised approach to automatically detect creative video, with promising results, showing that it is necessary to model both aesthetic value and novelty to achieve optimal classification accuracy.

* 8 pages, 1 figures, conference IEEE CVPR 2014 
Viaarxiv icon

People are Strange when you're a Stranger: Impact and Influence of Bots on Social Networks

Jul 30, 2014
Luca Maria Aiello, Martina Deplano, Rossano Schifanella, Giancarlo Ruffo

Figure 1 for People are Strange when you're a Stranger: Impact and Influence of Bots on Social Networks
Figure 2 for People are Strange when you're a Stranger: Impact and Influence of Bots on Social Networks
Figure 3 for People are Strange when you're a Stranger: Impact and Influence of Bots on Social Networks
Figure 4 for People are Strange when you're a Stranger: Impact and Influence of Bots on Social Networks

Bots are, for many Web and social media users, the source of many dangerous attacks or the carrier of unwanted messages, such as spam. Nevertheless, crawlers and software agents are a precious tool for analysts, and they are continuously executed to collect data or to test distributed applications. However, no one knows which is the real potential of a bot whose purpose is to control a community, to manipulate consensus, or to influence user behavior. It is commonly believed that the better an agent simulates human behavior in a social network, the more it can succeed to generate an impact in that community. We contribute to shed light on this issue through an online social experiment aimed to study to what extent a bot with no trust, no profile, and no aims to reproduce human behavior, can become popular and influential in a social media. Results show that a basic social probing activity can be used to acquire social relevance on the network and that the so-acquired popularity can be effectively leveraged to drive users in their social connectivity choices. We also register that our bot activity unveiled hidden social polarization patterns in the community and triggered an emotional response of individuals that brings to light subtle privacy hazards perceived by the user base.

* 10 pages, 9 figures, Proceedings of the 6th International AAAI Conference on Weblogs and Social Media, Dublin, IR, 2012 
Viaarxiv icon