1
|
Haddadan S, Menghini C, Riondato M, Upfal E. Reducing polarization and increasing diverse navigability in graphs by inserting edges and swapping edge weights. Data Min Knowl Discov 2022. [DOI: 10.1007/s10618-022-00875-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
2
|
Gatmiry K, Gomez-Rodriguez M. The Network Visibility Problem. ACM T INFORM SYST 2022. [DOI: 10.1145/3460475] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/01/2022]
Abstract
Social media is an attention economy where broadcasters are constantly competing for attention in their followers’ feeds. Broadcasters are likely to elicit greater attention from their followers if their posts remain visible at the top of their followers’ feeds for a longer period of time. However, this depends on the rate at which their followers receive information in their feeds, which in turn depends on the broadcasters they follow. Motivated by this observation and recent calls for fairness of exposure in social networks, in this article, we look at the task of recommending links from the perspective of visibility optimization. Given a set of candidate links provided by a link recommendation algorithm, our goal is to find a subset of those links that would provide the highest visibility to a set of broadcasters. To this end, we first show that this problem reduces to maximizing a nonsubmodular nondecreasing set function under matroid constraints. Then, we show that the set function satisfies a notion of approximate submodularity that allows the standard greedy algorithm to enjoy theoretical guarantees. Experiments on both synthetic and real data gathered from Twitter show that the greedy algorithm is able to consistently outperform several competitive baselines.
Collapse
|
3
|
Krafft PM, Shmueli E, Griffiths TL, Tenenbaum JB, Pentland AS. Bayesian collective learning emerges from heuristic social learning. Cognition 2021; 212:104469. [PMID: 33770743 DOI: 10.1016/j.cognition.2020.104469] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2018] [Revised: 09/14/2020] [Accepted: 09/16/2020] [Indexed: 11/28/2022]
Abstract
Researchers across cognitive science, economics, and evolutionary biology have studied the ubiquitous phenomenon of social learning-the use of information about other people's decisions to make your own. Decision-making with the benefit of the accumulated knowledge of a community can result in superior decisions compared to what people can achieve alone. However, groups of people face two coupled challenges in accumulating knowledge to make good decisions: (1) aggregating information and (2) addressing an informational public goods problem known as the exploration-exploitation dilemma. Here, we show how a Bayesian social sampling model can in principle simultaneously optimally aggregate information and nearly optimally solve the exploration-exploitation dilemma. The key idea we explore is that Bayesian rationality at the level of a population can be implemented through a more simplistic heuristic social learning mechanism at the individual level. This simple individual-level behavioral rule in the context of a group of decision-makers functions as a distributed algorithm that tracks a Bayesian posterior in population-level statistics. We test this model using a large-scale dataset from an online financial trading platform.
Collapse
Affiliation(s)
- P M Krafft
- Creative Computing Institute, University of Arts London, London, England, United Kingdom.
| | - Erez Shmueli
- Department of Industrial Engineering, Tel-Aviv University, Tel-Aviv, Israel
| | | | | | | |
Collapse
|
4
|
Sun W, Nasraoui O, Shafto P. Evolution and impact of bias in human and machine learning algorithm interaction. PLoS One 2020; 15:e0235502. [PMID: 32790666 PMCID: PMC7425868 DOI: 10.1371/journal.pone.0235502] [Citation(s) in RCA: 23] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2019] [Accepted: 06/17/2020] [Indexed: 12/22/2022] Open
Abstract
Traditionally, machine learning algorithms relied on reliable labels from experts to build predictions. More recently however, algorithms have been receiving data from the general population in the form of labeling, annotations, etc. The result is that algorithms are subject to bias that is born from ingesting unchecked information, such as biased samples and biased labels. Furthermore, people and algorithms are increasingly engaged in interactive processes wherein neither the human nor the algorithms receive unbiased data. Algorithms can also make biased predictions, leading to what is now known as algorithmic bias. On the other hand, human's reaction to the output of machine learning methods with algorithmic bias worsen the situations by making decision based on biased information, which will probably be consumed by algorithms later. Some recent research has focused on the ethical and moral implication of machine learning algorithmic bias on society. However, most research has so far treated algorithmic bias as a static factor, which fails to capture the dynamic and iterative properties of bias. We argue that algorithmic bias interacts with humans in an iterative manner, which has a long-term effect on algorithms' performance. For this purpose, we present an iterated-learning framework that is inspired from human language evolution to study the interaction between machine learning algorithms and humans. Our goal is to study two sources of bias that interact: the process by which people select information to label (human action); and the process by which an algorithm selects the subset of information to present to people (iterated algorithmic bias mode). We investigate three forms of iterated algorithmic bias (personalization filter, active learning, and random) and how they affect the performance of machine learning algorithms by formulating research questions about the impact of each type of bias. Based on statistical analyses of the results of several controlled experiments, we found that the three different iterated bias modes, as well as initial training data class imbalance and human action, do affect the models learned by machine learning algorithms. We also found that iterated filter bias, which is prominent in personalized user interfaces, can lead to more inequality in estimated relevance and to a limited human ability to discover relevant data. Our findings indicate that the relevance blind spot (items from the testing set whose predicted relevance probability is less than 0.5 and who thus risk being hidden from humans) amounted to 4% of all relevant items when using a content-based filter that predicts relevant items. A similar simulation using a real-life rating data set found that the same filter resulted in a blind spot size of 75% of the relevant testing set.
Collapse
Affiliation(s)
- Wenlong Sun
- Department of Computer Engineering and Computer Science, University of Louisville, Louisville, Kentucky, United States of America
| | - Olfa Nasraoui
- Department of Computer Engineering and Computer Science, University of Louisville, Louisville, Kentucky, United States of America
| | - Patrick Shafto
- Department of Mathematics and Computer Science, Rutgers University - Newark, Newark, New Jersey, United States of America
| |
Collapse
|
5
|
Olteanu A, Castillo C, Diaz F, Kıcıman E. Social Data: Biases, Methodological Pitfalls, and Ethical Boundaries. Front Big Data 2019; 2:13. [PMID: 33693336 PMCID: PMC7931947 DOI: 10.3389/fdata.2019.00013] [Citation(s) in RCA: 83] [Impact Index Per Article: 16.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2019] [Accepted: 05/27/2019] [Indexed: 11/24/2022] Open
Abstract
Social data in digital form-including user-generated content, expressed or implicit relations between people, and behavioral traces-are at the core of popular applications and platforms, driving the research agenda of many researchers. The promises of social data are many, including understanding "what the world thinks" about a social issue, brand, celebrity, or other entity, as well as enabling better decision-making in a variety of fields including public policy, healthcare, and economics. Many academics and practitioners have warned against the naïve usage of social data. There are biases and inaccuracies occurring at the source of the data, but also introduced during processing. There are methodological limitations and pitfalls, as well as ethical boundaries and unexpected consequences that are often overlooked. This paper recognizes the rigor with which these issues are addressed by different researchers varies across a wide range. We identify a variety of menaces in the practices around social data use, and organize them in a framework that helps to identify them. "For your own sanity, you have to remember that not all problems can be solved. Not all problems can be solved, but all problems can be illuminated." -Ursula Franklin.
Collapse
Affiliation(s)
- Alexandra Olteanu
- Microsoft Research, New York, NY, United States
- Microsoft Research, Montreal, QC, Canada
| | - Carlos Castillo
- Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | | | | |
Collapse
|
6
|
Ciampaglia GL, Nematzadeh A, Menczer F, Flammini A. How algorithmic popularity bias hinders or promotes quality. Sci Rep 2018; 8:15951. [PMID: 30374134 PMCID: PMC6206065 DOI: 10.1038/s41598-018-34203-2] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2017] [Accepted: 10/01/2018] [Indexed: 11/19/2022] Open
Abstract
Algorithms that favor popular items are used to help us select among many choices, from top-ranked search engine results to highly-cited scientific papers. The goal of these algorithms is to identify high-quality items such as reliable news, credible information sources, and important discoveries–in short, high-quality content should rank at the top. Prior work has shown that choosing what is popular may amplify random fluctuations and lead to sub-optimal rankings. Nonetheless, it is often assumed that recommending what is popular will help high-quality content “bubble up” in practice. Here we identify the conditions in which popularity may be a viable proxy for quality content by studying a simple model of a cultural market endowed with an intrinsic notion of quality. A parameter representing the cognitive cost of exploration controls the trade-off between quality and popularity. Below and above a critical exploration cost, popularity bias is more likely to hinder quality. But we find a narrow intermediate regime of user attention where an optimal balance exists: choosing what is popular can help promote high-quality items to the top. These findings clarify the effects of algorithmic popularity bias on quality outcomes, and may inform the design of more principled mechanisms for techno-social cultural markets.
Collapse
Affiliation(s)
| | - Azadeh Nematzadeh
- School of Informatics, Computing, and Engineering, Indiana University, Bloomington, Indiana, USA
| | - Filippo Menczer
- Indiana University Network Science Institute, Bloomington, Indiana, USA.,School of Informatics, Computing, and Engineering, Indiana University, Bloomington, Indiana, USA
| | - Alessandro Flammini
- Indiana University Network Science Institute, Bloomington, Indiana, USA.,School of Informatics, Computing, and Engineering, Indiana University, Bloomington, Indiana, USA
| |
Collapse
|
7
|
Abstract
At a time when information seekers first turn to digital sources for news and opinion, it is critical that we understand the role that social media plays in human behavior. This is especially true when information consumers also act as information producers and editors through their online activity. In order to better understand the effects that editorial ratings have on online human behavior, we report the results of a two large-scale in vivo experiments in social media. We find that small, random rating manipulations on social media posts and comments created significant changes in downstream ratings, resulting in significantly different final outcomes. We found positive herding effects for positive treatments on posts, increasing the final rating by 11.02% on average, but not for positive treatments on comments. Contrary to the results of related work, we found negative herding effects for negative treatments on posts and comments, decreasing the final ratings, on average, of posts by 5.15% and of comments by 37.4%. Compared to the control group, the probability of reaching a high rating ( ⩾ 2,000) for posts is increased by 24.6% when posts receive the positive treatment and for comments it is decreased by 46.6% when comments receive the negative treatment.
Collapse
|
8
|
Centeno R, Hermoso R. Estimating global opinions by keeping users from fraud in online review systems. Knowl Inf Syst 2017. [DOI: 10.1007/s10115-017-1089-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
9
|
|
10
|
Burghardt K, Alsina EF, Girvan M, Rand W, Lerman K. The myopia of crowds: Cognitive load and collective evaluation of answers on Stack Exchange. PLoS One 2017; 12:e0173610. [PMID: 28301531 PMCID: PMC5354439 DOI: 10.1371/journal.pone.0173610] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2016] [Accepted: 02/22/2017] [Indexed: 11/18/2022] Open
Abstract
Crowds can often make better decisions than individuals or small groups of experts by leveraging their ability to aggregate diverse information. Question answering sites, such as Stack Exchange, rely on the “wisdom of crowds” effect to identify the best answers to questions asked by users. We analyze data from 250 communities on the Stack Exchange network to pinpoint factors affecting which answers are chosen as the best answers. Our results suggest that, rather than evaluate all available answers to a question, users rely on simple cognitive heuristics to choose an answer to vote for or accept. These cognitive heuristics are linked to an answer’s salience, such as the order in which it is listed and how much screen space it occupies. While askers appear to depend on heuristics to a greater extent than voters when choosing an answer to accept as the most helpful one, voters use acceptance itself as a heuristic, and they are more likely to choose the answer after it has been accepted than before that answer was accepted. These heuristics become more important in explaining and predicting behavior as the number of available answers to a question increases. Our findings suggest that crowd judgments may become less reliable as the number of answers grows.
Collapse
Affiliation(s)
- Keith Burghardt
- Dept of Computer Science, University of California at Davis, Davis, CA, United States of America
- Dept of Political Science, University of California at Davis, Davis, CA, United States of America
- * E-mail:
| | | | - Michelle Girvan
- Dept of Physics, University of Maryland, College Park, MD, United States of America
- Santa Fe Institute, Santa Fe, NM, United States of America
| | - William Rand
- Department of Business Management, North Carolina State University, Raleigh, NC, United States of America
| | - Kristina Lerman
- Information Sciences Institute, University of Southern California, Marina del Rey, CA, United States of America
| |
Collapse
|
11
|
Information Is Not a Virus, and Other Consequences of Human Cognitive Limits. FUTURE INTERNET 2016. [DOI: 10.3390/fi8020021] [Citation(s) in RCA: 39] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
|
12
|
Lamprecht D, Lerman K, Helic D, Strohmaier M. How the structure of Wikipedia articles influences user navigation. NEW REV HYPERMEDIA M 2016; 23:29-50. [PMID: 28670171 PMCID: PMC5468769 DOI: 10.1080/13614568.2016.1179798] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2015] [Accepted: 04/14/2016] [Indexed: 11/03/2022]
Abstract
In this work we study how people navigate the information network of Wikipedia and investigate (i) free-form navigation by studying all clicks within the English Wikipedia over an entire month and (ii) goal-directed Wikipedia navigation by analyzing wikigames, where users are challenged to retrieve articles by following links. To study how the organization of Wikipedia articles in terms of layout and links affects navigation behavior, we first investigate the characteristics of the structural organization and of hyperlinks in Wikipedia and then evaluate link selection models based on article structure and other potential influences in navigation, such as the generality of an article's topic. In free-form Wikipedia navigation, covering all Wikipedia usage scenarios, we find that click choices can be best modeled by a bias towards article structure, such as a tendency to click links located in the lead section. For the goal-directed navigation of wikigames, our findings confirm the zoom-out and the homing-in phases identified by previous work, where users are guided by generality at first and textual similarity to the target later. However, our interpretation of the link selection models accentuates that article structure is the best explanation for the navigation paths in all except these initial and final stages. Overall, we find evidence that users more frequently click on links that are located close to the top of an article. The structure of Wikipedia articles, which places links to more general concepts near the top, supports navigation by allowing users to quickly find the better-connected articles that facilitate navigation. Our results highlight the importance of article structure and link position in Wikipedia navigation and suggest that better organization of information can help make information networks more navigable.
Collapse
Affiliation(s)
- Daniel Lamprecht
- Knowledge Technologies Institute, Graz University of Technology, Graz, Austria
| | - Kristina Lerman
- Information Sciences Institute, University of Southern California, Los Angeles, CA, USA
| | - Denis Helic
- Knowledge Technologies Institute, Graz University of Technology, Graz, Austria
| | - Markus Strohmaier
- Department of Computer Science, University of Koblenz-Landau, Mainz, Germany.,GESIS - Leibniz Institute for the Social Sciences, Cologne, Germany
| |
Collapse
|
13
|
Abeliuk A, Berbeglia G, Cebrian M, Van Hentenryck P. The benefits of social influence in optimized cultural markets. PLoS One 2015; 10:e0121934. [PMID: 25831093 PMCID: PMC4382093 DOI: 10.1371/journal.pone.0121934] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2014] [Accepted: 02/09/2015] [Indexed: 11/19/2022] Open
Abstract
Social influence has been shown to create significant unpredictability in cultural markets, providing one potential explanation why experts routinely fail at predicting commercial success of cultural products. As a result, social influence is often presented in a negative light. Here, we show the benefits of social influence for cultural markets. We present a policy that uses product quality, appeal, position bias and social influence to maximize expected profits in the market. Our computational experiments show that our profit-maximizing policy leverages social influence to produce significant performance benefits for the market, while our theoretical analysis proves that our policy outperforms in expectation any policy not displaying social signals. Our results contrast with earlier work which focused on showing the unpredictability and inequalities created by social influence. Not only do we show for the first time that, under our policy, dynamically showing consumers positive social signals increases the expected profit of the seller in cultural markets. We also show that, in reasonable settings, our profit-maximizing policy does not introduce significant unpredictability and identifies "blockbusters". Overall, these results shed new light on the nature of social influence and how it can be leveraged for the benefits of the market.
Collapse
Affiliation(s)
- Andrés Abeliuk
- Optimization Research Group, National ICT Australia, Melbourne, Victoria, Australia
- Melbourne School of Engineering, The University of Melbourne, Melbourne, Victoria, Australia
| | - Gerardo Berbeglia
- Optimization Research Group, National ICT Australia, Melbourne, Victoria, Australia
- Centre for Business Analytics, Melbourne Business School, The University of Melbourne, Melbourne, Victoria, Australia
| | - Manuel Cebrian
- Optimization Research Group, National ICT Australia, Melbourne, Victoria, Australia
- Melbourne School of Engineering, The University of Melbourne, Melbourne, Victoria, Australia
| | - Pascal Van Hentenryck
- Optimization Research Group, National ICT Australia, Melbourne, Victoria, Australia
- Research School of Computer Science, The Australian National University, Canberra, ACT, Australia
- * E-mail:
| |
Collapse
|