Reflection #6 – [2/8] – Jiameng Pu

Danescu-Niculescu-Mizil, C., Sudhof, M., Jurafsky, D., Leskovec, J., & Potts, C. (2013). A computational approach to politeness with application to social factors. arXiv preprint arXiv:1306.6078.

Summary:

The paper focuses on exploring the content politeness on social platforms like Wikipedia, Stack Exchange. They conduct linguistic analysis exhibiting politeness on the corpus of requests and extract positive and negative politeness strategies from the context. In the experiment, two classifiers, a bag of words classifier (BOW) and a linguistically informed classifier (Ling.), are compared to illustrate the effectiveness of their theory-inspired features. They also show the relationship between politeness levels and social power/reputation. In general, we see a negative correlation between politeness and power on Stack Exchange or Wikipedia. They also add that there are differences in politeness levels across factors of interest, such as communities, geographical regions, and gender.

Reflection:

In the paper, they use a bag of words classifier as a strong baseline for the new classification task. However, it didn’t mention whether the BOW classifier is the state-of-the-art, thus the advancement of BOW needs to be discussed or some other classifiers are expected to be listed in the experiment. With a more effective classifier, where can we take advantage of it to benefit users or virtual communities? For instance, we may attach users in the community with a ‘politeness’ feature, hostile users will be automatically muted or banned if their politeness is lower than the floor value. We can also use it to evaluate the politeness of online customer service extensively used in the electronic business, which would benefit the environment of online shopping.

 

Voigt, R., Camp, N. P., Prabhakaran, V., Hamilton, W. L., Hetey, R. C., Griffiths, C. M., … & Eberhardt, J. L. (2017). Language from police body camera footage shows racial disparities in officer respect. Proceedings of the National Academy of Sciences, 201702413.

Summary:

Similar to the politeness-level difference analyzed in the above paper, one of the most non-negligible topics is the respect-level between different races(even gender, nationality). The paper collects footage from body-worn cameras, we extract the respectfulness of police officer language toward white and black community members by applying computational linguistic methods on transcripts. Disparities are shown in the speaking way of officers toward black versus white community members.

Reflection:

For the data collection, the source of datasets is not geographically diverse. Given that the issue of racial discrimination in different regions may not be of different levels of severity, I think it might be a better choice to collect footage in more than one place in the United States, not just Oakland, which would provide a holistic point of view of how officer treat white and black drivers. On the other hand, the race of officer is a key control factor since I’m curious about how black officer treats black and white drivers in their routine traffic stops, which is not discussed in detail though.
I noticed that the data they extracted were footage from body-worn cameras, transcripts were mainly used in the research. With the development of computer vision technology, extracting footage for the frame detection as an additional feature may help us to better understand the relation pattern of between police officers and drivers. For example, the frequency of conflicts between officers and drivers is also an important measure of respect degree.

Questions:

  • For these topic-similar paper, is there a clear difference between being polite and being respectful?
  • Since it is not easy for social scientists to measure how police officers communicate with the public, researchers use body-worn cameras to collect data. However, with body-worn cameras capturing interactions every day, how do we guarantee police officers would act exactly the same as the way when they don’t wear cameras? Based on this, keeping data collection procedure agnostic to police officers might be an alternative choice for researchers?

Read More

Reflection #6 – [2/8] – Pratik Anand

Paper 1 : A computational approach to politeness with application to social factors

Paper 2 : Language from police body camera footage shows racial disparities in officer respect

Both the papers provide interesting ideas related to respect and politeness in communication. While the second paper deals with a very hot topic of unfair treatment of minority races by the police in real world, the first paper takes a more academic approach towards behavior in online space.
The authors in paper 1 develop a framework to measure politeness in online conversation which is a very valuable resource. It has words with their politness scores. They painstakingly comb through interactions of editors in Wikipedia and annotate it as polite or impolite conversation. I really liked the fact that they acknowledged that these filtering might be biased as politeness is subjective. So, they took extra measure to choose only those words which are unanimously marked by all the annotators. Once completed, they used it to apply to StackExchange conversations and found that people who were polite earlier, became less polite once they gained a certain status or power after admin elections. It is a very fascinating result. Similarly, people who lost in those elections became more polite. What could be causing this ? Is this a general with power comes arrogance argument or something else ? Did those people start getting more involved and hence, got tired of being careful with words ? Also the paper doesn’t say about the people who got popular votes but didn’t apply for admin elections. Does their behavior change too ? The paper also mentions that people from certain part of US are more polite than the others. It brings an interesting point of culture. Some people are culturally polite in conversation by US and western standards. Other people are not fluent in English vocabulary enough and could be more direct and hence, appear to be impolite. Also, sarcasm plays a big role in conversations.Yeah, right” is a less polite phrase made up of more polite ones. The paper doesn’t talk about such limitations in its study.

The second paper is short but provides a crucial insight on how race based biases can influence a conversation. It analyses the transcripts from video feed of police officers who stopped someone and use it to determine how much polite they were in the talk. The test subjects are not revealed the race or sex of the people being investigated or the police officer but their conclusion shows that Non-Caucasians are talked less politely by the police irrespective of the race of the police officer. The paper measure in form of Respect and Formality. Though, formality is maintained for white as well black people. But it is drastically less for black people. An interesting observation is from Fig 5 that formality goes down with more time passing. Respect has a minima but increases on both ends of time of interaction. This trend can be explained by the fact that greetings are given at the start and the end of a conversation. A further direction of this paper should be facial and body gesture of the police officer as well as the apprehended user.

Overall, both papers bring attention to change in politeness either due to power and influence or racial difference which opens up new fields of study.

Read More

Reflection #6 – [02/08] – [John Wenskovitch]

This pair of papers discussed computational mechanisms and studies for determining politeness, both in online communities (Danescu-Niculescu-Mizil et al.) and in interactions with law enforcement (Voigt et al.).  In the DNM et al. paper, the authors use requests from both Wikipedia and Stack Exchange to build a politeness corpus, using Mechanical Turkers to annotate these requests on a spectrum of politeness.  They then build two classifiers to predict politeness, testing the classifiers both in-domain (training and testing data from the same source) and cross-domain (training on Wikipedia and testing on Stack Exchange, and vice versa).  The Voigt et al. paper used transcribed data from Oakland Police Department body cameras for several studies regarding racial disparity in officer behavior.  These studies included measuring respectfulness and politeness perceptions, identifying linguistic features to model respect, and measuring racial disparities in respect.

I have generalizability concerns about both papers because of the choices made in data collection.  In the DNM paper, both the bag of words classifier and the linguistically informed classifier performed worse than the human reference percentages of classification.  This was true both in-domain and cross-domain, though cross-domain is more valid for this concern.  As a result, I suspect that any new corpus that is added as a source will have classification rates similar to the cross-domain accuracy.  Further, their use cases that focus on requests provide a further bias – I suspect that introducing any new corpus not focused on requests will have similar performance, if not worse performance.  I wonder if their classifier accuracies might improve if they consider more than two sentences of text with each request, to acquire additional contextual information.

The Voigt paper used only transcribed body camera audio from a city police department, in a city well-known for crime rates.  As a result, their results may not generalize to interactions with law enforcement in rural communities (where crime rates are different), near the national borders (where demographics are different), and in safer communities (where criminal behavior is less prevalent).  Further, the behavior of the officers may differ with the knowledge that they are wearing body cameras.  I’m curious to know if patterns found in transcribed audio from police cruiser dashboard cameras (in situations when the officers aren’t wearing body cameras) are any better or worse than the results shown in this study.

In general, I also felt that the discussion sections of the papers were the most interesting parts.  The DNM paper looks at specific cases within the corpus, such as changes in the behavior of Wikipedia moderators when they become administrators and no longer have to be as polite (and who have to be particularly polite in the timeframe leading up to their election).  The Voigt paper discussion notes that while their work demonstrates that racial disparities in levels of officer respect exist, the root causes of those disparities are less clear, perhaps an ideal target for a follow-up study on a broader range of interaction transcriptions. 

 Another potential follow-up study to the Voigt paper could consider the effect of seasons on officer politeness.  All of the data from the Oakland Police Department was from interactions that occurred in April.  Are officers more likely to be polite when the weather is nicer, or less polite in the depths of winter?  And if there are seasonal or weather-related changes, does the racial disparity grow or shrink?

I found the distributions from Figure 1 of the DNM paper to be intriguing.  I’m curious why the Stack Exchange politeness spectrum seems to mimic the Gaussian distribution that you would expect to see, but the Wikipedia politeness spectrum seems to plateau just above the mean.  Trying to understand the difference in these distributions would be yet another interesting follow-up study – is the difference a result of inflated semi-polite interaction frequency because of the moderators trying to become administrators, or is it a result of the language in interactions on Wikipedia being more formal than the informal Stack Exchange, or some other reason entirely?  I’m curious to hear the thoughts of anyone else in the class.

Read More

Exposure to ideologically diverse news and opinions on Facebook

Summary:

The paper discusses about the influence of social networks on spreading diverse news. It analyzes the network and user connections on popular social  networking site Facebook and observe the ideologically diverse shared contents among user and their friends. Their study finds out some interesting analysis about the proportion of shares of news by diverse ideological friends of a user.

 

Reflection:

This paper studies the influence of exposure of a news shared or clicked by users of diverse contents. They find out not only users of same diverse contents share news among them, also their friends of different ideology also click the shared contents. Also, its been observed that conservative contents have highest proportion of shares. In my opinion, there are studies on users and shared contents of diverse ideological news by them, but the paper did not take age of users as a factor, which I think could be a good observation for better study the analysis.

Read More

Reflection #5 – [02-06] – [Patrick Sullivan]

Exposure to Ideologically Diverse News and Opinion on Facebook” by Bakshy, Messing and Adamic is exploring how social media sites like Facebook give political information to users and how it affects their political perspective.

This research shows the influence that social media can have on polarizing opinions on certain topics. Since social media algorithms are tuned to recommend content that the user ‘likes’, then most content will match up to the user’s political leaning, polarizing the political landscape even more.

Perhaps this could be remedied by adjusting the name ‘like’ to a different action word that has less problematic outcomes. Perhaps a user study could be used to both replicate this study and test out some alternatives and their effects. A breakthrough in this area would allow social media sites to create a more transparent showcase of what this new action name means. For instance, if users now ‘nod’ instead of ‘like’ articles on Facebook then users might take this as ‘they endorse this content as correct‘ instead of the currently existing ‘they like the content, want more of it, and want their friends to know they like the content‘.

This also is a good chance for social media sites to adjust algorithms to take into account political leanings, and to continually work on tuning a recommendation engine that is as unbiased as possible. This would have a great impact on the political exposure and perspective social media users experience, and could lead to more moderate views being expressed and more nuanced arguments being used as support.

In addition, I wonder if there is a measurable difference of user’s accuracy in self-reporting as a minority political party in countries where dissenting political opinion is actively repressed? Could this be used to determine if other countries have a repression of opinion that is implicitly known among the population instead of explicit and visible political threats?

Echo Chambers Online?: Politically Motivated Selective Exposure Among Internet News Users” by Garrett is an investigation into if users prefer support of their political stance over counterarguments to an opposing argument.

I believe that this result might be caused by users having a stronger background on the information their political party promotes, and thus have a better understanding of the support article that the user appreciates? Participants in the study could be finding it difficult to parse the article that counters the view of an political opponent, since they could be less familiar with that viewpoint.

Could the user who only selects articles that support their viewpoint be considered self-censorship? Forcing users out of this behavior would likely qualify as a violation of freedom of speech and freedom of press. Perhaps incentivizing users to read articles that contain a conversation between opposing political perspectives or a read from a less biased news source is more viable.

Read More

Reflection #4 – [02/6] – [Md Momen Bhuiyan]

Paper #1: Echo Chambers Online?: Politically Motivated Selective Exposure among Internet News Users
Paper #2: Exposure to ideologically diverse news and opinion on Facebook

Summary #1:
This paper checks whether selective exposure plays any significant role in people’s choice of a news article in high-choice information environment like the Internet. The traditional approach of opinion reinforcement seeking and opinion challenge avoidance doesn’t make sense on the Internet. This paper suggests that this is due to people’s tendency to more focus on opinion reinforcement than filtering opinion challenging information. The study was run on 700+ people recruited online. Each of the participants is shown a set of articles related to a politically relevant topic and asked their interest in reading those articles in two phase. Thier reading time of the articles is also recorded. Based on the result it appears that people tend to select more stories with opinion reinforcement while spending more time on opinion challenging articles.

Summary #2:
This paper compares the influence of algorithm and user’s clickthrough in exposure to cross-cutting stories on Facebook. The network structure in Facebook is different than other social media due to connections in different offline context (school, family, social activity etc.). This is probably why homophily structure in Facebook is different. The authors use 10 million Facebook users who self-reported their political affiliation and their click activity on 7 million web URLs shared to analyze the influence. The analysis was limited to hard contents that have political significance. User’s news feed algorithm depends on various things on Facebook including how many times they visit the site daily, how much they interact with their friends, how often they click certain web URLs etc. The authors’ found that algorithmic exposure to cross-cutting content for conservative users was significantly higher than liberal users while the random probability of seeing cross-cutting content was the opposite. The click rate further limits their exposure.

Reflection #1:
This study was designed in a very simplistic way. The author has already noted that population was selected from two partisan sites which might have influenced the result. By selecting a different set of users it is easy to check if people with more partisan view are more likely to want to see the opposing view. Author’s inference of the non-existence of echo chamber is a bit of a stretch. One thing interesting in the result was that people had more interest in “gay marriage” than “civil liberty” while their reading time on “gay marriage” is less than that of “civil liberty”. This conflicts with the common notion that people spend more time things of interest. People seem to find most stories more opinion challenging than opinion reinforcing. The variance in this opinion seems to increase when people have read the stories. Another variation of this could be checking how much difference it makes in case of just headlines of a story is presented to users. The study didn’t ask if the users already read a story. This could make a significant difference given that news aggregator tend to show most popular stories first. This could also account for less time people spend on stories of interest. One thing I didn’t understand was why use log of the time to fit a linear regression model as the time span was in the range of 4sec – 9min. Another particular odd inference was author thought the correlation of reading time with Age was not significant with coefficient .01 for p < .01 while it was significant in case of opinion reinforcement (retrospective) with coefficient .02 with p < .05.

Reflection #2:
This paper checks what is the role of users click through in the limiting exposure to diverse content. Although the user selection process had a self-reporting bias, the proportion of liberal and conservative users were very close surprisingly. Figure 3 has several implications given certain condition is met. They failed to mention the number of average shares by both conservative and liberal users. Although they mentioned the total proportion of different types of shares by both conservative and liberal users, the average number of shares could skew the result. Similar relation could be found in the proportion of liberal and conservative contents in the 226,000 hard contents. They also didn’t mention users’ living area (city or rural) which could skew the percentage of ties. The authors did find a correlation between the position of a story in the news feed and click through ratio which might have affected the inference. They used a technique to calculate the risk ratio (results not shown in the supplementary text) but provided a partial method for calculation.

Read More

Reflection #5 – [02/06] – Vartan Kesiz-Abnousi

Papers

[1]. Garrett, R. Kelly. “Echo chambers online?: Politically motivated selective exposure among Internet news users.” Journal of Computer-Mediated Communication2 (2009): 265-285.

[2]. Bakshy, Eytan, Solomon Messing, and Lada A. Adamic. “Exposure to ideologically diverse news and opinion on Facebook.” Science6239 (2015): 1130-1132.

Summary

In Baksy’s et al 2015 paper, the main question is: how do online networks influence exposure to perspectives that cut across ideological lines? They focus only on Facebook. To this end, they assemble data shared by U.S. users over a 6-month period between 7 July 2014 and 7 January 2015. Then, they measure ideological homophily in friend networks and examined the extent to which heterogeneous friends could potentially expose individuals to cross-cutting content They construct an indicator score, “Alignment”, that ranks sources based on their ideological affiliation. They quantify the extent to which individuals encounter comparatively more or less diverse content while interacting via Facebook’s algorithmically ranked News Feed and further studied users’ choices to click through to ideologically discordant content. They find that compared with algorithmic ranking, individuals’ choices played a stronger role in limiting exposure to cross-cutting content.

In Garret’s 2009 paper, the author examines whether the desire for opinion reinforcement may play a more important role in shaping individuals’ exposure to online political information than an aversion to opinion challenge. In doing so, data collected via a web administered behavior-tracking study over a six-week period in 2005. The subjects were recruited from the readership of 2 partisan online news sites. The results demonstrate that opinion-reinforcing information promotes news story exposure while opinion-challenging information makes exposure only marginally less likely.

Reflection

Baksy et al published in 2015, is dealing with contemporary events that are widely discussed. The role of social media has been at the epicenter of due to its, according to some, significant impact in shaping the political landscape of the United States. Compared with algorithmic ranking, individuals’ choices played a stronger role in limiting exposure to cross-cutting content Subsequently, one question that begs an answer is whether these results remain the same now

Baksy et al specifically focused on Facebook, there is a caveat that I wonder whether the research took into account on whether people actually “Follow” their friends. For instance, what about the information users are exposed to by the things their Friends like, or share? Is this considered part of the “News Feed”? I would argue that this might have more significant effect on “Echo Chambers” than the algorithmically ranked news feed. Even in this case, you only see a subset of what your “friends” share, the friends that you actually “follow”. In addition, how about the information that you receive by the “groups” that you “follow”? I am not sure if the paper addressed this issue.

In addition, individuals may read the summaries of articles that appear in the News Feed and therefore be exposed to some of the articles’ content without clicking through.

I also find it interesting that for both liberals and conservatives the median proportion of friendships of the people on the opposite spectrum is roughly the same, around 20% following the 20/80 Pareto principle.

In Garret’s I found the experimental design particularly thoughtful an interesting, including the fact that they had screening questions. However, it should be stressed that they examine the issue of echo chambers only the in the context of politics. The dependent variable is the “use of issue related news”. It should be noted that they measured the dependent variable by their “interest in reading” and “read time”. From modeling perspective, it appears that Garret is using a “mixed model”, which is a statistical model containing both “fixed effects” and “random effects”. It might have been a good idea to include control for time trends, since there is a time wedge of six weeks in the study. Possibly including a dummy for each week would make the results more robust, although I can understand that from the author’s view 6 weeks is a short period. Still, individual fixed-effects control only for factors that remain constant between those 6 weeks.

For the logit model that examines “story selection”, the “opinion challenge” variable in the logit model has a p-value 5% to 10%. Therefore, for that model I believe the readers should place more emphasis on the results regarding “opinion-reinforcement” variable.

I wonder whether there should be some guidebook or manual that would warn and instruct us how to search information on the web-search engines. However, I don’t believe that opinion-reinforcing web searches belief is tantamount to an echo chamber. To put it bluntly, if the users are conscious that they conducting an opinion-reinforcing web-search, then what’s wrong with that? Nothing. There are degrees of “echo chambers” and whether there is critical threshold by which an echo chamber is harming the user is yet to be seen. For instance, opinion-reinforcing web search on medical issues has serious ramification on public health. Such qualitative factors impart a different contextual meaning of what an “echo-chamber” is and its ramifications. Moreover, such qualitative factors should be taken into account. The quantitative models must include more information distinctions than just “hard” or “soft” news.

Finally, I have a broader skepticism that transcends the main goal of the Baksy et al paper. Facebook, and other social media websites, are under fire for bolstering “fake news”. This criticism -financial incentive to force social media platforms to rank them higher than smaller media outlets. This is called the “principal-agent” problem in corporate finance. There should healthy skepticism.

Questions

  1. The study was conducted in 2015. Would these results hold today?
  2. How about effect of “Following” your friends? Does this take into account the information that you are exposed to?
  3. What are the factors the about the order which users see stories in the News Feed? Do they weigh the same way?
  4. There is a qualitative aspect on “echo-chambers”. The distinction of information should be more than “hard” or “soft” news. There might be an “echo-chamber” on information related to politics but not on medical/health issues. When the stakes are high, i.e. your health, are any of the hypotheses listed on Garret’s paper validated? My hunch is not. This qualitative heterogeneity is not addressed properly. I believe this requires further investigation.

Read More

Reflection #5 – [02/06] – Hamza Manzoor

[1]. Garrett, R. Kelly. “Echo chambers online?: Politically motivated selective exposure among Internet news users.” Journal of Computer-Mediated Communication2 (2009): 265-285.

[2]. Bakshy, Eytan, Solomon Messing, and Lada A. Adamic. “Exposure to ideologically diverse news and opinion on Facebook.” Science6239 (2015): 1130-1132.

The theme of both these papers is online echo chambers that people read news or information, which favor their opinions ignoring the views of the opposing ideology. The Garrett et al. conducted a study with users recruited from two news websites, AlterNet (left-leaning) and WorldNetDaily (right-leaning). Their study was a web administered behavioral study in which participants selected articles to read in 15 minutes. The findings of the study supported the author’s hypotheses that users are more likely to view opinion-reinforcing information and will spend more time on opinion-reinforcing information.

On the other hand, Bakshy et al. performed analysis on around 10.1 million active users of Facebook who self-report their political affiliations. They examined how users interact with shared news articles by their friends as well as by algorithmically ranked news feed. Their findings were that compared with algorithmic ranking, individuals’ choices play a strong role in limiting exposure to cross-cutting content.

The Garrett et al. study was very thorough and the paper was very well written. The hypotheses were clear and the results presented backed up those hypothesis. But they assumed that everyone who visited AlterNet was left-leaning and WorldNetDaily visitors were right-leaning without any evidence to support this claim. Secondly, this sample does not ensure that it is representative of the population. Also, the length of news article should have been mentioned because the read time per story ranged from 4 seconds to 122 seconds. How can it be 4 seconds? Also, can we make a claim about read time from such study? It is highly likely that general reading patterns of people can be vastly different than the ones under study because they know that they are under experiment. It will be interesting to compare the results with people who aren’t aware that they are under study. Finally, the study is focused on three issues (gay marriage, social security reform, and civil liberties), which can have political affiliation, but will this study hold true for sports or other benign issues?

The Bakshy et al. study was interesting but had a huge limitation that it recorded engagements based on clicks. They mention that users might be reading the displayed summaries but not clicking it. This limitation makes me question the results for a reason that just because user is not clicking on opinion-challenging information does not mean that it will create an echo chamber. It is possible that user is aware of the other side of story but it is only natural that people read more of what they like and hence clicking on opinion-reinforcing information. I also felt that study on neutral people should have been more thorough and one out of context research idea that popped in my mind while reading the paper was that it will be interesting to see how ranking algorithms perform in case of neutral people? Do ranking algorithms play a role in creating echo chambers? It is highly likely that a neutral person has majority of left (or right) leaning friends and therefore, will encounter respective information and as a result ranking algorithm might confuse him with left (or right) leaning person. This seems like an interesting research direction to me.

Read More

Reflection #5 – [02/06] – Jiameng Pu

Garrett, R. K. (2009). Echo chambers online?: Politically motivated selective exposure among Internet news users. Journal of Computer-Mediated Communication, 14(2), 265-285.

Summary:

The Internet offers users abundant access to information from around the world, which expose people to more diverse viewpoints, including both similar perspective and attitude-challenging information. With so much information, voice comes up that whether the Internet leads people to adopt increasingly insular political news exposure practices, i.e., “echo chambers” and “filter bubbles”, which refer to cases that individuals tend to be exposed only to information from like-minded individuals or materials similar to previous reading due to the influence of algorithms the news sites employed. This paper tests the viewpoint by conducting a web-administered behavior-tracking study. They find that the effect of opinion-challenging information is small and only marginally significant, which means worry that the Internet will lead to an increasingly fragmented society appears to have been overstated.

Reflection:

The paper shows readers how to conduct an exhausted and comprehensive experimental design when studying a specific problem, from selecting the research subjects to designing the experimental setting/workflow, e.g., using a purpose-built software administrates the study. Appropriate groups should be chosen to study according to the problem itself — two partisan online news services, left AlterNet and right WorldNet Daily, are able to bring immediate benefits to the study with the inherent engagement in selective exposure. Another thing I am concerned about is that there are not enough samples participating in the experiment, i.e., a total of 727 subjects (358 Alternet readers and 369 WorldNetDaily readers respectively), which possibly means the generalization of our observation result is not universal enough. Although the author mentions in the limitation that “when viewing these results is that the sample is not representative of the U.S. population at large”, I think it is very difficult to represent groups with hundreds of subjects, not to mention specific group like U.S. population at large. Why do not we find some volunteers on university campuses or research institutes, e.g., students and faculties, which is easier to get more volunteers?

Bakshy, E., Messing, S., & Adamic, L. A. (2015). Exposure to ideologically diverse news and opinion on Facebook. Science, 348(6239), 1130-1132.

Summary:

Based on the same problem background as the first paper, e.g., “echo chambers” and “filter bubbles”, the paper attempts to explore how this modern online Internet environment will influence people’s exposure to diverse content. Past empirical attempts turn out to be limited because of difficulties in measurement and show mixed results. To overcome previous limitations, they use a large, comprehensive Facebook dataset to measure the extent of ideological homophily and content heterogeneity in friend networks. They found that individual choice plays a stronger role than algorithmic ranking does in content exposure.

Reflection:

One heuristic point to take away is to classify stories as either “hard” (such as national news, politics, or world affairs) or “soft” content (such as sports, entertainment, or travel) by training a support vector machine, which reduces the data size while keeping the most efficient part of the dataset since “hard” content tends to be more controversial than soft content. However, some of the “soft” content is also practicable and appropriate for the task, i.e., hot and controversial topics in sports and entertainment fields. Thus the author may consider augmenting the size of the dataset to find out the result.

Same as the vague distinction between exposure and consumption mentioned in the limitation analysis, I also think how to clarify the line among conservative, neutral, liberal? Can we possibly give a uniform and appropriate criteria for the classification of those three categories?

Read More

Reflection #5 – [02/06] – Jamal A. Khan

Both the papers ask questions along similar lines, however, the question raised by the Kelly is more subtle and in a sense a follow up to the one posed by Bakshy et al.

  1. Garrett, R. Kelly. “Echo chambers online?: Politically motivated selective exposure among Internet news users.” Journal of Computer-Mediated Communication2 (2009): 265-285.
  2. Bakshy, Eytan, Solomon Messing, and Lada A. Adamic. “Exposure to ideologically diverse news and opinion on Facebook.” Science6239 (2015): 1130-1132.

The first paper asks quite an interesting question that deals with the consequences of technology  i.e. “Will selective new exposure or the ability to limit one’s exposure to articles or sources of a certain kind lead to a dystopian future where democracy will collapse?” The answer  to which was, unsurprisingly, no.

The results of the paper suggest that prospective opinion-reinforcing information greatly influences article selection and that immediately raises a follow up question, “how does fake news framed as opinion-reinforcing content come into play in shaping perception of truth among people?”. If people are more likely to consume articles that support their ideological point of view, then such articles can serve as good tools for derailing the truth. Since there is potential of the truth being manipulated, “does this lead people to very extreme, one-sided and tunneled vision i.e. can people be indirectly be programmed to adopt a certain point of view?if so, how devastating can the effects be?”

The second question has more to do with the design of the study, “does voluntarity (if that’s a word) exaggerate behaviors?” i.e. does passive vs active instrumentation of people lead to different results? will a person’s choice of articles be different if they knew that they were being monitored?

The third question has to do with the choice of the population. People chosen from well known left or right leaning sights are generally quite enthusiastic and/or as compared to the general public. Hence the results that were reported here might be an artifact of the choice of the population. So i guess the question i’m asking is “whether people who read news from partisan sources are democrats for the sake of being democrats or republicans for the sake of being republicans?”. Hence the generalization of the results to the general public is uncertain.

The second paper asks a different question of whether people get isolated from a opinion-challenging information even if they choose to. Given, the amount of data available to the authors i’m surprised the paper is as fleshed out as it could be, but the paper does a good job of quantitatively proving that the fear is unfounded. An interesting thing that i noted was that the neutral group tended to have a smaller proportion of friends who were neutral as compared to liberal of conservative friend, the difference is almost ~20%.

So is it that neutral people have these proportions because they are neutral in leaning, or is that them being neutral is an affect of having equal proportions of conservative or liberal friends?

Another aspect that i think could’ve been easily studied in the paper but was overlooked, is that given hard news that was opinion challenging, how likely are people to share it? This statistic would give deeper insight into how open are each of the three groups to change and/or how likely they are to peruse the truth of the matter and whether they value truth over ego.

Perhaps, this study could also look into different but specific topics e.g. climate change, and look at similar statistics to determine if  “micro echo chambers” (topic based echo chambers) are formed. This sounds certainly seems like an interesting direction to me!

Read More