01/29/20 – Dylan Finch – The Future of Crowd Work

Summary of the Reading

This paper looks to build a model of crowd work to better understand some of the major challenges that the future of crowd work will face. The paper also collects data and makes suggestions on how to make the future of crowd work better for all those involved. 

One of the main contributions of this paper is building a model to represent crowd work. They do this by using not only theoretical data, but also empirical data that is gathered from both requesters and workers on crowd work websites. This model helps the authors to visualize some of the problems that come from crowd work and allow them to better understand possible solutions.

The paper then goes into depth on 12 research foci that are part of the model. They consider the future of the crowd work processes, crowd computation, and crowd workers. After this, they synthesize these foci into concrete recommendations for the future of the industry, including creating career ladders, improving task design and communication, and facilitating learning.

Reflections and Connections

I think that this paper looks at a very important part of crowd work. Crowd work is a very powerful tool, but if we don’t do it right, it could lead to a terrible future where work is cheepened and workers are not respected. We need to make sure that while this industry is still young that it is shaped in a way that will allow it to be sustainable well into the future. Articles like these are important so that we can know what we need to change now, when the industry is still young so that when it grows up, it can be a great part of the economy. 

I also like that the article has a very optimistic view of crowd work. The technology is very powerful and can allow people to do things that never could have been done without it, or could only have been done at a tremendous cost. If we put in the work to make this industry work well for its users, then we will be able to use crowd work to accomplish many great things and many people will be able to use it to make ends meet or to get an extra income on their own time. Crowd work unlocks new, more efficient ways to accomplish many old goals. 

I also think that the goals laid out in this article will be very good for the industry. The idea of creating career ladders would be extremely beneficial to everyone involved. Many people in traditional jobs only try their best because of the possibility of moving up in the company. So, a job where you can’t move up will make people less invested and less motivated. The idea of introducing better communication would also be great for the industry. There will always be times when a task is poorly worded and a quick email can save hours of wasted time and effort. The facilitation of learning also seems extremely beneficial to everyone involved and would also increase the quality of tasks and results. 

Questions

  1. Considering this article was published in 2013, do you think the industry is moving in the right direction?
  2. How practical do you think these recommendations are? Are they too ambitious?
  3. Which of the recommendations do you think is most important for the industry to implement? 

Read More

01/22/2020 | PALAKH MIGNONNE JUDE | GHOST WORK

SUMMARY

‘Ghost work’ is a new world of employment that encompasses the work done behind-the-scenes by various unnamed online workers that help to power mobile apps, websites, and AI systems. These workers are the ‘humans-in-the-loop’ contributing their creativity, innovation, and rational judgement to cases where machines are unable to make decisions accurately. The importance of these ‘gigs’ has increased exponentially over the past few years – especially in order to provide better training data for Machine Learning systems (such as the ImageNet challenge that involved annotating millions of images), cleaning-up social media pages (ensuring that there is little to no abusive content), as well as providing accurate descriptions for product and restaurant reviews. This work covers various online platforms like Amazon’s Mechanical Turk, Microsoft’s UHRS, LeapGenius, and Amara – each catering to slightly varied tasks ranging from micro-tasks (tasks that can be done quickly, but require many people) to macro-tasks (larger projects such as copyediting a newsletter, linking video to captions). The work also illustrates the lives of four such ghost workers, and their experiences on these platforms, giving some insight into the ‘human’ behind the seemingly interchangeable worker represented by a pre-defined worked ID.

REFLECTION

As a researcher working with Machine Learning, I think that I often forget the ‘human-side’ of Machine Learning – especially with respect to the generation of labels for supervised tasks. I found the description about the lives of Joan, Kala, Zaffar, and Karen to be particularly interesting and insightful. This was interesting to me as it helped me better understand the motivation of people engaging in crowd work and helped me better appreciate the efforts that need to be taken in order to gain a good ‘gold standard’. I also found it interesting to learn that on-demand workers produce higher-quality work when compared to the quality of work produced by full-time employees. While the reading posits that a potential reason for this is the competitive nature of on-demand jobs, I wonder if incentive-based bonuses in the case of full-time employees could have an impact on the quality of work provided by these employees.

The reading reminded me of another paper entitled ‘Dirty Jobs: The Role of Freelance Labor in Web Service Abuse’ which focusses on freelance work done via Freelancer.com. This paper discusses the abuse of various web services including solving CAPTCHAs, Online Social Networks Linking, etc. This reading talks about vetting processes done for the workers, but it made me wonder about the type of vetting done for ‘requesters’ and the moderation of the type of tasks that are posted on such ghost work platforms.

QUESTIONS

  • What is the main motivation for ghost workers with a Master’s degree to work on these platforms? Is it generally due to an ailing family member? Additionally, Amara has a larger percentage of women, is there any reason behind this?
  • In the case of a platform such as Amara, where workers perform tasks in teams, how do they handle cases of harassment (if any were to occur)? Are they any policies that exist to deal with such situations?
  • Is there any form of moderation of the type of tasks that are posted on these sights? Are the ghost workers allowed to flag tasks that might seem to contribute to web service abuse? (For example, multiple account creation by using ghost workers to solve CAPTCHAs in real-time).

Read More

01/22/2020 – Lee Lisle – Ghost Work

Summary

            Ghost work’s introduction and first chapter cover a birds-eye look at the new gig economy of crowd intelligence tasks. They cover several anecdotes of people working on these tasks in various situations, mostly in the United States and India. The introduction also wants to get across that these types of tasks are for problems that AI can’t solve or needs training to be able to solve. The text also tries to impart that there is nothing to fear from AI – automation has always happened through new technologies and that there will always be more work generated by the blind spots of the newer technologies. The first chapter then goes through several different scenarios for these workers, starting with the worst working conditions and then working up to the “best.” Lastly, it pointed out that there are possible moral issues with the whole setup, using a lawsuit of workers for a specific company arguing that they were essentially being paid minimum wage working full time with no benefits.

Personal Reflection

            I thought it was interesting to better understand how these gig workers came into being and why they’re needed.  However, I couldn’t stop thinking about the human element. Yes, the text seems to drive you towards that direction, but it’s not until the last 2-3 pages of the book that it ever actually asks the question “Is this right?” The first few specific anecdotes in the first chapter were chilling – these were people with not just undergraduate degrees but post-graduate degrees who were working for a paycheck that put them below the poverty line. Arguably only the last 2 companies mentioned, Upwork and Amara, were even close to acceptable living conditions. LeadGenius was close, as there were tiers and “promotions” that could be earned through working, but they still seemed to pay very little for quality work. MTurk being the worst really outshone the others.  A talented worker as the example was, she was only earning $16,000 a year working (according to the intro) 10 hours a day, and she was happy it was better than the $4,400 she earned the first year. The text even (insultingly) points out $4400 is more than earning $0. Futhermore, working as an mTurk required additional work to figure out what good HITs would open and what requesters to avoid as well as learn the tips and tricks of the trade. At least in a Starbucks you get paid the full amount during your training. This text and the whole ghost work gig-economy industry feels like share-cropping, where the workers are cheated out of their proper valuation.

Questions

  1. How could the ghost-work/gig-economy be regulated? Is self-regulation as shown by the mechanical turk forums and reddits enough?
  2. Now knowing what life is like for these workers, could you ethically use this service? Are the rosy-stories of “I can fill in gaps on my resume” or “I can’t work standard hours so the flexibility is nice” or “I have another source of income so this is just free money” enough to counteract the underpayment of these workers?
  3. Which of the businesses that setup gig-workers seems like the best tradeoff for requesters and workers? Why?
  4. What do you think about the idea of having to screen “employers” on Mechanical Turk? How can this impact the pay rate?

Read More

01/22/2020 – Bipasha Banerjee – Ghost Work

Paper: Ghost Work, How to Stop Silicon Valley from Building a New Global Underclass

Summary: 

The book highlights one of the major players of the artificial intelligence realm that is the “people.” By people, we often think about the customers. However, ghost workers are people in the system who work behind the scenes. These are on-demand workers who are working to keep the system running without any hiccups. The work is unique for these people. Any company who needs a certain task that requires humans to complete, they can request for such services. The task can be anything from flagging adult content, verifying human credentials, labeling images to create training data. On-demand workers are individuals who are not considered full-time or hourly wage workers. They are paid according to the task at hand. Hence, they need to be vigilant to take up tasks as soon as they appear and not lose the job to others who are looking. Tasks can be classified as a micro and macro task. Micro-tasks are small tasks that take less amount of time, but lots of humans to complete the task. Macro-task, on the other hand, are larger projects like developing webpages or building apps. Some common platforms of on-demand works are Amazon Mechanical Turks, Universal Human Relevance System, Lead genius, and Amara.

Reflection 

The topic was fascinating and gave a unique perspective on “gig-workers.” I was aware of Amazon Mechanical Turks and that humans were an integral part of the automation process. However, this gave a more in-depth idea of what role humans play in this. It was intriguing to know how such individuals support themselves and their families. However, the lack of regulated law makes it difficult to accurately estimate how much such people earn and how exactly do they support their lives. 

One important thing to note here is for all of the companies about most of the workers were in the age-group 18-37, and most had a college degree. This means that younger people find it comfortable to use modern technology to depend on such a unique method of earning money. The workers are generally not paid well enough, and they often need to be extremely skilled to make this their sole livelihood. 

The book-chapters made it very clear that in various ways, such employees are an integral part of the technical system, and they are, at times, better performers than regular employees. Does this mean that companies can shift their work model from hiring more and more full-time employees to move to the on-demand model completely? That is still debatable. I do not think that such a model would be liable when it comes to long term commitments to deliver projects. Although, for full-time employees, job security might be the reason they are comfortable. However, having no financial guarantee, in the long run, is bound to be detrimental for a person. In my opinion, the need to perform “flawlessly” would cause people to reach their breaking point eventually. I firmly believe that such roles make people live an unhealthy work-life balance as they would continuously search for tasks to get more gigs.

Questions

  1. Can use more “gig workers” prove to be efficient/profitable for companies?
  2. Other than personal reasons, what makes people take up such jobs? Can someone aspire for such roles, or necessity drives them?
  3. Would older people be able to do such tasks? If yes, how can we measure the efficiency and compare the productivity against the average workforce age of 18-37?

Read More

01/22/20 – Vikram Mohanty – Ghost Work

Summary

In the opening chapter of this book, the authors introduce the world of ghost workers, or gig workers, who have been thriving, almost unnoticed, in the shadows of modern-day software. Here, the authors give us an insight about how the gig economy has grown over the years, in terms of number of workers, where they come from, the different kinds of on-demand platforms and the features they offer, and the different kinds of jobs on these platforms. The highlight of these chapters is how these gig workers sit at the core of Artificial Intelligence (AI), either by providing training labels to build the AI models, or by filling when AI falls short of the job. The authors break down the hype of robots rising, and present the last-mile paradox of AI i.e. the race towards automation will never converge, nor take away all our jobs, but will result in the shifting of jobs from full-time employment towards gig work. They also draw historical analogies to factory work, piece work and outsourcing. Posting tasks for gig workers, instead of hiring people full-time, seems more appealing to companies and requesters for multiple reasons including automated hiring, evaluation and payment, low costs and overhead, and higher-quality work in a short time. The authors discuss the future of employment, which would look more like a world of ghost workers and not robots, and therefore, make the case for focusing our attention on improving this world of on-demand work.

Reflection

  1. The hype about AI has been majorly propelled by our media [3]. Companies, especially the rising number of AI start-ups, are partly to blame, as they have to portray themselves as an AI start-up for getting investors onboard. As they explore complex open-ended problems, these standard AI systems are almost certainly bound to fall short, leaving room for the “last-mile” to be filled in by on-demand online workers [1].
    1. Essentially, the efforts of these ghost workers are passed off as the genius of AI, without any credit, and thus propels the hype further. This adds on to the “algorithmic cruelty” of the ghost work platform APIs. 
    2. Sometimes, companies acknowledge the contributions of the experts who are responsible for training and shaping their AI models, but fall short of acknowledging the inevitable presence of these human experts in a predominantly AI workflow [2]. This, once again, can be attributed towards a need to portray themselves as an AI company. 
  2. Instead of the usual rhetoric about AIs taking over jobs, this book paints a realistic portrait about the future of employment, which involves a shift towards on-demand crowd work. This necessitates the need to redesign on-demand infrastructure with a goal of enhancing the quality of workers’ lives (illustrated in great detail in the later chapters). The chapters briefly discuss organizational hierarchy and worker collaboration, two important factors contributing towards LeadGenius’s success. These factors are also echoed by Kittur et al [4] as crucial for enhancing the value and meaning of crowd work. 
    1. It’s good to see politicians addressing issues related to gig economy [6, 7] and AI disrupting the traditional workplace [8]. 
  3. The gig workers need to be hypervigilant in order to get more work, and earn more money. Employers no longer need to actively assign jobs to employees. Therefore, the cost is now borne through a worker’s time and effort to seek and eventually get the job. This is one area where algorithms could help by connecting workers to jobs on these on-demand platforms [9]. Going one step further, ML models can be used to automatically assign workers to tasks where they can contribute [4].  
  4. The AI hype will inevitably blow up, and result in AI-infused systems to invariably rely on human intelligence for complementing (and completing) the shortcomings of AI. Human intelligence, employed either by on-demand labor or any other source, is an invaluable resource, and therefore, shouldn’t suffer the brunt of algorithmic cruelty. To address this, we need human-centered design of tools and platforms. Jeff Bigham, in his “The Coming AI Autumn” post, points out some areas where HCI research and practice could help in designing intelligent systems and making them useful for people. 
  5. These chapters illustrate many real-world examples where humans are needed to “backfill decision-making with their broad knowledge of the world” to make up for the AI’s limitations e.g. spike in search terms during a disaster, identifying hate speech, finding a great springtime wedding venue, face verification. Partially to blame is the predominant use of artificial/synthetic datasets for training AIs, which calls for designing problems rooted in the real world. 

Questions

  1. Let’s say, we have an organization with a human workforce using an AI system for solving a certain problem. The whole workflow involves a feedback channel from the human experts that is also used to train certain aspects of the AI systems, which going forward, may reduce the role of these human experts. Some of these experts may be AI critics as well. Should this organization be collecting data from these experts? If so, how should the workflow be designed? What are some of trade-offs?
  2. The book chapters raises some interesting points about global labor arbitrage and localization of data. Most of the AI systems being built are almost always deficient of data coming in from places/regions/countries with lower labor costs, and therefore, may be biased towards non-US data (face recognition, speech translation, etc.). Why is that the case? How should this be addressed?
  3. The API isn’t designed to listen to Ayesha”. (How does Ghost Work Work?) Has anyone been on the receiving end of algorithmic cruelty? What kind of systems or intelligent user interfaces did you wish for, if any? 
  4. How should journalists cover AI? How should AI claims be fact-checked? 

References

  1. The rise of ‘pseudo-AI’: how tech firms quietly use humans to do bots’ work https://www.theguardian.com/technology/2018/jul/06/artificial-intelligence-ai-humans-bots-tech-companies
  2. AI at the Speed of Real Time: Applying Deep Learning to Real-time Event Summarization https://www.dataminr.com/blog/ai-at-the-speed-of-real-time-applying-deep-learning-to-real-time-event-summarization
  3. An Epidemic of AI Misinformation https://thegradient.pub/an-epidemic-of-ai-misinformation/
  4. Kittur, A., Nickerson, J. V., Bernstein, M., Gerber, E., Shaw, A., Zimmerman, J., … & Horton, J. (2013, February). The future of crowd work. In Proceedings of the 2013 conference on Computer supported cooperative work (pp. 1301-1318).
  5. AI and automation will disrupt our world — but only Andrew Yang is warning about it https://thehill.com/opinion/technology/469750-ai-and-automation-will-disrupt-our-world-but-only-andrew-yang-is-warning
  6. Elizabeth Warren Takes On the ‘Gig Economy’ https://www.thenation.com/article/elizabeth-warren-takes-on-the-gig-economy/
  7. Pete Buttigieg just called out Uber and McDonald’s for their treatment of workers — and said beefing up unions is the best way to protect them https://www.businessinsider.com/pete-buttigieg-plan-to-overhaul-the-gig-economy-2019-7
  8. The Coming AI Autumn https://jeffreybigham.com/blog/2019/the-coming-ai-autumnn.html
  9. Prolific. https://www.prolific.co/

Read More

01/28/2020 | Palakh Mignonne Jude | The Future of Crowd Work

SUMMARY

This paper aims to define the future of crowd work in an attempt to ensure that future crowd workers will share the same benefits as those currently shared by full-time employees. The authors define a framework keeping in mind various factors such as workflow, assignment of tasks, real-time response to tasks, etc. The future that the paper envisions includes worker considerations such as providing timely feedback, and job motivation as well as requester considerations such as quality assurance and control, task decomposition. The research foci mentioned in the paper broadly consider the future of work processes, integration of crowd work and computation, supporting the crowd workers of the future in terms of job design, reputation and credentials, motivation and rewards. With respect to the future of crowd computation, the paper suggests a hybrid human-computer system that would capitalize on the best of both human intelligence and machine intelligence. The authors mention two such strategies – crowds guiding AI and AIs guiding crowds.  As a set of future steps that can be undertaken to ensure environment for crowd workers, the authors describe three design goals – creation of career ladders, improving task design through better communication, facilitating learning.

REFLECTION

I found it interesting to learn about the framework proposed by the authors in order to ensure a better working environment in the future for crowd workers. I like the structure of paper wherein the authors mentioned a brief description about the research foci followed by some prior work and then some potential research that can be performed in each of these foci.

I particularly liked the set of steps that the authors proposed – such as the creation of a career ladder. I believe that the creation of such a ladder, will help workers stay motivated as they will have the ability to work towards a larger goal as promotions can be a good incentive to foster a better and more efficient working environment. I also found it interesting to learn how often times, the design of the tasks cause ambiguity which makes it difficult for the crowd workers to perform their tasks well. I think that having sample tests of these designs with some of the better performing workers (as indicated in the paper) is a good idea as it will allow the requesters to gain feedback on their task design since many of the requesters may not realize that these tasks are not as easy to understand as they might believe.

QUESTIONS

  1. While talking about crowd-specific factors, the authors mention how crowd workers can leave tasks incomplete with fewer repercussions as compared to traditional organizations. Perhaps having a common reputation system in order to maintain a history of employment (associated with some common ID) in order to maintain recommendation letters, work histories might help to keep track of all the platforms with which a crowd worker was associated as well as their performance?
  2. Since the crowd workers interviewed were from Amazon Mechanical Turk alone, wouldn’t the responses collected from the workers as part of this study be biased? The opinion these workers would give would be specific to AMT alone and these opinions might be different among workers that are part of different platforms.
  3. Do any of these platforms perform a thorough vetting for the requesters? Have any measures been taken to move towards the development of a better system in order to ensure that the tasks posted by requesters are not harmful/abusive in nature (CAPCTHA solving, reputation manipulation, etc)?

Read More

01/29/2020 – Bipasha Banerjee – An Affordance-Based Framework for Human Computation and Human-Computer Collaboration

Summary

The paper elaborates on an affordance-based framework for human computation and human-computer collaboration. It was published in 2012 in IEEE Transactions on Visualization and Computer Graphics. Affordances is defined as “opportunities provided to an organism by an object or environment”. They reviewed 1271 papers on the area and formed a collection of 49 documents that have state-of-the-art research work. They have grouped them into machine and human based affordances.

In human affordance, they talk about all the skills that humans have to offer namely, visual perception, visuospatial thinking, audiolinguistic ability, sociocultural awareness, creativity and domain knowledge. In machine affordances they discussed about large-scale data manipulation, collecting and storing large amount of data, effective data movement, bias-free analysis. There also is a separate case where a system makes use of multiple affordances like the reCAPTCHA and the PatViz projects. They have included some possible extensions that include human adaptability and machine sensing. The paper also describes the challenges in measuring complexity of visual analytics and the best way to measure work.

Reflection

Affordance is a new concept to me. It was interesting how the authors defined human vs machine affordance-based system along with systems that make use of both. Humans have a special ability that outperforms machines that is creativity and comprehension. Nowadays, machines have the capability to classify data, but this requires a lot of training samples. Recent neural network-based architectures are “data hungry” and using such system are extremely challenging when proper labelled data is lacking. Additionally, humans have a good capability of perception, where distinguishing audio, images, video are easy for them. Platforms like Amara do take advantage of this and employ crowd-workers to caption a video. Humans are effective when it comes to domain knowledge. Jargons specific to a community e.g., chemical names, legal domain, medical domains are difficult for machines to comprehend. Named entity recognizers help machines in this aspect. However, the error is still high. The paper does succeed in highlighting the positives of both systems. Humans are good in various aspects as mentioned before but are often prone to error. This is where machines outperform humans and can be used effectively by systems. Machines are good when dealing with a large quantity of data. Machine-learning based algorithms are useful to classify, cluster data or other services as necessary. Additionally, not having perception acts as a plus as humans do tend to get influenced from certain opinion. If it is a task that require political angle, it would be extremely difficult for humans to have an-unbiased opinion. Hence, both humans and machines have a unique advantage over the other. It is the task of the researcher to utilize them effectively.

Questions

  1. How to effectively decide which affordance is the best for the task at hand? Human or machine?
  2. How to evaluate the effectiveness of the system? Is there any global evaluation metric that can be implemented?
  3. When using both the systems how to separate task effectively?

Read More

01/28/2020 | Palakh Mignonne Jude | Beyond Mechanical Turk: An Analysis Of Paid Crowd Work Platforms

SUMMARY

In this work, the authors perform a study that extends to crowd work platforms beyond Amazon’s Mechanical Turk. They state that this is the first research that has attempted to perform a more thorough study of the various crowd platforms that exist. Given that prior work has mainly focused on Mechanical Turk, a large number of the issues faced by both requesters and workers has been due to the idiosyncrasies associated with this platform in particular. Thus, the authors aim to broaden the horizon for crowd work platforms in general and present a qualitative analysis of various platforms such as ClickWorker, CloudFactory, CrowdComputing Systems, CrowdFlower, CrowdSource, MobileWorks, and oDesk. The authors identify the key criteria to distinguish between the various crowd platforms as well as identify key assumptions in crowd sourcing that maybe caused due to a narrow vision of AMT.

The limitations of AMT, as described by the authors, include inadequate quality control (caused due to a lack of gold standards, lack of support for complex tasks), inadequate management tools (caused due to lack of details about a worker’s skills, expertise; lack of focus on worker ethics and conditions), missing support to detect fraudulent activities, and a lack of automated tools for routing of tasks. In order to compare and assess the seven platforms selected for this study, the authors focus on four broad categories – quality concerns, poor management tools, missing support to detect fraud, and lack of automated tools. These broad categories further map to various criteria such as identifying distinguishing features of each platform, identifying if the crowd platform maintains its own workforce or relies on other sources for its workers as well as if the platform allows for/offers a private workforce, the amount and type of demographic information provided by the platform, platform support for routing of tasks, support for effective and efficient communication among workers, the incentives provided by the platform, organizational structures and processes for quality assurance, existence of automated algorithms to help human workers, and the existence of an ethical environment for workers.

REFLECTION

I found it interesting to learn that prior research in this field was done mainly using AMT. I agree that the research that was performed with only AMT as the crowd platform would have led to conclusions that were biased due to this narrow vision of crowd platforms in general. I believe that the qualitative analysis performed by this paper is an important contribution to this field at large as it will help future researchers to select a platform that is best suited for their task at hand due to a better awareness of the distinguishing features of each of the platforms considered in this paper. I think that the analogy about the Basic programming language aptly describes the motivation for the study performed in this paper.

I also found the categories selected by the authors to be interesting and relevant for requesters when they are considering a platform to be chosen. However, I think that it may have also been interesting (as a complementary study) for the authors to have included information about reasons why a crowd worker may join a certain platform – this would give a more holistic perspective and an insight into these crowd working platforms beyond Mechanical Turk. For example, the book on Ghost Work, included information about platforms such as Amara, UHRS, and LeadGenius in addition to AMT. Such a study coupled with a list of limitations of AMT from a workers’ perspective as well as a similar set of criteria for platform assessment would have been interesting.

QUESTIONS

  1. Considering that many of the other crowd platforms such as CrowdFlower (2007), ClickWorker (2005) have existed before the date of publication of this work, is there any specific reason that prior research with crowd work did not explore any these platforms? Was there some hinderance to the usage and study of these platforms?
  2. The authors mention that one of the limitations of AMT was its poor reputation system – this made me wonder why AMT did not take any measures to remedy this poor reputation system?
  3. Why is that AMT’s workforce is focused in U.S. and India? Do these different platforms have certain distinguishing factors that cause certain demographics of people to be more interested in one over the other?
  4. The paper mentions that oDesk provides payroll and health-care benefits to its workers. Does this make requesting on oDesk more expensive due to this additional cost? Are requesters willing to pay a higher fee to ensure such benefits exists for crowd workers?

Read More

01/29/2020 – Bipasha Banerjee – The Future of Crowd Work

Summary

The paper discusses crowd work and was presented at CSCW (Computer-supported cooperative work) in 2013. It proposes a framework that takes ideas from organizational behavior and distributed computing along with workers’ feedback. The authors of the paper consider the crowd sourcing platform to be a distributed system’s platform where each worker is considered to be analogous to a node in distributed system. This would help in partitioning tasks like parallel computing does. The ways shared resources can be managed, and allocation is also discussed well in this paper. The paper provides deep analysis on the kind of work crowd workers end up doing, the positives and the negatives of such work.

The paper outlines and identifies 12 research areas that form their model. This takes into account broadly, the future of crowd work processes, crowd computation and the crowd workers. Each of the broad topics addressed various subtopics from quality control to collaboration between workers. The paper also talks about how to create leaders in such systems, the importance of better communication and that learning, and assessment should be an integral part of such systems.

Reflection

It was an interesting read on the future of the crowd work. The approach to define the system as a distributed system was fascinating and a novel way to look at the problem. Workers do have a capability to act as “parallel processors” which make the system more efficient and would enable to do intensive tasks (like application development) effectively. Implementing theories from organizational behavior is interesting that it allows the system to better manage and allocate resources. The authors address various subtopics that talk about various issues in depth. It was a very informative read on where they incorporated background work on each of the research areas. I will be discussing some of the topics or problems that stood out to me.

Firstly, they spoke about processes. Assignment of work, management turns out to be a challenging task. In my opinion, a universal structure or hierarchy is not the way to go. In certain kinds of work or tasks it is needed to have a structure where hierarchy would prove to be useful. Work like software development, would benefit from a structure where the code is reviewed, and the quality is assessed by a separate person. Such work also needs a synchronous as people might have tasks dependent on each other.

Secondly, the paper discussed the future of crowd-computation. This included the discussion of AIs and how they can be used in the future to guide crowd working. AI in recent years have proved to be an important tool. Automatic text summarization can be used to help create “Gold standards”. Similarly, other NLP techniques could very well be used to extract information, annotate, summarize and provide other automatic services that can be used to integrate with the current human framework. This would create a human-in the loop system.

Lastly, the future of crowd workers is also an important topic to ponder. Crowd workers are often not compensated well. Similarly, requesters are often delivered sub-par work. The paper did mention that background verification is not always done properly for such “on-demand worker” as it is done for full-time employees from transcripts to interviews. This is a challenge. However, on-demand workers can be validated like Coursera does to validate students. They can be asked to upload documents for tasks that require specialization. This is in itself a task that can be carried out by contractors who verify documentation or create a turk job for the same.

Overall, this was an interesting read and research should be conducted in each of the areas to see how the system and work improves. It has the potential to create more jobs in the future with recruiters being able to hire people instantaneously.

Questions

  1. The authors only considered AMT and ODesk to define the framework. Would other platforms (like Amara, LeadGenuis) have greater/lesser issue which differ from the current needs?
  2. They mentioned about “oDesk Worker Diary” which takes snapshots of workers’ computer screen. How is the privacy and security addressed?
  3. Can’t credentials be verified digitally for specialized tasks?

Read More

01/29/20 – Lee Lisle – Human Computation: A Survey and Taxonomy of a Growing Field

Summary

In the paper, Quinn and Bederson reflect on the current state of human computation research and define a framework for current and future research in the field. They make sure to impart to the reader that human computation is not crowdsourcing, nor collective intelligence – rather, it is a space where human effort is used where computer may be able to solve the problems in the future. They then define several dimensions on how to classify a human computation study; these are motivation (which can include pay or altruism among others), quality control (or how the study ensures reliable results), how the study aggregates the data, what human skill is used (visual perception etc.), process order (how the tasks are deployed) and task-request cardinality (how many tasks are deployed for how many requests). Using these dimension definitions, the authors define new research areas for growth, through pointing out uncombined dimensions or by creating new dimensions to explore.

Personal Reflection

I read this paper after reading the human computation/human computer collaboration affordances survey, and it was interesting to compare and contrast the two papers for how they approached very similar problems in different ways. This paper did a good job in defining dimensions rather than research areas. It was much easier to understand how one can change the dimensions of research as a sort of toggle on how to tackle the issues they purport to solve.

Also, the beginning seemed to be on a tangent about what human computation is really defined as, but I thought this section helped considerably narrow the scope of what they wanted to define. I had thought of human computation and crowdsourcing as synonyms, so getting them separated early on was a good way of setting the scene for the rest of the paper.

Also, this paper opened my eyes to see how wide the dimensions could be. For example, while I had known of a few methods for quality control, I hadn’t realized there were so many different options.

Lastly, I am very happy they addressed the social issues (in my opinion) plaguing this field of research in the conclusion. Treating these workers as faceless mercenaries is dehumanizing at best. I wish there was a little more interaction between the two parties than there is currently, but it is being at least thought of in these survey studies.

Questions

  1. What dimension do you think has the most promising potential for new growth, and why?
  2. Do you think you can start a new research project by just choosing a set of 6 choices (1 for each dimension) and then design a project?
  3. If a project has the same collection of dimensions as another proven study, is there merit in researching it? Or should it just work?
  4. Can you think of any study that might fit under two different discrete values of the same dimension? I.E., is there a many (studies) to one dimensional value relationship, or is it many to many?

Read More