Academic Crowdsourcing in the Humanities
eBook - ePub

Academic Crowdsourcing in the Humanities

Crowds, Communities and Co-production

  1. 190 pages
  2. English
  3. ePUB (mobile friendly)
  4. Available on iOS & Android
eBook - ePub

Academic Crowdsourcing in the Humanities

Crowds, Communities and Co-production

About this book

Academic Crowdsourcing in the Humanities lays the foundations for a theoretical framework to understand the value of crowdsourcing, an avenue that is increasingly becoming important to academia as the web transforms collaboration and communication and blurs institutional and professional boundaries. Crowdsourcing projects in the humanities have, for the most part, focused on the generation or enhancement of content in a variety of ways, leveraging the rich resources of knowledge, creativity, effort and interest among the public to contribute to academic discourse. This book explores methodologies, tactics and the "citizen science" involved.- Addresses crowdsourcing for the humanities and cultural material- Provides a systematic, academic analysis of crowdsourcing concepts and methodologies- Situates crowdsourcing conceptually within the context of related concepts, such as 'citizen science', 'wisdom of crowds', and 'public engagement'

Frequently asked questions

Yes, you can cancel anytime from the Subscription tab in your account settings on the Perlego website. Your subscription will stay active until the end of your current billing period. Learn how to cancel your subscription.
No, books cannot be downloaded as external files, such as PDFs, for use outside of Perlego. However, you can download books within the Perlego app for offline reading on mobile or tablet. Learn more here.
Perlego offers two plans: Essential and Complete
  • Essential is ideal for learners and professionals who enjoy exploring a wide range of subjects. Access the Essential Library with 800,000+ trusted titles and best-sellers across business, personal growth, and the humanities. Includes unlimited reading time and Standard Read Aloud voice.
  • Complete: Perfect for advanced learners and researchers needing full, unrestricted access. Unlock 1.4M+ books across hundreds of subjects, including academic and specialized titles. The Complete Plan also includes advanced features like Premium Read Aloud and Research Assistant.
Both plans are available with monthly, semester, or annual billing cycles.
We are an online textbook subscription service, where you can get access to an entire online library for less than the price of a single book per month. With over 1 million books across 1000+ topics, we’ve got you covered! Learn more here.
Look out for the read-aloud symbol on your next book to see if you can listen to it. The read-aloud tool reads text aloud for you, highlighting the text as it is being read. You can pause it, speed it up and slow it down. Learn more here.
Yes! You can use the Perlego app on both iOS or Android devices to read anytime, anywhere — even offline. Perfect for commutes or when you’re on the go.
Please note we cannot support devices running on iOS 13 and Android 7 or earlier. Learn more about using the app.
Yes, you can access Academic Crowdsourcing in the Humanities by Mark Hedges,Stuart Dunn in PDF and/or ePUB format, as well as other popular books in Social Sciences & Sociology. We have over one million books available in our catalogue for you to explore.

Information

Chapter 1

Introduction

Academic crowdsourcing from the periphery to the centre

Abstract

This chapter gives a brief introduction to the volume, and an historical overview of academic crowdsourcing. Beginning with the definition of crowdsourcing which we first proposed in 2012, The origins of academic crowdsourcing are outlined, and we consider the arguments as to whether it can be said to have evolved from a means of producing digital resources to a research methodology. We discuss the basis of the research drawn upon in the following chapters, the material gathered and the methodologies used to interrogate it.

Keywords

Academia; Connectivity; Engagement; Methodology; Social media

Introduction

Crowdsourcing is the process of leveraging public participation in or contributions to projects and activities. It has become a familiar term, and a concept that has gained increasing attention in many spheres over the last decade. Government, industry, and commercial enterprises are developing crowdsourcing practices as a means to engage their audiences and readerships, to improve and enrich their own data assets and services, and to address supposed disconnects between the public and professional sectors (Boudreau & Lakhani, 2013). At a time when the Web is simultaneously transforming the way in which people collaborate and communicate, and merging the spaces that the academic and nonacademic communities inhabit, it has never been more important to consider the role that public communities – connected or otherwise – have come to play in academic humanities research. Public involvement in the humanities can take many forms – transcribing handwritten text into digital form; tagging photographs to facilitate discovery and preservation; entering structured or semi-structured data; commenting on content or participating in discussions; or recording one’s own experiences and memories in the form of oral history – and the relationship between the public and the humanities is convoluted and poorly understood.
This book explores this diverse field, and focuses on crowdsourcing as a research method. We consider where, in purely semantic terms, the boundaries of what is considered to be academic crowdsourcing should lie. Since humanities crowdsourcing is at an emergent stage as a research method, there is a correspondingly emergent field of academic literature dealing with its application and outcomes, which allows some assessments to be made about its potential to produce academically credible knowledge. The problematization of method, academic credibility, value and knowledge outputs is familiar from the history of the Digital Humanities. In 2002, Short and McCarty proposed a ‘methodological commons’, common ways of doing things that linked subject areas, digital research methods and domains: ‘computational techniques shared among the disciplines of the humanities and closely related social sciences, e.g., database design, text analysis, numerical analysis, imaging, music information retrieval, communications’ (McCarty, 2003). In McCarty’s terms, this commons formed a combination of ‘collegial service’ and ‘research enterprise’ which both made provision for existing research activities, and expanded them. The principal purpose of this book is to develop a similar ‘methodological commons’ for academic crowdsourcing. We contend that just as (say) the application of text processing technologies in history enhances the study of history and provokes new questions about the past, and can inform the development of processing technologies for (again, say) music; so can methods of leveraging public participation in museums form and relate to participation elsewhere in the humanities. What is needed is a demarcation of the kinds of material involved, the ‘assets’, the types of task available to the public, the processes that undertaking those tasks involve, and the outputs. In other words, we seek to apply to crowdsourcing in academia the kind of formal structure of value and review that crowdsourcing has implicitly acquired in many other domains.
Academia, however, has always been something of a special case. It is worth spending some time reflecting on why this is so. Long before crowdsourcing was ever known by that name, researchers in especially the natural sciences were engaging in ‘citizen science’, a set of practices in which unpaid volunteers provided input to professionally coordinated research projects. This has been going on in domains such as field ecology, conservation and habitat studies since at least the 17th century, when in any case the role of professional scientist did not exist, at least in its 21st century form (Miller-Rushing, Primack, & Bonney, 2012). Networks, collaborations and codependencies developed within and across professional boundaries, leading to the production of original knowledge that passed all the thresholds of academic peer review and credibility.
The most significant changes to these networks and collaborations can be traced to the mid- and late 2000s. The Galaxy Zoo project, for example, one of the largest and most successful citizen science projects, and to which we return later, was launched on July 11, 2007, with the Zooniverse suite of collaborations coming 2 years later. Shortly before this, in 2006, Jeffrey Howe coined the term ‘crowdsourcing’ in a now-famous article in Wired. In this, he stated:
‘All these companies grew up in the Internet age and were designed to take advantage of the networked world. It doesn’t matter where the laborers are—they might be down the block, they might be in Indonesia—as long as they are connected to the network … technological advances in everything from product design software to digital video cameras are breaking down the cost barriers that once separated amateurs from professionals. … The labor isn’t always free, but it costs a lot less than paying traditional employees. It’s not outsourcing; it’s crowdsourcing.’
Howe (2006)
This definition and its timing are critical to the thesis of this book. The year 2006 was a period when the World Wide Web was becoming ubiquitous and Hypertext was established as its main medium, and it was the time when social media and the interactive Web started to emerge. Twitter was launched in 2006, Facebook in 2008. The emergence of increasingly fluid digital networks of communication spawned crowdsourcing as both a term and a concept, and brought a range of challenges and opportunities to an academic environment already familiar with the traditions of citizen science. The Galaxy Zoo project was an early adopter, using the affordances of the Internet to engage the public in the task of classifying images of galaxies from the Sloan Digital Sky Survey – a job that is straightforward for the human eye, but impossible for even the most sophisticated automated image processing – with now-legendary success (Bamford et al., 2008). Early crowdsourcing projects in the humanities (such as Transcribe Bentham – see Chapter 3) engaged with the concept of crowdsourcing to operationalize in a similar manner tasks of a larger size and scale than was previously possible using unpaid labour, such as mass transcription tasks. Between the mid-2000s and the present day, this paradigm of academic crowdsourcing underwent a shift in perception. It is now acknowledged that it is not a ‘cheap’ alternative to paid-for labour, as suggested by the use of unpaid volunteers, and by Howe’s contextualization with outsourcing. Rather, it is a set of processes, as we argue in Chapter 3. There is merit in some of these processes being better described as ‘methodologies’ – methods, extrapolated and grounded, which allow academic teams and institutions, and bodies such as libraries, museums and archives, to function in different ways in terms of their relationships with the public. This draws explicitly on the themes of ‘citizen science’, which has long been acknowledged as a distinct set of traditions within the epistemologies of science itself (see above).

Crowdsourcing, citizen science and engagement

In citizen science, a useful distinction can be drawn between tasks which are ‘delegative’, i.e., where data is processed, digitized or otherwise enhanced, and those which are ‘democratizing’, i.e., in which participants outside the core research team are involved in setting the research agenda and asking the research questions (Savio, Prainsack, & Buyx, 2016). These are different kinds of engagement.
An immediate assumption underlying the word ‘engagement’ itself is an ontological separation of one entity in to two or more further entities: for there to be engagement, one entity must engage with another. While the 20th and 21st centuries viewed post-Victorian academia in North America and Europe as an ‘ivory tower, or Matthew Arnold’s ‘Dreaming Spires’, detached from the humdrum concerns of day-to-day life, the truth is that this disconnect has always been rather more complex. Drawing from the traditions of citizen science and public engagement discussed above, different fields of academic research have engaged with crowdsourcing in different ways. As with politics, industry and commerce, there are parts of the academic sphere which have different histories of, and motivations for, engaging with the public. The history of mass contribution to academic research and the building of the most august scholarly resources must include the Oxford Dictionary of National Biography, the British Museum’s Bronze Age Index (Bonacchi et al., 2014), as well as Wikipedia and the Zooniverse suite of citizen science projects (Simpson, Page, & De Roure, 2014), all of which represent different traditions, which are nonetheless rooted in the institutional, economic, social and political distinction between ‘professional research’ and ‘non-professional research’. A primary contention of this book, as suggested above, is that the Internet and World Wide Web (WWW) have altered all aspects of this relationship, and that a ‘methodological commons’ which articulates what crowdsourcing can bring to academia must take account of this.
However, the types (and value) of knowledge produced by academic crowdsourcing activities – by the exposure of professional academia to the Internet age – have not been widely addressed by the literature on the subject. In many cases this is because ‘academic knowledge’, in the pure sense of new understandings that can withstand professional peer scrutiny and critical analysis, has not been sought or expected by those academics consciously engaging with crowdsourcing, except inasmuch as they will write papers in the way that they do during and after every research project. Rather, the focus on crowdsourcing, certainly in the humanities, has been on the improvement and transformation of content from one type to another, the description of objects and the synthesis of information from different sources (Ridge, 2014, p. 23). This may be seen largely as a refinement of Howe’s definition concerned with value: ‘breaking down the cost barriers that once separated amateurs from professionals’. At the same time, analysts such as Darren Brabham situated it as a productive enterprise, a means of ‘doing profitable business’ (Brabham, 2008, p. 82). In many of the early instances of academic crowdsourcing of the early 2010s, it was seen as a method of approaching digitization of (very) hard-to-digitize humanities research assets (Dunn & Hedges, 2013).
However, in the same period, one of the key sectors of the humanities where public engagement is mission-critical, the cultural memory sector (museums, archives, libraries etc.), much – but not all – of the focus has been on the production of unstructured knowledge content, such as blogs, social media and user-generated content, and the seeking of feedback through social media platforms such as Twitter. This in itself has been hailed as a democratizing paradigm in these sectors (Russo, Watkins, Kelly, & Chan, 2008). However, as with crowdsourcing elsewhere in the humanities it is important to draw a distinction between democratization, and the freeing of content that would otherwise by stored with limited accessibility within an institutional framework; between the production of new (or better) systems of cataloguing and/or documentation, and the production of new knowledge based on that content. For this reason, it is also true that the purpose of academic crowdsourcing, at least in its earlier guises, has not been angled at the critical production of knowledge, as it would be understood in any of the pedagogical literature in the humanities and cultural heritage.
The idea that crowdsourcing is primarily concerned with production was implicit in the typology of Dunn and Hedges (2012), which argued that academic crowdsourcing could be conceived as a set of workflow tasks, assets, processes and outputs. The ‘Asset’ category for the typology concerned the different kinds of content with which humanists work, such as text, image, video and audio. The ‘Output’ category of the typology was described as ‘the thing an activity produces as the result of the application of a process, using tasks of a particular task type, to an asset’ (2012, p. 37). In many cases, this involved dependencies of task types: for example, an output of the type ‘structured data’ would likely be a result of a project employing the task types of ‘collaborative tagging’ or ‘linking’. This is perhaps inevitable, as the very act of structuring implies an act...

Table of contents

  1. Cover image
  2. Title page
  3. Table of Contents
  4. Series Page
  5. Copyright
  6. About the Authors
  7. Preface
  8. Acknowledgements
  9. Chapter 1. Introduction: Academic crowdsourcing from the periphery to the centre
  10. Chapter 2. From citizen science to community co-production
  11. Chapter 3. Processes and products: A typology of crowdsourcing
  12. Chapter 4. Crowdsourcing applied: Case studies
  13. Chapter 5. Roles and communities
  14. Chapter 6. Motivations and benefits
  15. Chapter 7. Ethical issues in humanities crowdsourcing
  16. Chapter 8. Crowdsourcing and memory
  17. Chapter 9. Crowds past, present and future
  18. Bibliography
  19. Index