Skip to main content
Email Print Share


News Release 16-056

Crowd-augmented cognition

NSF early career awardee develops tools that combine human and machine intelligence to accelerate learning

illustration showing a bubble filled with faces of people


Crowdsourcing enlists the services of many people to accomplish a task.
Credit and Larger Version


May 11, 2016

Crowdsourcing has brought us Wikipedia and ways to understand how HIV proteins fold. It also provides an increasingly effective means for teams to write software, perform research or accomplish small repetitive digital tasks.

However, most tasks have proven resistant to distributed labor, at least without a central organizer. As in the case of Wikipedia, their success often relies on the efforts of a small cadre of dedicated volunteers. If these individuals move on, the project becomes difficult to sustain.

Scientists funded by the National Science Foundation (NSF) are finding new solutions to these challenges.

Aniket Kittur, an associate professor in the Human-Computer Interaction Institute at Carnegie Mellon University (CMU), designs crowdsourcing frameworks that combine the best qualities of machine learning and human intelligence, in order to allow distributed groups of workers to perform complicated cognitive tasks. Those include writing how-to guides or organizing information without a central organizer.

At the Computer-Human Interaction conference in San Jose this week, Kittur and his collaborators Nathan Hahn and Joseph Chang (CMU), and Ji Eun Kim (Bosch Corporate Research), will present two prototype systems that enable teams of volunteers, buttressed by machine learning algorithms, to crowdsource more complex intellectual tasks with greater speed and accuracy (and at a lower cost) than past systems.

"We are trying to scale up human thinking by letting people build on the work that others have done before them," Kittur said.

The Knowledge Accelerator

One piece of prototype software developed by Kittur and his collaborators, called the Knowledge Accelerator empowers distributed workers to perform information synthesis.

The software combines materials from a variety of sources, and constructs articles that can provide answers to commonly sought questions -- questions like: "How do I get my tomato plant to produce more tomatoes?" or "How do I unclog my bathtub drain?"

To assemble answers, individuals identify high-value sources from the Internet, extract useful information from those sources, cluster clips into commonly discussed topics, and identify illustrative images or video.

With the Knowledge Accelerator, each crowd worker contributes a small amount of effort to synthesize online information to answer complex or open-ended questions, without an overseer or moderator.

The researchers' challenge lies in designing a system that can divide assignments into short microtasks, each paying crowd workers $1 for 5-10 minutes of work. The system then must combine that information in a way that maintains the article flow and cohesion, as if it were written by a single author.

The researchers showed that their method produced articles judged by crowd workers as more useful than pages that were in the top five Google results from a given query. Those top Google results are typically created by experts or professional writers.

"Overall, we believe this is a step towards a future of big thinking in small pieces, where complex thinking can be scaled beyond individual limits by massively distributing it across individuals," the authors concluded.

Alloy

A related problem that Kittur and his team tackled involved clustering -- pulling out the patterns or themes among documents to organize information, whether Internet searches, academic research articles or consumer product reviews.

Machine learning systems have proven successful at automating aspects of this work, but their inability to understand distinctions in meaning among similar documents and topics means that humans are still better at the task. When human judgement is used in crowdsourcing, however, individuals often miss the full context that allows them to do the task effectively.

The new system, called Alloy, combines human intelligence and machine learning to speed up clustering using a two-step process.

In the first step, crowdworkers identify meaningful categories and provide representative examples, which the machine uses to cluster a large body of topics or documents. However, not every document can be easily classified, so in the second step, humans consider those documents that the machines weren't able to cluster well, providing additional information and insights.

The study found that Alloy, using the two-step process, achieved better performance at a lower cost than previous crowd-based approaches. The framework, researchers say, could be adapted for other tasks such as image clustering or real-time video event detection.

"The key challenge here is trying to build a big picture view when each person can only see a small piece of the whole," Kittur said. "We tackle this by giving workers new ways to see more context and by stitching together each worker's view with a flexible machine learning backbone."

On the path to knowledge

Kittur is conducting his research under an NSF Faculty Early Career Development (CAREER) award, which he received in 2012. The award supports junior faculty who exemplify the role of teacher-scholars through outstanding research, excellent education and the integration of education and research within the context of the mission of their organization. NSF is funding his work with $500,000 over five years.

The work advances the understanding and design of crowdsourcing frameworks, which can be applied to a variety of domains, he says.

"It has the potential to improve the efficiency of knowledge work, the training and practice of scientists, and the effectiveness of education," Kittur says. "Our long-term goal is to produce a universal knowledge accelerator: capturing a fraction of the learning that every person engages in every day, and making that benefit later people who can learn faster and more deeply than ever before."

-NSF-

Media Contacts
Aaron Dubrow, NSF, 703-292-4489, adubrow@nsf.gov

Principal Investigators
Aniket Kittur, Carnegie Mellon University, 412-268-7505, nkittur@cs.cmu.edu

Related Websites
Big Thinking in Small Pieces: Computer Guides Humans in Crowdsourced Research: http://www.cmu.edu/news/stories/archives/2016/may/knowledge-accelerator.html

The National Science Foundation (NSF) is an independent federal agency that supports fundamental research and education across all fields of science and engineering. In fiscal year (FY) 2017, its budget is $7.5 billion. NSF funds reach all 50 states through grants to nearly 2,000 colleges, universities and other institutions. Each year, NSF receives more than 48,000 competitive proposals for funding and makes about 12,000 new funding awards.

 Get News Updates by Email 

Useful NSF Web Sites:
NSF Home Page: https://www.nsf.gov
NSF News: https://www.nsf.gov/news/
For the News Media: https://www.nsf.gov/news/newsroom.jsp
Science and Engineering Statistics: https://www.nsf.gov/statistics/
Awards Searches: https://www.nsf.gov/awardsearch/

 

Aniket Kittur
Aniket Kittur is an associate professor in the Carnegie Mellon Human-Computer Interaction Institute.
Credit and Larger Version

graphic showing a machine and text knowledge, prioritizing, clustering, aggregating, finding, judgin
The Knowledge Accelerator uses a machine-learning program to sort and organize information.
Credit and Larger Version

diagram showing the intersection of cognition, computation and crowds
Research in Kittur's lab is on augmenting human cognition using crowds and computation.
Credit and Larger Version

screenshot fo a wikipedia page about tomatoes
The final output of the Knowledge Accelerator system.
Credit and Larger Version

graphic showing a conceptual overview of how the Alloy system works between crown and machine
A conceptual overview of the Alloy system.
Credit and Larger Version