Jane Hsu is a professor and department chair of Computer Science and Information Engineering at National Taiwan University. Her research interests include multi-agent systems, intelligent data analysis, commonsense knowledge, and context-aware computing. Prof. Hsu is the director of the Intel-NTU Connected Context Computing Center, featuring global research collaboration among NTU, Intel, and the National Science Council of Taiwan. She serves on the editorial board of Journal of Information Science and Engineering (2010-), International Journal of Service Oriented Computing and Applications (Springer, 2007-2009) and Intelligent Data Analysis (Elsevier/IOS Press, 1997-2002). She is actively involved in many key international AI conferences as organizers and members of the program committee. In addition to serving as the President of Taiwanese Association for Artificial Intelligence (2013-2014), Prof. Hsu has been a member of AAAI, IEEE, ACM, Phi Tau Phi, and an executive committee member of the IEEE Technical Committee on E-Commerce (2000) and TAAI (2004-current).
Effects of Smartphone Addiction on the Academic Performances of Grades 9 to 1...
許永真/Crowd Computing for Big and Deep AI
1. Jane Hsu
Computer Science & Information Engineering
Intel-NTU Connected Context Computing
National Taiwan University
Crowd Computing for Big and Deep AI
DRAFT
6. Key Challenges of Labelling Big Data
Scalability
Label Quality
- High cost for obtaining large amount of high quality labels
- Lack of qualified annotators
- Human bias
- Concept ambiguity
13. Wisdom of Crowds
Vox Populi by Sir Francis Galton [1907]
Nature, No. 1949,Vol. 75, 450-451
Weight Judging Competition
W of English Fat Stock & Poultry Exhibition
787 votes
Middlemost vote: 1207 lb
Actual weight: 1198 lb
15. Here is the original image Here are all clicks received Here is the consensus
NASA’s Clickworkers (2000)
- NASA showed that public volunteers can perform science tasks that
would normally require months of work by scientists or graduate
students
- During one year period (Nov. 2000~Jan. 2002), they had 101,000
clickworkers contributing 14,000 work hours, 612,832 sessions, and
2,378,820 crater entries
http://nasaclickworkers.com/classic/age-maps.html
22. “Crowdsourcing represents the act of a company or institution
taking a function once performed by employees and outsourcing
it to an undefined (and generally large) network of people in the
form of an open call.”
- Jeff Howe (2006)
28. Crowd Algorithms
Find-Fix-Verify
[Bernstein et al., 2010]
Price-Divide-Solve
[Kulkarni et al., 2012]
Iterative and Parallel Process
[Little et al., 2010]
Context Trees
[Verroios and Bernstein, 2014]
33. Ask workers: “I am motivated to do HITs on Mechanical
Turk...”
- to kill time
- to make extra money
- for fun
- because it gives me a sense of purpose
Motivations [Antin and Shaw, 2012]
34. Incentives
- Does paying more money produce better work?
- More work, but not higher-quality work
[Mason and Watts, 2009]
- Does feedbacks produce better work?
- Self-assessment and expert assessment both
improve the quality of work
[Dow et. al, 2011]
38. Crowdsourcing - Rad Lab Talk - UC Berkeley Fall 2010 30
VizWiz [Bigham et. al, 2010]
To help blind people, Vizwiz offers a new alternative to answering visual
questions in nearly real-time — asking multiple people on the web.
42. Crowd-Powered System
Soylent
Scribe
A word processor with a crowd inside
- Shortn: a text shortening service
- Crowdproof: a human-powered spelling and
grammar checker
- The Human Macro: an interface for offloading
arbitrary word processing tasks
[Bernstein et. al, 2010]
Real-time captioning by non-experts
ToolScape
Extracting step-by-step information from
how-to video
[Lasecki et. al, 2012]
[Kim et. al, 2014]
51. Mobile design mining from mockup
Navigation bar
Grid view
Tab bar
Image
Image
Image
Photo gallery
- show multiple images
tab bar
grid item grid item
image image
buttonbutton text button
image text
root
navigation bar grid view
Design structure
55. Crowd Collaboration
Individual: work individually
Sequential: work individually but perceive others’ feedbacks
Simultaneous: paired workers work simultaneously and
communicate immediately
59. Reference
[1] M. S. Bernstein, G. Little, R. C. Miller, B. Hartmann, M. S. Ackerman, D. R. Karger, D. Crowell, and K. Panovich. Soylent:
a word processor with a crowd inside. In Proceedings of the 23nd annual ACM symposium on User interface software and
technology, UIST ’10, pages 313–322, New York, NY, USA, 2010. ACM.
[2] J. P. Bigham, C. Jayant, H. Ji, G. Little, A. Miller, R. C. Miller, R. Miller, A. Tatarowicz, B. White, S. White, and T. Yeh.
Vizwiz: nearly real-time answers to visual questions. In Proceedings of the 23nd annual ACM symposium on User interface
software and technology, UIST ’10, pages 333–342, New York, NY, USA, 2010. ACM.
[3] S. Cooper, F. Khatib, A. Treuille, J. Barbero, J. Lee, M. Beenen, A. Leaver-Fay, D. Baker, Z. Popovíc, and F. players.
Predicting protein structures with a multiplayer online game. Nature, 466(7307):756–760, Aug 2010.
[4] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image database. In IEEE
Computer Vision and Pattern Recognition, CVPR ’09, pages 248–255, 2009.
[5] K. Heimerl, B. Gawalt, K. Chen, T. Parikh, and B. Hartmann. Communitysourcing: Engag- ing local crowds to perform
expert work via physical kiosks. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI
’12, pages 1539–1548, New York, NY, USA, 2012. ACM.
[6] J. Kim, P. T. Nguyen, S. Weir, P. J. Guo, R. C. Miller, and K. Z. Gajos. Crowdsourcing step-by-step information extraction
to enhance existing how-to videos. In Proceedings of the 32Nd Annual ACM Conference on Human Factors in Computing
Systems, CHI ’14, pages 4017–4026, New York, NY, USA, 2014. ACM.
[7] G. Laput, W. S. Lasecki, J. Wiese, R. Xiao, J. P. Bigham, and C. Harrison. Zensors: Adaptive, rapidly deployable,
human-intelligent sensor feeds. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing
Systems, CHI ’15, pages 1935–1944, New York, NY, USA, 2015. ACM.
[8] W. Lasecki, C. Miller, A. Sadilek, A. Abumoussa, D. Borrello, R. Kushalnagar, and J. Bigham. Real-time captioning by
groups of non-experts. In Proceedings of the 25th Annual ACM Symposium on User Interface Software and Technology,
UIST ’12, pages 23–34, New York, NY, USA, 2012. ACM.
[9] V. Verroios and M. S. Bernstein. Context trees: Crowdsourcing global understanding from local views. In Proceedings of
the Second AAAI Conference on Human Computation and Crowdsourcing (HCOMP-2014), 2014.