SlideShare a Scribd company logo
Classsourcing: Crowd-Based Validation
of Question-Answer Learning Objects
Jakub Šimko, Marián Šimko, Mária Bieliková,
Jakub Ševcech, Roman Burger

12.9.2013

jsimko@fiit.stuba.sk

ICCCI ’13
This talk
• How can we use crowd of students to
reinforce the learning process?
• What are the upsides and downsides of using
student crowd?
• And what are the tricky parts?
• Case of a specific method: interactive exercise
featuring text answer correctness validation
Using students as a crowd
• Cheap (free)
• Students can be motivated
– The process must benefit them
– Secondarily reinforced by teacher’s points

• Heterogeneity (in skill, in attitude)
• Tricky behavior
Example 1: Duolingo
• Learning language by translating real web
• Translations and ratings also support the learning
itself
Example 2: ALEF
• Adaptive LEarning Framework
• Students crowdsourced for highlights, tags,
external resources
Our method: motivation
• Students like online interactive exercises
– Some as a preferred form of learning
– Most as self-testing tool (used prior to exams)

• … but these are limited
– They require manually-created content
– Automated evaluation is limited for certain answer
types
• OK with (multi)choice questions, number results, …
• BAD with free text answers, visuals, processes, …

• … limited to certain domains of learning content
Method goal
• Bring-in interactive online exercise, that
1. Provides instant feedback to student
2. Goes beyond knowledge type limits
3. Is less dependent on manual content creation
Method idea
Instead of answering a question with free text,
student evaluates an existing answer…

The question-answer combination is our
learning object.
… like this:
This form of exercise
• Uses answers of student origin
– Difficult and tricky to be evaluated, thus challenging

• Enables to re-use existing answers
– Plenty of past exam questions and answers
– Plenty of additional exercises done by students

• Feedback may be provided
– By existing teacher evaluations
– By aggregated evaluations of other students (average)
Deployment
•
•
•
•
•

Integrated into ALEF learning framework
2 weeks, 200 questions (each 20 answers)
142 students
10 000 collected evaluations
Greedy task assignment
– We wanted 16 evaluations for each questionanswer (in the end, 465 reached this).
– Counter-requirement: one student can’t be
assigned with the same question for some time.
1
9
17
25
33
41
49
57
65
73
81
89
97
105
113
121
129
137

500
450
400
350
300
250
200
150
100
50
0

Some students are more motivated
than others: expect a long tail
Crowd evaluation: is the answer
correct or wrong?
• Our first thought: (having a set of individual
evaluations – values between 0 and 1):
– Compute average
– Split the interval in half
– Discretize accordingly

• … didn’t work well
– “trustful student effect”
Example of a trustful student
120
100
80
60
40
20
0
0,1

0,2

0,3

0,4

0,5

0,6

0,7

0,8

0,9

1

0

0,1

0,2 0,3 0,4 0,5 0,6 0,7
Estimated correctness (intervals)

0,8

0,9

True ratio of correct and wrong answers in the data set was 2:1
Example question and answer
Question:
“What is the key benefit of software modeling?”
Seemingly correct answer:
“We use it for communication with customers and
developers, to plan, design and outline goals”
Correct answer:
“Creation of a model cost us a fraction of the whole
thing”
Interpretation of the crowd
• Wrong answer
0

1

• Correct answer
0

1

• Correctness computation
– Average
– Threshold
– Uncertainty interval around threshold
Evaluation: crowd correctness
• We trained threshold (t) and uncertainty interval (ε)
• Resulting in precision and “unknown cases” ratios
t
0.55

ε = 0.0
79.60 (0.0)

ε = 0.05
83.52 (12.44)

ε = 0.10
86.88 (20.40)

0.60

82.59 (0.0)

86.44 (11.94)

88.97 (27.86)

0.65

84.58 (0.0)

87.06 (15.42)

91.55 (29.35)

0.70

80.10 (0.0)

88.55 (17.41)

88.89 (37.31)

0.75

79.10 (0.0)

79.62 (21.89)

86.92 (46.77)
Aggregate distribution of student
evaluations to correctness intervals
Conclusion
• Students can work as a cheap crowd, but
–
–
–
–

They need to feel benefits of their work
They abuse/spam the system, if this benefits them
Be more careful with their results (“trustful student”)
Expect long-tailed student activity distribution

• Interactive exercise with immediate
feedback, bootstrapped from the crowd
– Future work:
• Moving towards learning support CQA
• Expertise detection (spam detection)

More Related Content

What's hot (20)

PPT
academic effect of technology in math learning
bayani domingo
 
PPTX
Instructional software
GregAndrade
 
PPTX
Instructional software presentation
tdsparks3
 
PDF
MagicBox Digital Video Assessments Brochure
MagicBox
 
PPT
Fine tuning assessment for teaching and learning
Roy Williams
 
PPTX
High School Maths
Zadok Olinga
 
PPTX
Integrating Problem Solving And Educational Software
herikah
 
PPTX
Intelligent Uses of Intelligent Agents, New and Improved
D2L Barry
 
PPTX
Quality Matters and Competency-Based Education
Teresa Potter
 
PPTX
TCSiON Digital Prep Test Series
SLR-34
 
PPT
Learning-Based Evaluation of Visual Analytic Systems.
BELIV Workshop
 
PPTX
Instructional software
tbteacher77
 
PPT
M. Brown-edtech 541 Instructional software presentation
mollibrown
 
PPT
Qcl 14-v3 [cause and effect]-[nitie]_[ravali preethi]
ravali preethi
 
PPTX
Sample Problems for Common Core State Standard - K.CC
Priyanka Reddy
 
PPTX
eTest Solution
Srdjan Verbić
 
PPTX
Module 5 integrating technology into the curriculum
cbgill38
 
PPTX
Accessing student performance by nlp
SimranAgrawal16
 
PPTX
Opening up multiple choice - assessing with confidence
Jon Rosewell
 
PPTX
Instructional software Presentation
chadum07
 
academic effect of technology in math learning
bayani domingo
 
Instructional software
GregAndrade
 
Instructional software presentation
tdsparks3
 
MagicBox Digital Video Assessments Brochure
MagicBox
 
Fine tuning assessment for teaching and learning
Roy Williams
 
High School Maths
Zadok Olinga
 
Integrating Problem Solving And Educational Software
herikah
 
Intelligent Uses of Intelligent Agents, New and Improved
D2L Barry
 
Quality Matters and Competency-Based Education
Teresa Potter
 
TCSiON Digital Prep Test Series
SLR-34
 
Learning-Based Evaluation of Visual Analytic Systems.
BELIV Workshop
 
Instructional software
tbteacher77
 
M. Brown-edtech 541 Instructional software presentation
mollibrown
 
Qcl 14-v3 [cause and effect]-[nitie]_[ravali preethi]
ravali preethi
 
Sample Problems for Common Core State Standard - K.CC
Priyanka Reddy
 
eTest Solution
Srdjan Verbić
 
Module 5 integrating technology into the curriculum
cbgill38
 
Accessing student performance by nlp
SimranAgrawal16
 
Opening up multiple choice - assessing with confidence
Jon Rosewell
 
Instructional software Presentation
chadum07
 

Viewers also liked (8)

PPTX
Jsimko wikt2012 v01
Jakub Šimko
 
PDF
Health Affairs
TekDozDijital
 
PPTX
Html
Mohammed Awad
 
PPT
Familo prezentacja
Jakub Szotek
 
PPTX
comp2
franzneri
 
PPT
Familo prezentacja
Jakub Szotek
 
PPT
comp1
franzneri
 
PPT
Familo prezentacja
Jakub Szotek
 
Jsimko wikt2012 v01
Jakub Šimko
 
Health Affairs
TekDozDijital
 
Familo prezentacja
Jakub Szotek
 
comp2
franzneri
 
Familo prezentacja
Jakub Szotek
 
comp1
franzneri
 
Familo prezentacja
Jakub Szotek
 
Ad

Similar to Classsourcing: Crowd-Based Validation of Question-Answer Learning Objects @ ICCCI 2013 (20)

PDF
Universitat de Girona' RESEARCH | Collaborative learning
TECNIO Centre EASY & Smart Cities Master
 
PPT
Dbdm presentation updated_june_27
Maria Tazelaar
 
PDF
Crowd Teaching with Imperfect Labels
collwe
 
PDF
On Quality Control and Machine Learning in Crowdsourcing
Matthew Lease
 
PPTX
Use of various online platforms to conduct examination.pptx
Dr. Chetan Bhatt
 
PDF
Reflexive learning, socio-cognitive conflict and peer- assessment to improve ...
Franck Silvestre
 
PPTX
SERF: een gestructureerde opgavenbank met feedback voor OO (Java-)programmeer...
SURF Events
 
PDF
The College Classroom Fa15 Meeting 6: Peer Instruction
Peter Newbury
 
PDF
Pilot experience applying an active learning methodology in a Software Engine...
Grial - University of Salamanca
 
PPTX
16 tarek18
afacct
 
PPTX
ACB4 tec pre - p4 - presenting a technical paper
Nisansa de Silva
 
PPT
Cognitive Apprenticeship in Accounting Education
Nona Press
 
PDF
Innovative-Teaching-Methods.pdf
Dr. Manjunatha. P
 
PPTX
Cit 2013 - Badging / Micro-credentialling
Eileen O'Connor
 
PPT
Agilemind presentation
teufelsdroch
 
PDF
ACM ITICSE 2014 - Talk on Motivational Active Learning
Johanna Pirker
 
PDF
Collaborative Projects And Self Evaluation Within A Social Reputation Based E...
MegaVjohnson
 
KEY
Agile2011
Caelum
 
PDF
Setsuya Kurahashi: Teaching Simulation on Collaborative Learning, Ability Gro...
Saxion University of Applied Sciences
 
PDF
Learning theories
William McIntosh
 
Universitat de Girona' RESEARCH | Collaborative learning
TECNIO Centre EASY & Smart Cities Master
 
Dbdm presentation updated_june_27
Maria Tazelaar
 
Crowd Teaching with Imperfect Labels
collwe
 
On Quality Control and Machine Learning in Crowdsourcing
Matthew Lease
 
Use of various online platforms to conduct examination.pptx
Dr. Chetan Bhatt
 
Reflexive learning, socio-cognitive conflict and peer- assessment to improve ...
Franck Silvestre
 
SERF: een gestructureerde opgavenbank met feedback voor OO (Java-)programmeer...
SURF Events
 
The College Classroom Fa15 Meeting 6: Peer Instruction
Peter Newbury
 
Pilot experience applying an active learning methodology in a Software Engine...
Grial - University of Salamanca
 
16 tarek18
afacct
 
ACB4 tec pre - p4 - presenting a technical paper
Nisansa de Silva
 
Cognitive Apprenticeship in Accounting Education
Nona Press
 
Innovative-Teaching-Methods.pdf
Dr. Manjunatha. P
 
Cit 2013 - Badging / Micro-credentialling
Eileen O'Connor
 
Agilemind presentation
teufelsdroch
 
ACM ITICSE 2014 - Talk on Motivational Active Learning
Johanna Pirker
 
Collaborative Projects And Self Evaluation Within A Social Reputation Based E...
MegaVjohnson
 
Agile2011
Caelum
 
Setsuya Kurahashi: Teaching Simulation on Collaborative Learning, Ability Gro...
Saxion University of Applied Sciences
 
Learning theories
William McIntosh
 
Ad

Recently uploaded (20)

PPTX
Simple and concise overview about Quantum computing..pptx
mughal641
 
PDF
Presentation about Hardware and Software in Computer
snehamodhawadiya
 
PPTX
AVL ( audio, visuals or led ), technology.
Rajeshwri Panchal
 
PDF
How Open Source Changed My Career by abdelrahman ismail
a0m0rajab1
 
PDF
Researching The Best Chat SDK Providers in 2025
Ray Fields
 
PPTX
Applied-Statistics-Mastering-Data-Driven-Decisions.pptx
parmaryashparmaryash
 
PDF
State-Dependent Conformal Perception Bounds for Neuro-Symbolic Verification
Ivan Ruchkin
 
PDF
Data_Analytics_vs_Data_Science_vs_BI_by_CA_Suvidha_Chaplot.pdf
CA Suvidha Chaplot
 
PDF
Peak of Data & AI Encore - Real-Time Insights & Scalable Editing with ArcGIS
Safe Software
 
PDF
Tea4chat - another LLM Project by Kerem Atam
a0m0rajab1
 
PDF
Build with AI and GDG Cloud Bydgoszcz- ADK .pdf
jaroslawgajewski1
 
PPTX
What-is-the-World-Wide-Web -- Introduction
tonifi9488
 
PDF
Make GenAI investments go further with the Dell AI Factory
Principled Technologies
 
PDF
GDG Cloud Munich - Intro - Luiz Carneiro - #BuildWithAI - July - Abdel.pdf
Luiz Carneiro
 
PDF
Research-Fundamentals-and-Topic-Development.pdf
ayesha butalia
 
PDF
Market Insight : ETH Dominance Returns
CIFDAQ
 
PDF
Brief History of Internet - Early Days of Internet
sutharharshit158
 
PDF
CIFDAQ's Market Wrap : Bears Back in Control?
CIFDAQ
 
PDF
MASTERDECK GRAPHSUMMIT SYDNEY (Public).pdf
Neo4j
 
PPTX
Farrell_Programming Logic and Design slides_10e_ch02_PowerPoint.pptx
bashnahara11
 
Simple and concise overview about Quantum computing..pptx
mughal641
 
Presentation about Hardware and Software in Computer
snehamodhawadiya
 
AVL ( audio, visuals or led ), technology.
Rajeshwri Panchal
 
How Open Source Changed My Career by abdelrahman ismail
a0m0rajab1
 
Researching The Best Chat SDK Providers in 2025
Ray Fields
 
Applied-Statistics-Mastering-Data-Driven-Decisions.pptx
parmaryashparmaryash
 
State-Dependent Conformal Perception Bounds for Neuro-Symbolic Verification
Ivan Ruchkin
 
Data_Analytics_vs_Data_Science_vs_BI_by_CA_Suvidha_Chaplot.pdf
CA Suvidha Chaplot
 
Peak of Data & AI Encore - Real-Time Insights & Scalable Editing with ArcGIS
Safe Software
 
Tea4chat - another LLM Project by Kerem Atam
a0m0rajab1
 
Build with AI and GDG Cloud Bydgoszcz- ADK .pdf
jaroslawgajewski1
 
What-is-the-World-Wide-Web -- Introduction
tonifi9488
 
Make GenAI investments go further with the Dell AI Factory
Principled Technologies
 
GDG Cloud Munich - Intro - Luiz Carneiro - #BuildWithAI - July - Abdel.pdf
Luiz Carneiro
 
Research-Fundamentals-and-Topic-Development.pdf
ayesha butalia
 
Market Insight : ETH Dominance Returns
CIFDAQ
 
Brief History of Internet - Early Days of Internet
sutharharshit158
 
CIFDAQ's Market Wrap : Bears Back in Control?
CIFDAQ
 
MASTERDECK GRAPHSUMMIT SYDNEY (Public).pdf
Neo4j
 
Farrell_Programming Logic and Design slides_10e_ch02_PowerPoint.pptx
bashnahara11
 

Classsourcing: Crowd-Based Validation of Question-Answer Learning Objects @ ICCCI 2013

  • 1. Classsourcing: Crowd-Based Validation of Question-Answer Learning Objects Jakub Šimko, Marián Šimko, Mária Bieliková, Jakub Ševcech, Roman Burger 12.9.2013 [email protected] ICCCI ’13
  • 2. This talk • How can we use crowd of students to reinforce the learning process? • What are the upsides and downsides of using student crowd? • And what are the tricky parts? • Case of a specific method: interactive exercise featuring text answer correctness validation
  • 3. Using students as a crowd • Cheap (free) • Students can be motivated – The process must benefit them – Secondarily reinforced by teacher’s points • Heterogeneity (in skill, in attitude) • Tricky behavior
  • 4. Example 1: Duolingo • Learning language by translating real web • Translations and ratings also support the learning itself
  • 5. Example 2: ALEF • Adaptive LEarning Framework • Students crowdsourced for highlights, tags, external resources
  • 6. Our method: motivation • Students like online interactive exercises – Some as a preferred form of learning – Most as self-testing tool (used prior to exams) • … but these are limited – They require manually-created content – Automated evaluation is limited for certain answer types • OK with (multi)choice questions, number results, … • BAD with free text answers, visuals, processes, … • … limited to certain domains of learning content
  • 7. Method goal • Bring-in interactive online exercise, that 1. Provides instant feedback to student 2. Goes beyond knowledge type limits 3. Is less dependent on manual content creation
  • 8. Method idea Instead of answering a question with free text, student evaluates an existing answer… The question-answer combination is our learning object.
  • 10. This form of exercise • Uses answers of student origin – Difficult and tricky to be evaluated, thus challenging • Enables to re-use existing answers – Plenty of past exam questions and answers – Plenty of additional exercises done by students • Feedback may be provided – By existing teacher evaluations – By aggregated evaluations of other students (average)
  • 11. Deployment • • • • • Integrated into ALEF learning framework 2 weeks, 200 questions (each 20 answers) 142 students 10 000 collected evaluations Greedy task assignment – We wanted 16 evaluations for each questionanswer (in the end, 465 reached this). – Counter-requirement: one student can’t be assigned with the same question for some time.
  • 13. Crowd evaluation: is the answer correct or wrong? • Our first thought: (having a set of individual evaluations – values between 0 and 1): – Compute average – Split the interval in half – Discretize accordingly • … didn’t work well – “trustful student effect”
  • 14. Example of a trustful student 120 100 80 60 40 20 0 0,1 0,2 0,3 0,4 0,5 0,6 0,7 0,8 0,9 1 0 0,1 0,2 0,3 0,4 0,5 0,6 0,7 Estimated correctness (intervals) 0,8 0,9 True ratio of correct and wrong answers in the data set was 2:1
  • 15. Example question and answer Question: “What is the key benefit of software modeling?” Seemingly correct answer: “We use it for communication with customers and developers, to plan, design and outline goals” Correct answer: “Creation of a model cost us a fraction of the whole thing”
  • 16. Interpretation of the crowd • Wrong answer 0 1 • Correct answer 0 1 • Correctness computation – Average – Threshold – Uncertainty interval around threshold
  • 17. Evaluation: crowd correctness • We trained threshold (t) and uncertainty interval (ε) • Resulting in precision and “unknown cases” ratios t 0.55 ε = 0.0 79.60 (0.0) ε = 0.05 83.52 (12.44) ε = 0.10 86.88 (20.40) 0.60 82.59 (0.0) 86.44 (11.94) 88.97 (27.86) 0.65 84.58 (0.0) 87.06 (15.42) 91.55 (29.35) 0.70 80.10 (0.0) 88.55 (17.41) 88.89 (37.31) 0.75 79.10 (0.0) 79.62 (21.89) 86.92 (46.77)
  • 18. Aggregate distribution of student evaluations to correctness intervals
  • 19. Conclusion • Students can work as a cheap crowd, but – – – – They need to feel benefits of their work They abuse/spam the system, if this benefits them Be more careful with their results (“trustful student”) Expect long-tailed student activity distribution • Interactive exercise with immediate feedback, bootstrapped from the crowd – Future work: • Moving towards learning support CQA • Expertise detection (spam detection)