skip to main content
10.1145/2501988.2502028acmconferencesArticle/Chapter ViewAbstractPublication PagesuistConference Proceedingsconference-collections
research-article

Crowd-scale interactive formal reasoning and analytics

Published:08 October 2013Publication History

ABSTRACT

Large online courses often assign problems that are easy to grade because they have a fixed set of solutions (such as multiple choice), but grading and guiding students is more difficult in problem domains that have an unbounded number of correct answers. One such domain is derivations: sequences of logical steps commonly used in assignments for technical, mathematical and scientific subjects. We present DeduceIt, a system for creating, grading, and analyzing derivation assignments in any formal domain. DeduceIt supports assignments in any logical formalism, provides students with incremental feedback, and aggregates student paths through each proof to produce instructor analytics. DeduceIt benefits from checking thousands of derivations on the web: it introduces a proof cache, a novel data structure which leverages a crowd of students to decrease the cost of checking derivations and providing real-time, constructive feedback. We evaluate DeduceIt with 990 students in an online compilers course, finding students take advantage of its incremental feedback and instructors benefit from its structured insights into course topics. Our work suggests that automated reasoning can extend online assignments and large-scale education to many new domains.

References

  1. Coursera support documentation. http://support.coursera.org.Google ScholarGoogle Scholar
  2. Bennett, R. E., and Bejar, I. I. Validity and automated scoring: It's not only the scoring. Educational Measurement: Issues and Practice 17, 4 (1998), 9--17.Google ScholarGoogle ScholarCross RefCross Ref
  3. Bennett, R. E., Steffen, M., Singley, M. K., Morley, M., and Jacquemin, D. Evaluating an automatically scorable, open-ended response type for measuring mathematical reasoning in computer-adaptive tests. Journal of Educational Measurement 34, 2 (1997), pp. 162--176.Google ScholarGoogle ScholarCross RefCross Ref
  4. Burstall, R. Proveeasy: Helping people learn to do proofs. In Proc. ENTCS 2000 (2000), 16--32.Google ScholarGoogle Scholar
  5. Cheang, B., Kurnia, A., Lim, A., and Oon, W.-C. On automated grading of programming assignments in an academic institution. Comput. Educ. 41, 2 (2003), 121--131. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Clavel, M., Durán, F., Eker, S., Lincoln, P., Mart-Oliet, N., Meseguer, J., and Quesada, J. Maude as a metalanguage. In Proc. WRLA 1998 15 (1998).Google ScholarGoogle Scholar
  7. Corbett, A., and Anderson, J. Knowledge tracing: Modeling the acquisition of procedural knowledge. In Proc. UMUAI 1994 (1994), 253--278.Google ScholarGoogle Scholar
  8. Corbett, A. T., and Anderson, J. R. Locus of feedback control in computer-based tutoring: impact on learning rate, achievement and attitudes. In Proc. CHI '01 (2001). Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Gallien, T., and Oomen-Early, J. Personalized versus collective instructor feedback in the online courseroom: Does type of feedback affect student satisfaction, academic performance and perceived connectedness with the instructor? International Journal on E-Learning 7, 3 (2008), 463--476.Google ScholarGoogle Scholar
  10. Hearst, M. The debate on automated essay grading. Intelligent Systems and their Applications, IEEE 15, 5 (2000), 22--37. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Heffernan, N. T., Koedinger, K. R., and Razzaq, L. Expanding the model-tracing architecture: A 3rd generation intelligent tutor for algebra symbolization. Int. J. Artif. Intell. Ed. (2008), 153--178. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Hernan-Losada, I., Pareja-Flores, C., and Velazquez-Iturbide, A. Testing-based automatic grading: A proposal from bloom's taxonomy. In Proc. ICALT 2008 (2008), 847--849. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Huang, S.-W., and Fu, W.-T. Enhancing reliability using peer consistency evaluation in human computation. In Proc. CSCW 2013 (2013), 639--648. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Kaindl, H., and Kainz, G. Bidirectional heuristic search reconsidered. Journal of Artificial Intelligence Research 7 (1997), 283--317. Google ScholarGoogle ScholarCross RefCross Ref
  15. Kaufmann, M., and Moore, J. S. An industrial strength theorem prover for a logic based on common lisp. IEEE Trans. Softw. Eng. 23, 4 (1997), 203--213. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Kirsh, D., and Maglio, P. P. On Distinguishing Epistemic from Pragmatic Action. Cognitive Science 18, 4 (1994), 513--549.Google ScholarGoogle ScholarCross RefCross Ref
  17. Kulkarni, C., Pang, K., Le, H., Chia, D., Papadopoulos, K., Cheng, J., Koller, D., and Klemmer, S. Peer and self assessment in massive online design classes. ACM TOCHI (2013).Google ScholarGoogle Scholar
  18. Lapets, A., Skowyra, R., Bassem, C., Kfoury, A., and Bestavros, A. Towards an infrastructure for integrated accessible formal reasoning environments. In Proc. UITP 2012.Google ScholarGoogle Scholar
  19. Mart-Oliet, N., and Meseguer, J. Rewriting logic: Roadmap and bibliography. J. Log. Algebr. Program. 81 (2001).Google ScholarGoogle Scholar
  20. Nielsen, J. Usability Engineering. Morgan Kaufmann, 1993. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Nipkow, T., Wenzel, M., and Paulson, L. C. Isabelle/HOL: a proof assistant for higher-order logic. Springer-Verlag, Berlin, Heidelberg, 2002. Google ScholarGoogle ScholarCross RefCross Ref
  22. Pappano, L. Massive open online courses are multiplying at a rapid pace. http://www.nytimes.com/2012/11/04/education/edlife/massive-open-online-courses-are-multiplying-at-a-rapid-pace.html.Google ScholarGoogle Scholar
  23. Paulin-Mohring, C. Inductive definitions in the system coq rules and properties. TLCA 1993 (1993). Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Paulson, L. C. The foundation of a generic theorem prover. Journal of Automated Reasoning 5 (1989). Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Ritter, S., Towle, B., Murray, R., Hausmann, R., and Connelly, J. A cognitive tutor for geometric proof. In Prof. ITS 2010 (2010), 453--453. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Schleimer, S., Wilkerson, D. S., and Aiken, A. Winnowing: local algorithms for document fingerprinting. In Proc. ACM SIGMOD 2003 (2003), 76--85. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Suppes, P. The next generation of interactive theorem provers. 7th International Conference on Automated Deduction 170 (1984), 303--315. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Suppes, P. Student use of an interactive theorem prover. Contemporary Mathematics 29 (1984).Google ScholarGoogle Scholar
  29. Tosic, M., and Nejkovic, V. Trust-based peer assessment for virtual learning systems. In Proc. SocInfo 2010 (2010), 176--191. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. VanLehn, K. The relative effectiveness of human tutoring, intelligent tutoring systems, and other tutoring systems. Educational Psychologist 46, 4 (2011), 197--221.Google ScholarGoogle ScholarCross RefCross Ref
  31. Windsteiger, W. Theorema 2.0: A graphical user interface for a mathematical assistant system. CEUR Workshop Proceedings (2012), 73--81.Google ScholarGoogle Scholar

Index Terms

  1. Crowd-scale interactive formal reasoning and analytics

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      UIST '13: Proceedings of the 26th annual ACM symposium on User interface software and technology
      October 2013
      558 pages
      ISBN:9781450322683
      DOI:10.1145/2501988

      Copyright © 2013 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 8 October 2013

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      UIST '13 Paper Acceptance Rate62of317submissions,20%Overall Acceptance Rate842of3,967submissions,21%

      Upcoming Conference

      UIST '24

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader