Mostrar el registro sencillo del ítem
dc.contributor.author | Insa Cabrera, David | es_ES |
dc.contributor.author | Silva, Josep | es_ES |
dc.date.accessioned | 2016-06-06T12:18:42Z | |
dc.date.available | 2016-06-06T12:18:42Z | |
dc.date.issued | 2015-07-06 | |
dc.identifier.isbn | 978-1-4503-3440-2 | |
dc.identifier.uri | http://hdl.handle.net/10251/65314 | |
dc.description | © ACM 2015. This is the author's version of the work. It is posted here for your personal use. Not for redistribution. The definitive Version of Record was published in http://dx.doi.org/10.1145/2729094.2742615 | es_ES |
dc.description.abstract | Automated marking of multiple-choice exams is of great interest in university courses with a large number of students. For this reason, it has been systematically implanted in almost all universities. Automatic assessment of source code is however less extended. There are several reasons for that. One reason is that almost all existing systems are based on output comparison with a gold standard. If the output is the expected, the code is correct. Otherwise, it is reported as wrong, even if there is only one typo in the code. Moreover, why it is wrong remains a mystery. In general, assessment tools treat the code as a black box, and they only assess the externally observable behavior. In this work we introduce a new code assessment method that also verifies properties of the code, thus allowing to mark the code even if it is only partially correct. We also report about the use of this system in a real university context, showing that the system automatically assesses around 50% of the work. | es_ES |
dc.description.sponsorship | This work has been partially supported by the EU (FEDER) and the Spanish Ministerio de Economíay Competitividad (Secretaría de Estado de Investigación, Desarrollo e Innovación) under grant TIN2013-44742-C4-1-R and by the Generalitat Valenciana under grant PROMETEOII2015/013. David Insa was partially supported by the Spanish Ministerio de Educación under FPU grant AP2010-4415. | es_ES |
dc.format.extent | 6 | es_ES |
dc.language | Inglés | es_ES |
dc.publisher | ACM | es_ES |
dc.rights | Reserva de todos los derechos | es_ES |
dc.subject | Computer-managed instruction | es_ES |
dc.subject.classification | LENGUAJES Y SISTEMAS INFORMATICOS | es_ES |
dc.title | Semi-automatic assessment of unrestrained Java code: a Library, a DSL, and a workbench to assess exams and exercises | es_ES |
dc.type | Comunicación en congreso | es_ES |
dc.identifier.doi | 10.1145/2729094.2742615 | |
dc.relation.projectID | info:eu-repo/grantAgreement/MINECO//TIN2013-44742-C4-1-R/ES/VALIDACION ASISTIDA DE PROGRAMAS MEDIANTE METODOS PRECISOS Y RIGUROSOS PARA UNA INGENIERIA DEL SOFTWARE ROBUSTA/ | es_ES |
dc.relation.projectID | info:eu-repo/grantAgreement/GVA//PROMETEOII%2F2015%2F013/ES/SmartLogic: Logic Technologies for Software Security and Performance/ | es_ES |
dc.relation.projectID | info:eu-repo/grantAgreement/MECD//AP2010-4415/ES/AP2010-4415/ | es_ES |
dc.rights.accessRights | Abierto | es_ES |
dc.contributor.affiliation | Universitat Politècnica de València. Departamento de Sistemas Informáticos y Computación - Departament de Sistemes Informàtics i Computació | es_ES |
dc.description.bibliographicCitation | Insa Cabrera, D.; Silva, J. (2015). Semi-automatic assessment of unrestrained Java code: a Library, a DSL, and a workbench to assess exams and exercises. ACM. https://doi.org/10.1145/2729094.2742615 | es_ES |
dc.description.accrualMethod | S | es_ES |
dc.relation.conferencename | ACM-SIGCSE Annual Conference on Innovation and Technology in Computer Science Education (ITiCSE 2015) | es_ES |
dc.relation.conferencedate | July 6–8, 2015 | es_ES |
dc.relation.conferenceplace | Vilnius, Lithuania | es_ES |
dc.relation.publisherversion | http://dl.acm.org/citation.cfm?id=2742615 | es_ES |
dc.type.version | info:eu-repo/semantics/publishedVersion | es_ES |
dc.relation.senia | 286705 | es_ES |
dc.contributor.funder | Ministerio de Economía y Competitividad | es_ES |
dc.contributor.funder | European Regional Development Fund | es_ES |
dc.contributor.funder | Generalitat Valenciana | es_ES |
dc.contributor.funder | Ministerio de Educación, Cultura y Deporte | es_ES |
dc.description.references | K. A Rahman and M. Jan Nordin. A review on the static analysis approach in the automated programming assessment systems. In National Conference on Programming 07, 2007. | es_ES |
dc.description.references | K. Ala-Mutka. A survey of automated assessment approaches for programming assignments. In Computer Science Education, volume 15, pages 83--102, 2005. | es_ES |
dc.description.references | C. Beierle, M. Kula, and M. Widera. Automatic analysis of programming assignments. In Proc. der 1. E-Learning Fachtagung Informatik (DeLFI '03), volume P-37, pages 144--153, 2003. | es_ES |
dc.description.references | J. Biggs and C. Tang. Teaching for Quality Learning at University : What the Student Does (3rd Edition). In Open University Press, 2007. | es_ES |
dc.description.references | P. Denny, A. Luxton-Reilly, E. Tempero, and J. Hendrickx. CodeWrite: Supporting student-driven practice of java. In Proceedings of the 42nd ACM technical symposium on Computer science education, pages 09--12, 2011. | es_ES |
dc.description.references | R. Hendriks. Automatic exam correction. 2012. | es_ES |
dc.description.references | P. Ihantola, T. Ahoniemi, V. Karavirta, and O. Seppala. Review of recent systems for automatic assessment of programming assignments. In Proceedings of the 10th Koli Calling International Conference on Computing Education Research, pages 86--93, 2010. | es_ES |
dc.description.references | H. Kitaya and U. Inoue. An online automated scoring system for Java programming assignments. In International Journal of Information and Education Technology, volume 6, pages 275--279, 2014. | es_ES |
dc.description.references | M.-J. Laakso, T. Salakoski, A. Korhonen, and L. Malmi. Automatic assessment of exercises for algorithms and data structures - a case study with TRAKLA2. In Proceedings of Kolin Kolistelut/Koli Calling - Fourth Finnish/Baltic Sea Conference on Computer Science Education, pages 28--36, 2004. | es_ES |
dc.description.references | Y. Liang, Q. Liu, J. Xu, and D. Wang. The recent development of automated programming assessment. In Computational Intelligence and Software Engineering, pages 1--5, 2009. | es_ES |
dc.description.references | K. A. Naudé, J. H. Greyling, and D. Vogts. Marking student programs using graph similarity. In Computers & Education, volume 54, pages 545--561, 2010. | es_ES |
dc.description.references | A. Pears, S. Seidman, C. Eney, P. Kinnunen, and L. Malmi. Constructing a core literature for computing education research. In SIGCSE Bulletin, volume 37, pages 152--161, 2005. | es_ES |
dc.description.references | F. Prados, I. Boada, J. Soler, and J. Poch. Automatic generation and correction of technical exercices. In International Conference on Engineering and Computer Education (ICECE 2005), 2005. | es_ES |
dc.description.references | M. Supic, K. Brkic, T. Hrkac, Z. Mihajlovic, and Z. Kalafatic. Automatic recognition of handwritten corrections for multiple-choice exam answer sheets. In Information and Communication Technology, Electronics and Microelectronics (MIPRO), pages 1136--1141, 2014. | es_ES |
dc.description.references | S. Tung, T. Lin, and Y. Lin. An exercise management system for teaching programming. In Journal of Software, 2013. | es_ES |
dc.description.references | T. Wang, X. Su, Y. Wang, and P. Ma. Semantic similarity-based grading of student programs. In Information and Software Technology, volume 49, pages 99--107, 2007. | es_ES |