versión On-line ISSN 0717-5000
CLEIej vol.15 no.1 Montevideo abr. 2012
Previous work has demonstrated that the use of software testing concepts for teaching introductory programming courses may be a good practice for the students. Indeed, these courses provide limited opportunities for the teacher regarding dynamic activities that could help students in the learning process. This paper describes an approach in which test sets are reused in the context of teaching introductory programming courses, as an alternative to increase the quality of the programs generated by students. An experimental study was carried out to investigate the impact of reusing test cases during the programming learning. The objective is to evaluate if the use of test cases might improve the quality of programs implemented by the students. Sixty undergraduate students participated in the experiment, implementing programs in the domain of vectors. A set of reference programs was used to generate test cases, based on functional testing, to be reused by the students to test their programs. Considering a range of , the programs’ quality increased from to using this approach. The results provide evidences that the reuse of test cases during introductory programming courses may help to increase the quality of the programs generated by students, motivating them to apply software testing during the development of the programs.
Keyworks: Experimental study, Functional testing, Reuse, Teaching introductory programming courses.
Resumo: Estudos anteriores mostraram que conceitos de teste de software aplicados ao ensino de disciplinas introdutórias de programação podem contribuir com a aprendizagem dos alunos. Um problema que ocorre com os professores que lecionam essas disciplinas é a limitação na realização de atividades dinâmicas e interativas que possivelmente contribuiriam com a aprendizagem. Este artigo descreve uma abordagem baseada no reúso de casos de teste no contexto de disciplinas introdutórias de programação como alternativa para aumentar a qualidade dos programas gerados pelos alunos. Um estudo experimental foi realizado para investigar o impacto do reúso de casos de teste durante a aprendizagem de conceitos de programação. O objetivo foi avaliar o uso de casos de teste como alternativa à melhoria da qualidade dos programas implementados pelos alunos. Sessenta estudantes de graduação participaram do experimento implementando programas no domínio de vetores. Um conjunto de programas de referência foi usado para geração dos casos de teste com base no teste funcional para ser reusado pelos estudantes. Os programas gerados pelos estudantes foram avaliados considerando notas de 1 a 10 e observou-se que a média das notas aumentou de 5.3 para 7.4 com o uso da abordagem proposta. Os resultados forneceram evidências de que o reúso de casos de teste em cursos introdutórios de programação pode aumentar a qualidade dos programas gerados pelos alunos, motivando-os a aplicarem teste de software durante o desenvolvimento dos programas.
Palavras-chave: Estudo experimental, Teste funcional, Reúso, Ensino em cursos introdutórios de programação
Received: 2011-08-01 Revised: 2011-12-30 Accepted: 2011-12-30
The teaching of introductory programming courses differs from the teaching of other subjects, because in order to teach programming, first it is necessary to teach how to solve problems and build ideas (1). Often, the students’ inexperience in programming logic implies that teachers need to find ways to improve their ability to understand the logic behind the mere codification.
In traditional teaching methods, students initially receive the rules (programming language structures) so that problems can be solved. Then the teacher provides the student the program specification and (s)he waits for the student to build a program able to solve the proposed problem. In this case, the student has got just the program specification and, based on it, (s)he should achieve the expected result, which is the code of a correct program. In such approach, the student creates and applies some tests which (s)he judges enough and admits that the program is correct.
This paper presents a study where we propose an approach for teaching in which the student receives from the teacher not only the problem specification, but also a test set that should be used during the testing. The test set is suitable for a reference program which is equivalent to the one that will be created by the student. Two programs and are defined as equivalent if the same output resulting from processing the input to is obtained by the program considering the input [2). However, the functionality in equivalent programs can be implemented in different ways. Based on this idea, when the program is built, the student can reuse the test cases and evaluate whether the program under construction is free of defects. Thus, the method contributes to improving the quality of the programs generated by students. Quality here refers to the conformancy of the program implementation with its specification, i.e., the program implementation is totally compliant with its specification. Another aspect of this approach is that, as the student learns to program, at the same time (s)he understands the importance of software testing activity. This is an activity with a small portion allocated in the Computer Science curriculum in comparison to other activities of the software development process [3).
The experimental study performed in this work involved two programming themes: vectors and matrix manipulation and was conducted with undergraduate students of two computer science courses at ICMC/USP (Institute of Mathematics and Computer Science/University of São Paulo). In each classroom, students were randomly divided into two groups, in which a group of students received only the program specification and the other group received the program specification and the test sets. Each group was instructed to use the specification to build a program in conformance to it. The group that has received the test set was instructed to use it to validate their programs and to record the results of this validation. At the end, programs were implemented by the students. The ANOVA statistical tests were applied to evaluate the experiment hypotheses.
This paper is structured as follows. In Section 2, we describe related work on the reuse of software testing and experimental studies. In Section 3, we present the important concepts about software testing that are required to understand the present work. In Section 4, we describe details about our experimental study and, in Section 5, the results obtained are discussed. In Section 6, we present the validity threats of our study. In Section 7, we outline a proposal for a data repository to ease the use of testing during programming learning. Finally, in Section 8 we present our conclusions and future work.
Although there are papers reporting empirical studies about reuse of test cases [4, 5, 6), in the teaching context there are many works about the insertion of software testing as a resource to teach programming.
In Schulte et al. [7) the authors compare and contrast the way in which different models conceptualize program comprehension from an education point of view. Jones [8, 9, 10, 11) is the most outspoken educator on this area who suggests systematically incorporating testing practices across the undergraduate curriculum, thinking that the students can be more prepared for real world testing tasks upon graduation. Stephen [12, 13) describes the use of software testing to teach programming introdutory courses with the purpose of encouraging students to write tests as they code (in the spirit of TDD – Test-Driven Development).
Few works report empirical studies about reuse of test cases in teaching introductory programming courses. In Souza and Barbosa (14) it is presented an environment for submission and evaluation of practical work performed by students. The students submit their program and their test cases and the environmet analyses them using a reference program and a test case set, defined by instructor. The objective is to evaluate the coverage of the student´s test cases in relation to reference program and the coverage of the instructor´s test cases in relation to student´s program. Thus, the tool allows the reuse of test cases generated by the instructor to verify the correctness of the student’s program. Differently of our study, this approach is defined to evaluate the program correctness, mainly related to the coverage of testing criteria. Currently, this environment is configured for Java programs. In (15) is presented a tool, in which the student’s programs are tested using test cases generated by the instructor. The tool MARMOSET (16) uses an approach that is similar to our work as it uses test cases generated by the instructor, but it allows students to create their own tests. The tests generated by the instructor are delivered to the students in three different moments: before they start to code their programs, during the implementation (with limitation in how many tests they can perform) and after they finish the program implementation. The authors believe that if part of the tests are a limited resource, this will encourage students to start tests earlier in their projects. This approach is also defined to evaluate the program correctness and the generation of test cases by students.
Campanha et al. (17) describe an experimental study where test case sets are reused to validate alternative versions of a program. In this study, two master students developed reference programs for sorting algorithms, such as bubble sort, selection sort and insertion sort, and subsequently, they generated, for each program, test cases sets adequated to the Mutation Testing. In an introductory programming course, undergraduate students implemented different programs to solve the data sorting problem. These programs were tested using the test case sets generated from the reference programs, evaluating if this sets are adequated to test equivalents programs. The authors concluded that the reuse of test sets adequated to Mutation Testing to validate equivalent programs can be truly effective, reducing the testing activity costs. Our study, presented in this paper, is inspired on the Campanha et al. (17) study.
The software development process involves the use of methods, techniques and tools, but this does not prevent the occurrence of errors in the product in development. In order to reduce these errors, there are VV&T - Software Verification, Validation and Testing activities, which are employed during the software development cycle. Verification is the process of identifying whether the software construction is being done in the right way. Validation involves evaluating whether the software developed meets the requirements proposed by the client. The testing activity involves discovering the presence of defects, if they exist (18).
The testing activity should be present throughout the development cycle, and it has the following testing phases: 1) unit testing, in which the smallest units of the software are tested, such as procedures, functions or methods and classes; 2) integration testing, where the integration between modules, components or other forms of units of the software is tested, and; 3) system testing, in which the system is tested as a whole. The tests applied at this phase include stress testing, security testing, load testing, among others. In addition to these three phases, there is also the regression testing performed when the program is modified and it is necessary to guarantee that previous tests are still applied, and also to test new requirements introduced by modification (19).
A test case is a pair formed by one or more input values and an expected output for this entry. A test criterion defines which properties or requirements need to be tested to evaluate the test cases quality (20). The selection of test cases should be performed based on testing criteria, which are divided in different testing techniques: structural, functional and error-based testing (19). In the functional technique, the software is considered a black box and the test cases are built based on the software specification. For the structural testing, the internal structures of software are tested, so the test cases are built based on the software implementation. In the error-based testing technique, the most common faults that can be made during the development process are used for selecting the test cases.
The Equivalence Partitioning Functional criterion divides the domain of program entry data in classes, in which the test cases are derived (21, 18). In this criterion, the division is carried out into valid and invalid equivalence classes and the objective is to produce test cases that cover each class, reducing the total number of test cases needed. The principle is that, if a test case of a particular equivalence class reveals an error, any other test case in this class should reflect the same error. The Boundary Value Analysis Functional criterion complements the Equivalence Partitioning through test cases that exercise boundary values, for instance, the biggest value of a variable. Boundary values refer to situations that occur within an equivalence class, directly above or below.
The experiment described in this paper uses the functional testing criteria to generate the test cases for the reference programs. As the programs developed by students were equivalent to the reference programs, we believe that these testing criteria are the best choice to generate the test cases for this experiment. The functional test sets are simpler to generate, facilitating the reuse and the use of this approach in introductory programming courses.
The experimental study was defined based on the Wohlin et al. (22) process. Next, we summarize the experiment planning.
To investigate the reuse of test case sets as an mechanism to teach programming, checking whether or not this mechanism allows an increase in the quality of the programs implemented by students in introductory programming courses.
Evaluation of programs for vectors and matrices manipulation, implemented in language by undergraduate students of computer courses. The master student who is leading this experiment defined the reference programs and the test case sets to be reused by the subjects. The experiment was performed by two groups of students (subjects) randomly selected from each classroom; two programs for vector and matrices manipulation and and functional test case sets and adequate to test the reference programs. During a practice class, half the subjects received only the program specification, while the other half additionally received the test case sets and to validate their implementation. The test case sets were given only after the students finished the program implementation. The evaluation of the programs developed by the subjects was done by applying a score, ranging from to , considering the correctness of the program in relation to test case set.
- Null Hypothesis : Students who received test case sets and will produce programs whose quality is equal to or lower than students who received only the program specifications.
- Alternative Hypothesis : Students who received test case sets and will produce higher quality programs than students who received only the program specifications.
Although our interest is more focused on the effect of test case sets in the quality of the generated programs, hypotheses considering other factors were also taken into consideration. They are described in Section 5.
Program specifications, reference programs, test cases sets, questionnaires and programs implemented by the students.
Score of the programs implemented by the students and answers to the questionnaire.
Sixty undergraduate students from Introduction to Computer Science courses of the Institute of Mathematics and Computer Science of University of São Paulo, divided in three groups: classrooms I ( students) and II ( students) formed by students of the course Bachelor of Computer Science and classroom III ( students) formed by students of the course Bachelor in Informatics. Each of these students developed his/her individual version of the programs to be evaluated. A master student who is leading this experiment generated the test cases sets for a reference program set.
The programming topic used in the experiment (vectors and matrices manipulation) was previously taught for the students, in this way, it was considered that the students had the necessary knowledge to develop the required programs. We did not provide training on software testing. During the experiment, subjects from each classrooms were randomly separated into two groups. (in this case, following the alphabetical order).
All subjects in the classroom completed a questionnaire, providing information about sex, background and previous experience in programming, which was used in the analysis. Then, one group received only the programs specification and the other one received also the test cases sets (these were distributed only after the programs were implemented). Each subject was responsible for implementing each program from specification. The subjects that received also the test case sets, had to use them to test the programs and take notes of the results of the tests (if they fail, then they fixed the program and tested again). The subjects that did not receive the test case sets could test the programs in an ad-hoc way.
This activity was carried out during a classroom practice, under teacher supervision. The duration of the classroom is as follows: classroom I - 1 hour, classroom II - 2 hours and classroom III - 2 hours. In this phase of the experiment it was not allowed the communication between the students and the consultation of other materials. Before this activity, the master student implemented the reference programs, based on the specifications available in (23). Then, she generated test case sets to be reused by the subjects, based on the Equivalence Partitioning and Boundary Value Analysis functional criteria. The experiment material were all available electronically and hence could be accessed easily (http://www.icmc.usp.br/~masbrit/TestCaseExperiment.rar). Table 1 shows the programs selected for the experiment, the programs size (lines of code) and the number of test cases generated for each one. The table also shows which programs are used in each classroom.
In the second part of the experiment, the implementations and questionnaires completed by the subjects were analyzed. The master students compiled each program and executed it using the test cases sets generated. Thus, the correctness was evaluated and a score, from to , was assigned for each program. To evaluate the hypotheses, statistical tests were performed, using analysis of variance (ANOVA) to cases that the sample was normal and Kruskal-Wallis tests when the sample was not normalized.
Data analysis was conducted considering the programs written by students and the answers of the questionnaires. Table 2 presents the population involved in the experiment, showing for each classroom the number of programs and the LOC, in average. In classroom I, the students received both specifications (vector and matrix), however, due to the short time (1 hour) available, only of the students completed the two programs on time. Due to this, the matrix implementation was not considered in the analysis. In classrooms II and III the two specifications were implemented by the students. Comparing with Table 1, the implementations of the students have, in average, more lines of code than the reference programs.
The students programs were analysed, observing their correctness and a score from to was assigned to each program (Table 3). As can be seen in the scores of Table 3, the students groups that reused the test case sets (TC) have obtained better scores than the students that did not reuse the test case sets.
The students reported some defects revealed by the test case sets. In some cases, the proposed problem was simplified and, hence, the solution did not match with the program specification. As an example, considers the following specification: the program must read a set of numbers and insert these numbers in a set . Then, the program must print: 1) all the even numbers of and their sum; and 2) all the prime numbers of and their sum. In some programs implemented by the students, the test to check if the number is even or prime is done when the number is read, without storing them in a set.
Table 4 presents the main defects found in the programs generated by the students. Most of them are related to defects in programming logic and in the design of the solutions.
Table 5 shows the data extracted from the questionnaires filled by students, which were also considered in the analysis of the results. For each classroom and group (not using TC and using TC) it is presented the quantity of males and females, the number of students from public and private school, and, finally, the number of students with previous experience in computer programming and students without previous experience. In the following, these results are statistically analyzed and discussed.
During the data analysis, the influence of some factors in the programs quality was analysed. Examples of factors analysed are classroom and previous programming experience. We have conducted two analysis of variance, through two-factor models for unbalanced data. At first, it was studied the influence of the group and classroom factors. Then, it was studied the influence of scores and previous experience. In both cases, it was also checked if there was a significant relationship between these factors.
The models for the variance analysis are described in the following form:
- is the index related to the group, where indicates non-reuse of TC and indicates reuse of TC;
- is the index related to second specific factor of each analysis; in the first analysis, is the classroom , or . In the second analysis, is the previous experience, equal for students without previous experience or for students with previous experience;
- is the number of replicas for the score, considering the factors and ;
- is the index related to number of replicas, with the value between 1 to ;
- is the value of score variable for , and index;
- is the mean value of score;
- is the effect of factor group on the mean;
- is the effect of the second factor of the analysis (class or experience) on the mean;
- is the effect of interaction between analysis factors;
- is the error attributed for the score variable considering the index , and . These errors are normal, independent and identically distributed, with mean and unknown variance ;
- , is the mean value of replicas of the score variable, for the fixed index and ;
It was assumed a confidence level of 95% (therefore rejecting the hypothesis for p-values ). In several studies, permutations were performed considering the variables studied to obtain evidence on the results of the study and the degree of suitability of each variable.
Before carrying out the hypothesis test, we have performed some descriptive analysis of the scores obtained by students considering the different levels of each factor. It is possible to observe that the mean score of the students belonging to group , which have reused test case sets, is higher than the mean score of the students in group (Table 6 and Figure 1).
The analysis has shown that there is a difference between the mean scores obtained by students from different classrooms, mainly from classroom II to the others, as can be seen in Table 7 and Figure 3. However, the difference in the means was not meaningful in the variance analysis.
The results obtained through the analysis of variance for the group and classroom factors are shown in Table 8. It can be observed that the influence of the group factor in the mean score was significant, i.e., it can be stated that the reuse of test cases provided a significant increase in the students scores. The relationship between group and classroom factors was not considered significant.
Based on the descriptive analysis of the scores according to the experience factor (Table 9 and Figure 2) we have evidence that the mean score may differ for the two levels of the factor. The second analysis of variance showed that the mean score was significantly different (Table 10). The hypothesis that previous programming experience does not produce an increase in score was rejected, i.e., students with previous experience obtained a highest score compared to those without previous experience. Considering the interaction between the factors, it was possible to conclude that their relationship is not significant. The influence of the group factor was also considered significant to the increase of score in this analysis.
In addition, an analysis was performed to verify if the use of test cases has significant influence on the scores of the subjects, considering the group without previous experience in programming. This analysis was conducted using the Kruskal-Wallis test, because it was not possible to obtain normality in the data by transformations. It was observed that the reuse of test cases provided a significant increase in scores of the subjects without previous experience (). Therefore, the main objective of our approach has been reached, i.e., our approach contributes to leverage the programming knowledge of unexperienced students.
After finishing the analysis of tests based on groups, it was obtained a (Table 8 and Table 10), allowing us to reject the null hypothesis (formulated at Section 4) and to conclude about the existence of significant differences between groups. Thus, there is evidence that the reuse of test cases generated based on reference programs can contribute to improve the quality of equivalent programs implemented by students in introductory programming courses. As expected, we have also observed that the previous experience in computer programming can contribute to improve the programs quality. Other tests were performed considering the sex and background variables, but we have not found evidence that these variables can contribute to test the hypothesis for the sample considered.
Internal validity threats concern external factors that may affect an independent variable. Appropriate analysis - i.e., ANOVA and Kruskal-Wallis - were performed to analyse the effect of these factors. However, it was not possible to say that the students’ experience, for example, properly discerns the effect of results.
Another important threat is the possibility of fraud, where identical implementations could be delivered by undergraduate students. Precautions were taken during the execution of the experiment so that subjects did not communicate while implementing the programs. In addition, the codes were examined individually and no evidence of identical programs was found.
Another threat is the simplicity of the tested programs, as it can be assumed that any test case sets would be sufficient to achieve a satisfactory level of correctness for the programs. Treatment was not found for this threat, since the level of the students did not allow very complex programs and the time available for implementation was also reduced.
Construct validity threats concern the relationship between theory and observation. In particular, this threat is related to how the students carried out the tests. To avoid any subjective evaluation, in the moment of the experiment the master student was present and answered the subjects’ doubts during the experiment. The great majority of subjects had no practice in software testing, so it avoided that a subject could favor some technique. In the classroom where the study was applied twice, the subjects and groups were interspersed, i.e. students who used test case sets in first study did not used them in the second study.
We did not provide treatment to the threat in which people try to look better when being observed, because it would require isolating each student, with only the presence of the instructor, but this is not feasible for this kind of study.
The conclusion validity is concerned with issues that affect the ability to draw the correct conclusion about relations between the treatment and the results of an experiment. In this case the quality that is being measured is subjetive and contains human judgement, so the quality was measured by the scores of the subjects’ programs. The master student who conducted the experiment analysed the programs and assigned them a score. To improve this reliability, she has used a criteria list to give scores to the subjects programs. Attention was paid to not violating assumptions made by statistical tests. In situations that allowed to use of parametric test, ANOVA was used, and in situations that did not allow it, the Kruskal-Wallis test was used. The size of subjects set was selected based on classes where it was possible to carry out the study, this is a threat, but this set must be large and more representative to generalization of the study.
The external validity of the experiment refers to the ability to generalize the results. Since the programs studied were in the specific domain of vectors and matrices manipulation, the results cannot be applied to other areas of programming. The generalization can be achieved if the experiment is replicated in other classrooms for other levels of undergraduate students and other domains of programs.
Motivated by the results of the experimental study described above, in this section we propose the implementation of a repository to store the reference programs and the associated test case sets, so that they can be easily retrieved to be used during computer programming classes.
Repositories are information systems containing features that allow to import, store, preserve, extract, and export digital objects. The current increased of use of repositories has been possible due to the reduction of physical storage prices, the use of standards such as metadata protocol of the Open Archives Initiative (OAI-PMH), and advances in the development of metadata standards that support the open communication model. The term learning object repository comprises a database in which learning objects are stored to facilitate their search and reuse. Learning objects are any digital resources that can be reused to assist with learning and distributed network on demand (24).
The DSpace Institutional Digital Repository System is a system of electronic publications management developed in a collaboration between MIT (Massachusetts Institute of Technology) and HP (Hewlett-Packard Co.). The DSpace has BSD Open Source license and is available since November . It is currently the most used software for building institutional repositories (25). With this software, the user can manage the technical and scientific information, ensuring a permanent access to this information (26).
Considering the results of the experimental study described in this paper, we propose a repository that will be implemented with the instantiation of DSpace. The objective is to enable reuse of teaching materials generated, such as: sets of reference programs and its associated test case sets, helping to optimize the teaching process.
Two phases are necessary to implement the repository: 1) exploratory and descriptive theoretical basis for the use of DSpace; and 2) experimental research consisting on evaluation of the software configured. These two phases can be applied in the steps: installation and configuration of the tool, metadata customization, maintenance of the external digital object, operation of the data provided, study of the license distribution model. In this direction, we are building a repository that can be used to store programs and test case sets, while they are being generated, aiming at reuse and improvement of the materials quality. For this proposal, we require a special feature, which is a free access provided only to teachers previously registered on the system, who can download the material available in the repository.
A schema for the proposed repository is shown in Figure 4. The complete deployment and configuration of DSpace is planned as a future work. The repository is configured to store test case sets and the associated reference programs, considering topics taught in introductory programming disciplines. The repository is organized to enable user interaction, so that it does not act merely as a storage compartment; users can add new learning objects and, for each download done, they should provide feedback about the use of the downloaded object. The teacher can look for materials in the repository using a sort key, as for example the name of the program. The information initially available in the repository are related to the discipline of Introduction to Computer Science, such as: program structures: decision and repetition statements, primitive data type, modularization of programs, procedures, functions and parameter passing; compound data types: vectors, matrices, strings, records, dynamic structures (pointers and files). When the user selects a particular topic, the system will retrieve all the programs registered in the repository that are related to the chosen topic. Selecting a program, the user obtains the following information: program specification, programming language, courses where it can be used, level of difficulty and estimated time for implementing the exercise. Information about the test case sets related to the program selected are also avaliable, such as: testing criteria used, number of test cases, the test cases (input and output data) and coverage obtained.
This paper presented an experimental study to evaluate the reuse of test case sets to help the teaching of introductory programming courses. The objective is improving the quality of the programs generated by the students. It is not expected that every aspect of the student’s program be tested, but that test cases are able to detect typical errors made by new students in computer programming and, thus may contribute to their learning.
The development of the proposal followed a methodology based on experimentation, according to Wohlin (22). The conducted experimental study aimed to validate this proposal. The ANOVA test provided evidence that sets of functional test cases reused can help the student in obtaining an improvement in the quality of the programs. It is expected that the experiment has motivated the participating students to use testing activities during the program development and, as a consequence, that they will give a greater importance to this activity. It was also analysed the effect of previous experience factor, which indicated evidence of positive influence of the reuse of test cases by subjects with no previous experience in programming. This is the biggest benefit of the approach, because it shows that newbie students can develop better quality programs.
One limitation of our proposed approach is that the test cases are generated considering reference programs and they are reused by the students to test equivalent programs. Thus, it is not expected that all aspects (for instance, all statements) of the student’s program are tested, but that the test case sets are able to validate only functional aspects of the programs.
As future work we intend to conduct the same experiment using other courses and periods of the computer science courses and thereby obtain more general results about the hypotheses studied. Other studies can be developed considering covariates (extra information) possibly explaining the relationship between the results of the statistical tests.
The authors would like to thank the Brazilian funding agencies CAPES and FAPESP, for the financial support, under CAPES process DS-7109935/M and FAPESP processes: 2009/04517-2.
(1) E. W. Dijkstra, “Structured programming,” O. J. Dahl, E. W. Dijkstra, and C. A. R. Hoare, Eds. London, UK, UK: Academic Press Ltd., 1972, ch. Chapter I: Notes on structured programming, pp. 1–82.
(2) E. J. Weyuker, “Axiomatizing software test data adequacy,” IEEE Transactions on Software Engineering, vol. 12, no. 12, pp. 1128–1138, 1986.
(3) T. Shepard, M. Lamb, and D. Kelly, “More testing should be taught,” Commun. ACM, vol. 44, pp. 103–108, June 2001.
(4) Y. D., Y. W., M. Lau, and S. yu L., “Experiments on test case reuse of test coverage criteria,” in Ubiquitous Intelligence Computing and 7th International Conference on Autonomic Trusted Computing (UIC/ATC), 7th International Conference on, 2010, pp. 277–281.
(5) L. Lima, J. Iyoda, A. Sampaio, and E. Aranha, “Test case prioritization based on data reuse an experimental study,” in 3rd International Symposium on Empirical Software Engineering and Measurement, ser. ESEM ’09, 2009, pp. 279–290.
(6) M. Mirzaaghaei, “Automatic test suite evolution,” in 19th ACM SIGSOFT symposium and the 13th European conference on Foundations of software engineering, ser. ESEC/FSE ’11, 2011, pp. 396–399.
(7) C. Schulte, T. Clear, A. Taherkhani, T. Busjahn, and J. H. Paterson, “An introduction to program comprehension for computer science educators,” in 2010 ITiCSE working group reports on Working group reports, ser. ITiCSE-WGR ’10, 2010, pp. 65–86.
(8) E. L. Jones, “Software testing in the computer science curriculum – a holistic approach,” in Australasian conference on Computing education, ser. ACSE ’00, 2000, pp. 153–157.
(9) ——, “Grading student programs - a software testing approach,” in fourteenth annual consortium on Small Colleges Southeastern conference, ser. CCSC ’00, 2000, pp. 185–192.
(10) ——, “Integrating testing into the curriculum – arsenic in small doses,” in Thirty-second SIGCSE technical symposium on Computer Science Education, ser. SIGCSE ’01, 2001, pp. 337–341.
(11) E. Jones, “An experiential approach to incorporating software testing into the computer science curriculum,” in Frontiers in Education Conference, 2001. 31st Annual, vol. 2, 2001.
(12) S. H. Edwards, “Teaching software testing: Automatic grading meets test-first coding,” in Addendum to the 2003 Proceedings of the Conference on Object-oriented Programming, Systems, Languages, and Applications, 2003, pp. 318–319.
(13) A. Tinkham and C. Kaner, “Experiences teaching a course in programmer testing,” in Agile Development Conference, 2005, pp. 298–305.
(14) D. M. Souza and E. F. Barbosa, “Progtest: An environment for the submission and automatic evaluation of programming assignments based on testing activities,” in Simpósio Internacional de Iniciação Científica da USP, 2009, (in Portuguese).
(15) S. H. Edwards and M. A. Perez-Quinones, “Web-cat: automatically grading programming assignments,” in 13th annual conference on Innovation and technology in computer science education, ser. ITiCSE ’08, 2008, pp. 328–328.
(16) J. Spacco, D. Hovemeyer, W. Pugh, F. Emad, J. K. Hollingsworth, and N. Padua-Perez, “Experiences with marmoset: designing and using an advanced submission and testing system for programming courses,” in 11th annual SIGCSE conference on Innovation and technology in computer science education, ser. ITICSE ’06, 2006, pp. 13–17.
(17) D. N. Campanha, S. R. S. Souza, O. A. Lemos, E. F. Barbosa, and J. C. Maldonado, “Reuse of test case sets: A study of domain of algorithms sorting,” in VI Experimental Software Engineering Latin American Workshop, 2009, pp. 114–123.
(20) H. Zhu, P. A. V. Hall, and J. H. R. May, “Software unit test coverage and adequacy,” ACM Comput. Surv., vol. 29, no. 4, pp. 366–427, 1997.
(24) D. A. Wiley, “Connecting learning objects to instructional design theory: A definition, a metaphor, and a taxonomy,” The Instructional Use of Learning Objects Online Version, vol. 2830, no. 435, pp. 1–35, 2000.
(25) C. L. M. Viana, M. A. M. A, and M. Shintaku, “Institutional repositories in science and technology: an experience of customizing of the dspace,” in Simpósio de Bibliotecas Digitais, São Paulo, SP, Brasil, 2006, (in Portuguese).
(26) M. A. A. Mardero, “Dspace repositorie and digital preservation,” in Encontro de Informação em Ciências da Comunicação, 2004, p. 14, (in Portuguese).
(27) G. R. B. Fachin, J. Stumm, R. L. Comarella, F. A. P. Fialho, and N. Santos, “Knowledge management and cognitive view of institutional repositories,” Perspect. ciênc. inf., vol. 14, no. 2, pp. 220–236, 2009, (in Portuguese).