Physics is not only theory, concepts, laws, and formulas. It is also experimental science. Laboratory work is at the heart of physics. In Turkey, although the National Curriculum includes experiments, physics classes often do not include experimentation. Requirements for high-stake exams restrict teachers’ instruction time available to teach lab work.
1. Journal of Physics: Conference Series PAPER • OPEN ACCESS You may also like - Conceptual understanding procedure to Learning how to teach experiments in the school elicit metacognition with pre-service physics teachers physics laboratory Jared Carpendale and Rebecca Cooper - School physics for the 1980s Ken Thomas To cite this article: O Gkioka 2019 J. Phys.: Conf. Ser. 1286 012016 - Language gap between college introductory physics textbooks and high school physics textbooks Eunjeong Yun View the article online for updates and enhancements. This content was downloaded from IP address 203.115.91.138 on 10/01/2022 at 19:35
2. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 Learning how to teach experiments in the school physics O Gkioka Department of Secondary Mathematics and Science Education, Bogazici University, Istanbul, Turkey olga.gkioka@boun.edu.tr Abstract. The focus of this research is in the broader area of physics teacher education. The aim is to investigate physics teachers’ efforts to learn how to teach and assess experiments, and in particular, concepts related to scientific evidence in the school physics laboratory. The study has looked at the participants as both learners and prospective teachers. It has taken place within the context of the course “Secondary Science Lab Applications” within a pre-service teacher education program in a Department of Secondary Science and Mathematics Education. Twenty- four students participated. The participants: a) revised the main concepts related to scientific practices (in particular, experimental validity and reliability of measurement), b) developed lesson plans, teaching and assessment methods and, c) taught school physics experiments. Interviews were conducted with the participants during their preparation for teaching and after teaching. They were observed when teaching and all classes were videotaped. A qualitative and quantitative data analysis identified particular trends among the participants. Students’ difficulties while they designed, carried out experiments and wrote lab reports were identified, as well as the difficulties they experienced when teaching experiments. 1. Introduction Physics is not only theory, concepts, laws and formulas. It is also an experimental science. Laboratory work is at the heart of physics. In Turkey, although the National Curriculum includes experiments, physics classes often do not include experimentation. Requirements for high-stake exams restrict teachers’ instruction time available to teach lab work. In addition, in pre-service teacher education, major emphasis is given to conceptual understanding of physics and concepts across the curriculum. In our department, an initiative was taken to prepare our student teachers to teach experiments effectively in the school physics laboratory: to address the knowledge and understandings required to teach scientific skills and methods explicitly. The goal has been to elicit pre-service teachers’ understandings of scientific evidence and, through targeted instruction, to support them to revise such understandings and learn how to write laboratory reports. 2. Research into secondary students’ difficulties with laboratory work In physics education, many secondary students experience difficulties in the laboratory. The literature provides a range of difficulties that secondary pupils encounter in the school science laboratory [1–5]. Many research studies support the lack of a fundamental understanding of scientific evidence in secondary students. For example, studies on performance on lab work questioned the assumption that Content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI. Published under licence by IOP Publishing Ltd 1
3. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 students understand the nature of measurement and experimentation [6,7]. This work has led to the development and refinement of various laboratory teaching materials [8]. After completing a traditional laboratory course, the majority of students have ideas about measurement that are inconsistent with the generally accepted scientific model. For example, as Séré and her team found, a large proportion of students view the ideal outcome of a single measurement as an “exact” or “point-like” value. Only if a measurement is considered really “bad” would it be reported in terms of an interval [6]. Boudreaux et al investigated the ability of students to reason on the basis of the control of variables [9]. The participants were undergraduate physics students and they were asked to decide whether or not a given variable influences the behaviour of a system. They found out that although most of the students recognized the need to control variables, many had significant difficulty with the underlying reasoning. It was argued that teachers should be prepared for how to teach experiments and develop science practices in secondary school students. Yet, little is known about pre-service teachers’ understanding of scientific evidence. In physics departments it is often assumed that the participation of physics students in regular undergraduate science laboratory courses provides them with the required knowledge and skills to teach in a school laboratory [10]. This assumption stands in parallel with the belief that the more rigid and difficult modules one attends in undergraduate programs, the better and well-prepared the physics teacher will be [11]. There is a need to develop ‘special’ courses for physics teachers’ preparation since teachers need to have deeper knowledge than their students. The aim of this research study was to investigate physics teachers’ efforts to learn how to teach and assess experiments, and in particular, concepts related to scientific evidence in the school physics laboratory. The study looked at the participants as both learners [12,13] and prospective teachers [14,15]. 3. Context of the research study and research questions In Turkey, the secondary science curriculum gives much attention to conceptual knowledge, leaving little space for teaching of experiments. Although the National Curriculum includes experiments, physics classes often do not include experimentation. Requirements for high-stake exams restrict teachers’ instruction time available to teach lab work. In our department, an initiative was taken to prepare our student teachers for how to teach experiments effectively in the school physics laboratory: to address the knowledge and understandings required to teach scientific skills and methods explicitly. The goal has been to develop a good understanding of what makes ‘good scientific evidence’ firstly, in teachers and secondly, to their secondary school students. Two research questions have guided the study: 1) What are the understandings of scientific evidence that pre-service physics teachers demonstrate when they conduct experiments? 2) What are the difficulties and challenges that they experience when they teach and assess laboratory work? The study has taken place within the context of the course “Secondary Science Lab Applications” within a pre-service teacher education program in a Department of Secondary Science and Mathematics Education. Twenty-four pre-service physics teachers participated for one academic year. The participants firstly revised the main concepts related to scientific practices (experimental validity, planning and design of experiments, identification of variables, controlled experiment, measurement and uncertainty, identification of errors and sources of errors, reliability of measurement) taught in physics lab classes. Secondly, they carried out school experiments (i.e., Hooke’s law, simple pendulum motion, insulation experiment, experiments with simple electrical circuits, refraction and reflection, friction, motion on an inclined plane and electromagnetic induction). Then, they learned about effective teaching and assessment methods (criteria and feedback) in the lab, informed by research. Finally, they taught a 40-minute lesson including an experiment to their peers. 2
4. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 4. Research methodology Individual interviews based on tasks specifically designed to look at ideas related to scientific evidence were conducted. Students’ difficulties while they design, carry out experiments and write lab reports have been identified. All experiments and discussion in the lab were videotaped. Mid-term and final exam items are other important sources of data to answer the first research question. In the second part of the semester, the participants have the opportunity to prepare lesson plans and teach one lesson to their peers. Lessons were videotaped and interviews (during their preparation and after their teaching) were conducted to identify prospective teachers’ difficulties. Their lesson plans and the tasks they developed for their teaching and assessment were collected to be analysed and answer the second research question. The participants were given samples of actual secondary school students’ work and they were asked to give written feedback to them. Research studies give much importance on teachers’ ability to give constructive feedback so that learners understand what they need to do to improve [16,17]. Teachers should be skillful in assessing lab skills and, more generally, students’ performance in the laboratory [18–21]. In addition, a research study by Waren Little et al reported on the benefits for teachers’ learning and professional development when teachers look at actual student work and think about students’ performance [22]. Also, Crespo found evidence that by examining students’ work, teachers have the opportunity to learn about students’ thinking and practices [23]. 4.1. Data collection and analysis The reliability of the analysis is based on the triangulation of the methods employed. Qualitative data analysis [24] was conducted to identify particular trends among the participants. We generated initial categories from interviews, observation transcripts and exam item responses of each teacher. We constantly compared new data from the interviews, and from the observations and exam items (mid- term and final) with the current categories, and refined them. When clarification was needed, we collected more data by conducting focused conversations with the teachers. The data were analyzed by comparing the responses for each question both across and the interviewees and through each interview in order to identify key categories and features among teachers and possible changes within the same teacher. In the next phase, analysis results of the individuals were compared to identify common themes. Triangulation was also applied by comparing and discussing the interpretations by the main researcher and the two project assistants. 5. Results 5.1. Student understanding of the fair test Students had difficulties in planning a fair test or a controlled experiment. One of the difficulties was that they could not identify all relevant variables involved and then decide which variables should be independent, constant, or dependent in their design. The ability to control some variables to design a fair test and related understandings was documented by asking the participants to design a controlled experiment and by looking at their reports when experiments were performed. Most importantly, participants were presented with tasks and asked to evaluate other experimenters’ conclusions. Such tasks were used as research instruments. For example, the following task asks students to evaluate conclusions by previous experimenters who performed an experiment. Notice that students were asked to interpret conclusions based on the design of a controlled experiment. Thus, they were not explicitly asked to examine whether the experiment is a fair test; a basic requirement for the evaluation. 3
5. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 Task 1: Fair Test Ali and Selin wanted to find out the relationship between the resistance and the cross-sectional area of a constantan wire. They performed a series of measurements for six pieces of constantan wire with different cross sectional areas, and reported these in the table above. a) What are the variables involved in this experiment? b) Based on their data, Ali and Selin concluded that the resistance is inversely proportional to the cross- section. Do you agree? c) What is the total percentage error involved in the measurement of the resistance? d) How can this experiment be improved? The respondents gave various answers along with different explanations. The majority of students agree with the conclusion (of Ali and Selin), without examining whether the experiment is controlled or a fair test. Many students did not mention the idea of the fair test and they did not consider whether the design is that of a fair test along with variables, which are involved in the experiment. In fact, they do not mention that there is no information given about length of the wire, temperature and material type. Thus, most of them replied: “I trust the data, because their conclusion is correct; according to the theory, according to what I know”. Similarly, another student wrote: “I agree with their conclusion, because their conclusion matches with the theory”; “I agree with this conclusion, because I learned it from theory”. Those students do not look at the data and the procedure followed to examine whether the experiment is a controlled one. For improvement, they only recommend that they need to repeat without explaining how repetition of measurements improve the results. When students are asked to evaluate a claim that is consistent with what they have been taught (theory), they agree without examining the design of the test and in fact, the validity of the experiment (whether it is actually an adequate fair test). Without examining whether the experiment is a controlled experiment, they report that they trust the data. Thus, although one student mentioned that more variables (as length) should be included, she did not examine whether it is a fair test or not, because their conclusion is consistent with theory. “I agree with this claim since I know it from the theory. If someone does not know the theory, he cannot conclude this from data set because it is not a fair test”. “It is not a fair test, but still I agree with the conclusion, because I know from theory”. Only a few students reported that they do not agree with the conclusion because there is missing information about the length of the wire and the temperature. Such information is crucial when one makes judgements about whether the experiment is a controlled experiment: “We do not have 4
6. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 information about the variable of length. We do not know that it is a fair test. We need information about length and that length is constant (so that it is a controlled experiment)”. “We cannot conclude this because the experiment is not a fair test, they did not report on the length of the wires, which is also an independent variable to determine its resistance …”; “This is not a fair test and it should be a fair test in order to be able to make conclusions”; “We need to have a fair test. Length also needs to be considered”; or, better, “This experiment is not valid because information about length of the wire is missing from our data table. Without it, as a variable, we cannot conclude with any statement like this, since length could be different, and this would affect the conclusion. I do not agree with the two students”. On this basis they suggest that the experiment needs to be controlled. For the twenty-four students the results are shown in Table 1. These categories are not mutually Table 1: Student understanding of fair test in the first two weeks of the semester. categories no of students no mention of the concept of fair test 22 no mention of all the variables involved (length of wires and 19 temperature) “We know from theory that the resistance is inversely 23 proportional to cross-sectional area” This is not a fair test. They should make it a fair test. 2 Student understanding of fair test improved only after targeted instruction and work on more tasks with secondhand data asking them to evaluate the procedure and experimental evidence. In later tasks, they think about the controlled experiment and answer that they need more information about some variables so that they know whether the experiment is a controlled one or not. 5.2. Student difficulties in designing a controlled experiment Our students demonstrated difficulty in recognizing the variables that contribute to an experimental result. For example, this is the case when they are asked about the variables that are involved in the insulation experiment. Task 2: The insulation experiment You, as a physics student, are investigating how the temperature of hot water drops in three similar cans each wrapped with three different insulating materials. 1. What variables do you think affect the temperature of hot water in a can wrapped with an insulating material? 2. How does each variable affect the temperature of hot water while it is cooling down and its rate of change? I would like you to design an experiment that would allow you to decide which of the three insulating materials is best. After you are finished, you need to describe how you plan your experiment, obtain evidence, analyze and explain your data, as well as how you evaluate your experiment. Student difficulties in designing the fair test are recalled in their lab reports, too. For example, they did not include all the variables involved and did not control all the variables in the insulation experiment. This is due to the fact that when performing experiments in lab classes in the physics department, students are always provided with guidelines and instructions on how to proceed. In the insulation experiment, they need to include the following variables: amount of water, temperature drop/ time, starting temperature, material of the container, room temperature, number of layers, and thickness of insulation layer. 5
7. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 Figure 1: The insulation experiment. 5.3. The participants giving feedback to laboratory reports as prospective teachers The participants were asked to give feedback to secondary student lab reports. The reports were actual reports taken from secondary schools. Task 3: You, giving feedback as a teacher Give your written feedback to a secondary school student (grade 11) who submitted the attached laboratory report (the report is attached). In giving feedback, the participants would give information about what needs to get improved and how: • “You need to draw a graph and use it”. • “You need to collect better data / results so that you draw a better graph and be able to look at the pattern”; “Theory is missing”, or, “You need to improve your theory”. • “You need to describe your method. What are the variables involved in your experiment?” • “Nice description of the procedure” “The fair test is OK, but more measurements are needed”; “The data table is good enough”; “Why this number of measurements?” • “Variables (depended and independent) are missing”; “Points are plotted properly”. • “You need to draw a graph”; “What does the graph tell you about the whole pattern?” • “Analysis is missing. What are the different rates of cooling?” • “You need to write a statement about the relationship between the two variables”; “You need to give the graph a title”. Giving feedback helps pre-service teachers develop a good understanding of the quality of a lab report and what is included in each section. Also, the process helps them make the transition from undergraduate student to practicing teacher. They reported that they enjoy grading and giving feedback. It is a process through which they learn how to improve the quality of their own lab reports. The assessment criteria have been developed during the semester. Through practice and by time, they can give detailed feedback. Figure 2. Interaction between giving feedback to student lab reports and improvement of lab reports. 6
8. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 5.4. Our students as prospective teachers: difficulties in preparing and planning to teach Preparing to teach in the laboratory presents a great challenge. Our students are not confident to teach laboratory skills and their plans are weak. The difficulty to prepare a lesson plan with learning goals related to the development of particular experimental skills is clear. Tasks like the following were distributed in the class, as homework and in the exam papers. Such tasks were developed in order to develop teaching methods and assessment methods closely related to experimental skills. Task 4: You, as a teacher Prepare a lesson plan to teach your students how to plan (planning part) and take measurements (obtaining evidence) in the insulation experiment. Task 5: You, as a teacher Write a set of rubrics for assessment of the design an experiment and for making a prediction of the outcome of the experiment. Task 6: You, as a physics teacher You need to write a set of rubrics for assessment of the “analysis and explanation of experimental results” in an experiment and the lab report. The set of rubrics should help teachers in their teaching and in the assessment of the “analysis and explanation” part. In addition, secondary students will use the set of rubrics in peer- and self- assessment of analysis and explanation. When working on the insulation task (Task 2), the participants talked only about the theory of insulation, heat and temperature, heat capacity and Newton’s law of cooling. They also sketched the graph of the cooling down of hot water. Similarly, while they were preparing to teach Ohm’s law experiment, students wrote that the topic is about Ohm’s law (not about the teaching of experimental skills). And the lesson objective is: “By the end of the lesson, students will have understood the relation between voltage and current and the associated Ohm’s law”. Similarly, “Students will be able to identify and determine the difference between heat and temperature”. And, for the insulation and the insulation experiment: “Students will learn what an insulator is”; “Understand how insulators work”. They confuse the context (theory) with the experimental/laboratory skills they want to teach and develop in their secondary students. Instead, we wanted them to talk about the preliminary experiment and the class discussion, in which, students talk in groups about planning, variables and the planning of a controlled experiment. Similarly, in their lesson plans they describe the different parts of an experiment, when they need to write what the teacher and the students will do (teaching methods and strategies) in the laboratory or in the classroom. Only a few and senior students (not more than 3 from the total 24) answered that they will teach the theory of the experiment (Newton’s cooling law) in the introductory lesson. They also said that they are going to review each part of the experiment and what to include in the lab report. “Use instruments to make measurements of temperature”; “Carry out a controlled experiment”. Other teaching goals and student/ learning objectives are: “Students will be able to draw a best fit line/ best fit curve”. Most students had difficulties in preparing a lesson plan to teach students how to plan and design a fair test or, in preparing a lesson to help them improve the analysis of evidence. Through practice they • “Teach to judge the range of measurements they need to take, when and why these need to be repeated, and how to deal with anomalous or discrepant results” • “Teach students to examine evidence for validity and reliability by considering questions of accuracy, error and discrepancy” • “Teach students how to draw the best fit line or best fit curve” • “Teach how to use the graph in the analysis of results” • “How to design a fair test”; “How to collect and record data” 7
9. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 • “How to analyze data”; “How to make a detailed analysis of results by using the plotted graph and make the calculation of slopes”. Students experienced difficulties in writing learning goals and objectives and teaching goals related to the development of lab skills. This is mainly because in all education and physics education classes, they have had training and practice in developing lessons plans to teach only concepts of physics (theory). At the end of the semester were they are able to articulate and write some teaching goals: • “Students learn how to plan the insulation experiment” • “How to take measurements while performing the experiment” • “In the preliminary experiment, they will make decisions about...” • “Students to be able to plan the insulation experiment, to identify the variables involved, to carry out the preliminary experiment and how to design a controlled experiment” They also developed assessment methods; they wrote assessment goals, were introduced to performance assessment, and developed assessment criteria for feedback and rubrics for grading. When it comes them to prepare a lesson to address specific weaknesses (as demonstrated in a lab report), they do not know how to match the learning goals with laboratory activities. One senior student with considerable teaching experience in internship schools wrote in a reflection journal: “Planning a lesson was difficult for me. I needed some guidance to design and plan my lesson. I keep the theory simple, so that I teach them laboratory skills. I learned about progression in teaching (teaching sequences)”. One of the important issues in their teaching is that pre-service teachers do not seem to understand the common components of controlled experiments. There is also an interplay between theory (the context of the experiment) and the development of lab skills, which makes them attach more importance to the underlying theory than to the development of experimental skills, when asked to evaluate Pre-service teachers have difficulties in connecting what they learn with teaching practice. Our participants were not successful in applying what they learned in the course to prepare their lesson plans and do the actual teaching. They experienced particular difficulties when preparing lessons plans for teaching in the lab, when writing lesson objectives and so on. Only a few of from the total of twenty- four students had prepared adequate lesson plans. 6. Discussion of results and conclusions This study has provided strong evidence about the difficulties and needs of our participants. The aim was to document pre-service physics teachers’ difficulties in understanding scientific evidence and then, difficulties in teaching and assessing lab skills and experiments in the school laboratory. Tasks were developed to elicit such understandings and challenges. Conversely, the findings guided the development of tasks and teaching materials to explicitly address such deficiencies. In addition, the study confirmed research results from previous studies, but also elicited new findings. Firstly, the results show that teachers have a limited understanding of experimental design and concepts of validity and reliability. Similar to findings from two studies [6, 25] the participant students have a limited understanding of the uncertainty that is inherent to each measurement. There is strong belief in making one measurement carefully [1]. The participants struggled with many aspects of scientific evidence and with writing laboratory reports. Through discussion in the interviews and feedback by the instructor, students improve the versions of lab reports and submit lab reports of better quality for subsequent experiments. The structure of the reports and what is included in each section improves across subsequent lab experiments for each participant. Secondly, the results indicate serious shortcomings in the preparation of future physics teachers. The study revealed a lack of knowledge related to experimental skills. Pre-service teachers have difficulties in connecting what they learn with teaching practice. Our participants were not successful in applying what they learned in the course to prepare their lesson plans and do the actual teaching. They experienced particular difficulties when preparing lessons plans for teaching in the lab, when writing lesson 8
10. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 objectives and so on. Only a few of them had prepared adequate lesson plans. It is also true that the participants have had different teaching experiences prior to joining the research study. One strength of the project is that with various research instruments and targeted instruction succeeded to elicit student understanding along with the underlying reasoning. In terms of research methodology, the use of only interviews may not explore the full range of difficulties and related reasoning. Moreover, the combination of interviews, targeted instruction, classroom discussion as well as work on tasks brought up various misunderstandings and difficulties that students experience when they design controlled experiments and they evaluate scientific evidence. Secondly, an important achievement of the study has been to document student reasoning in enough detail to enable our participants to improve through the design of effective instruction and then, to help them practice effective instruction addressed to secondary school students. This project presented and analyzed the types of understandings that physics teachers should have developed in order to teach and assess experiments in the school lab. All these understandings constitute a body of knowledge for teaching in the lab to help prospective physics teachers become familiar with “rules” of evaluating experimental procedures and evidence [26]. Those types of understandings will support teachers in developing better concepts of evidence and further, become confident with teaching in the lab. This study raises important questions for physics teacher educators. Pre-service teachers need to develop confidence in teaching physics experiments and in developing assessment goals and methods related to laboratory skills. They need repeated practice in lab skills. There are implications for pre- service physics teacher education, because many of the participants are expected to teach lab work in secondary schools. The development of special courses for the preparation of physics teachers is necessary. Adequate preparation of science teachers is vital to ensuring good teaching of science in secondary schools. The standard lab classes in the physics departments are not enough. They do not address the needs that they will have as teachers. Most importantly, our physics pre-service teachers are not the best undergraduate students. Instead, it is usually the case that the low achievers come to the faculty of education to become physics teachers. Standard instruction often does not improve student understanding of scientific evidence. Trainee teachers experience difficulties and have needs which must be considered in planning and implementing their initial education. Instruction that explicitly addresses the underlying reasoning can have a significant effect to the development of scientific evidence. The study was carried out as a start-up project funded by BAP 10800. [1] Allie S and Buffler A 1998 A course in tools and procedures for Physics I Am. J. Phys. 66(7) 613–24. DOI: 10.1119/1.18915 [2] Hofstein A and Kind P M 2012 Learning in and from science laboratories. In: Fraser B, Tobin K and McRobbie C (eds) Second International Handbook of Science Education. Springer International Handbooks of Education, vol 24. Springer, Dordrecht (pp. 189–207). DOI: 10.1007/978-1-4020-9041-7_15 [3] Lunetta V N, Hofstein A and Clough M 2007 Learning and teaching in the school science laboratory: An analysis of research, theory and practice. In: Lederman N and Abel S (eds) Handbook of research on science education (pp. 393-441), Mahawah, NJ : Lawrence Erlbaum. [4] Séré M-G 1999 Learning Science in the Laborator: Issues raised by the European Project “Labwork in Science Education”. In: Bandiera M, Caravita S, Torracca E, Vicentini M (eds) Research in Science Education in Europe. Springer, Dordrecht pp. 165–74. DOI: 10.1007/978- 94-015-9307-6_21 [5] Séré M-G, Journeaux R and Winther J 1998 Enquête sur la pratique des enseignants de lycée dans le domaine des incertitudes [Investigate into the practice of the teachers of high school in the field of uncertainties] Bulletin d’Union des Physiciens 801 247–254. [6] Lubben F, Campbell B, Buffler A and Allie S 2001 Point and set reasoning in practical science 9
11. GIREP-ICPE-EPEC 2017 Conference IOP Publishing IOP Conf. Series: Journal of Physics: Conf. Series 1286 (2019) 012016 doi:10.1088/1742-6596/1286/1/012016 measurement by entering university freshmen Sci. Educ. 85(4) 311–327. DOI: 10.1002/sce.1012 [7] Séré M-G, Journeaux R and Larcher C 1993 Learning the statistical analysis of measurement errors. Int. J. Sci. Educ. 15(4) 427–438. DOI: 10.1080/0950069930150406 [8] Allie S, Buffler A, Kaunda L, Campbell B and Lubben F 1998 First-year physics students’ perceptions of the quality of experimental measurements Int. J. Sci. Educ. 20(4) 447–459. DOI:10.1080/0950069980200405 [9] Boudreaux A, Shaffer P S, Heron P R L and McDermott L C 2008 Student understanding of control of variables: Deciding whether or not a variable influences the behavior of a system Am. J. Phys. 76(2) 163–70. DOI: 10.1119/1.2805235 [10] McDermott L C 1990 A perspective on teacher preparation in physics and other sciences: The need for special science courses for teachers Am. J. Phys. 58(8) 734–4. DOI: 10.1119/1.16395 [11] McDermott L C 2014 Melba Newell Phillips Medal Lecture 2013: Discipline-Based Education Research—A View From Physics Am. J. Phys. 82(8) 729–41. DOI: 10.1119/1.4874856 [12] Bransford J D, Brown A and Cocking R 1999 How people learn: Mind, brain, experience, and school (Washington DC: National Research Council) [13] Darling-Hammond L, Bransford J, LePage P, Hammerness K and Duffy H 2007 Preparing teachers for a changing world (Jossey-Bass). [14] National Research Council 2005 America’s Lab Report: Investigations in High School Science. Washington, DC: The National Academies Press. [15] Wilson S et al (Eds) 2015 Science Teachers’ Learning: Enhancing Opportunities, Creating Supportive Contexts (Washington, DC: The National Academies Press) [16] Black P, Harrison C, Lee C, Marshall B and Wiliam D 2003 Assessment for Learning- Putting it into practice (Maidenhead, U.K: Open University Press) [17] Wiliam D and Leahy S 2015 Embedding formative assessment: Practical techniques for K-12 Classrooms (West Palm Beach, FL: Learning Sciences International) [18] Brown J H and Shavelson R J 1996 Assessing hands-on science. Corwin Press: Thousand Oaks, CA [19] Gioka O 2006 Assessment for learning in physics investigations: assessment criteria, questions and feedback in marking Phys. Educ. 41(4) 342–6 [20] Keys C W, Hand B, Prain V and Collins S 1999 Using the Science Writing Heuristic as a Tool for Learning from Laboratory Investigations in Secondary Science J. Res. Sci. Teach. 36(10) 1065–84. DOI: 10.1002/(SICI)1098-2736(199912)36:10<1065::AID-TEA2>3.0.CO;2-I [21] Mintzes J J, Wandersee J H and Novak J D (Eds) 2005 Assessing science understanding: A human constructivist view. London: Elsevier Academic Press [22] Waren Little J, Gearhart M, Curry M and Kafka J 2003 Looking at student work for teacher learning, teacher community and school reform Phi Delta Kappan 185–92 [23] Crespo S 2000 Seeing More Than Right and Wrong Answers: Prospective Teachers’ Interpretations of Students’ Mathematical Work J. Math. Teach. Educ. 3(2) 155–81. DOI: 10.1023/A:1009999016764 [24] Miles M B and Huberman A M 1994 Qualitative data analysis: An expanded sourcebook. Sage Publications, Inc., Thousand Oaks, CA [25] Varelas M 1997 Third and fourth graders’ conceptions of repeated trials and best representatives in science experiments J. Res. Sci. Teach. 9 853–872. DOI: 10.1002/(SICI)1098- 2736(199711)34:9<853::AID-TEA2>3.0.CO;2-T [26] Thames M H and Ball D L 2010 What mathematical knowledge does teaching require? Knowing mathematics in and for teaching Teach. Child. Math. 17(4) 220–229 10