Do you want to publish a course? Click here

Relative impacts of different grade-scales on student success in introductory physics

192   0   0.0 ( 0 )
 Added by David Webb
 Publication date 2019
  fields Physics
and research's language is English




Ask ChatGPT about the research

In deciding on a students grade in a class, an instructor generally needs to combine many individual grading judgments into one overall judgment. Two relatively common numerical scales used to specify individual grades are the 4-point scale (where each whole number 0-4 corresponds to a letter grade) and the percent scale (where letter grades A through D are uniformly distributed in the top 40% of the scale). This paper uses grading data from a single series of courses offered over a period of 10 years to show that the grade distributions emerging from these two grade scales differed in many ways from each other. Evidence suggests that the differences are due more to the grade scale than to either the students or the instructors. One major difference is that the fraction of students given grades less than C- was over 5 times larger when instructors used the percent scale. The fact that each instructor who used both grade scales gave more than 4 times as many of these low grades under percent scale grading suggests that the effect is due to the grade scale rather than the instructor. When the percent scale was first introduced in these courses in 2006, one of the authors of this paper, who is also one of the instructors in this data set, had confidently predicted that any changes in course grading would be negligible. They were not negligible, even for this instructor.



rate research

Read More

Covariational reasoning -- reasoning about how changes in one quantity relate to changes in another quantity -- has been examined extensively in mathematics education research. Little research has been done, however, on covariational reasoning in introductory physics contexts. We explore one aspect of covariational reasoning: ``goes like reasoning. ``Goes like reasoning refers to ways physicists relate two quantities through a simplified function. For example, physicists often say that ``the electric field goes like one over r squared. While this reasoning mode is used regularly by physicists and physics instructors, how students make sense of and use it remains unclear. We present evidence from reasoning inventory items which indicate that many students are sense making with tools from prior math instruction, that could be developed into expert ``goes like thinking with direct instruction. Recommendations for further work in characterizing student sense making as a foundation for future development of instruction are made.
Laboratory courses are key components of most undergraduate physics programs. Lab courses often aim to achieve the following learning outcomes: developing students experimental skills, engaging students in authentic scientific practices, reinforcing concepts, and inspiring students interest and engagement in physics. Some of these outcomes can be measured by the Colorado Learning Attitudes about Science Survey for Experimental Physics (E-CLASS), a research-based assessment that measures students views about experimental physics. We used E-CLASS at the University of Colorado Boulder to measure learning outcomes during a course transformation process in which views about experimental physics were reflected in the learning goals. We collected over 600 student responses per semester from the large introductory laboratory course, both before and after implementing the course transformation. While we observed no statistically significant difference in overall post-instruction E-CLASS scores before and after the transformation, in the transformed course, student responses to three E-CLASS items that related to the goals of the transformation were more favorable than in the original course.
While laboratory instruction is a cornerstone of physics education, the impact of student behaviours in labs on retention, persistence in the field, and the formation of students physics identity remains an open question. In this study, we performed in-lab observations of student actions over two semesters in two pedagogically different sections of the same introductory physics course. We used a cluster analysis to identify different categories of student behaviour and analyzed how they correlate with lab structure and gender. We find that, in lab structures which fostered collaborative group work and promoted decision making, there was a task division along gender lines with respect to laptop and equipment usage (and found no such divide among students in guided verification labs).
Computational Thinking (CT) is still a relatively new term in the lexicon of learning objectives and science standards. There is not yet widespread agreement on the precise definition or implementation of CT, and efforts to assess CT are still maturing, even as more states adopt K-12 computer science standards. In this article we will try to summarize what CT means for a typical introductory (i.e. high school or early college) physics class. This will include a discussion of the ways that instructors may already be incorporating elements of CT in their classes without knowing it. Our intention in writing this article is to provide a helpful, concise and readable introduction to this topic for physics instructors. We also put forward some ideas for what the future of CT in introductory physics may look like.
Proficiency with calculating, reporting, and understanding measurement uncertainty is a nationally recognized learning outcome for undergraduate physics lab courses. The Physics Measurement Questionnaire (PMQ) is a research-based assessment tool that measures such understanding. The PMQ was designed to characterize student reasoning into point or set paradigms, where the set paradigm is more aligned with expert reasoning. We analyzed over 500 student open-ended responses collected at the beginning and the end of a traditional introductory lab course at the University of Colorado Boulder. We discuss changes in students understanding over a semester by analyzing pre-post shifts in student responses regarding data collection, data analysis, and data comparison.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا