Effects of Educator Training on New Instructional Behaviour


This paper can be an academics critique of an article written by de Jager, Reezigt, and Creemers (2002) titled: The effects of teacher training on new instructional behaviour in reading understanding. The creators undertook a research study to examine the results of tutor inservicing on sensible professor behaviours. My examination systematically targets specific areas of this article in conditions of process and validity of research methods and results. I have attempted to develop a cohesive and unified justification which not only expounds the particulars of the research but which also formulates an obvious interpretation of this research throughout. I suggest that how big is the sample and the technique of selecting themes for the experimental categories makes the research externally invalid and thus greatly lowering generalizability to the best, and perhaps even to the immediate, human population.

Quantitative Research Article Critique

In their article, The effects of teacher training on new instructional behaviour in reading understanding, de Jager, Reezigt, and Creemers (2002) put together a quasi-experimental research design regarding three sample organizations (two experimental and one control) that have been drawn from an instantaneous society of eighty-three primary school instructors in the north area of the Netherlands. In their introduction to this article, the authors declare that "teachers need suitable instructional models that provide them with suggestions for new types of training and they should have access to inservice training, which helps them to successfully implement these models in their regular lessons" (de Jager et al. , 2002, p. 832). This affirmation basically outlines the premise behind the study - it isn't a research question but a affirmation of belief for the authors; a assertion of belief upon which they attract framing the purpose and emphasis of their research. The writers articulate their acceptance to the fact that educating must be targeted around the principles of constructive, effective and student-based learning where in fact the instructor facilitates and courses his/her pupils with their own understandings. In addition they recognize the actual fact that while educational theory has advanced to meet the higher needs of the current paradigm, educational practice is often significantly less than up to date. The study itself is based on exploring the opportunity of reconciling educational theory and educational practice in pragmatically.

Research Problem

In their research, de Jager et al. (2002) concentrate on answering a particular research question which is discussed clearly in the analysis on webpage 832. They offer their problem as a statement somewhat as an interrogative so, for clearness, I shall paraphrase: Can teachers in primary classes learn in using the models of either cognitive apprenticeship or direct education? Of particular importance to the writers is the integration of metacognitive skills (in reading) into these coaching models. They note that, through research such as that conducted by Muijs & Reynolds (2001) among others, it has been proven that training in Direct Education (DI) is effective for enhancing the development basic skills. On top of that, they show that lab tests have proven the potency of Cognitive Apprenticeship (CA) on the development of metacognitive skills in managed situations (de Jager et al. , 2002, p. 832). Due to these facts, it is practical to assume that similar tests (as well as perhaps similar results) can be conducted and analysed in real class room situations. Indeed, the take action of replicating methods (even - and perhaps actually -- with some changes) reaches the heart of small range quantitative research: "Quantitative research assumes the probability of replication" (Cohen, Manion & Morrison, 2003). Thus, the likelihood of training and implementation is researchable, as it builds upon pre-existing research.

The creators propose a functional necessity because of this research problem to be explored. As advised above, a theory foundation because of this research is pre-existing and, as such, the authors plainly plan to provide significant results in the form of practical application of these theoretical concepts. In the same way it is suggested that the research is pragmatic in its character and implications, it is clear that the research is also significant. Essentially, the research is dependant on changing the instructional strategy of teachers to meet current final results - benefits which seem to be to be on par with those of the current system in Newfoundland. That's, they appear to be college student centered and centered on the development of metacognitive thinking skills. The authors make reference to De Corte (2000) regarding this matter: "Learning is a constructive, cumulative, self-regulated, goal-oriented, situated, collaborative, and independently different process of knowledge building and meaning building" (de Jager et al. , 2002, p. 831). Teachers must be dynamic and able to adjust their methods and coaching styles to support the shifts in the curricula and pedagogy inherent in today's paradigm. Therefore, research in to the opportunity of adapting to new instructional models is very significant - indeed, what could be more significant?

The creators explicitly state that their research study is "a quasi-experiment with a pre-test, post-test, control group design" (de Jager et al. , 2002, p. 835). Presumably, the best population of the study will be all teachers in the whole world. The immediate people includes 83 primary university instructors in the northern part of the Netherlands. Not only is it identifiable by locale and occupation, the populace is also particular for the reason that each individual within the group was previously acquainted with and had been utilizing I UNDERSTAND What I Read - a particular curriculum for reading comprehension (de Jager et al. , 2002, p. 835). That is an important factor which could introduce a possible problem in conditions of exterior validity. That's, assuming that the authors desire to make their findings generalizable beyond the immediate human population (whether through following studies of their own or others), the prior use of the specific curriculum could be viewed as an independent variable. Regardless, the immediate people for the analysis is clearly indicated by the authors.

The recognition of indie and dependent parameters is of extreme importance in research it is the groundwork of any test or similar activity. De Jager et al. (2002) explored the likelihood of changing educational methods through training. So, the independent parameters in this research are the training of two experimental organizations (one group in DI and one in CA) and the based mostly variable is, therefore, the addition of instructional methods which can be clearly designed to develop metacognitive reading skills. While this is actually the case, these factors are not suggested clearly within the study. Nowhere does the terms of independent and dependent variables occur explicitly. On top of that, beyond these parameters, other significant extraneous parameters exist that could be looked at as independent parameters, like the pre-existing knowledge of a specific curriculum model (mentioned above), the age of the participants, the educational training and experience of the immediate people and research sample, the utilization of alternative curriculum materials, etc. I'll package with these matters in another section.

Review of the Literature

The authors pull on pre-existing research to formulate the purpose of their own review. Additionally, they appear to have attracted on a comprehensive list of options throughout the analysis. For instance, under section 3 with their record, they include an comprehensive background for the development of their in-servicing models. Through the execution of the self-employed variables (training in DI and CA) they offer evidence which reveals that they meticulously followed the results layed out in the books on inservice training. While this is actually the case, in addition they note a particular difficulty inherent in this mass of research books, specifically that no information exists to indicate whether instructors should be inserviced via the techniques that they are expected to use (eg: if the CA group have been inserviced using the CA model?)

Beyond their extensive use of source referencing in conditions of the arrangements for the study, the writers also appear to rely greatly on pre-existing literature in the development of the observational instrument and in the justification of the methods in conditions of sampling. In conditions of the latter example, the support attracted from their mention of Anders, Hoffman, & Duffy (2000) is fallacious in that it generally does not truly validate their approach to sampling (I offer with this explicitly in selecting Content section below). Within the past, the development of this tool in two parts (Low Inference Observation and High Inference Observation) relied seriously on the work of Booij, Houtveen, & Overmars (1995); Sleipen & Reitsma (1993); and Burgess et al. (1993). Specifically, in their use of High Inferece Observation, that could possibly have launched problems of interior validity, the writers reference Gower's congruence strategy (1971) in order to dispel any notion of issues with interrater consistency (the probability of unintentional subjective bias on the part of trained observers).

Generally, the review of literature seems extensive. The authors guide past studies on the adaptation of instructional models, specifically on instructor trained in the execution of the DI model (de Jager et al. , 2002, p. 832). Somewhat troubling is the actual fact that no clear reference point emerges for the lab-testing of the CA model. Resnick (1989) is cited in terms of introducing the concept of "constructivist models such as cognitive apprenticeship" (de Jager et al. , 2002, p. 832) but the actual experiments are not cited.

No emphasis appears to be placed on major sources. Interviews, artifacts, etc. weren't involved in the study or its development beyond the framework of the inservice training where members were free to discuss and socialize openly. As this is not a source but a dynamic and inclusive area of the study, this cannot be seen as a principal source.

In terms to be up-to-date, the overview of the books seems valid. That's, the article carries a well-organized bibliographical reference list of 35 studies and 13 of those studies were released during or after 1995. A the greater part of those studies took place after 1990. Additionally, as can be seen from the discourse above, the literature appears to be straight related to the development of this study and is also mixed up in development of the study hypotheses (mentioned explicitly within the next section).

Research Hypothesis

Unlike the self-employed and dependent factors, the authors of the study obviously, consisely, and explicitly signify their hypotheses about the outcomes of the study.

Particularly, these hypotheses are the following:

After appropriate trained in which they learn how to apply one of these instructional models, educators will more and more show the primary characteristics of cognitive apprenticeship or immediate instruction.

Teachers in both experimental communities will increase the general quality of these instructional behaviour.

The professors in both experimental groups will learn to target more on understanding skills and metacognitive skills (thus they'll spend more lessons time on these skills) (de Jager et al. , 2002, p. 834).

These hypotheses clearly follow from the books cited within the section of the article working with the theoretical backdrop of the Direct Teaching and Cognitive Apprenticeship types of instructions. Implicit within these hypotheses is the suggestion a true causal romantic relationship will exist between the independent and dependent variables. That's, the training in CA and DI will bring about teachers being with the capacity of applying these models. While this seems to be a simplistic recommendation initially, if proven true, the implications of the study on the inservicing of new curricula and instructional methods will be significant. In essence, the study will prove the validity of the notion that instructors not only need inservicing but they can benefit from it. What is perhaps most effective about these hypotheses is they are clearly testable via the observational tool developed for the analysis. Characteristics of lessons can be discovered, and lesson time/concentrate can be measured. The second of the three hypotheses is the only person which really presents a difficulty in terms of measurement as it depends on an research of a very subjective subject matter - basic instructional quality. While this is actually the case, the observational tool includes 16 items of high inference observation (though it is arguable that this is insufficient). Thus, the creators have made specific provision for examining all three hypotheses.

Selection of Subjects

The immediate human population of the research is evidently and explicitly discovered by the writers in section 4. 1 of the article (I have mentioned the immediate population of this analysis in the Release and Research Problem parts of this paper). With the 83 individuals in the immediate society, three sample teams were made from 20 volunteer members/subjects. Of these three groups, two were experimental and one was used as a control group. The first of the two experimental groups consisted of 8 teachers who have been inserviced in CA; the next experimental group consisted of 5 teachers who had been inserviced in DI; the control group contains the rest of the 7 professors. The subjects are shown to be equivalent in terms of years educating experience (average 22. 1, SD=6) basically in their experience with the precise reading curriculum (average 2. 8, SD=1. 9).

Much dialogue regarding these facts has took place among my co-workers in Education 6100. As the sample groups appear to be comparable, there are some serious questions of validity which must be increased here. First of all, one must note that 95% of the things will statistically have ten or even more years educating experience. That is significant because the type of the study study is such that it attempts to assess teachers' ability to change their instructional methods. In their section on teacher training, the authors cite Galton & Moon (1994) where it is stated that "More capable teachers could find it even more complicated to improve than novice professors" (de Jager et al. , 2002, p. 834). If this is indeed the case, it seems as though the sample groupings (being quite experienced) will see it specifically difficult to improve their methods and, so, this may distort the research findings. An attempt must have been designed to include teachers of various experience levels in order to make the conclusions more generalizable and for that reason more externally valid.

In addition, the technique of sampling employed by the writers seems significantly less than valid. That is, the authors made no try to control the parsing of topics into the three sample groupings. What's worse, they allowed the topics themselves to choose between the teams. The writers cite Anders, Hoffman, & Duffy (2000) and their comments on voluntary involvement as a good affect on inservice tutor training (de Jager et al. , 2002, p. 835). There's a deep difference in taking volunteers and allowing those volunteers to choose their experimental sample groups. Like a instructor myself, it is clear that those teachers who 'volunteered' for the CA inservice training were predisposed (or at the minimum, more open up) compared to that method of instructions; also with those for the DI group. Beyond this, 20 volunteers is hardly a reliable sample from 83 individuals - corresponding to Container 4. 1 on site 94 of the course words, were this a random sample, a society size of 80 subjects would need a test size of 66 subject matter (Cohen et al. , 2003). Thus, the sample size is quite disproportional to the immediate population from whence it arrived. Besides this, the content who consent to the research project will necessarily have significantly more in keeping with the other person than they are doing with the remaining associates of the immediate society (if little or nothing else, the actual fact they are volunteers in the analysis). It really is in the sampling of the experimental and control teams that the inherent flaw of the review rests. The authors freely admit that this method (if one could call it that) was useful for pragmatic reasons (de Jager et al. , 2002, p. 835). It was clearly a calculated compromise, but the one that, in my opinion, invalidates their work.


All experimental and quasi-experimental research must include some method(s) or tool(s) for evaluating the effects of the unbiased adjustable(s) on the reliant adjustable(s) - otherwise, there would be no indicate the analysis. In conditions of the method for evaluation in this specific study, the writers developed a survey style observation tool which was applied four times for each experimental group in support of twice for the control group. As the device was designed specifically for the measurement of these groups, it falls into the group of non-parametric testing. This assertion is verified by the actual fact that review/questionnaire research methods deliver non-parametric data (Cohen et al. , 2003, p. 77) and also by the actual fact that the equivalence of the communities before treatment was "tested with the Kruskal-Wallis One-way examination of variance for independent organizations" - an ordinal rank test (de Jager et al. , 2002, p. 837). Ordinal scales produce non-parametric data (Cohen et al. , 2003, p 77). Another ordinal scaling test ( The Mann-Whitney U test) was utilized by the authors in the evaluation of the data generated by both low and the high inference servings of the observation tool. Thus there can be without doubt that the observation tool is non-parametric. This is significant because, like so many other components of this analysis, this creates problems of exterior validity as a non-parametric tests "do not make any assumptions about how precisely normal, even and regular the distributions of scores will be" (Cohen et al. , 2003, p. 318).

The inner validity of the test seems to be quite positive. Indeed, the experts take great pains to ensure this and they explicitly summarize their initiatives in the development of their observation tool, especially about the use of low and high inference observations. Additionally, they point out that while five individuals were trained to hire the observational tool, the interrater stability was quite high (0. 81).

The observation tool originated to include two sections. The low inference section seems to have consisted of a checklist that was obtained at two minute intervals. This list simply indicated which activities were occurring at each interval; while helpful for gathering quantitative data, it did not allow for qualitative observation. Therefore, the next section of the observational tool consisted of a higher inference evaluation in the form of a Likert Size (a score tool normally having gradable/equivalent range of reactions to a prompt for observation - usually on a scale of 1-5). The use of the Likert (score) Scale could introduce problems in conditions of inner validity in that observers might not wish to reveal extremes (eg: circling a 1 or a 5) but instead they may adhere to the mid-range of the level. This is practical, as the circling/sign of your extreme value on the ranking scale is comparable to a dichotomous level (binary in aspect).

The primary benefit to this sort of evaluation in terms of inner validity appears to be in the actual fact that the topic will not complete the survey but instead, s/he is detected by an (ideally) purpose third party. Of additional relevance this is actually the fact that the observers are that - they are not the researchers. The opportunity of observation bias (the likelihood of inadvertently - or otherwise - 'doctoring' the results in an attempt to verify research hypotheses) is greatly reduced. While this is the case however, the technique or amount of observer training is not mentioned it is merely explained that five people were trained (de Jager et al. , 2002, p. 837).


The design of the research study is explicitly suggested as "being quasi-experimental with a pre-test, post-test control group design" (de Jager et al. , 2002, p. 835). The indie factors in this research are plainly manipulated for the reason that the inservicing of both control groups was methodologically carried out. Also, each treatment was conducted individually of the other even though both dealt with metacognitive skills, each originated to attain those skills using completely different methods. The control group received absolutely no training or coaching aids, and had not been affected by the studies beyond the actual fact that during two lessons within a school season, each member was observed during the delivery of an lessons on reading understanding. While this is the case, there are always a wealth of extraneous factors which are not (and cannot) be studied into consideration here. Firstly, the knowledge of the topics in the sample groups can't be denied to be significant (see above). On top of that, the fact that both experimental groups received completely alternate materials has serious repercussions on the validity of the research findings. The study was supposedly based on the likelihood of adapting instructional methods. It would have been more valid if the test groups all made use of the same curriculum materials because, as it stands, we can not be sure that any differing leads to the observations of the teams are a result of the training or the use of different (perhaps superior) materials. Who is to say that the control group, outfitted with the different materials, wouldn't normally experienced similar or equal results?

Another extraneous problem, beyond sampling and various curricula rests in the actual fact that as the study involves educator training, it always includes student engagement (as well as perhaps achievement) as well. The students are not indicated as being comparative in this research. With a group of 20 teachers, we're able to be discussing a sample populace of students comprising up to 500 (or even more) students. While it might be argued that bears no relavence to the problem, I hold that it does. What can be achieved with one class may be considered a nightmarish impossibility with another; what comes easily for the main one group, may be quite difficult for the next. Consider the idea of group work, which is included in the observation tool developed for this study. Some classes are more likely to accept this style of instruction than others. In short, the students must have been considered.


The results of this research article are proven to demonstrate the hypotheses of the creators to be correct. A relative change in instructional strategy happened within both experimental categories and it is attributed by the writers to the indie factors: "Differences in the observation a the finish of the institution yr can be attributed to the experimental treatments" (de Jager et al. , 2002, p. 837). The statistical results of the quasi-experiment on both experimental test groups is described and displayed in separate sections and separate tables. This emphasizes the fact that there was no intent for the researchers to compare the DI and CA types of training (although they do evidently declare that both were evenly difficult/easy to use). While this is actually the case, it could have perhaps been useful to see the results combined in a single chart with two unbiased variables and two experimental organizations, one cannot help but to question about human relationships among the effects on the dependent variables. However, what is perhaps most troubling about the representation of the info in these charts is usually that the charts do not correspond. I suppose that this is due to the fact that they represent the execution of two particular instructional methods, however the amount of concurrence in terms of high inference appears to be quite lower in my estimation.

Whereas the authors state that significant differences exist between your pre-test and post-test of the experimental categories, these dissimilarities are somewhat limited in scope. In conditions of the CA group, "only four of the 13 indications show significant variations and with the DI group only four distinctions on the high inference indications are significant" (de Jager et al. , 2002, p. 838). While this holds true, additionally it is reported that on most indicators the instructors of both experimental teams show more favourable behaviours than the control group. As stated in the previous section however, this may be related to extraneous variables and not necessarily to the 3rd party variables (ie: tutor training).

On the complete, the results of the research do not appear to be very valid or indicative. Yes, there are some differences, but there is no certainty as to whether this is related to the inservice trained in either the DI or CA instructional models. Beyond this, as layed out above, an inadequate number of content was sampled from the immediate society for the leads to be statically valid. While appropriate statistical tests, such as the Mann-Whitney U test, were used, the actual fact that the figures are insufficient delimits the importance of these testing. More often than not, the significance of the differences which can be purportedly the consequence of the indie variable appears to be overrated.

Discussion and Conclusion

For the most part, the conclusion of the article can take the ideas shown in the Results section and makes value statements based after them. Thus, the interpretation of the results is reserved for this section (although some interpretation is undertaken in the last section). Generally, the results are not discussed with regards to previous research studies. It is mentioned that the professors who were been trained in DI and CA demonstrated more characteristics of these models than the teachers who were not inserviced in the execution of these models - this is actually the case. Exhibiting characteristics of a model of training, however, does not constitute success.

In terms of success, it could very well be beneficial to reiterate the study question because of this article: Can instructors in primary classes learn in using the types of either cognitive apprenticeship or direct training? The authors point out that the project was a success: "teachers, when they are correctly trained and coached, can transform their behaviour in accordance with relatively new ideas about learning and educating predicated on constructivist theories" (de Jager et al. , 2002, p. 839). This plainly indicates that the results of the research conducted by the creators is constant with the results of both previous research in DI style of training and the lab test tests with the CA model of instruction. While this is the case, the creators also signify that the ratings of the experimental organizations "did not differ significantly from the control instructors" (de Jager et al. , 2002, p. 839). This means that that the experts, while securing to the idea of success, acknowledge the fact that their results are non-conclusive in and of themselves. They recommend amendments to the study methods which they used throughout their review and ask others to reproduce the test.

Personal Analytic Statement

As a complete, I found this informative article to be missing merit in conditions of reverberationary quality - exterior validity. That is, although it is interesting to see how the training in different instructional methods may possibly affect the teaching ways of regular classroom professors, it is relatively superfluous in that it generally does not offer any real generalizable results. The sample groups were terribly selected and didn't represent the immediate society. Therefore, the findings of the research are particular to the individuals researched. Of course, as regular educators which were shown to be equivalent (though, in my opinion, not convincingly so) the examples can be viewed as to be possibly representative of other individuals, hence the creators' call for replication of the study. Beyond this, the results themselves do not seem to be conclusive as there is no real general factor between the test and the control categories. Of particular interest if you ask me was the fact that on the main one hand, the authors clearly point out that educators who are incredibly experienced in the profession will have a far more difficult time in accepting and incorporating changes with their instructional methods and on the other hand, they decide on a group which statistically has 95% of its topics enter the experiment with 10 or even more many years of experience. If this is intentional (ie: use the most difficult to show the power of inservicing), the authors fail to reveal their motives.

  • More than 7,000 students prefer us to work on their projects
  • 90% of customers trust us with more than 5 assignments
submit a project

Latest posts

Read more informative topics on our blog
Shiseido Company Limited Is A Japanese Makeup Company Marketing Essay
Marketing Strength: Among the main talents of Shiseido is its high quality products. To be able to satisfy customers, the company invested a great deal...
Fail To Plan You Plan To Fail Management Essay
Management This report will concentrate on two aspects of project management, their importance within the overall project management process. The report...
Role of High-protein Diet in Weight Management
Nursing Structured Representation: Probably one of the most wide-spread and popular problems on earth is the weight problems that people is suffering...
Waste To Prosperity Program Environmental Sciences Essay
Environmental Sciences Urban and rural regions of India produce very much garbage daily and hurting by various kinds of pollutions which are increasing...
Environmental Studies Pollution Introduction Many people across the world can remember having walked on the street and seen smoke cigars in the air or...
Soft System Methodology
Information Technology Andrzej Werner Soft System Methodology can be described as a 7-step process aimed to help provide a solution to true to life...
Strategic and Coherent methods to Recruiting management
Business Traditionally HRM has been regarded as the tactical and coherent method of the management of the organizations most appreciated assets - the...
Religious Healthcare Organisation
Health Religious Health Care Introduction I help the firm of consulting. Spiritual HEALTHCARE of Middleville community have appointed us to identify and...
Enterprise Rent AN AUTOMOBILE Case Analysis Business Essay
Commerce With a massive network of over 6,000 local rental locations and 850,000 automobiles, Organization Rent-A-Car is the greatest rental car company...
Check the price
for your project
we accept
Money back
100% quality