Quantitative Research Manuscript Critique
R E S E A R C H A R T I C L E
Model based assessment of learning dependent change within a two semester class
Nadine Schlomske Æ Pablo Pirnay-Dummer
Published online: 13 May 2009 � Association for Educational Communications and Technology 2009
Abstract The following empirical study examines the acquisition of expertise. The model structures in the domain of empirical methodologies were examined in a time period
of two semesters. At each measurement point (N = 5), the model structures of the ref-
erence groups were compared with those of the group of learners. The group of learners
consisted of (N = 31) students who did not have any experience with research methods
and their theoretical foundations. The results indicate that the external criteria enabled a
precise assessment of learning dependent model change.
Keywords Acquisition of expertise � Learning dependent change � Learning and instruction � Mental models � Knowledge assessment
Knowledge and excellence are acquired by students as they learn. Assessing the acquisition
of expertise requires valid and reliable instruments. Existing instruments already in use at
schools and universities are often very time consuming and difficult to handle in procedure
and analysis. The present study examines whether MITOCAR (Model Inspection Trace of
Concepts and Relations) is an appropriate tool for predicting the acquisition of expertise.
MITOCAR is based on the theory of mental models (c.f. Seel 1991) and is an automated
instrument for the analysis of knowledge and expertise (Ifenthaler 2008; Pirnay-Dummer
2006, 2007; Pirnay-Dummer and Spector 2008; Pirnay-Dummer and Walter 2008).
Moreover, the present study is interested in whether or not the models identified with the
help of this instrument can be drawn on as external criteria in order to predict changes of
this kind in learners’ models. According to Johnson-Laird (1983), learners change the
N. Schlomske (&) Department of Educational Science, University of Jena, Am Planetarium 4, 07743 Jena, Germany e-mail: [email protected]
P. Pirnay-Dummer Department of Educational Science, University of Freiburg, Rempartstraße 11, 79098 Freiburg, Germany e-mail: [email protected]
123
Education Tech Research Dev (2009) 57:753–765 DOI 10.1007/s11423-009-9125-x
structures of their models as they progress from novices to experts. Within our theoretical
framework we work with words as basic semantic references to the world. We do not
exclude the possibility that there may be sub-symbolic levels of knowledge, but we do not
need this level for our theorems at this time. Words, mainly nouns, are connected pairwise
by a function (e.g., a causal relation) and a level of association (a strength). Although we
are interested in both connections, we have made more coherent progress with the latter—
especially for automated assessment. A model structure is a set of pairwise connections
with a focus on either a specific subject domain, a problem solving or reasoning task, or
both. It contains the strongest connections between higher symbols (words, nouns) as
regards associations.
The group of learners in the present study consists of students who are at the beginning
of their studies and do not have any prior experience in research methods and who can thus
be seen as novices. Their model structures will be examined at five different points of
measurement and will be compared with the model structures of two reference groups
which have been empirically identified beforehand with the same instrument and have
taken the same seminar.
The present paper is structured in two main parts: First, we discuss the theoretical
background and the selection of MITOCAR as an appropriate instrument. Second, we
explain how we conducted the study, the measurements, and the statistical analysis.
Finally, we discuss the empirical results and embed them in the theoretical background.
Learning evokes models which change over time
Mental models
The theory of mental models (Johnson-Laird 1983; Seel 1991) posits a change in the
learner’s knowledge of the world and thus also a change in model structures after learning
has taken place. In order to negotiate one’s way in a changing environment, one needs the
ability to permanently update one’s knowledge (Pirnay-Dummer 2006). As a consequence,
the internal states of knowledge need to be approximated to the external world (van der
Meer 1996) because decisions are made on the basis of present knowledge. This implies
that the internal states of knowledge be changed so that learning can take place. Individuals
create mental models to explain the world on the basis of their knowledge (Ifenthaler
2006). Mental models contain only elements of knowledge which are relevant for
explaining one’s present situation (Stachowiak 1973). The statements represented in a
person’s knowledge must be plausible according to his or her knowledge of the world,
regardless of whether they are true in reality or not (Seel 1991). According to Pirnay-
Dummer (2006) mental models are only constructed if a person cannot explain a particular
situation on the basis of his or her present knowledge of the world. In addition, he argues
that there cannot be instruments that measure mental models reliably because of their
instability. The models identified in the present study are regarded as novice model
structures that change to become more advanced models during the course of learning.
Expertise
Novices are characterized as having no prior experience in a particular domain (Gruber
1994). Experts differ from novices not only in the amount of knowledge available to them,
754 N. Schlomske, P. Pirnay-Dummer
123
but also in their knowledge structures (Pirnay-Dummer 2006). Gruber (1994) makes a
distinction between four different activities in which high achievement can be recognized
by experts: (1) manual activities, (2) mental and academic activities, (3) complex activities,
and (4) artistic activities (Gruber 1994). The second and third parts are of special interest in
the present study.
The learning and instruction dependent acquisition of expertise
Learning and instruction at schools or universities have the aim of bringing learners to a
higher level of competence in a particular domain (Gruber and Mandl 1996). Learners
change their models by means of constructing mental models. This implies that they have
to be influenced by teaching in such a way that they approximate expert models (Seel
1991). Otherwise, the novices hold on to their ‘‘incorrect’’ models as long as they can still
explain the world on the basis of their invalid knowledge. Only when their knowledge no
longer suffices to explain the world will they change their invalid knowledge (Ifenthaler
2006). Finally, novice models change to become expert models through learning (Johnson-
Laird 1989). Learning—and finally the change of learners’ models—takes place on three
levels: first, on a declarative level, second, on a procedural level, and third, on a semiotic
level (Ifenthaler 2006). The amount of epistemic statements available to the cognitive
system in a particular domain is permanently updated on the declarative level. Moreover,
models constructed by experts are more elaborated than those made by novices. Heuristics
are improved on the procedural level, while symbolic signs are updated on the semiotic
level depending on how the knowledge can be represented (Ifenthaler 2006). Johnson-
Laird (1983) calls this change of models and the processes involved in it ‘‘fleshing-out,’’
while Seel (1991) speaks of it as ‘‘successive model change.’’ An external criterion is
needed in order to demonstrate a learning dependent change and, ultimately, to reveal the
acquisition of expertise. In the present study, we draw on novice models in order to
represent the present state of knowledge of the target group and advanced learners’ models,
which may be considered expert models, in order to represent the target state of
knowledge.
Re-representation of knowledge
Bruner (1964) distinguishes between three different formats of knowledge representations
(cf. Ifenthaler 2006). (1) the enactive format of knowledge representation, (2) the iconic
format of knowledge representation, and (3) the symbolic format of knowledge repre-
sentation. The first indicates activities that have a particular aim, such as ‘‘knowing what to
do in order to send an email.’’ The iconic format of knowledge representation indicates
activities by means of pictures, for instance having a map in order to get from point A to
point B. The last format of knowledge representation indicates symbols, such as signs that
are important for re-representing one’s knowledge.
According to Seel (1991), a symbolic characteristic is realized in internal representa-
tions (Seel 1991) because a symbolic system is necessary in order to represent one’s
knowledge. For assessment and observation, the internal structure has to be externalized
again, which can be described as a re-representation. In the present study, we examine
re-representations of knowledge. In the following section, we will discuss the examination
of such external structures.
Model based assessment of learning dependent change 755
123
Research question
The question driving this study is: Can the acquisition of expertise be assessed with already
proven instruments?
The central issues in the present study were, first, whether MITOCAR is able to identify
the acquisition of expertise, and second, what kind of learning dependent model change
can be detected by learners of empirical methodologies using MITOCAR.
Hypotheses
H1a: The model structures of the learners correlate with the model structures of the reference groups (novices and advanced learners).
H0a: The model structures of the learners do not correlate with the model structures of the reference groups (novices and advanced learners).
While a decrease is expected in the correlation coefficients between the novices, a
learning dependent increase is expected between the advanced learners.
H1b: There is a correlation between the model structures of the novices/advanced learners and the measurement points.
H0b: There is no correlation between the model structures of the novices/advanced learners and the measurement points.
Model guided assessment
Instruments are required that reliably and validly assess the construction of ‘knowledge’
(Eckert 1998). According to the theory of mental models (Johnson-Laird 1983; Seel 1991),
learners change the structures of their models as they progress from novice to expert. In
order to examine this process, it is thus necessary to employ an instrument that identifies
novices and experts and distinguishes them from one another. As already pointed out,
internal representations can never be assessed in a direct way, but rather only their
re-represented realization. This implies that they can be re-represented by a system. Thus,
the instrument must be able to re-represent the model structures of novices and experts
(Eckert 1998). Janetzko and Strube (2000) points out how important it is to consider the
relational connections of model structures when modeling knowledge procedures (Janetzko
and Strube 2000). This can be done when novices and experts relate concepts to each other
graphically. This allows one to infer their cognitive structures (Mandl and Fischer 2000).
Existing instruments that consider this aspect of analyzing knowledge are already being
used in pedagogical diagnostics (Eckert 1998). Ifenthaler (2006) provides an overview of
available instruments for assessing model structures.
In order to assess learning dependent changes in model structures, methodological
requirements have to be aligned precisely to the theoretical assumptions (Ifenthaler 2006).
Several measurements are desirable to reveal changes from novice models to advanced
learning models. Therefore, a design is required that identifies learning dependent model
changes between the individual time points on a quantitative level. In addition, it is
desirable for the parameter value of the dependent variable to remain constant at each time
point to enable a comparison of the values (Ifenthaler 2006). Finally, it is desirable to set
external criteria that describe both the beginning state of knowledge and the target state of
756 N. Schlomske, P. Pirnay-Dummer
123
knowledge. The target state of knowledge enables an especially precise assessment of how
far away the learners are from the learning goals. The present study is particularly inter-
ested in assessing changes in models after learning has taken place.
Methods
Experiment
The design of the study is an intra-subject design which shows the learners’ progress. The group of learners consisted of 31 students who were at the beginning of their studies and
did not have any experience in empirical research methods and theoretical fundamentals
and who could thus be seen as novices. During their first semester they were familiarized
with the descriptive statistics as well as the basics of scientific hypotheses. This involved
having them run little projects in which they investigated everyday problems. In the
following semester they learned about inferential statistics by conducting little empirical
studies with educational relevance and learning to choose the appropriate statistical
analysis for their research questions. In addition to the seminar, they participated in
tutorials (Table 1).
As external criteria, we compared the group of learners to two further groups which had
been assessed before by the same instrument: a group of novices (N = 26) (equivalent to
the beginning state of knowledge of the group of learners) and a group of advanced
learners (N = 32) (equivalent to the target state of knowledge of the group of learners).
Both reference groups had participated in the same seminar in the past but with different
teachers. Their models had been assessed during another study (Pirnay-Dummer 2006).
Within our study, their models were used only to compare with the models constructed by
the learner groups.
The group of learners was measured at five different points within a period of 7 months:
first, at the beginning of their first semester, second, in the middle of their first semester
right before Christmas break and, third, at the end of their first semester. Finally, they were
measured at the beginning of their second semester and in the middle of their second
semester right before Pentecost break. At each measurement point, the model structures of
the reference groups were compared with those of the group of learners. We expected a
change in the structure of the learner group’s models in the course of the learning process
away from the models of the first reference group and toward those of the second reference
group.
The experiment took place in the computer rooms of the Department of Education in
Freiburg. The data was collected anonymously. After the subjects had logged in with a web
browser, they entered their number, the group and block number, and their gender and age.
Finally, they read the instructions and worked through the experiment on their own. This
prevented investigator effects and thus guaranteed a high objectivity. The group and block
Table 1 Intra-subject design
Winter semester Summer semester
2004 Novices (November) Advanced learner (July)
2006/07 lst MP 2nd MP 3rd MP 4th MP 5th MP
November December January April May
Model based assessment of learning dependent change 757
123
numbers identified the particular experiment. The former number referred to the group of
learners, whereas the latter referred to the stable sequences of the experiment.
The verification mode from MITOCAR was taken to test the subjects’ models against
the comparison models. Already existing models of the reference groups were drawn on as
external criteria. The comparison models consisted of 30 pairs of concepts each. The paired
concepts were rated by the subjects for their closeness, contrast, and for how sure the
subjects were about their rating.
Instrument (MITOCAR)
MITOCAR is a software toolset developed and introduced by Pirnay-Dummer (2006,
2007). It is based strictly on mental model theory (cf. Seel 1991) and has proven to deliver
valid, homogeneous, and reliable results. MITOCAR is an acronym for ‘‘Model Inspection
Trace of Concepts and Relations.’’ It measures properties of language re-representations of
a realization by a group. The re-representation is called the group consensus model:
MITOCAR also measures whether there is sufficient agreement within the group
(homogeneity).
To produce the consensus model of the graph, all the subjects need to do is go through a
two phase web-based assessment procedure which takes approximately 1.5 h for a whole
group. Afterwards, MITOCAR generates automated reports which not only display the
knowledge structure in a concept map-like format but also calculate and interpret several
tests, e.g., multidimensional scaling and homogeneity (within a group), and provide
additional descriptive measures and graphs which help the subjects to find answers within
the knowledge structure (cf. Pirnay-Dummer 2007).
Identification, Review, Construction, Verification, and Confrontation are the modules
which are presented separately to and used by the subjects. While Identification and
Verification are mandatory for the functioning of MITOCAR, all of the other modules can
be used to improve the quality of the knowledge assessment. All of the other steps are
calculated automatically by the software and handled and stored on a database. The
identification mode is the first phase of MITOCAR and is a simple collection of statements on a given subject domain. Between the first and the second phase, a concept parser filters nouns (with and without attached adjectives) and compiles a list of the most frequent
concepts from the ‘‘mini-corpus.’’ The second phase consists of the review, the con-
struction, the verification, and the confrontation. In the review mode, every group member rates all expressions of the group for plausibility and for their relatedness to the subject
domain. In the construction mode, the subjects categorize concepts into groups which can be processed into model information using Markov chains (Pirnay-Dummer 2006). Veri- fication and confrontation are both modes for a pairwise comparison of concepts. Table 2 has been translated to be better understandable. The original instrument was applied in the
German language (Table 2).
Paired concepts are rated by the subjects in the second phase of MITOCAR for their
closeness and contrast. Additionally, the subjects rate how sure they are about their rating.
The three basic measures and meaningful combinations of them can be used for the
graphical reconstruction of the model later on. All items are rated on a 5 point Likert scale
on screen by the subjects.
1. Closeness: The item of closeness describes how closely related two concepts are rated as being by the subjects.
758 N. Schlomske, P. Pirnay-Dummer
123
2. Contrast: For the item of contrast the subjects rate how different two concepts are or to what extent they exclude each other (e.g., fire and water).
3. Combined: This measure combines the items of closeness s and contrast k. It is calculated by |(s - 1) - (k - 1)| ? 1 = |s - k| ? 1. High contrast with low close-
ness and low contrast with high closeness both generate high combined values. The
closer contrast and closeness become the lower the combined value will be. The scale
remains the same as for closeness and contrast.
4. Confidence: The confidence rating 1 measures how sure the subjects are of their ratings of contrast and closeness. To save space in titles and headers, all measures which are
weighted by the confidence rating are indicated by a (?) sign.
The MITOCAR software takes six pairwise related model representation measures into
account:
1. Closeness: The model is constructed only on the basis of the closeness rating s. 2. Contrast: The model is constructed only on the basis of the contrast rating k. 3. Closeness?: The model is constructed on the basis of closeness and weighted by
confidence: k � 1. If the subjects rate the relation of concepts with more confidence, they will also be more likely to become a part of the model.
4. Contrast?: The model is constructed on the basis of contrast and weighted by confidence: s � 1. If the subjects rate the relation of concepts with more confidence, they will also be more likely to become a part of the model.
5. Combined: The model is constructed on the basis of the combined measure |(s - 1) - (k - 1)|.
6. Combined?: The model is constructed on the basis of the combined measure and weighted by confidence: |(s - 1) - (k - 1)| � 1. If the subjects rate the relation of concepts with more confidence, they will also be more likely to become a part of the
model.
Depending on the quality of the data (which is tested before re-representation), different
measures may be used. E.g., if the combined item has too much deviance or is inhomo-
geneous within a group it will be excluded from re-representation. This is automatically
tested and reported by the MITOCAR software. In this study the data quality sufficed for
the combined measure (all measures had a good quality). Verification and confrontation modes differ only in the pairs of terms which are rated. In the verification mode subjects
rate the terms which come from their own group (utilizing their own power of language),
while in the confrontation mode they rate pairs from another group (typically from a group
which they are being compared to). This information is used to build (re-represent) the
knowledge structure in the form of a concept map.
Table 2 Ratings of the concepts
Please rate the following concept pair
Statistics empirical research
1. The concepts are closely related to each other
Not at all Completely
2. The concepts are mutually exclusive
Not at all Completely
3. In my rating I’m
Unconfident Confident
Model based assessment of learning dependent change 759
123
In this study the combined? measure was taken to compare the learners’ models to the reference models.
Results
Learning dependent change
For the combined measures, a decrease was revealed in the model structures of the novices,
whereas an increase was observed in those of the advanced learners. In addition, a sta-
bilization was observed as compared to the advanced learner model, whereas no such
stabilization was observed as compared to the novice model. Moreover, the table above
illustrates that the model structures may approximate each other (Tables 3 and 4).
All correlations with an asterisk (*) are statistically significant (tMP2Nov = 2.90,
dfMP2Nov = 49, pMP2Nov = 0.016 \ 0.05, dMP2Nov = 0.85; tMP5Nov = 2.01, dfMP5Nov = 31, pMP5Nov = 0.030 \ 0.05, dMP5Nov = 0.75; tMP3AdvL = 2.59, dfMP3AdvL = 47, pMP3AdvL = 0.025 \ 0.05, dMP2Nov = 0.77; tMP4AdvL = 2.65, dfMP4AdvL = 33, pMP4AdvL = 0.009 \ 0.05, dMP4AdvL = 0.95). The correlation coefficients reveal a decrease in similarity in comparison to the novice models (except for at the second and the last measurements points)
and an increase as compared to the advanced learners’ models, except for at the last point.
Statistical analyses
In order to examine whether or not the models of the present study were homogenous, we
tested the variances within each item for all measurement points. Only the first mea-
surement point had homogeneity problems, which shows that the subjects may have had
Table 3 Combined measures during the measurements of change
MP M SD
Novices Advanced learner Novices Advanced learner
1 0.78 0.56 0.09 0.17
2 0.70 0.56 0.08 0.08
3 0.70 0.59 0.07 0.07
4 0.68 0.62 0.08 0.07
5 0.67 0.63 0.08 0.07
Table 4 Correlations (between the learners and the reference groups)
rs
MP Novices Advanced learners
1 0.12 0.25
2 0.39(*) 0.26
3 0.30 0.36(*)
4 0.25 0.43(*)
5 0.35(*) 0.27
760 N. Schlomske, P. Pirnay-Dummer
123
some problems with the rating environment as complete novices. However, all groups of
novices in three previous studies had a sufficiently high homogeneity. The homogeneities
were tested with an intra-item ANOVA (see Appendix).
The correlations (illustrated above) revealed that the novice models correlated signifi-
cantly with the reference model at the second and at the fifth measurement point, whereas
the advanced learner models correlated significantly with the reference models at the third
and at the fourth measurement point (Fig. 1).
The figure above illustrates that the measurement points predict the means of the
combined measures with a high correlation. A correlation of -0.99 was identified for the
novices and a correlation of 1.00 for the advanced learners. Both correlations are statis-
tically significant (p \ 0.05).
Discussion
The results indicate that the external criteria may help to predict learning dependent model
change beyond curricular variations. The measurements identified a correlation coefficient
(Spearman) of -0.99 for the novices and a coefficient of 1.00 for the advanced learners.
The learners in the present study showed a traceable change in their model structures as
they progressed from novices to advanced learners. This indicates that the reference
models provide appropriate indicators for predicting the development of expertise.
An examination of the learning dependent changes in the correlations revealed that the
novices changed in accordance to the theoretical expectations, except for the change
between the first and second measurements and at the last measurement point. The cor-
relations of the advanced learners changed in line with the theoretical expectations, except
for at the last measurement point. Ifenthaler (2006) has found a similar unexpected
development of learning dependent change at the last measurement point. His empirical
findings revealed the same phenomena, and his empirical data thus did not confirm his
theoretical expectations. It may be assumed that knowledge which was plausible to the
learner at a former point in time cannot easily be revised through learning and instruction.
Obviously, the knowledge that is represented before learning has occurred and instructions
have been given is still available to the learner even though it has been changed. This
Fig. 1 Measurement point as predictor
Model based assessment of learning dependent change 761
123
finding underlines Seel’s (1991) and Pirnay-Dummer’s (2006) hypothesis that invalid
models are not easy to revise. Regarding the standard deviations, a stabilization was
revealed in the advanced learner models. Obviously, the instructions evoked learning, thus
stabilizing the target knowledge about empirical research methods. Something which is
still unclear is the unusually high correlation of the learner progression trajectory. This
would even be surprising if the testing had been applied to the very same group, e.g., to
account for retest reliability. Of course, we would have expected a correlation worth
mentioning in this case. However, the dataset correlates with both models almost func-
tionally and with precise linearity. At the moment, we can not explain such a high effect on
a proper theoretical basis. And as much as researchers may like to find high effects, we will
certainly need at least a replication to uncover the mechanisms which led to such results.
Conclusion
Even with a completely new group and a new instructor, the reference models of the
previous group could be used to predict the learners’ progress over time. In some cases this
study shows that is possible to predict a group’s learning behavior and progress. This
further strengthens the position that—with a proper analysis—instruction can be planned
on a systematic basis. The knowledge-based approach used in this study may provide a
basis for continuing research planning and design. Undoubtedly, a good analysis will also
require components other than knowledge assessment, but knowledge assessment can help
extensively in planning curricula for groups (also see Pirnay-Dummer and Nußbickel
2008). The second conclusion is that automated knowledge assessment tools are in the
process of becoming a methodologically sound means of tracking changes over time. Thus,
the results from the study can be interpreted as another validation of the tools used in our
study. Our future studies will therefore concentrate on two key aspects: How to utilize the
predictability and validity for performance testing (particularly grading) and how to design
and develop tools for automated self-assessment for individual learners. If it turns out that
performance testing and self-assessment for individual learners can be assessed in an
automated way and measured objectively, reliably, and validly with theory-based instru-
ments, this research may lead to innovations in the educational field.
Acknowledgment An earlier version of this paper was presented at the CELDA 2008 meeting in Freiburg, Germany and published in the CELDA 2008. Proceeding edited by Kinshuk, D. G. Sampson, J. M. Spector, P. Isası́as, & D. Ifenthaler.
Appendix
Homogeneity in the reference model of the novices
The number of compared single ratings for the novices is n = 3518 (Tables 5 and 6).
• The group produced sufficiently homogeneous ratings. A model derived from the closeness rating (and combinations with it) can be interpreted as a model of the group.
• The group produced sufficiently homogeneous ratings. A model derived from the contrast rating (and combinations with it) can be interpreted as a model of the group.
• The group produced sufficiently homogeneous ratings. A model derived from the combined measure (and combinations with it) can be interpreted as a model of the group.
762 N. Schlomske, P. Pirnay-Dummer
123
• The group produced sufficiently homogeneous ratings. A model weighted by the confidence rating can be interpreted as a model of the group.
The derivation within the ratings of the novices is small enough to interpret their model
using all ratings and combinations thereof. We can understand their model to be a con-
sensus model.
Homogeneity in the reference model of the advanced learners
The number of compared single ratings for the advanced learners is n = 4208 (Tables 7
and 8).
• The group produced sufficiently homogeneous ratings. A model derived from the closeness rating (and combinations with it) can be interpreted as a model of the group.
• The group produced sufficiently homogeneous ratings. A model derived from the contrast rating (and combinations with it) can be interpreted as a model of the group.
• The group produced sufficiently homogeneous ratings. A model derived from the combined measure (and combinations with it) can be interpreted as a model of the
group.
Table 5 Description of the measures
Measures M SD Variance
Closeness 3.75 0.88 0.7762
Contrast 1.07 0.92 0.8390
Combined 3.32 1.24 1.5304
Confidence 3.83 0.85 0.7190
Table 6 ANOVA Measures df F q
Between subject
Closeness 3045
Contrast 3045
Combined 3045
Confidence 3045
Within subject
Closeness 434 0.2520 1.000
Contrast 434 0.4055 1.000
Combined 434 0.2632 1.000
Confidence 434 0.3640 1.000
Table 7 Descriptions of the measures
Measures M SD Variance
Closeness 3.71 0.94 0.8819
Contrast 2.01 0.96 0.9135
Combined 3.36 1.34 1.7893
Confidence 3.64 1.05 1.0971
Model based assessment of learning dependent change 763
123
• The group produced sufficiently homogeneous ratings. A model weighted by the confidence rating can be interpreted as a model of the group.
The derivation within the ratings of the novices is small enough to interpret their model
using all ratings.
References
Bruner, J. S. (1964). The course of cognitive growth. American Psychologist, 19, 1–16. Eckert, A. (1998). Kognition und Wissensdiagnose. Die Entwicklung und empirische Überprüfung des
computerunterstützten wissensdiagnostischen Instrumentariums Netzwerk-Elaborierungs- Technik (NET). Lengerich: Pabst.
Gruber, H. (1994). Expertise. Modelle und empirische Untersuchungen (Vol. 34). Opladen: Westdeutscher Verlag.
Gruber, H., & Mandl, H. (1996). Das entstehen von expertise. In J. Hoffmann & W. Kintsch (Eds.), Enzyklopädie der Psychologie, Themenbereich C, Theorie und Forschung, Serie II, Kognition (Vol. 7, pp. 583–615). Göttingen: Hogrefe.
Ifenthaler, D. (2006). Diagnose lernabhängiger Veränderung mentaler Modelle. Veränderungsmessungen als Verfahren der empirischen Lehr-Lern-Forschung (University dissertation, Freiburg: FreiDok).
Ifenthaler, D. (2008). Relational, structural, and semantic analysis of graphical representations and concept maps. Educational Technology Research and Development. doi:10.1007/s11423-008-9087-4.
Janetzko, D., & Strube, G. (2000). Knowledge tracking–Eine neue Methode zur Diagnose von Wissensstrukturen. In H. Mandl & F. Fischer (Eds.), Wissen sichtbar machen. Wissensmanagement mit Mapping-Techniken (pp. 199–217). Göttingen: Hogrefe Verlag.
Johnson-Laird, P. (1983). Mental models toward a cognitive science of language. Inference and language. Cambridge: Cambridge University Press.
Johnson-Laird, P. N. (1989). Mental models. In M. I. Posner (Ed.), Foundations of cognitve science (pp. 469–499). Cambridge, MA: MIT Press.
Mandl, H., & Fischer, F. (2000). Wissen sichtbar machen. Wissensmanagement mit Mapping-Techniken. Göttingen: Hogrefe.
Pirnay-Dummer, P. N. (2006). Expertise und Modellbildung. MITOCAR (University dissertation, Freiburg: FreiDok).
Pirnay-Dummer, P. (2007). Model inspection trace of concepts and relations. A heuristic approach to language-oriented model assessment. Paper presented at the AERA 2007, Division C, TICL SIG, Chicago, April 2007.
Pirnay-Dummer, P., & Nußbickel, M. (2008). New ways to find out what is needed to know. Using the latest tools for knowledge elicitation in the processes of needs assessment. Paper presented to the AERA 2008, Devision I, Education in the Professions.
Table 8 ANOVA Measures df F q
Between subject
Closeness 4350
Contrast 4350
Combined 4350
Confidence 4350
Within subject
Closeness 434 0.1646 1.000
Contrast 434 0.2465 1.000
Combined 434 0.2242 1.000
Confidence 434 0.4516 1.000
764 N. Schlomske, P. Pirnay-Dummer
123
Pirnay-Dummer, P., & Spector, J. M. (2008). Language, association, and model re-representation. How features of language and human association can be utilized for automated knowledge assessment. Paper presented to the AERA 2008, TICL SIG.
Pirnay-Dummer, P., & Walter, S. (2008). Bridging the world’s knowledge to individual knowledge. Using latent semantic analysis and web ontologies to complement classical and new knowledge assessment technologies. Paper presented to the AERA 2008, TICL SIG.
Seel, N. M. (1991). Weltwissen und mentale Modelle. Göttingen: Hogrefe. Stachowiak, H. (1973). Allgemeine Modelltheorie. Wien: Springer. van der Meer, E. (1996). Gesetzmäßgikeiten und Steuerungsmöglichkeiten des Wissenserwerbs. In
Enzyklopädie der Psychologie (Vol. 2, pp. 208–248). Göttingen: Hogrefe.
Nadine Schlomske is PhD-student at the Department of Educational Science at the University of Jena, Germany. Her research interests are learning and the automated assessment of performance.
Pablo Pirnay-Dummer is assistant professor at the University of Freiburg, Germany. His main research interests are in learning, cognition and instruction with a focus on mental models, knowledge and knowledge assessment as well in technology enhanced learning environments and knowledge management.
Model based assessment of learning dependent change 765
123