Fumiko Samejima’s research while affiliated with The University of Tennessee Medical Center at Knoxville and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (31)


Graded response model based on the logistic positive exponent family of models for dichotomous responses
  • Article

February 2008

·

64 Reads

·

22 Citations

Psychometrika

Fumiko Samejima

Samejima (Psychometrika 65:319–335, 2000) proposed the logistic positive exponent family of models (LPEF) for dichotomous responses in the unidimensional latent space. The objective of the present paper is to propose and discuss a graded response model that is expanded from the LPEF, in the context of item response theory (IRT). This specific graded response model belongs to the general framework of graded response model (Samejima, Psychometrika Monograph, No. 17, 1969 and No. 18, 1972; Handbook of modern item response theory, Springer, New York, 1997; Encyclopedia of Social Measurement, Academic Press, San Diego, 2004), and, in particular to the heterogeneous case (Samejima, Psychometrika Monograph, No. 18, 1972). Thus, the model can deal with any number of ordered polytomous responses, such as letter grades (e.g., A, B, C, D, F), etc. For brevity, hereafter, the model will be called the LPEF graded response model, or LPEFG. This model reflects the opposing two principles contained in the LPEF for dichotomous responses, with the logistic model (Birnbaum, Statistical theories of mental test scores, Addison Wesley, Reading, 1968) as their transition, which provide a reasonable rationale for partial credits in LPEFG, among others.


Nonparametric Online Item Calibration

January 2007

·

14 Reads

Abstract In estimating the operating characteristic (OC) of an item, in contrast to parametric estimation, nonparametric estimation directly approaches the entire conditional probabilty curve, without assuming any mathematical form. Samejima,proposed several nonparametric methods,in the 1970s and 1980s, under several multi-year research contracts with the Office of Naval Research. Later, one of them, the conditional p.d.f. approach, was adapted to the environment of computerized adaptive testing, utilizing its strengths. In the present research, the truncated logistic model, which leads to higher accuracy in estimating the item characteristic functions (ICFs) of dichotomous items, was used. The results of simulations showed that this method faithfully depicted even complicated changes in nonmonotonic ICFs using as few as 1,202 hypothetical examinees.


Logistic Positive Exponent Family of Models: Virtue of Asymmetric Item Characteristic Curves

February 2000

·

216 Reads

·

65 Citations

Psychometrika

The paper addresses and discusses whether the tradition of accepting point-symmetric item characteristic curves is justified by uncovering the inconsistent relationship between the difficulties of items and the order of maximum likelihood estimates of ability. This inconsistency is intrinsic in models that provide point-symmetric item characteristic curves, and in this paper focus is put on the normal ogive model for observation. It is also questioned if in the logistic model the sufficient statistic has forfeited the rationale that is appropriate to the psychological reality. It is observed that the logistic model can be interpreted as the case in which the inconsistency in ordering the maximum likelihood estimates is degenerated. The paper proposes a family of models, called the logistic positive exponent family, which provides asymmetric item chacteristic curves. A model in this family has a consistent principle in ordering the maximum likelihood estimates of ability. The family is divided into two subsets each of which has its own principle, and includes the logistic model as a transition from one principle to the other. Rationale and some illustrative examples are given.


Efficient Nonparametric Approaches for Estimating the Operating Characteristics of Discrete Item Responses

February 1998

·

9 Reads

·

25 Citations

Psychometrika

Rationale and the actual procedures of two nonparametric approaches, called Bivariate P.D.F. Approach and Conditional P.D.F. Approach, for estimating the operating characteristic of a discrete item response, or the conditional probability, given latent trait, that the examinee's response be that specific response, are introduced and discussed. These methods are featured by the facts that: (a) estimation is made without assuming any mathematical forms, and (b) it is based upon a relatively small sample of several hundred to a few thousand examinees. Some examples of the results obtained by the Simple Sum Procedure and the Differential Weight Procedure of the Conditional P.D.F. Approach are given, using simulated data. The usefulness of these nonparametric methods is also discussed.


Departure from Normal Assumptions: A Promise for Future Psychometrics with Substantive Mathematical Modeling

February 1997

·

18 Reads

·

41 Citations

Psychometrika

Normal assumptions have been used in many psychometric methods, to the extent that most researchers do not even question their adequacy. With the rapid advancement of computer technologies in recent years, psychometrics has extended its territory to include intensive cognitive diagnosis, etcetera, and substantive mathematical modeling ha become essential. As a natural consequence, it is time to consider departure from normal assumptions seriously. As examples of models which are not based on normality or its approximation, the logistic positive exponent family of models is discussed. These models include the item task complexity as the third parameter, which determines the single principle of ordering individuals on the ability scale.


Graded Response Model

January 1997

·

843 Reads

·

1,081 Citations

The graded response model represents a family of mathematical models that deals with ordered polytomous categories. These ordered categories include rating such as letter grading, A, B, C, D, and F, used in the evaluation of students’ performance; strongly disagree, disagree, agree, and strongly agree, used in attitude surveys; or partial credit given in accordance with an examinee’s degree of attainment in solving a problem.


Evaluation of mathematical models for ordered polytomous responses

January 1996

·

16 Reads

·

57 Citations

Behaviormetrika

In this paper, mathematical modeling is treated as distinct from curve fitting. Considerations of psychological reality behind our data are emphasized, and criteria such as additivity in a model, its natural generalization to a continuous response mode], satisfaction of the unique maximum condition and orderliness of the modal points of the operating characteristics of the ordered polychotomous responses are proposed. Strengths and weaknesses of mathematical models for ordered polychotomous responses that include the normal ogive model, the logistic model, the acceleration model and the family of ordered polychotomous models developed from Bock's nominal model are observed and discussed in terms of such criteria. It was concluded that it will be better to leave Bock's model as a nominal model as he intended it to be, without expanding it to ordered polychotomous models.


Acceleration Model in the Heterogeneous Case of the General Graded Response Model

February 1995

·

18 Reads

·

60 Citations

Psychometrika

A new model, called acceleration model, is proposed in the framework of the heterogenous case of the graded response model, based on processing functions defined for a finite or enumerable number of steps. The model is expected to be useful in cognitive assessment, as well as in more traditional areas of application of latent trait models. Criteria for evaluating models are proposed, and soundness and robustness of the acceleration model are discussed. Graded response models based on individual choice behavior are also discussed, and criticisms on model selection in terms of fitnesses of models to the data are also given.


Nonparametric Estimation of the Plausibility Functions of the Distractors of Vocabulary Test ltems

March 1994

·

8 Reads

·

6 Citations

Applied Psychological Measurement

The Level 11 Vocabulary Subtest of the Iowa Tests of Basic Skills was analyzed using a two-stage latent trait approach and an empirical dataset of 2,356 examinees. First, each of the 43 multiple-choice test items was scored dichotomously; then, assuming the (two-parameter) normal ogive model the item parameters were estimated. The operating characteristics of the correct answer and of the three distractors were estimated using a nonparametric approach called the simple sum procedure of the conditional probability density function approach combined with the normal approach method. Differential information was provided by the distractors, and these operating characteristics were named the plausibility functions of the distractors. The operating characteristic of the correct answer of each item estimated by assuming the normal ogive model was compared with the nonparagnetrically estimated operating characteristic for model validation. It was concluded that the nonparametric approach leads to efficient estimation of the latent trait. Index terms: distractors, item response theory, latent trait models, multiple-choice test items, nonparametric estimation, plausibility functions of distractors.


Some critical observations of the test information function as a measure of local accuracy in ability

February 1994

·

9 Reads

·

13 Citations

Psychometrika

The test information function serves important roles in latent trait models and in their applications. Among others, it has been used as the measure of accuracy in ability estimation. A question arises, however, if the test information function is accurate enough for all meaningful levels of ability relative to the test, especially when the number of test items is relatively small (e.g., less than 50). In the present paper, using the constant information model and constant amounts of test information for a finite interval of ability, simulated data were produced for eight different levels of ability and for twenty different numbers of test items ranging between 10 and 200. Analyses of these data suggest that it is desirable to consider some modification of the test information function when it is used as the measure of accuracy in ability estimation.


Citations (26)


... However, the measure exhibited inadequate fit indices in this study (CFI = 0.769; TLI = 0.711; SRMR = 0.111), prompting a more detailed analysis. To critically assess the quality of the scale items, we employed Item Response Theory (IRT), specifically utilising the Graded Response Model (GRM; Samejima 1969). The analysis was conducted using the 'mirt' package in R software (Chalmers 2012), focusing on calculating item parameters: 'a' (indicating discrimination, with a threshold of > 0.50) and 'b' (representing difficulty, within the range of −3.0-3.0) as recommended by Barker (2001) and Baker and Kim (2017) (see Supporting Information for detailed analyses). ...

Reference:

The More I See It the Less I Want It: Sociotropic Corruption and (In)tolerance Towards Corruption in Contexts of Perceived Economic Crisis
Estimation of Latent Ability Using a Response Pattern of Graded Scores
  • Citing Article
  • March 1969

Psychometrika

... respectively. Figure 1 shows the density functions of these six distributions of 8. [From Samejima (1994). Reprinted by Permission of Kluwer Academic Publishers] The hypothetical test consisted of 30 equivalent dichotomous items, which followed the logistic model (Equation 26) with a -1.0, ~ -0.0, c~ = 0.0~ and I~ =1.7. ...

Roles of Fisher Type Information in Latent Trait Models
  • Citing Chapter
  • January 1994

... As this first step concluded to a better fit of the bi-factor model (as detailed below), we then pursued IRT analyses to explore the characteristics of the different items. As items are measured on a several-point scale, we relied on a polytomous model, specifically the graded response model (GRM, Samejima, 1968Samejima, , 2010. The GRM assumes unidimensionality (which we had just verified), local independence, monotonicity, and a normally distributed latent trait (Edwards, 2009;Wirth & Edwards, 2007). ...

Graded Response Model
  • Citing Chapter
  • January 1997

... Ngoài ra còn có thể có những yếu tố khác góp phần vào khả năng phân biệt năng lực người học, yếu tố đặc điểm tiềm tàng của thí sinh ít được tìm hiểu trong các nghiên cứu đánh giá chất lượng bài kiểm tra. Đặc điểm tiềm tàng là những khả năng, phẩm chất, hay năng lực ẩn sâu trong mỗi con người, tuy không thể nhìn thấy bằng mắt thường nhưng lại đóng vai trò quan trọng trong việc định hình tương lai của mỗi người (7) . Các điểm tiềm tàng có thể là khả năng học tập, thời điểm tốt nghiệp và môi trường học tập (8) . ...

Comprehensive Latent Trait Theory
  • Citing Article
  • July 1988

Behaviormetrika

... Second, Chi-squaTe analyses were run within each distractor set to investigate which kinds of distractor tended to attract test-takers more strongly, and discussion fbllows in order to find the cornmonalities of significantly chosen distractors based on six question types based on Shimizu (2005): Paraphrase, Inference, Thematic, Reference, Vbcal)ulary, and Tbxt Organization. In both a multiple regression and Chi-square analyses, only distractors were included in the analyses, fbllowingSamejima (1994)who studied estimation of plausibility fUnctions only for the distractors but not fbr the correct answers.3.4 CodingDistractors were classified by two raters based onDavey (1988)and Freedle and Kostin (1993) fbr regression analysis. ...

Nonparametric Estimation of the Plausibility Functions of the Distractors of Vocabulary Test ltems
  • Citing Article
  • March 1994

Applied Psychological Measurement

... Non-academic reproduction requires payment of royalties through the Copyright Clearance Center, http://www.copyright.com/ reduced to When the MLE of 0 is used, let 8y or 6 denote the MLE of 0 based on the response pattern V. Samejima (1977aSamejima ( , 1979 observed that even with a relatively small number of test items the conditional distribution of 6, given 0, can be approximated by the normal distribution N{8,[7(6)]~} if two conditions hold. Condition 1 is that 11 must be practically conditionally unbiased for the 8 interval of interest. ...

Effects of Individual Optimization in Setting the Boundaries of Dichotomous Items on Accuracy of Estimation
  • Citing Article
  • January 1977

Applied Psychological Measurement

... Therefore, MFI may not always select the most informative item given an inaccurate trait estimator, especially in the early stage of the test (Chen et al., 2000). On the other hand, the distribution of the estimated trait level will gradually converge as the test proceeds, but this updated distribution is not considered in selecting the item (Samejima, 1977). The problem of MFI has inspired abundant research on improvements in strategies. ...

A Use of the Information Function in Tailored Testing
  • Citing Article
  • March 1977

Applied Psychological Measurement

... I will show that another reason to elicit survey responses about belief certainty is that the resulting scales are more sensitive to partial or uncertain knowledge. From the perspective of IRT, a superior survey instrument includes items with high discrimination so as to reduce measurement error (Samejima, 1990). In other words, a more precise instrument yields more information. ...

Predictions of Reliability Coefficients and Standard Errors of Measurement Using the Test Information Function and Its Modifications
  • Citing Article
  • July 1990

... In this research, we propose a new model for modeling multiple choice test items which we refer to as the monotone multiple choice (MMC) model. Similar to the multiple choice model introduced by Samejima (1979), the model not only models the probability for answering each item correctly, but also the probability for selecting each incorrect option (distractor) of an item over various levels of the latent trait. This provides a tool for test creators to evaluate the quality of the distractors for each item. ...

A New Family of Models for the Multiple-Choice Item
  • Citing Article
  • December 1979