scholarly journals Student ability assessment based on two IRT models

2005 ◽  
Vol 2 (2) ◽  
Author(s):  
Silvia Cagnone ◽  
Roberto Ricci

The aim of this work is to analyze a part of the data collected in the Computer Science Department during the Informatics exams in the year 2003. Two different Item Response Theory models for ordered polytomous variables are considered in order to get an evaluation of student ability. Ordered polytomous variables are used for a problem solving process that contains a finite number of steps so that the ability of a student can be evaluated on the basis of the step achieved, namely, higher steps achieved are related to higher ability. The models considered are the Partial Credit Model and the Graded Response Model. The choice of these models has been dictated by the fact that although they are defined into different theoretical frameworks, the former belongs to the Rasch family (Masters, 1982) and the latter can be viewed as a Generalized Linear Latent Variable Model (Bartholomew and Knott, 1999), and hence they present different properties, both of them allow to treat ordinal observed variables. The analysis of the real data set through the two approaches allows to highlight their advantages and disadvantages.

Author(s):  
Alexander Baturo ◽  
Johan A. Elkink

Abstract How can one assess which countries select more experienced leaders for the highest office? There is wide variation in prior career paths of national leaders within, and even more so between, regime types. It is therefore challenging to obtain a truly comparative measure of political experience; empirical studies have to rely on proxies instead. This article proposes PolEx, a measure of political experience that abstracts away from the details of career paths and generalizes based on the duration, quality and breadth of an individual's experience in politics. The analysis draws on a novel data set of around 2,000 leaders from 1950 to 2017 and uses a Bayesian latent variable model to estimate PolEx. The article illustrates how the new measure can be used comparatively to assess whether democracies select more experienced leaders. The authors find that while on average they do, the difference with non-democracies has declined dramatically since the early 2000s. Future research may leverage PolEx to investigate the role of prior political experience in, for example, policy making and crisis management.


2020 ◽  
Vol 44 (6) ◽  
pp. 465-481
Author(s):  
Carl F. Falk

We present a monotonic polynomial graded response (GRMP) model that subsumes the unidimensional graded response model for ordered categorical responses and results in flexible category response functions. We suggest improvements in the parameterization of the polynomial underlying similar models, expand upon an underlying response variable derivation of the model, and in lieu of an overall discrimination parameter we propose an index to aid in interpreting the strength of relationship between the latent variable and underlying item responses. In applications, the GRMP is compared to two approaches: (a) a previously developed monotonic polynomial generalized partial credit (GPCMP) model; and (b) logistic and probit variants of the heteroscedastic graded response (HGR) model that we estimate using maximum marginal likelihood with the expectation–maximization algorithm. Results suggest that the GRMP can fit real data better than the GPCMP and the probit variant of the HGR, but is slightly outperformed by the logistic HGR. Two simulation studies compared the ability of the GRMP and logistic HGR to recover category response functions. While the GRMP showed some ability to recover HGR response functions and those based on kernel smoothing, the HGR was more specific in the types of response functions it could recover. In general, the GRMP and HGR make different assumptions regarding the underlying response variables, and can result in different category response function shapes.


2012 ◽  
Vol 36 (4) ◽  
pp. 81-94 ◽  
Author(s):  
Emmanouil Benetos ◽  
Simon Dixon

In this work, a probabilistic model for multiple-instrument automatic music transcription is proposed. The model extends the shift-invariant probabilistic latent component analysis method, which is used for spectrogram factorization. Proposed extensions support the use of multiple spectral templates per pitch and per instrument source, as well as a time-varying pitch contribution for each source. Thus, this method can effectively be used for multiple-instrument automatic transcription. In addition, the shift-invariant aspect of the method can be exploited for detecting tuning changes and frequency modulations, as well as for visualizing pitch content. For note tracking and smoothing, pitch-wise hidden Markov models are used. For training, pitch templates from eight orchestral instruments were extracted, covering their complete note range. The transcription system was tested on multiple-instrument polyphonic recordings from the RWC database, a Disklavier data set, and the MIREX 2007 multi-F0 data set. Results demonstrate that the proposed method outperforms leading approaches from the transcription literature, using several error metrics.


2000 ◽  
Vol 25 (3) ◽  
pp. 253-270 ◽  
Author(s):  
John G. Baker ◽  
James B. Rounds ◽  
Michael A. Zevon

Two multiple category item response theory models are compared using a data set of 52 mood terms with 713 subjects. Tellegen’s (1985) model of mood with two independent, unipolar dimensions of positive and negative affect provided a theoretical basis for the assumption of unidimensionality. Principle components analysis and item parameter tests supported the unidimensionality assumption. Comparative model data fit for the Samejima (1969) logistic model for graded responses and the Masters (1982) partial credit model favored the former model for this particular data set. Theoretical and practical aspects of the comparative application of multiple category models in the measurement of subjective well-being or mood are discussed.


2005 ◽  
Vol 30 (4) ◽  
pp. 369-396 ◽  
Author(s):  
Eisuke Segawa

Multi-indicator growth models were formulated as special three-level hierarchical generalized linear models to analyze growth of a trait latent variable measured by ordinal items. Items are nested within a time-point, and time-points are nested within subject. These models are special because they include factor analytic structure. This model can analyze not only data with item- and time-level missing observations, but also data with time points freely specified over subjects. Furthermore, features useful for longitudinal analyses, “autoregressive error degree one” structure for the trait residuals and estimated time-scores, were included. The approach is Bayesian with Markov Chain and Monte Carlo, and the model is implemented in WinBUGS. They are illustrated with two simulated data sets and one real data set with planned missing items within a scale.


2021 ◽  
Author(s):  
◽  
Rajbir Singh Nirwan

Machine Learning (ML) is so pervasive in our todays life that we don't even realise that, more often than expected, we are using systems based on it. It is also evolving faster than ever before. When deploying ML systems that make decisions on their own, we need to think about their ignorance of our uncertain world. The uncertainty might arise due to scarcity of the data, the bias of the data or even a mismatch between the real world and the ML-model. Given all these uncertainties, we need to think about how to build systems that are not totally ignorant thereof. Bayesian ML can to some extent deal with these problems. The specification of the model using probabilities provides a convenient way to quantify uncertainties, which can then be included in the decision making process. In this thesis, we introduce the Bayesian ansatz to modeling and apply Bayesian ML models in finance and economics. Especially, we will dig deeper into Gaussian processes (GP) and Gaussian process latent variable model (GPLVM). Applied to the returns of several assets, GPLVM provides the covariance structure and also a latent space embedding thereof. Several financial applications can be build upon the output of the GPLVM. To demonstrate this, we build an automated asset allocation system, a predictor for missing asset prices and identify other structure in financial data. It turns out that the GPLVM exhibits a rotational symmetry in the latent space, which makes it harder to fit. Our second publication reports, how to deal with that symmetry. We propose another parameterization of the model using Householder transformations, by which the symmetry is broken. Bayesian models are changed by reparameterization, if the prior is not changed accordingly. We provide the correct prior distribution of the new parameters, such that the model, i.e. the data density, is not changed under the reparameterization. After applying the reparametrization on Bayesian PCA, we show that the symmetry of nonlinear models can also be broken in the same way. In our last project, we propose a new method for matching quantile observations, which uses order statistics. The use of order statistics as the likelihood, instead of a Gaussian likelihood, has several advantages. We compare these two models and highlight their advantages and disadvantages. To demonstrate our method, we fit quantiled salary data of several European countries. Given several candidate models for the fit, our method also provides a metric to choose the best option. We hope that this thesis illustrates some benefits of Bayesian modeling (especially Gaussian processes) in finance and economics and its usage when uncertainties are to be quantified.


Author(s):  
Roman Tkachenko ◽  
Ivan Izonin ◽  
Ivanna Dronyuk ◽  
Mykola Logoyda ◽  
Pavlo Tkachenko

Background: Today, using of systems on the base of Internet of Things (ІоТ) devices is very widespread in various applications. Intellectual analysis of the data collected by similar devices is an important task for efficient and successful functioning of such systems. In particular, the reliability of such kind of analysis has greatly influence on the ability to partially or fully automate certain processes or subsystems. However, imperfect devices of data collection, transportation errors, etc. cause data missing to appear. A number of limitations cause this problem, and in the work, they makes it impossible an effective intellectual analysis for specific use. That is why the scientific and applied problem of effectively filling the missing in the data collected by the sensors of specific characteristics should be considered. Methods: The authors propose a new prediction method for solving this problem based on the use of General Regression Neural Networks (GRNN). Results: The possibility of approximation and partial elimination of the error of computational intelligence of this type has been analytically proved. A cascade of two sequentially connected GRNN was developed. The optimal parameters of the developed cascade were selected. The simulation of its work was performed to solve the problem of recover missing sensor data in the dataset for monitoring the state of air environment. A high number of missing for one reason or another characterizes this real data set, collected by IoT device. Conclusion: High accuracy of cascade operation in comparison with existing methods of this class is inserted. All advantages and disadvantages are described. Perspectives of further research are outlined.


2013 ◽  
Vol 300-301 ◽  
pp. 848-852
Author(s):  
Zong Hai Sun ◽  
Osman Osman

Data sets of high–dimensional spaces are problematic when it comes to classification, compression, and visualization. The main issue is to find a reduced dimensionality representation that corresponds to the intrinsic dimensionality of the original data. In this paper we try to investigate a practical Bayesian method for feature extracting problem, in particular we will apply Gaussian Process Latent Variable Model (GPLVM) to a real world data set. Feature extraction experiments were performed on a cancer treatments’ components data set using GPLVM, then we used PCA on the same data set for comparison of the results.


Entropy ◽  
2021 ◽  
Vol 23 (5) ◽  
pp. 552
Author(s):  
Hamid Mousavi ◽  
Mareike Buhl ◽  
Enrico Guiraud ◽  
Jakob Drefs ◽  
Jörg Lücke

Latent Variable Models (LVMs) are well established tools to accomplish a range of different data processing tasks. Applications exploit the ability of LVMs to identify latent data structure in order to improve data (e.g., through denoising) or to estimate the relation between latent causes and measurements in medical data. In the latter case, LVMs in the form of noisy-OR Bayes nets represent the standard approach to relate binary latents (which represent diseases) to binary observables (which represent symptoms). Bayes nets with binary representation for symptoms may be perceived as a coarse approximation, however. In practice, real disease symptoms can range from absent over mild and intermediate to very severe. Therefore, using diseases/symptoms relations as motivation, we here ask how standard noisy-OR Bayes nets can be generalized to incorporate continuous observables, e.g., variables that model symptom severity in an interval from healthy to pathological. This transition from binary to interval data poses a number of challenges including a transition from a Bernoulli to a Beta distribution to model symptom statistics. While noisy-OR-like approaches are constrained to model how causes determine the observables’ mean values, the use of Beta distributions additionally provides (and also requires) that the causes determine the observables’ variances. To meet the challenges emerging when generalizing from Bernoulli to Beta distributed observables, we investigate a novel LVM that uses a maximum non-linearity to model how the latents determine means and variances of the observables. Given the model and the goal of likelihood maximization, we then leverage recent theoretical results to derive an Expectation Maximization (EM) algorithm for the suggested LVM. We further show how variational EM can be used to efficiently scale the approach to large networks. Experimental results finally illustrate the efficacy of the proposed model using both synthetic and real data sets. Importantly, we show that the model produces reliable results in estimating causes using proofs of concepts and first tests based on real medical data and on images.


2017 ◽  
Vol 14 (4) ◽  
pp. 172988141770357 ◽  
Author(s):  
Lei Tai ◽  
Shaohua Li ◽  
Ming Liu

The exploration problem of mobile robots aims to allow mobile robots to explore an unknown environment. We describe an indoor exploration algorithm for mobile robots using a hierarchical structure that fuses several convolutional neural network layers with decision-making process. The whole system is trained end to end by taking only visual information (RGB-D information) as input and generates a sequence of main moving direction as output so that the robot achieves autonomous exploration ability. The robot is a TurtleBot with a Kinect mounted on it. The model is trained and tested in a real world environment. And the training data set is provided for download. The outputs of the test data are compared with the human decision. We use Gaussian process latent variable model to visualize the feature map of last convolutional layer, which proves the effectiveness of this deep convolution neural network mode. We also present a novel and lightweight deep-learning library libcnn especially for deep-learning processing of robotics tasks.


Sign in / Sign up

Export Citation Format

Share Document