Analyzing Reliability of Interpretable Parameters in Deep Learning Language Models
Skip to main content
eScholarship
Open Access Publications from the University of California

Analyzing Reliability of Interpretable Parameters in Deep Learning Language Models

Creative Commons 'BY' version 4.0 license
Abstract

Deep-learning has made remarkable progress in recent years. However, its parameter estimation methods are very different from those of traditional statistical models. They are typically iteration-update formula, such as Adam, designed to avoid falling into a local solution even when there are a large number of parameters. While deep-learning-based models still have interpretable parameters such as difficulty or ability, how reliable are these parameters estimated by such estimation methods? In this study, we compared the estimation methods used in conventional statistical models, such as Marginalized Maximum Likelihood Estimation and MCMC, with those used in deep learning, such as Adam, for simple item response models. In experiments, the parameter values estimated by the methods using Adam and others and the parameter values estimated by the methods used in conventional statistical packages showed statistically significant correlation. This result supports that the parameter estimates by deep learning can be trusted to some extent.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View