Educational Resources for AI and Machine Learning — Glossary
Artificial Intelligence (AI)
Artificial Intelligence is the science and engineering of building machines that perform tasks normally associated with human intelligence.
Related terms: Machine Learning (ML)
Computational/Model Bias
The various forms of mathematical bias that arise due to the choice of model and parameters of interest for a given dataset. These biases include but are not limited to: the bias-variance tradeoff, frequency bias, and/or inductive bias.
Related terms: Data Bias, Decision-Making Bias, Bias-Variance Tradeoff, Frequency Bias, Inductive Bias
When developing a model for some dataset, we may initially choose a simple model (e.g., a linear model) to describe the data (upper left hand panel). Simple models, though, fail to capture regularities in the data and tend to underfit the data (high bias, low variance). To capture regularities in a dataset often requires a low-bias prediction. However, lowering the prediction bias comes at the cost of increasing the variance of the model prediction. Low bias and high variance as shown on the right overfits the data, as it pays way too much attention to training data and doesn’t generalize well. In other words, increasing the prediction variance tends to increase overfitting to the training data. The bias variance-tradeoff differs from one learning algorithm to another.

A classic illustration of the bias-variance tradeoff (From Fig. 37.1 in Guerts 2010). Left, a linear model fitted to two random samples. Right, a neural network fitted to the same samples.
Data Bias
A class imbalance or distortion in the data from what we know is true based on meteorological and other knowledge of interest.
Related Terms: Computational/Model Bias
Data itself can contain biases, which affect the ML model training, evaluation, and deployment. These biases can be caused by underlying human biases (e.g., unintentional or intentional) or by sampling and selection of data. For example, a completely unintentional bias has emerged in which some populations are served by poorer coverage of monitoring equipment like weather radars (see top figure) or hail reports coming mainly along highways and cities with few reports in rural areas (see bottom figure).

From Jack Sillin @JackSillin: https://twitter.com/JackSillin/status/1372957704138981378?s=20

From Figure 11 in Allen and Tippet 2015: Point reports of hail >0.75 in (1.9 cm), over the Texas Panhandle and surroundings, with population choropleth of intercensal estimated population segregated by Jenks Natural Breaks: a) all reports 1955–2014, shown with mean population 1979–2012; b) hail reports 1955–1979 with 1979 population; c) as for (b) except hail reports 1955–1995 and 1995 population from the CIESIN gridded global population data; d) as for (c) except hail reports 1955–2005 and 2000 population from the CIESIN data. Primary interstates and highways are shown in red.
Resources and References
Decision-Making Bias
The biases that arise due to how people perceive, think, remember, and represent events and phenomena and use these heuristics in their problem-solving and decision-making.
Related Terms: Computational/Model Bias, Data Bias
Heuristics or preferences that we have in solving problems or making decisions can be a source of bias. Biases in problem solving can be described as stemming more fundamentally from how people perceive, think, remember, and represent events and phenomena – both strategically and inadvertently, keeping in mind that all of these are extremely context dependent, much more than it seems most people realize.
For example, people are likely to perceive motion and color, which are more salient to them than orientation, and may be limited in their understanding of complexity by working memory limits (5 plus or minus two), and their ability to perceive categories – which can be greatly facilitated by expertise and an understanding of hierarchical relationships within data.
Of relevance to the definition of biases, Kahneman has suggested that most common heuristics can be characterized as resulting from “attribute substitution” – for example, people commonly substitute feelings/emotions in part or entirely for cognitive assessments, or mental availability/accessibility (i.e., does it come readily to mind) for more cognitive or deliberative assessments of prevalence / relative frequency. Another common type of bias is anchoring, where subjective estimates can be disproportionately influenced by some familiar or salient number for example, and adjust their estimates insufficiently away from that anchor. An example of a heuristic that can be hugely influential is a preference to use models and tools with which you are familiar.
“Narrow bracketing” has been highlighted in the judgment and decision making literature as a weakness in intuitive problem solving approaches (related to smaller variance). Even more specifically, people often use lexicographic approaches to problem solving, considering only one aspect or one option at a time, which might similarly result in ‘narrow bracketing’ of problems.
Resources and References
- Kahneman, D., & Frederick, S. (2002). Representativeness revisited: Attribute substitution in intuitive judgment. Heuristics and biases: The psychology of intuitive judgment, 49, 81.
- Read, D., Loewenstein, G., Rabin, M., Keren, G., & Laibson, D. (1999). Choice bracketing. In Elicitation of preferences (pp. 171-202). Springer, Dordrecht.
Ethical AI
A responsible approach to AI that ensures certain properties such as fairness, robustness, privacy respecting, and transparency are upheld during model development, deployment, and monitoring.
Related terms: AI, Responsible AI, Robust AI, Trust, Trustworthiness, Transparency, Fairness, Privacy Respecting
When developing AI systems for environmental sciences, we need to recognize that many of our data sources contain biases, errors, and other issues. By taking an ethical approach to AI, we can attempt to address these issues during model development. Otherwise, these issues can propagate downstream and potentially lead to unethical models (e.g., Facial recognition algorithms often perform poorly for individuals with darker skin complexions). The figure below (from McGovern et al. 2022) demonstrates how ethical issues can arise in environmental sciences.

A nonexhaustive list of issues that can arise though the use of AI for environmental science applications.
Resources and References
- McGovern, I. Ebert-Uphoff, D.J. Gagne, A. Bostrom, Why we Need Ethical, Responsible, and Trustworthy Artificial Intelligence for Environmental Sciences, Environmental Data Science (2022), April 2022. https://doi.org/10.1017/eds.2022.5
Explainability
The degree to which a human can derive meaning from the entire model and its components through the use of post-hoc methods (e.g., verification, visualizations of important predictors).
Related terms: Interpretability
A distinction between interpretability and explainability is needed since some in the ML and statistics community favor producing interpretable models (i.e., restricting model complexity beforehand to impose interpretability; (Rudin et al. 2019, Rudin et al. 2021) while the general trend in the ML community is to continue developing partially interpretable and black box models and implementing post-hoc methods to explain them. A deep neural network (DNN) or dense random forest is uninterpretable, but through external explanation methods, can be approximately understood. Explanation methods must be approximate as they would otherwise be as incomprehensible as the black box model itself. This is not a limitation of explanation methods as suggested by other studies (e.g., Rudin et al. 2019, Rudin et al. 2021) since abstracting a complex model is required for human understanding. For example, it is common in the weather community to replace the full Navier-Stokes equation with conceptual models that are more understandable. The degree of explainability, though, is controlled by the model complexity (Molnar et al. 2020). As the number of features increases or their interactions become more complex, the explanations for the ML model behavior become less compact and possibly less accurate. At the moment, it is unclear how much improvement in understanding of high-dimensional, highly non-linear models current and future explanation methods will offer (Fig. 1).

Fig. 1 Illustration of the relationship between understandability and model complexity. Fully interpretable models have high understandability (with little to gain from explainability) while partially interpretable or simpler black box models have the most to gain from explainability methods. With increased dimensionality and non-linearity, explainability methods can improve understanding, but there is considerable uncertainty about the ability of future explanation methods to improve the understandability of high-dimensional, highly non-linear methods.
Resources and References
- Interpretable Machine Learning: A Guide for Making Black Box Models Explainable
- Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead, Rudin et al., 2019
- Interpretable Machine Learning: Fundamental Principles and 10 Grand Challenges, Rudin et al., 2021
- Quantifying Model Complexity via Functional Decomposition for Better Post-Hoc Interpretability, Molnar et al. 2020
Machine Learning (ML)
A field of study within Artificial Intelligence that involves using computer algorithms that can improve and adapt automatically through continual experience with data
Related terms: Artificial Intelligence (AI)
Interpretability
The degree to which a human can derive meaning from the entire model and its components without the aid of additional methods.
Related terms: Explainability
A fully interpretable model is one that has simulatability (the entire model can be contemplated at once), decomposability (each component of the model are human-understandable), and algorithmic transparency (one can understand how the model was trained) (Lipton 2016) while a partially interpretable model may only meet one of these criteria. As an example, low-dimensional linear regression is fully interpretable and a shallow decision tree is partially interpretable (see figure below).

Fig. 1 Illustration of the relationship between understandability and model complexity. Fully interpretable models have high understandability (with little to gain from explainability) while partially interpretable or simpler black box models have the most to gain from explainability methods. With increased dimensionality and non-linearity, explainability methods can improve understanding, but there is considerable uncertainty about the ability of future explanation methods to improve the understandability of high-dimensional, highly non-linear methods.
Resources and References
Responsible AI
The development of AI systems where the affected populations are included in model development and the model is applied to situations in which its output is understood.
Related terms: AI, Ethical AI, Robust AI, Trust, Trustworthiness, Transparency
Resources and References
- McGovern, I. Ebert-Uphoff, D.J. Gagne, A. Bostrom, Why we Need Ethical, Responsible, and Trustworthy Artificial Intelligence for Environmental Sciences, Environmental Data Science (2022), April 2022. https://doi.org/10.1017/eds.2022.5
Robust AI
An AI system that fails in a responsible and graceful way and is resistant to adversarial attacks.
Related terms: AI, Ethical AI, Responsible AI, Trust, Trustworthiness
Trust
The willingness to assume risk by relying on or believing in the actions of another party.
Related terms: Trustworthiness, Ethical AI, Interpretability, Explainability, Responsible AI, Robust AI
The definition of trust above is purposefully broad, so as to capture the many different definitions and related dimensions of trust. Fundamentally, trust is relational, as it involves someone’s perception of something or someone else. Trust is dynamic, evolves with interactions, and is easier to lose than gain.
Resources and References
- Jacovi, Alon, Ana Marasović, Tim Miller, and Yoav Goldberg. 2021. “Formalizing Trust in Artificial Intelligence: Prerequisites, Causes and Goals of Human Trust in AI.” In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 624–35. FAccT ’21. New York, NY, USA: Association for Computing Machinery.
- Meyer, R. C., J. H. Davis, and F. David Schoorman. “An integrative model of organizational trust.” Academy of management review 20.3 (1995): 709-734.
- Luhmann, N. (1989). Vertrauen: Ein Mechanismus der Reduktion sozialer Komplexität [Trust: A mechanism for the reduction of social complexity]. Stuttgart, Germany: Enke.
- Freudenburg, William R. “Risk and recreancy: Weber, the division of labor, and the rationality of risk perceptions.” Social forces 71.4 (1993): 909-932.
- Rousseau, Denise M., et al. “Not so different after all: A cross-discipline view of trust.” Academy of management review 23.3 (1998): 393-404.
Trustworthiness
A trustee’s evaluation, or perception, of whether, when, why, or to what degree someone or something should or should not be trusted.
Related terms: Trust, Interpretability, Explainability, Ethical AI, Responsible AI, Robust AI
The dimensions of trustworthiness are dependent on the values and needs of the trustor (Schneiderman, Human-Centered AI, 2022). A model deemed trustworthy by one user does not imply that it will be trustworthy to another user with their own set of values and needs. The common dimensions associated with evaluation of trustworthiness include: model performance and verification, trustor’s experience with model, source of information, and usability of model and information. Dimensions specific to ML are basic performance, reliability, human interaction, and aligned purpose (Varshney 2022). Note, the provided dimensions are meant to be representative examples but not exhaustive of all possible dimensions.
