In today’s world, we constantly face decisions in environments that are complex, fast-changing, and filled with uncertainty. Choices are rarely presented all at once—instead, they unfold over time and space, requiring us to explore, adapt, and act under pressure. Whether navigating overwhelming amounts of information, dealing with strict time constraints, or managing unpredictable changes, our ability to make sound decisions is continually put to the test.
At the Dynamic Decision Making Laboratory (DDMLab), our mission is to uncover how people make decisions in these dynamic settings. We develop robust theoretical models to explain the cognitive processes behind human choice and translate this knowledge into real-world applications. By bridging theory and practice, we aim to support and enhance decision-making across critical domains such as healthcare, emergency response, cybersecurity, and more.
At the DDMLab, we combine laboratory experiments with cognitive computational models to understand and improve human decision making in dynamic environments.
In our experiments, individuals and teams interact with dynamic decision making tasks that unfold over time and space. These studies reveal behavioral patterns and adaptive strategies people use under uncertainty.
We complement this with cognitive models that simulate the decision-making process through algorithms. These models operate in the same tasks as humans, allowing us to compare their behavior across multiple dimensions—such as learning, risk-taking, and optimality.
By aligning human data with model predictions, we gain deep insights into cognitive mechanisms and use these findings to develop practical applications in areas like cybersecurity, climate change, phishing prevention, and human-machine interaction.
At the DDMLab, we use a powerful cognitive theory called Instance-Based Learning Theory (IBLT) (Gonzalez, Lerch, & Lebiere, 2003) to understand how people make decisions based on past experiences. IBLT explains how we draw on memories of similar situations (i.e., instances) we've encountered before, to make decisions..
As we face new challenges, our brain searches through these stored instances to find ones that match the current situation. We then weigh the options based on past outcomes and select the one that seems most promising. After making a choice, we update our memory with the result, helping us improve future decisions.
This process is driven by a mathematical equation rooted in the ACT-R cognitive architecture, which captures how recent, frequent, and similar experiences influence what we remember—and how accurately we recall it.
To support researchers and developers, we’ve created PyIBL — a Python-based platform that brings IBLT to life. PyIBL makes it easy to build and test models that simulate human decision-making. To try it yourself visit our Cognitive Modeling page to access PyIBL, a detailed manual, and a library of example models.
To test and improve IBLT and other theories of decisions from experience, we rely on Model Comparisons. IBL models are judged according to how well they replicate human decisions.
In sequential decision making, the goal is often to maximize long-term rewards while making selections among explicitly provided options. Under this theme, we investigate uncertainty, risk, context, and constraints of decision making including: time limits, feedback delays, and cognitive workload.
Initial contributions to understanding the process by which people make Decisions from Experience (DfE) emerged from the use of complex, dynamic decision making tasks (i.e., Microworlds). Microworlds represent highly complex tasks in computer simulations, such as dynamic resource allocation in real time and under time constraints. Currently, we also use simplified decision tasks in our research. For example, repeated choice tasks such as binary choice or bandit tasks.
Initial contributions first emerged from the use of complex, dynamic decision making tasks (i.e., Microworlds). Microworlds represent highly complex tasks, such as dynamic resource allocation in real time and under time constraints, in computer simulations.
Dynamic decision making can be also conceived as a continuous control task rather than discrete sequential choice. In a control task, the goal of the decision maker is to keep a system in “balance.” In fact, many real-life situations demand that a decision maker reduces the gap between the current state of a system and a target state. For example, maintaining a healthy body weight requires that decision makers balance diet and exercise, inventory control requires managers to balance demand and supply, and maintaining the earth’s CO2 within acceptable ranges requires a balance between emissions and absorptions. In other words, control tasks are common at the societal, organizational, and individual levels.
We have examined extreme simplifications of dynamic and complex tasks in contexts, such as global warming and inventory control. For example, we have used graphical representations concerning the accumulation of a quantity (e.g., CO2 levels) over time through an inflow (e.g., carbon emissions) that increases the quantity and an outflow (e.g., carbon absorptions) that decreases it.