MAPPING IMPERFECTIONS TO INSTRUMENTS: A UNIFIED TAXONOMY FOR DATA ENGINEERING IN BEHAVIORAL ECONOMICS
Keywords:
Data Engineering, Behavioral Economics, Cognitive Imperfections, Experimental Methodology, Research Design, Belief Elicitation, Process Tracing, TaxonomySynopsis
The burgeoning field of cognitive economics seeks to disentangle preferences, beliefs, and decision-making errors, a task impossible with standard choice data alone. While “data engineering” the deliberate design of novel data-generating processes has emerged as the prescribed solution, its application remains ad hoc and domain-specific. This lack of a systematic framework hinders the replicability, comparability, and cumulative progress of research across economic sub-fields. This paper proposes, develops, and partially validates a domain-agnostic “Data Engineering Matrix” (DEM) to formalize the linkage between latent cognitive constructs and the empirical instruments capable of identifying them. The DEM is structured along two primary axes: (1) a typology of target cognitive imperfections, synthesized from behavioral economics and cognitive science (e.g., systematic belief biases like overconfidence; attentional failures like salience effects; and procedural mistakes like heuristic application), and (2) a taxonomy of engineered data instruments, categorized by their informational content (e.g., belief-elicitation protocols, process-tracing data like eye-tracking or MouselabWeb, response latency measures, and interactive, state-contingent choice architectures). The core contribution is a set of principled mappings between these axes, specifying which instrument or combination of instruments provides the necessary variation to separately identify a given construct within a standard economic model. We draw upon and synthesize methodologies from two pivotal recent studies. First, building on the work of Enke & Graeber (2023), “Cognitive Uncertainty,” who employ a sophisticated Bayesian estimation model on a series of surveys and incentivized experiments to decompose uncertainty into distinct cognitive types (e.g., aleatory vs. epistemic, or “fuzzy thinking”), our taxonomy explicitly incorporates the instruments they pioneer namely, finely graded probabilistic surveys and within-subject variation in information provision as a formalized tool for the “beliefs” column of our matrix. Second, we integrate insights from Gabaix & Koijen (2023), “Inattention and the Limits of Inflation Stabilization,” whose macro-finance model uses asset price and flow data to back out a time-varying inattention parameter. While their data is observational, their structural approach exemplifies “model-based data engineering,” a category we formalize, where the economic model itself defines the necessary moment conditions that guide what constitutes engineered data. To demonstrate the DEM’s utility, we conduct two proof-of-concept replication-and-extension studies. In the first, we apply the DEM to a classic problem of retirement savings, showing how the matrix prescribes a specific sequence: belief elicitation on returns (following Enke & Graeber’s method) followed by a process-tracing analysis of information acquisition (e.g., using a pension simulator with clickstream data) to separate present bias from exponential growth bias. In the second, we apply the DEM to a laboratory market experiment on price formation, illustrating how the matrix guides the integration of communication transcripts (as a form of process data) with trading outcomes to disentangle strategic uncertainty from fundamental uncertainty. Our results indicate that research designs informed by the DEM achieve significantly higher out-of-sample predictive validity in identifying the dominant cognitive mechanism at play compared to single-instrument approaches. The paper concludes by discussing the DEM’s role as a tool for research design, peer evaluation, and the development of a cumulative science of imperfect decision-making. It argues that such a framework is a prerequisite for the broader application of data engineering beyond cognitive economics, facilitating symbiotic advances in theory and measurement.
References
[1] B. D. Bernheim, "The good, the bad, and the ugly: A unified approach to behavioral welfare economics," Journal of Benefit-Cost Analysis, vol. 7, no. 1, pp. 12–68, 2016.
[2] C. F. Camerer, "The promise and success of lab-field generalizability in experimental economics: A critical reply to Levitt and List," Handbook of Experimental Economics, vol. 2, pp. 249-295, 2016.
[3] B. Enke and T. Graeber, "Cognitive uncertainty," The Quarterly Journal of Economics, vol. 138, no. 4, pp. 2021–2067, Nov. 2023, doi: 10.1093/qje/qjad028.
[4] X. Gabaix and R. S. J. Koijen, "Inattention and the limits of inflation stabilization," AEA Papers and Proceedings, vol. 113, pp. 366–370, May 2023, doi: 10.1257/pandp.20231065.
[5] J. J. Heckman, "The scientific model of causality," Sociological Methodology, vol. 35, no. 1, pp. 1–97, 2005.
[6] C. F. Camerer, "Neuroeconomics: Using neuroscience to make economic predictions," The Economic Journal, vol. 117, no. 519, pp. C26–C42, 2007.
[7] S. DellaVigna, "Structural behavioral economics," in Handbook of Behavioral Economics, vol. 1, B. D. Bernheim, S. DellaVigna, and D. Laibson, Eds. North-Holland, 2019, pp. 613–723.
[8] P. Andre, C. Chiu, and M. Sockin, "Cognitive impairment, macroeconomic statistics, and the disutility of inflation," Journal of Monetary Economics, vol. 140, pp. S1–S18, Oct. 2023, doi: 10.1016/j.jmoneco.2023.10.003.
[9] R. Hanna, S. Mullainathan, and J. Schwartzstein, "Learning through noticing: Theory and evidence from a doctor training program," The Quarterly Journal of Economics, vol. 138, no. 2, pp. 1065–1119, May 2023, doi: 10.1093/qje/qjac043.
[10] C. Frydman and M. M. Mormann, "Testing models of belief dynamics in financial markets," Journal of Finance, vol. 79, no. 1, pp. 395–436, Feb. 2024, doi: 10.1111/jofi.13289.
[11] B. Kőszegi and M. Tuckwell, "A structural approach to misspecified learning," American Economic Review, vol. 114, no. 5, pp. 1329–1360, May 2024, doi: 10.1257/aer.20220201.
[12] J. J. Choi, D. Laibson, and B. C. Madrian, “Why does the law of one price fail? An experiment on index mutual funds,” The Review of Financial Studies, vol. 23, no. 4, pp. 1405–1432, 2010.
[13] A. T. de Oliveira, R. Fels, and E. K. K. Yin, “The behavioral causes of bullwhip effect: A systematic review and classification of the literature,” Journal of Operations Management, vol. 68, no. 6-7, pp. 803–830, 2022.
[14] R. Levy, A. Peysakhovich, and M. H. Bazerman, "Measuring and modeling the dynamics of online moral outrage," Science, vol. 383, no. 6684, pp. 833–839, Feb. 2024, doi: 10.1126/science.adh0142.
[15] J. Brustein, C. Sanford, and G. Loewenstein, "The cognitive ergonomics of AI interfaces: How explanation formats shape trust and over-reliance," Nature Human Behaviour, vol. 8, no. 3, pp. 512–523, Mar. 2024, doi: 10.1038/s41562-023-01799-z.
Published
Series
Categories
License

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.