work in progress
Click here for abstracts
Integrating AI with Physical Measurements: Using AlphaFold2 to Accelerate X-Ray Crystallography
This paper examines the integration of predictive methods based on machine learning within measurement practices in the physical sciences. Traditional wisdom holds that predictive methods should be validated with experimental methods when a high degree of accuracy is required. I aim to show, however, that in some contexts the relationship between AI predictions and physical measurement is not always a one-way street in which experimental methods univocally verify predictions. Many cases of physical measurement involved inferring a measurement outcome from a set of indicators based on a theoretical model of a particular experimental setup. In some such cases, AI predictions can function as inputs in model-based corrections of derived physical measurements. In these cases, scientists treat AI predictions as epistemically on par with physical measurements methods. This suggests that assessing the epistemic status of AI predictions in science is a deeply contextual affair. I illustrate this phenomenon with a concrete case from computational crystallography, wherein AI model predictions are integrated into experimental methods of measuring the structure of a protein using x-ray crystallography via a procedure called molecular replacement.
Artiface of Objectivity: why algorithms are necessarily value-laden (in progress)
Algorithmic decision-making systems applied in social contexts drape value-laden solutions in an illusory veil of objectivity. I argue that these systems are necessarily value-laden and that this follows from the need to construct a quantifiable objective function. Many researchers have convincingly argued that machine learning systems learn to replicate and amplify pre-existing biases of moral import found in training data. But these arguments permit a strategic retreat for those who nevertheless maintain that algorithms themselves are value-neutral. Proponents of the value-neutrality of algorithms argue that while the existence of algorithmic bias is undeniable such bias is merely the product of bad data curation practices. On such a view, eliminating biased data would obliterate any values embedded in algorithmic decision-making. This position can be neatly summarized by the slogan “Algorithms aren’t biased, data is biased.” However, this attitude towards algorithms is misguided. Training machine learning algorithms involves optimization, which requires either minimizing an error function or maximizing an objective function by iteratively adjusting a model’s parameters. The objective function represents the quality of the solution found by the algorithm as a single real number. Training an algorithm thus aggregates countless indicators of predictive success into a single, automatically generated, weighted index. But deciding to operationalize a particular goal in this way is itself a value-laden choice. This is because many qualities we want to predict are qualitative concepts with multifaceted meanings. Such concepts like “health” or “job-applicant-quality” lack sharp boundaries and admit plural and context-dependent meanings. Collapsing concepts into a quantifiable ratio scale of predictive success flattens out their quality dimensions. This process is often underdetermined and arbitrary, but convenient for enterprises that rely on precise and unambiguous predictions. Hence, the very choice to use an algorithm in the first place reflects the values and priorities of particular stakeholders.
Recent research suggests that manifolds play an important role in neural computations. These manifolds are continuous, low-dimensional structures embedded in high-dimensional neural activity. Investigators purport to uncover these structures by using data-analytic techniques to reduce the dimensionality of patterns of neural activity and subsequently reveal the underlying dynamics that are functionally relevant to a specific task. However, the practice of uncovering low-dimensional structures with dimensionality reduction involves modeling choices that introduce a range of implicit assumptions which threaten to cloud our analysis. Yet, the theoretical importance of these modeling choices are rarely discusses by the modellers themselves. To what extent do these techniques license realist claims about the existence of neural manifolds and their role in representing and computing information in the brain? I argue that low-dimensional structures uncovered through data-analysis are akin to rational reconstructions of the underlying dimensionality on the basis of what is empirically accessible to us. These structures reflect only our analyses and are strictly speaking non-factive. However, we can evaluate competing analyses and adjudicate between cases of underdetermination by asking which analysis best reconstructs our best theoretical hypotheses concerning the neural task under investigation. This approach, however, demands a much more robust role for theoretical work concerning the nature and structure of neural computations.