Why GeoPACHA-AI?

Addressing the scalar limits of archaeological research through human-machine teaming

Scalar Limits in Archaeology

Archaeology faces persistent challenges of aligning granular field data with the massive scale of many social processes we study. While the Inca Empire (Tawantinsuyu) organized distinct social and ecological zones into a massive, integrated polity spanning two million square kilometers, archaeological surveys have traditionally been confined to "valley-sized silos." To date, we have lacked continuous, consistent data on settlement patterns and land use at the continental scale necessary to investigate imperial political economy, demography, and long-term human adaptations to climate change.

Our Solution: Human-Machine Teaming

GeoPACHA-AI addresses this challenge through a novel "expert-machine teaming" approach. Rather than relying solely on manual labor or fully autonomous automation, we combine the discernment of regional experts with the scalability of artificial intelligence.

The project began with a "brute force" manual imagery survey (GeoPACHA 1.0), where our network of collaborators and student teams painstakingly documented over 40,000 archaeological loci across eight survey zones. This high-quality, expertly curated dataset now serves as the foundation for the next phase: an AI-assisted survey spanning nearly the entirety of the Andes, from northern Ecuador through southern Chile. This area of about two million square kilometers encompasses the approximate historic footprint of Tawantinsuyu (the Inka Empire). Our goal is to enable new continental-scale perspectives on archaeological features. We are pursuing continental-scale understandings of land use and settlement patterns, so our models include both semantic segmentation models of agricultural field systems (including active and abandoned terrace complexes, for example) and features associated with human settlements, such as abandoned structures (archaeological buildings) and features related to pastoralist settlements in the high puna grasslands.

DeepAndes and Vision Transformers

At the core of the AI system are two models: DeepAndes, a custom Vision Transformer (ViT) model that we have created from a large sample (3 million image patches), and DeepAndesArch, the fine-tuned model based on the training data inputs from our collaborating teams. We are now in the process of generating a revised DeepAndes foundation model based on a much larger sample of the WorldView-2, WorldView-3 multispectral imagery the project is using, and enriched by elevation, land cover, and climatic rasters. This revised foundation model (DeepAndesV2) is enabled by the supercomputing resources of Oak Ridge National Laboratory.

Our interdisciplinary team is led by Steven Wernke (Professor and Chair, Department of Anthropology, Vanderbilt University), Parker VanValkenburgh (Associate Professor, Department of Anthropology, Brown University), Yuankai Huo (Assistant Professor of Computer Science, Vanderbilt University), and James Zimmer-Dauphinee (Postdoctoral Fellow, Department of Anthropology, Vanderbilt University). The project has also benefitted from major contributions by Junlin Guo (PhD candidate, Computer Science, Vanderbilt University) and Siqi Lu (PhD candidate, Computer Science, College of William and Mary). The computationally-intensive foundation model training required for DeepAndesV2 is made possible by our ongoing collaboration with Xiao Wang (Research Scientist, Oak Ridge National Laboratory).