Niantic’s AI Model to Navigate Physical World Using Pokémon Go Scans
Last week, Niantic announced plans to create an AI model for navigating the physical world using scans collected from players of its mobile games, such as Pokémon Go, and from users of its Scaniverse app, reports 404 Media.
Training Data from Unconventional Sources
All AI models require training data. So far, companies have collected data from websites, YouTube videos, books, audio sources, and more, but this is perhaps the first we’ve heard of AI training data collected through a mobile gaming app.
Building the Visual Positioning System (VPS)
"Over the past five years, Niantic has focused on building our Visual Positioning System (VPS), which uses a single image from a phone to determine its position and orientation using a 3D map built from people scanning interesting locations in our games and Scaniverse," Niantic wrote in a company blog post.
The Large Geospatial Model (LGM)
The company calls its creation a "large geospatial model" (LGM), drawing parallels to large language models (LLMs) like the kind that power ChatGPT. Whereas language models process text, Niantic’s model will process physical spaces using geolocated images collected through its apps.
Scale of Data Collection
The scale of Niantic’s data collection reveals the company’s sizable presence in the AR space. The model draws from over 10 million scanned locations worldwide, with users capturing roughly 1 million new scans weekly through Pokémon Go and Scaniverse. These scans come from a pedestrian perspective, capturing areas inaccessible to cars and street-view cameras.
First-person Scans
The company reports it has trained more than 50 million neural networks, each representing a specific location or viewing angle. These networks compress thousands of mapping images into digital representations of physical spaces. Together, they contain over 150 trillion parameters—adjustable values that help the networks recognize and understand locations. Multiple networks can contribute to mapping a single location, and Niantic plans to combine its knowledge into one comprehensive model that can understand any location, even from unfamiliar angles.
Conclusion
Niantic’s LGM has the potential to revolutionize the way we navigate the physical world, leveraging the collective efforts of millions of users to create a comprehensive and accurate understanding of our surroundings. The company’s innovative approach to AI training data collection demonstrates its commitment to pushing the boundaries of what is possible in the AR space.
Frequently Asked Questions
Q: What is the purpose of Niantic’s Large Geospatial Model (LGM)?
A: The LGM is designed to process physical spaces using geolocated images collected through Niantic’s apps, enabling users to navigate the physical world with greater accuracy and precision.
Q: How does the LGM differ from large language models (LLMs)?
A: While LLMs process text, Niantic’s LGM processes physical spaces using geolocated images.
Q: How many scans has Niantic collected for its LGM?
A: The company has collected over 10 million scanned locations worldwide, with users capturing roughly 1 million new scans weekly through Pokémon Go and Scaniverse.
Q: What are the implications of Niantic’s LGM for the AR space?
A: The LGM has the potential to revolutionize the way we navigate the physical world, enabling users to access previously inaccessible areas and providing a more accurate and comprehensive understanding of their surroundings.

