Meta-learning to find every needle in every haystack – Technology Org

The quest to efficiently identify regions of interest in geospatial exploration has recently taken a leap forward with visual active search (VAS). This modeling framework uses visual cues to guide exploration with potential applications ranging from wildlife poaching detection to search-and-rescue missions to identifying illegal trafficking activities.

Meta-learning to find every needle in every haystack – Technology Org

A new framework for visual active search developed by McKelvey engineers leverages reinforcement learning to improve search performance. An adaptive search approach initially makes a mistake but quickly adapts by learning from it. Image credit: Anindya Sarkar

A new approach to VAS developed at the McKelvey School of Engineering at Washington University in St. Louis combines deep reinforcement learning, where a computer can learn to make better decisions through trial and error, with traditional active search, where human searchers go out and verify what’s in a selected region. The team that developed the novel VAS framework includes Yevgeniy Vorobeychik and Nathan Jacobs, computer science and engineering professors, and Anindya Sarkar, a doctoral student in Vorobeychik’s lab. The team presented its work at the New Orleans Neural Information Processing Systems conference.

“VAS improves on traditional active search more or less depending on search task,” Jacobs said. “If a task is relatively easy, then improvements are modest. But if an object is very rare — for example, an endangered species that we want to locate for wildlife conservation — then VAS offers substantial improvements. Notably, this isn’t about finding things faster. It’s about finding as many things as possible given limited resources, especially limited human resources.”

The team’s VAS framework improves on previous methods by breaking down the search into two distinct modules. The framework first uses a prediction module to produce regions of interest based on geospatial image data and search history. Then a search module takes the resulting prediction map as an input and outputs a search plan. Each module can be updated as human explorers return results from physical searches in real time.

“Instead of an end-to-end search policy, decomposing into two modules allows us to be much more adaptable,” Sarkar said. “We can update our prediction module with the search results during the actual search. Then the search module can learn the dynamics of the prediction module — how it’s changing across search steps — and adapt. In this meta-learning strategy, the search module is basically learning how to search. It’s also human interpretable, so if the model isn’t working properly, the user can check it and debug as needed.”

The major strength of the framework comes from its ongoing incorporation of two deployment levels: the computational model predicts where to search, then humans go out into the world to search. The human component is vastly more expensive in terms of time and other resources required to explore large geospatial areas, so it makes sense to adapt and optimize the computer-generated search plan for a maximally efficient search.

Adaptability in the computer model is especially important when the object sought varies drastically from the objects the model is trained on. Experimental results showed marked improvement by Sarkar, Vorobeychik and Jacobs’ proposed VAS framework over existing methods in various visual active search tasks.

Source: Washington University in St. Louis