Zamba is a tool built in Python that uses machine learning and computer vision to detect and classify animals in camera trap videos.
You can use zamba to:
Zamba can automate time-intensive video processing tasks, enabling researchers to focus on interpreting the content and using the results.
Pretrained model | Recommended use | Training data |
---|---|---|
African species identification | Species classification in jungle ecologies | ~250,000 camera trap videos from Central, West, and East Africa |
European species identification | Species classification in non-jungle ecologies | The African species identification model is finetuned with ~13,000 additional videos from camera traps in Germany |
Blank vs. non-blank | Classifying videos as either blank or containing an animal, without species identification | All data from both African and European models (~263,000 videos) |
|
|
|
|
|
|
Example use cases:
Example use cases:
The retraining process starts with one of the official models (see table above). The model then continues training on any new data provided, including predicting completely new species — the world is your oyster! (We'd love to see a model trained to detect oysters.) The custom model can then be used to generate predictions on new, unlabeled videos.
Zamba Cloud is a web application where you can use the zamba algorithms by just uploading videos or pointing to where they are stored. Zamba Cloud is created for conservation researchers and wildlife experts who aren’t familiar with using a programming interface. If you’d like all of the functionality of zamba without any of the code, this is for you!
Explore Zamba Cloud →Zamba is also provided as an open source package that can be run in the command line or imported as a Python library. If you’d like to interact with zamba through a programming interface, hack away! Visit the zamba package documentation for details and user tutorials.
Explore the Zamba package →As part of the Pan African Programme: The Cultured Chimpanzee, over 8,000 hours of camera trap footage has been collected across various chimpanzee habitats from camera traps in 15 African countries. Labeling the species in this footage is no small task. It takes a lot of time to determine whether or not there are any animals present in the data, and if so, which ones.
To date, thousands of citizen scientists have manually labeled video data through the Chimp&See Zooniverse project. In partnership with experts at The Max Planck Institute for Evolutionary Anthropology (MPI-EVA), this effort fed into a well-labeled dataset of nearly 2000 hours of camera trap footage from Chimp&See's database.
Using this dataset, DrivenData and MPI-EVA ran a machine learning challenge where hundreds of data scientists competed to build the best algorithms for automated species detection. The top 3 submissions that were best able to predict the presence and type of wildlife across new videos won the challenge and received €20,000 in monetary prizes. The winning techniques developed from this challenge provided a starting point for the algorithms behind Project Zamba.
Zamba means forest in Lingala. Lingala is one of many Bantu languages of central Africa, and is spoken throughout the Democratic Republic of the Congo and the Republic of the Congo. The first ever Homo sapiens emerged in African forests and savannas, and African zambas may hold the keys to unlocking critical mysteries of human evolution.
The code developed for Project Zamba is openly available for anyone to learn from and use. You can find the latest version of the project codebase on GitHub.
Thanks to all the participants in the Pri-Matrix Factorization Challenge! Special thanks to Dmytro Poplovskiy (@dmytro), developer of the top-performing solution adapted for Project Zamba, the project team at the Max Planck Institute for Evolutionary Anthropology for organizing the competitions and the data, and to the ARCUS Foundation for generously funding this project.