This repo is intended to aggregate information regarding the automation of our ROV survey imagery processing. Our overarching hope is to train a model to process our ROV survey imagery for us, such that the output files are then ready for subsequent analyses to extract data. Specifically, we currently use Adobe Lightroom Classic (ALC) to batch process anywhere from 100 to 200 images at a time, from a single ROV survey. We apply the de-noise feature, edit white-balance, brightness, crop, etc., enabling us to extract as much detail as possible from our images. This thoroughly prepares them for our ML approach (CoralNet-Toolbox) to extract percent-cover and abundance data from them.
At present, this photo processing step is by far the most rate-limiting--it's the bottleneck preventing us from more rapidly translating ROV survey imagery --> processed and analyzed results. See here for a 1-pager description of the problem on our CCR_development repo.
The latest developments on this issue come from the Underwater Image Enhancer tool. The UIE allows us to leverage our set of over 6,000 hand-edited images to train in-house machine learning models that will automate image processing. Our training dataset and model iterations are accessible on Hugging Face here. See below examples of our progress thus far; and track model output developments here.
We have linked to image sets here to facilitate testing, model training, and workflow development.
- linked here is our standard 20-image testing set - including the original .GPR files, hand-edited (in ALC) .JPG files, and de-noised (in ALC) .TIF files. This smaller dataset is to enable prototyping.
- linked here, as an alternative to Hugging Face, are 5000 image pairs: the raw .GPR and the hand-edited .JPG files. This larger dataset is to enable full-scale modeling training/deployment.




