The dataset consists of approximately 40 thousand images collected underwater from 20 habitats in the marine-environments of tropical Australia.
The dataset originally contained only classification labels. Thus, we collected point-level and segmentation labels to have a more comprehensive fish analysis benchmark.
Videos for DeepFish were collected for 20 habitats from remote coastal marine environments of tropical Australia. These videos were acquired using cameras mounted on metal frames, deployed over the side of a vessel to acquire video footage underwater. The cameras were lowered to the seabed and left to record the natural fish community, while the vessel maintained a distance of 100 m. The depth and the map coordinates of the cameras were collected using an acoustic depth sounder and a GPS, respectively. Video recording was carried out during daylight hours and in relatively low turbidity periods. The video clips were captured in full HD resolution (1920 × 1080 pixels) from a digital camera. In total, the number of video frames taken is 39,766.
The DeepFish dataset and code are publicly available at https://alzayats.github.io/DeepFish/ and https://github.com/alzayats/DeepFish, respectively.
The full methodology is available in the Open Access publication from the Related publications link below.