Abstract
While the availability of large and diverse datasets has contributed to significant breakthroughs in autonomous driving and indoor applications, forestry applications are still lagging behind and new forest datasets would most certainly contribute to achieving significant progress in the development of data-driven methods for forest-like scenarios. This paper introduces a forest dataset called FinnWoodlands, which consists of RGB stereo images, point clouds, and sparse depth maps, as well as ground truth manual annotations for semantic, instance, and panoptic segmentation. FinnWoodlands comprises a total of 4226 objects manually annotated, out of which 2562 objects (60.6%) correspond to tree trunks classified into three different instance categories, namely “Spruce Tree”, “Birch Tree”, and “Pine Tree”. Besides tree trunks, we also annotated “Obstacles” objects as instances as well as the semantic stuff classes “Lake”, “Ground”, and “Track”. Our dataset can be used in forestry applications where a holistic representation of the environment is relevant. We provide an initial benchmark using three models for instance segmentation, panoptic segmentation, and depth completion, and illustrate the challenges that such unstructured scenarios introduce. FinnWoodlands dataset is available at https://github.com/juanb09111/FinnForest.git.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Adelson, E.H.: On seeing stuff: the perception of materials by humans and machines. In: IS &T/SPIE Electronic Imaging (2001)
Bac, C.W., Van Henten, E., Hemming, J., Edan, Y.: Harvesting robots for high-value crops: state-of-the-art review and challenges ahead. J. Field Robot. 31 (2014). https://doi.org/10.1002/rob.21525
Bechar, A., Vigneault, C.: Agricultural robots for field operations: concepts and components. Biosyst. Eng. 149, 94–111 (2016)
Behley, J., et al.: A dataset for semantic segmentation of point cloud sequences. CoRR abs/1904.01416 (2019). https://arxiv.org/abs/1904.01416
Brostow, G.J., Fauqueur, J., Cipolla, R.: Semantic object classes in video: a high-definition ground truth database. Pattern Recogn. Lett. (2008)
Bulò, S.R., Porzi, L., Kontschieder, P.: In-place activated batchnorm for memory-optimized training of DNNs. CoRR abs/1712.02616 (2017). https://arxiv.org/abs/1712.02616
Cabon, Y., Murray, N., Humenberger, M.: Virtual KITTI 2. CoRR abs/2001.10773 (2020). https://arxiv.org/abs/2001.10773
Caesar, H., et al.: nuscenes: A multimodal dataset for autonomous driving. CoRR abs/1903.11027 (2019). https://arxiv.org/abs/1903.11027
Che, Z., et al.: D2-city: a large-scale dashcam video dataset of diverse traffic scenarios. arXiv abs/1904.01975 (2019)
Chen, Y., Yang, B., Liang, M., Urtasun, R.: Learning joint 2D-3D representations for depth completion. CoRR abs/2012.12402 (2020). https://arxiv.org/abs/2012.12402
Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. CoRR abs/1604.01685 (2016). https://arxiv.org/abs/1604.01685
Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255 (2009). https://doi.org/10.1109/CVPR.2009.5206848
Everingham, M., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A.: The PASCAL visual object classes challenge 2012 (VOC2012) results (2012). https://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html
Fortin, J.M., Gamache, O., Grondin, V., Pomerleau, F., Giguère, P.: Instance segmentation for autonomous log grasping in forestry operations (2022). https://doi.org/10.48550/ARXIV.2203.01902, https://arxiv.org/abs/2203.01902
Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2012)
Geyer, J., et al.: A2D2: Audi autonomous driving dataset. CoRR abs/2004.06320 (2020). https://arxiv.org/abs/2004.06320
Grondin, V., Fortin, J.M., Pomerleau, F., Giguère, P.: Tree detection and diameter estimation based on deep learning. Forestry: Int. J. Forest Res. (2022)
He, K., Gkioxari, G., Dollár, P., Girshick, R.B.: Mask R-CNN. CoRR abs/1703.06870 (2017). https://arxiv.org/abs/1703.06870
Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. CoRR abs/1502.03167 (2015). https://arxiv.org/abs/1502.03167
Jiang, P., Osteen, P.R., Wigness, M.B., Saripalli, S.: RELLIS-3D dataset: data, benchmarks and analysis. CoRR abs/2011.12954 (2020). https://arxiv.org/abs/2011.12954
Juliani, A., Berges, V., Vckay, E., Gao, Y., Henry, H., Mattar, M., Lange, D.: Unity: a general platform for intelligent agents. CoRR abs/1809.02627 (2018). https://arxiv.org/abs/1809.02627
Lin, T., et al.: Microsoft COCO: common objects in context. CoRR abs/1405.0312 (2014). https://arxiv.org/abs/1405.0312
Mitra, A., et al.: Everything you wanted to know about smart agriculture. CoRR abs/2201.04754 (2022). https://arxiv.org/abs/2201.04754
Mohan, R., Valada, A.: EfficientPS: efficient panoptic segmentation. CoRR abs/2004.02307 (2020). https://arxiv.org/abs/2004.02307
Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference from RGBD images. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7576, pp. 746–760. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33715-4_54
Neuhold, G., Ollmann, T., Bulà, S.R., Kontschieder, P.: The mapillary vistas dataset for semantic understanding of street scenes. In: 2017 IEEE International Conference on Computer Vision (ICCV),. pp. 5000–5009 (2017). https://doi.org/10.1109/ICCV.2017.534
Ringdahl, O.: Automation in forestry: development of unmanned forwarders. Ph.D. thesis, Umeå University, May 2011
Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. CoRR abs/1409.0575 (2014). https://arxiv.org/abs/1409.0575
Sekachev, B., et al.: OpenCV/CVAT: v1.1.0, August 2020. https://doi.org/10.5281/zenodo.4009388
Shamshiri, R., et al.: Research and development in agricultural robotics: a perspective of digital farming. Int. J. Agric. Biol. Eng. 11, 1–14 (2018). https://doi.org/10.25165/j.ijabe.20181104.4278
da Silva, D.Q., dos Santos, F.N.: ForTrunkDet - forest dataset of visible and thermal annotated images for object detection. J. Imaging (2021)
Sinha, R.K., Pandey, R., Pattnaik, R.: Deep learning for computer vision tasks: a review. CoRR abs/1804.03928 (2018). https://arxiv.org/abs/1804.03928
Tan, M., Le, Q.V.: Efficientnet: rethinking model scaling for convolutional neural networks. CoRR abs/1905.11946 (2019). https://arxiv.org/abs/1905.11946
Wang, W., et al.: Tartanair: a dataset to push the limits of visual SLAM. CoRR abs/2003.14338 (2020). https://arxiv.org/abs/2003.14338
Wigness, M., Eum, S., Rogers, J.G., Han, D., Kwon, H.: A RUGD dataset for autonomous navigation and visual perception in unstructured outdoor environments. In: International Conference on Intelligent Robots and Systems (IROS) (2019)
Yu, F., Xian, W., Chen, Y., Liu, F., Liao, M., Madhavan, V., Darrell, T.: BDD100K: a diverse driving video database with scalable annotation tooling. CoRR abs/1805.04687 (2018). https://arxiv.org/abs/1805.04687
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Lagos, J., Lempiö, U., Rahtu, E. (2023). FinnWoodlands Dataset. In: Gade, R., Felsberg, M., Kämäräinen, JK. (eds) Image Analysis. SCIA 2023. Lecture Notes in Computer Science, vol 13885. Springer, Cham. https://doi.org/10.1007/978-3-031-31435-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-031-31435-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-31434-6
Online ISBN: 978-3-031-31435-3
eBook Packages: Computer ScienceComputer Science (R0)