Skip to main content

FinnWoodlands Dataset

  • Conference paper
  • First Online:
Image Analysis (SCIA 2023)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13885))

Included in the following conference series:

  • 543 Accesses

Abstract

While the availability of large and diverse datasets has contributed to significant breakthroughs in autonomous driving and indoor applications, forestry applications are still lagging behind and new forest datasets would most certainly contribute to achieving significant progress in the development of data-driven methods for forest-like scenarios. This paper introduces a forest dataset called FinnWoodlands, which consists of RGB stereo images, point clouds, and sparse depth maps, as well as ground truth manual annotations for semantic, instance, and panoptic segmentation. FinnWoodlands comprises a total of 4226 objects manually annotated, out of which 2562 objects (60.6%) correspond to tree trunks classified into three different instance categories, namely “Spruce Tree”, “Birch Tree”, and “Pine Tree”. Besides tree trunks, we also annotated “Obstacles” objects as instances as well as the semantic stuff classes “Lake”, “Ground”, and “Track”. Our dataset can be used in forestry applications where a holistic representation of the environment is relevant. We provide an initial benchmark using three models for instance segmentation, panoptic segmentation, and depth completion, and illustrate the challenges that such unstructured scenarios introduce. FinnWoodlands dataset is available at https://github.com/juanb09111/FinnForest.git.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Adelson, E.H.: On seeing stuff: the perception of materials by humans and machines. In: IS &T/SPIE Electronic Imaging (2001)

    Google Scholar 

  2. Bac, C.W., Van Henten, E., Hemming, J., Edan, Y.: Harvesting robots for high-value crops: state-of-the-art review and challenges ahead. J. Field Robot. 31 (2014). https://doi.org/10.1002/rob.21525

  3. Bechar, A., Vigneault, C.: Agricultural robots for field operations: concepts and components. Biosyst. Eng. 149, 94–111 (2016)

    Google Scholar 

  4. Behley, J., et al.: A dataset for semantic segmentation of point cloud sequences. CoRR abs/1904.01416 (2019). https://arxiv.org/abs/1904.01416

  5. Brostow, G.J., Fauqueur, J., Cipolla, R.: Semantic object classes in video: a high-definition ground truth database. Pattern Recogn. Lett. (2008)

    Google Scholar 

  6. Bulò, S.R., Porzi, L., Kontschieder, P.: In-place activated batchnorm for memory-optimized training of DNNs. CoRR abs/1712.02616 (2017). https://arxiv.org/abs/1712.02616

  7. Cabon, Y., Murray, N., Humenberger, M.: Virtual KITTI 2. CoRR abs/2001.10773 (2020). https://arxiv.org/abs/2001.10773

  8. Caesar, H., et al.: nuscenes: A multimodal dataset for autonomous driving. CoRR abs/1903.11027 (2019). https://arxiv.org/abs/1903.11027

  9. Che, Z., et al.: D2-city: a large-scale dashcam video dataset of diverse traffic scenarios. arXiv abs/1904.01975 (2019)

    Google Scholar 

  10. Chen, Y., Yang, B., Liang, M., Urtasun, R.: Learning joint 2D-3D representations for depth completion. CoRR abs/2012.12402 (2020). https://arxiv.org/abs/2012.12402

  11. Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. CoRR abs/1604.01685 (2016). https://arxiv.org/abs/1604.01685

  12. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255 (2009). https://doi.org/10.1109/CVPR.2009.5206848

  13. Everingham, M., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A.: The PASCAL visual object classes challenge 2012 (VOC2012) results (2012). https://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html

  14. Fortin, J.M., Gamache, O., Grondin, V., Pomerleau, F., Giguère, P.: Instance segmentation for autonomous log grasping in forestry operations (2022). https://doi.org/10.48550/ARXIV.2203.01902, https://arxiv.org/abs/2203.01902

  15. Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2012)

    Google Scholar 

  16. Geyer, J., et al.: A2D2: Audi autonomous driving dataset. CoRR abs/2004.06320 (2020). https://arxiv.org/abs/2004.06320

  17. Grondin, V., Fortin, J.M., Pomerleau, F., Giguère, P.: Tree detection and diameter estimation based on deep learning. Forestry: Int. J. Forest Res. (2022)

    Google Scholar 

  18. He, K., Gkioxari, G., Dollár, P., Girshick, R.B.: Mask R-CNN. CoRR abs/1703.06870 (2017). https://arxiv.org/abs/1703.06870

  19. Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. CoRR abs/1502.03167 (2015). https://arxiv.org/abs/1502.03167

  20. Jiang, P., Osteen, P.R., Wigness, M.B., Saripalli, S.: RELLIS-3D dataset: data, benchmarks and analysis. CoRR abs/2011.12954 (2020). https://arxiv.org/abs/2011.12954

  21. Juliani, A., Berges, V., Vckay, E., Gao, Y., Henry, H., Mattar, M., Lange, D.: Unity: a general platform for intelligent agents. CoRR abs/1809.02627 (2018). https://arxiv.org/abs/1809.02627

  22. Lin, T., et al.: Microsoft COCO: common objects in context. CoRR abs/1405.0312 (2014). https://arxiv.org/abs/1405.0312

  23. Mitra, A., et al.: Everything you wanted to know about smart agriculture. CoRR abs/2201.04754 (2022). https://arxiv.org/abs/2201.04754

  24. Mohan, R., Valada, A.: EfficientPS: efficient panoptic segmentation. CoRR abs/2004.02307 (2020). https://arxiv.org/abs/2004.02307

  25. Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference from RGBD images. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7576, pp. 746–760. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33715-4_54

    Chapter  Google Scholar 

  26. Neuhold, G., Ollmann, T., Bulà, S.R., Kontschieder, P.: The mapillary vistas dataset for semantic understanding of street scenes. In: 2017 IEEE International Conference on Computer Vision (ICCV),. pp. 5000–5009 (2017). https://doi.org/10.1109/ICCV.2017.534

  27. Ringdahl, O.: Automation in forestry: development of unmanned forwarders. Ph.D. thesis, Umeå University, May 2011

    Google Scholar 

  28. Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. CoRR abs/1409.0575 (2014). https://arxiv.org/abs/1409.0575

  29. Sekachev, B., et al.: OpenCV/CVAT: v1.1.0, August 2020. https://doi.org/10.5281/zenodo.4009388

  30. Shamshiri, R., et al.: Research and development in agricultural robotics: a perspective of digital farming. Int. J. Agric. Biol. Eng. 11, 1–14 (2018). https://doi.org/10.25165/j.ijabe.20181104.4278

  31. da Silva, D.Q., dos Santos, F.N.: ForTrunkDet - forest dataset of visible and thermal annotated images for object detection. J. Imaging (2021)

    Google Scholar 

  32. Sinha, R.K., Pandey, R., Pattnaik, R.: Deep learning for computer vision tasks: a review. CoRR abs/1804.03928 (2018). https://arxiv.org/abs/1804.03928

  33. Tan, M., Le, Q.V.: Efficientnet: rethinking model scaling for convolutional neural networks. CoRR abs/1905.11946 (2019). https://arxiv.org/abs/1905.11946

  34. Wang, W., et al.: Tartanair: a dataset to push the limits of visual SLAM. CoRR abs/2003.14338 (2020). https://arxiv.org/abs/2003.14338

  35. Wigness, M., Eum, S., Rogers, J.G., Han, D., Kwon, H.: A RUGD dataset for autonomous navigation and visual perception in unstructured outdoor environments. In: International Conference on Intelligent Robots and Systems (IROS) (2019)

    Google Scholar 

  36. Yu, F., Xian, W., Chen, Y., Liu, F., Liao, M., Madhavan, V., Darrell, T.: BDD100K: a diverse driving video database with scalable annotation tooling. CoRR abs/1805.04687 (2018). https://arxiv.org/abs/1805.04687

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Juan Lagos .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Lagos, J., Lempiö, U., Rahtu, E. (2023). FinnWoodlands Dataset. In: Gade, R., Felsberg, M., Kämäräinen, JK. (eds) Image Analysis. SCIA 2023. Lecture Notes in Computer Science, vol 13885. Springer, Cham. https://doi.org/10.1007/978-3-031-31435-3_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-31435-3_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-31434-6

  • Online ISBN: 978-3-031-31435-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics