Skip to main content

Semi-Automated Construction of Virtual Scenes from Cinematographic Data

  • Chapter
Digital Convergence: The Information Revolution

Abstract

The aim of the present work is to demonstrate the benefit of integrating advanced computer vision techniques into a new generation of compositing tools for the post-production industry. The recent MPEG-4 image description standard supporting the overlay of multiple objects and sprites corresponds closely to the composition process and underpins our wider goal of creating the virtual post-production studio: collaborative creation of multimedia content between studios. Essential to this aim is the creation of efficient and robust compositing tools: in particular, rotoscoping and mosaicking functionality. Rotoscoping involves the separation of elements within a scene: an intensively manual procedure that we automate using active surfaces supported by edge chains. Mosaicking is a new and particularly exciting technique offering the post-production operator the opportunity to easily modify viewing trajectories, introduce computer-generated elements, stabilize jumpy action or positionally justify action in the viewport. Deriving a sophisticated three-dimensional motion model, an optical flow framework is used to facilitate the seamless merging of frames to create a larger virtual scene.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. P Anandan, JR Bergen, KJ Hanna and R Hingorani (1993) Motion analysis and image sequence processing, in Hierarchical Model-Based Motion Estimation (eds. MI Sezan and RL Lagendijk), Kluwer Academic, Boston, pp. 1–22.

    Google Scholar 

  2. S Ayer, P Schroeter and J Bigün (1994) Segmentation of moving objects by robust motion parameter estimation over multiple frame, in Proc. European Conference on Computer Vision, Stockholm, pp. 316–27.

    Google Scholar 

  3. JL Barron, DJ Fleet and SS Beauchemin (1994) Performance of optical flow techniques, International Journal of Computer Vision, 12 (1), 43–77.

    Article  Google Scholar 

  4. M Bober and J Kittler (1994) Robust motion analysis, in Proc. IEEE Computer Soc. Conf on Computer Vision and Pattern Recognition, pp. 947–52.

    Google Scholar 

  5. J Canny (1986) A computational approach to edge detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, 6 (6), 679–98.

    Article  Google Scholar 

  6. L Cohen and I Cohen (1993) finite-element methods for active contour models and balloons for 2D and 3D images, IEEE Transactions on Pattern Analysis and Machine Intelligence, 15 (11), 1l31–47.

    Article  Google Scholar 

  7. JM Corridoni, A Del Bimbo, D Lucarella and WX He (1996) Multi-perspective navigation of movies, Journal of Visual Languages and Computing, 7 (4), 445–466.

    Article  Google Scholar 

  8. D Daneels, D Van Campenhout, W Niblack, W Equitz, R Barber, E Bellon and F Fierens (1993) Interactive outlining: an improved approach using active contours, in Storage and Retrieval for Image and Video Databases, Vol. 1908, pp. 226–33, Proceedings of the International Society of Optical Engineering.

    Google Scholar 

  9. B Duc, P Schroeter and J Bigün (1995) Spatio-temporal robust motion estimation and segmentation, in 6th Int. Conf Computer Analysis of Images and Patterns, Prague, September, Springer-Verlag, p. 238–45.

    Google Scholar 

  10. J Figue (1997) Nemesis project: advanced image analysis tools based on a new 3D wide sense object oriented data representation model for multimedia content creation and post-production applications, in Workshop on Image Analysis for Multimedia Interactive Services, Louvain-Ia-Neuve, Belgium, 24–25 June, pp. 117–26.

    Google Scholar 

  11. E François (1997) Rigid layers reconstruction based on motion segmentation, in Workshop on Image Analysis for Multimedia Interactive Services, Louvain-la-Neuve, Belgium, 24–25 June, pp. 81–6.

    Google Scholar 

  12. M Gelgon and P Bouthemy (1997) A hierarchical motion-based segmentation and tracking technique for video storyboard-like representation and content-based indexing, in Workshop on Image Analysis for Multimedia Interactive Services, Louvain-la-Neuve, Belgium, 24–25 June, pp. 93–8.

    Google Scholar 

  13. PR Giaccone, D Greenhill and GA Jones (1997) Recovering very large visual motion fields, in 10th Scandinavian Conf. on Image Analysis, Finland, June, pp. 917–22.

    Google Scholar 

  14. PR Giaccone and GA Jones (1997) Feed forward estimation of optical flow, in IEE Conf. on Image Processing and its Applications, Dublin, July.

    Google Scholar 

  15. PR Giaccone and GA Jones (1997) Spatio-temporal approaches to the computation of optical flow, in Proc. British Machine Vision Conference, Colchester, UK, September, pp. 420–9.

    Google Scholar 

  16. D Greenhill and GA Jones (1997) Supervised segmentation of cinematographic sequences using active surfaces, in Workshop on Image Analysis for Multimedia Interactive Services, Louvain-la-Neuve, Belgium, 24–25 June, pp. 39–44.

    Google Scholar 

  17. ISO (1994) Information technology - generic coding of moving pictures and associated audio information - Part 2: Video, Dis 13818–2, ISO/IEC.

    Google Scholar 

  18. ISO (1997) Coding of moving pictures and audio: MPEG-7 applications document, N1735, ISO/IEC, Stockholm MPEG Meeting..

    Google Scholar 

  19. GA Jones (1997) Matching corner features using edge contour data, Technical Report KUCSES-97–02, Computer Vision Research Group, Kingston University.

    Google Scholar 

  20. M Kass, A Witkin and D Terzopoulos (1988) Snakes: active contour models, International Journal of Computer Vision, 1, 321–31.

    Article  Google Scholar 

  21. L Kitchen and A Rosenfeld (1982) Graylevel corner detection, PRL, 1, 95–102.

    Article  Google Scholar 

  22. J Kreyss, M Roper, P Alshuth, T Hermes and O Herzog (1997) Video retrieval by still image analysis with ImageMiner, in Proc. International Society for Optical Engineering, Storage and Retrieval for Image and Video Databases, San Jose, 13–14 February, pp. 36–44.

    Google Scholar 

  23. E-P Ong and M Spann (1995) Robust computation of optical flow, in Proc. British Machine Vision Conference, Vol. 2, pp. 573–82, 1995.

    Google Scholar 

  24. M Otte and H-H Nagel (1995) Estimation of optical flow based on higher-order spatiotemporal derivatives in interlaced and non-interlaced image sequences, Artificial Intelligence, 78, 5–43.

    Article  Google Scholar 

  25. J Porril and J Ivins (1994) A semiautomatic tool for 3D medical image analysis using active contour models, Medical Informatics, 19 (1), 81–90.

    Article  Google Scholar 

Download references

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer-Verlag London Limited

About this chapter

Cite this chapter

Giaccone, P.R., Greenhill, D., Jones, G.A. (1999). Semi-Automated Construction of Virtual Scenes from Cinematographic Data. In: Vince, J., Earnshaw, R. (eds) Digital Convergence: The Information Revolution. Springer, London. https://doi.org/10.1007/978-1-4471-0863-4_17

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-0863-4_17

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-4471-1220-4

  • Online ISBN: 978-1-4471-0863-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics