Biomedical Modeling in Tele-Immersion

  • Zhuming Ai
  • Raymond Evenhouse
  • Jason Leigh
  • Fady Charbel
  • Mary L. Rasmussen
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4650)

Abstract

The major goal of this research is to develop a networked collaborative surgical system for tele-immersive consultation, surgical pre-planning, implant design, post operative evaluation and education. Tele-immersion enables users in different locations to collaborate in a shared, virtual, or simulated environment as if they are in the same room.

The process of implant design begins with CT data of the patient and the Personal Augmented Reality Immersive System (PARISTM). The implant is designed by medical professionals in tele-immersive collaboration. In the PARIS augmented reality system the user’s hands and the virtual images appear superimposed in the same volume so the user can see what he is doing. A haptic device supplies the sense of touch by applying forces to a stylus that the medical modeler uses to form the implant. After the virtual model of the implant is designed, the data is sent via network to a stereolithography rapid prototyping system that creates the physical implant model. After implant surgery, the patient undergoes a postoperative CT scan and results are evaluated and reviewed over the tele-immersive consultation system.

Keywords

Augmented Reality Haptic Device Volumetric Data Biomedical Modeling Augmented Reality System 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Dujovny, M., Evenhouse, R., Agner, C., Charbel, F.T., Sadler, L., McConathy, D.: Preformed prosthesis from computed tomography data: Repair of large calvarial defects. In: Benzel, E.C., Rengachary, S.R. (eds.) Calvarial and Dural Reconstructuion. American Association of Neurological Surgeons, Park Ridge, vol. Ill, pp. 77–88 (1999)Google Scholar
  2. 2.
    Benzel, E.C., Alberstone, C.D.: Calvarial and Dural Reconstructuion. In: Cranioplasty Materials. American Association of Neurological Surgeons, Park Ridge, vol. Ill (1999)Google Scholar
  3. 3.
    Agner, C., Dujovny, M., Evenhouse, R., Charbel, F.T., Sadler, L.: Stereolithography for posterior fossa cranioplasty. Skull Base Surg. 8(2), 81–86 (1998)CrossRefGoogle Scholar
  4. 4.
    Leigh, J., DeFanti, T., Johnson, A., Brown, M., Sandin, D.: Global tele-immersion: Better than being there. In: Proceedings of ICAT 1997, 7th Annual International Conference on Artificial Reality and Tele-Existence Virtual Reality Society of Japan, University of Tokyo, Japan, pp. 10–17 (December 1997)Google Scholar
  5. 5.
    Johnson, A., Sandin, D., Dawe, G., Qiu, Z., Thongrong, S., Plepys, D.: Developing the paris: Using the cave to prototype a new vr display. In: CDROM Proceedings of IPT 2000: Immersive Projection Technology Workshop, Ames, IA (June 2000)Google Scholar
  6. 6.
  7. 7.
    Scharver, C., Evenhouse, R., Johnson, A., Leigh, J.: Designing cranial implants in a haptic augment reality environment. Communications of the ACM 27(8), 32–38 (2004)CrossRefGoogle Scholar
  8. 8.
    Chen, H., Sun, H.: Real-time haptic sculpting in virtual volume space. In: Proceedings of the ACM symposium on Virtual reality software and technology, Hong Kong, China, pp. 81–88. ACM Press, New York (2002)Google Scholar
  9. 9.
    Dachille IX, F., Qin, H., Kaufman, A., El-Sana, J.: Haptic sculpting of dynamic surfaces. In: Proceedings of the 1999 symposium on Interactive 3D graphics, Atlanta, Georgia, United States, pp. 103–110. ACM Press, New York (1999)CrossRefGoogle Scholar
  10. 10.
    Jagnow, R., Dorsey, J.: Virtual Sculpting with Haptic Displacement Maps. In: Proc. Graphics Interface, pp. 125–132 (May 2002)Google Scholar
  11. 11.
    Iwata, H., Noma, H.: Volume haptization. In: IEEE 1993 Symposium on Research Frontiers in Virtual Reality, pp. 16–23 (October 1993)Google Scholar
  12. 12.
    Lundin, K., Ynnerman, A., Gudmundsson, B.: Proxy based haptic feedback from volumetric density data. In: EuroHaptics 2002, Edinburgh, UK (2002)Google Scholar
  13. 13.
    Petersik, A., Pflesser, B., Tiede, U., Höhne, K.-H., Leuwer, R.: Realistic haptic interaction in volume sculpting for surgery simulation. In: Ayache, N., Delingette, H. (eds.) IS4TM 2003. LNCS, vol. 2673, pp. 194–202. Springer, Heidelberg (2003)CrossRefGoogle Scholar
  14. 14.
    Ai, Z., Evenhouse, R., Rasmussen, M.: Haptic rendering of volumetric data for cranial implant modeling. In: The 27th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Shanghai, China (September 2005)Google Scholar
  15. 15.
    Ai, Z., Evenhouse, R., Leigh, J., Charbel, F., Rasmussen, M.: New tools for sculpting cranial implants in a shared haptic augmented reality environment. Stud. Health Technol Inform. 119, 7–12 (2006)Google Scholar
  16. 16.
    Ackerman, M.J.: Accessing the visible human project. D. Lib. Mag (October 1995)Google Scholar
  17. 17.
    Kajiya, J.T., Von Herzen, B.P.: Ray tracing volume densities. In: Proceedings of the 11th annual conference on Computer graphics and interactive techniques, pp. 165–174. ACM Press, New York (1984)Google Scholar
  18. 18.
    Rezk-Salama, C., Engel, K., Bauer, M., Greiner, G., Ertl, T.: Interactive volume rendering on standard pc graphics hardware using multi-textures and multi-stage-rasterization. In: Eurographics / SIGGRAPH Workshop on Graphics Hardware 2000, pp. 109–118. Addison-Wesley Publishing Company, Inc., Reading (2000)Google Scholar
  19. 19.
    Cabral, B., Cam, N., Foran, J.: Accelerated volume rendering and tomographic reconstruction using texture mapping hardware. In: ACM Symp. on Vol. Vis. (1994)Google Scholar
  20. 20.
    Magallon, M., Hopf, M., Ertl, T.: Parallel volume rendering using pc graphics hardware. In: Ninth Pacific Conference on Computer Graphics and Applications (PG 2001), Tokyo, Japan (2001)Google Scholar
  21. 21.
    Ai, Z., Dech, F., Rasmussen, M., Silverstein, J.C.: Radiological tele-immersion for next generation networks. Studies in Health Technology and Informatics 70, 4–9 (2000)Google Scholar
  22. 22.
    Dech, F., Ai, Z., Silverstein, J.C.: Manipulation of volumetric patient data in a distributed virtual reality environment. Studies in Health Technology and Informatics 81, 119–125 (2001)Google Scholar
  23. 23.
    SensAble Technologies, Inc. 3D Touch SDK - OpenHaptics Toolkit Programmer’s Guide (1999-2004)Google Scholar
  24. 24.
    Cruz-Neira, C., Sandin, D.J., De Fanti, T.A.: Surround-screen projection-based virtual reality: The design and implementation of the CAVE. In: Proc. Siggraph 1993, pp. 135–142. ACM Press, New York (1993)Google Scholar
  25. 25.
    Schroeder, W., Martin, K., Lorensen, B.: The Visualization Toolkit: An Object-Oriented Approach To 3D Graphics. Prentice Hall PTR, Englewood Cliffs (1996)Google Scholar
  26. 26.
  27. 27.
  28. 28.
    Leigh, J., Yu, O., Schonfeld, D., Ansari, R., et al.: Adaptive networking for tele-immersion. In: Proc. Immersive Projection Technology/Eurographics Virtual Environments Workshop (IPT/EGVE), Stuttgart, Germany (May 2001)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2008

Authors and Affiliations

  • Zhuming Ai
    • 1
  • Raymond Evenhouse
    • 1
  • Jason Leigh
    • 2
  • Fady Charbel
    • 3
  • Mary L. Rasmussen
    • 1
  1. 1.Virtual Reality in Medicine Lab Department of Biomedical and Health Information SciencesUniversity of Illinois at ChicagoChicagoUSA
  2. 2.Electronic Visualization LabUniversity of Illinois at ChicagoChicagoUSA
  3. 3.Department of NeurosurgeryUniversity of Illinois at ChicagoChicagoUSA

Personalised recommendations