About the project

Patti Brennan talks about the vizHOME project.

From the mid 1980s we've been in over a thousand homes in the Midwest to help people manage their health at home and study how different kinds of computer tools might help people take care of themselves. What we’ve learned over and over again is that tools built for business environments don’t fit in bedrooms or kitchens. As healthcare increasingly moves into the home, we’ve got to have devices that better support people there. The vizHOME project was designed to do just that.

What is vizHOME?

vizHOME is a 5-year project funded by AHRQ whose primary aim was to explore the impact of home environments on how people manage their health information at home. The AHRQ website describing vizHOME is available here.

Health care today relies greatly on helping people help themselves. From monitoring diets to managing medication, the typical person faces many personal health information management (PHIM) demands. Computer tools can assist with this information management, but they are often created and tested in a lab, not in a home where they will be used.

We have learned that people use their environment to assist them in their self-care. They rely on such things as the visibility of a reminder system for a weekly blood test and the proximity of a glucometer to the computer that stores its readings to help themselves with PHIM. Device designers need to understand where such things occur in order to make effective solutions that work for them. So vizHOME set out to achieve a better understanding of how the home environment affects how, where, and how well people use these types of health care devices.

Using LiDAR, a laser emitting scanner, we captured the interior of homes in 3D, and then visualized them in our virtual reality CAVE. We were then able to use these 3D models to learn about how other people would use the same space, and infer how space affected what they did.

Participants and their homes

In the first year of vizHOME, we recruited 20 people who had been told that they have diabetes. Participants signed an informed consent form that was approved by the University of Wisconsin-Madison Institutional Research Board (IRB).

For more details on the participants and their homes, go here.

We conducted intensive interviews that explored each participantʼs health concerns and health information management practices. The interior of their homes were scanned using LiDAR during the 1-3 visits to each home. The LiDAR scanner used provided both depth and color, allowing computers to create a 3D reconstruction of the space. Photographs and other static views of houses do not afford the experience of actually walking through the houses in which the PHIM tasks are performed. Virtual reality (VR) models of actual homes, rendered in CAVEs such as ours, address some of these limits.

We sought to help designers and clinicians and computer scientists understand very intimate spaces in the home, to understand that sometimes spaces are quite cramped and lack sufficient storage or work space. Using LIDAR, we captured a 360 degree scan of any space within each home. We processed those images and displayed them in the VR CAVE. So participants in later phases of the study were able to walk through each house without burden to the actual resident and as if they were actually in the house. Participants were able to understand what it would be like to be that person in their own house and tell us if any of the various features would help or interfere with their ability to give care to themselves

Because these virtual replicas are displayed to-scale in the CAVE, participants could look at such things as relative location of key features, the size of doorways or the height of a toilet seat. And future users e.g. device designers, health care providers or architects would benefit from knowing the dimensions of spaces and objects within the homes.

For more details on how LiDAR scanning works, go here.

PHIM in the Virtual Homes

In later phases of vizHOME we brought in groups composed of experts (professionals including nurses, computer scientists, engineers) and community-dwelling lay people with diabetes to navigate through these reconstructed virtual houses to help identify what aspects of the houses might help, or interfere with, their health information management. Lay people also explored a series of non-contiguous rooms from a subset of the houses to identify what features would be most useful to them for performing PHIM in order to reduce and refine our list of key features.

Through this process, we developed a checklist named the Assessment of the Context of Home Environments (ACHE) that could be used by designers, clinicians and home care specialists, and lay people to plan the best places to store health information, build in reminders for common tasks like taking medications, and better organize health information.

Finally, we've created this website to share the homes with others who may be interested in learning about them or leveraging them for a specific purpose including device designers, health care providers, computer scientists and educators.

Broader Impacts of the Project

The techniques developed for use in vizHOME and the insights the project generated enabled further work, in the field of health care and beyond. Projects that have resulted from this work include:

Including Homes in Health Records

https://healthit.ahrq.gov/events/national-web-conference-use-health-it-aging-adults

As the team saw the benefits of the the virtualization of home environment, a project was undertaken to explore the how these datasets could be included in the electronic health record. The AHRQ R03 project HS024623 Virtualized Homes: Tools for Better Discharge Planning, created a partnership between the University of Wisconsin Hospitals and Clinics and Epic Systems. This project, alongside the vizHOME project, was presented as part of a National Web Conference on Use of Health IT for Aging Adults for AHRQ in July of 2017.

Simulation Guide for Virtual Home Visits

https://care.nursing.wisc.edu/education/virtual-home-visit-guide/

Gail Casper, PhD, RN developed a guide to help nurse educators use the vizHOME homes for virtual home visits. This guide has been made available through CARE at the UW-Madison School of Nursing.

Age in Place

https://ageinplace.wiscweb.wisc.edu/

Employing and extending techniques used in the vizHOME project, our research team employed 3D scanning, tracking biomarkers, and housing accessibility assesment tools to identify, document, and visualize the everyday activities of older adults at home, common barriers that hinder their activities, and how home modifications influence health related behaviors.

Study of Homes for Crime Scene Investigation

https://3d-csi.discovery.wisc.edu/

As the team was scanning homes for the purposes of studying health, an unexpected collaboration emerged. After a homicide occurred in a nearby town, the team was contacted by the local sheriff's office and FBI to perform the same methodology developed in this project for the purposes of CSI. This collaboration was considered an overwhelming success and the team applied for funding opportunities to explore furthering the collaboration. In early 2017, the team was awarded the project funded by the Department of Justice, 2016-IJ-CX-0017 Analyzing the Impact of Virtual Reality and 3D Capture Technology on Crime Scene Investigation.

Cultural Heritage

The project team has also found applications in the areas of cultural heritage. The team first teamed with Taliesin Preservation Inc. and the Frank Lloyd Wright Foundation in 2014. The team scanned approximately one-third of the 40,000-square-foot estate capturing locations that are not available on regularly scheduled tours of the estate due to accessibility limitations. The model created from those scans contained nearly 2 billion points in a single binary data file almost 35 GB in size. The data set is the largest and most detailed 3D model of Taliesin East in existence.

To broaden the impact of this work, the lab hosted several outreach events showcasing the Taliesin scan. The team held a public event at the WID facility, featuring talks by members of the project team and the Taliesin Preservation and offering attendees the chance to experience the model generated from the data set in the CAVE system and showcased the model at two community libraries in Wisconsin (the Marathon County public library in Wausau, WI and the Spring Green public library in Spring Green, WI).

The team has also partnered with the Mount Horeb Area Historical Society (MHAHS) to 3D scan and showcase historic buildings from the geographic region surrounding Mount Horeb, WI. The team is completing a museum exhibit located at the newly renovated Driftless Historium in Mount Horeb, WI to showcase these 3D scans. The museum exhibit combines the rendering technology developed in vizHOME with a consumer grade head mounted display (the Oculus Rift) to allow users to experience the buildings in virtual reality and combines this with audio narration that discusses the buildings’ history and cultural significance. The exhibit currently contains one 3D scan of a historic church that is no longer in the area, and the final exhibit is expected to be complete by mid October 2018 with the addition of two buildings (a historic barn and an additional historic church).

Recently, members from the vizHOME team have partnered with the National Parks Service in efforts of capturing 3D models of underwater shipwrecks. Underwater video and images were taken of several shipwrecks located near the Isle Royale in Michigan. Members from the vizHOME team use photogrammetry techniques on these videos to reconstruct 3D models. In addition to the Isle Royale, recently a member of the team went to Dry Tortugas National Park in Florida to continue work with the National Park Service on reconstructing 3D models of shipwrecks in that geographic region.

The group's work in the area of cultural heritage was recognized on the cover of the National Historic Preservation’s Forum Journal.

New Intellectual Property

The project team quickly found that existing methods to render pointclouds in virtual reality were insufficient for the project. A new rendering system was created which enabled both a high level of interactivity while maintaining high quality visuals. This work was patented by the Wisconsin Alumni Research Foundation and in 2018 funding was given to broaden the impact of this work.

Acknowledgements

This project was supported by grant number R01 HS22548-01 from the Agency for Healthcare Research and Quality. The content is solely the responsibility of the authors and does not necessarily represent the official views of the Agency for Healthcare Research and Quality.

Publications

Presentations

vizHOME Team

Faculty & Academic Staff

  • Patricia Flatley Brennan
  • Markus Broecker
  • Pascale Carayon
  • Gail Casper
  • Peter Hoonakker
  • Dan Nathan-Roberts
  • Kevin Ponto
  • Catherine Arnott Smith
  • Ross Tredinnick
  • Nicole Werner
  • Susan Valentine
  • Susan Zahner

Project Assistants

  • Taylan Acar
  • Dustin Brockberg
  • Anna Jolliff
  • Ting-Lan Ma
  • Alex Marvin
  • Rosaleena Mohanty
  • Conor Murphy
  • Alex Peer
  • Jesse Perreault
  • Simon Smith
  • Naveen Anand Subramaniam

Undergraduate Research Assistants

  • Brady Boettcher
  • Nikhita Chawla
  • Saulo Couto
  • Cale Geffre
  • Thaddeus Gue
  • Angela Holecko
  • Jake Lundy
  • Raymond Luu
  • Patrick Moosai
  • Lia Moratelli
  • Andrew Morland
  • Dana Paz
  • Peter Procek
  • Madhusudan Rajendran
  • Anne Reagan
  • Juliana Reams
  • Kayla Sampson
  • Sam Solovy
  • Nathan Weinshenker