Visualization Tools to immerse yourself, interact with, color, selectively visualize, & share
Analysis Tools to precisely and intuitively mark, measure, classify, edit, and segment
A smooth, responsive, and comfortable VR viewing experience
Unique tools to track objects in time-series images
A collection of VR specific menus and tools to provide a seamless experience
Applicability to virtually any 3D imaging technique and across application domains
In development for more than two and a half years, arivis InViewR has been constantly refined to provide users with the most natural and comfortable experience. When viewing images in VR it is important to supply the user with frame rates of at least 75 frames per second per eye and react to head movements in under 20ms so that the user does not feel seasick when immersed in the data. arivis InViewR utilizes patent-pending volume rendering techniques to achieve this performance even with large datasets. arivis InViewR goes further positioning the user in a virtual room to make the user feel grounded and providing them context to “walk-around” in. If a user gets lost, a simple button push resets the view. Advanced users can even control their environment, changing its visibility, transparency, and background color.
arivis InViewR offers the user a choice of control options. A user can utilize Leap Motion’s hand tracking to control movement with their bare hands. For those users that wish to take advantage of advanced functions, hand controllers which provide extremely previse tracking and option buttons can be utilized. Regardless of control choice, learning to move and interact with data is an intuitive process with a short learning curve. arivis InViewR also provides the possibility to enable/disable tools and gestures to match the skill level or use case of the end-user.
VR menus, including context-sensitive help, have been designed to allow full control of all operations and tools while in the VR space to provide a seamless user experience. Users can interact with menus or VR objects (segments, measurements, and markers) by laser pointing at them or reaching out to touch them in VR space. Menus are designed to have a consistent layout and consistent placement to allow easy use.
All functinos to work with image data are located in one unobtrusive menu that sticks to one hand and is accessible. Icons on the controller in VR clearly indicate what tool is in use and what mode the tool is currently in. In tool-specific menus, designed for each tool, users can adjust tool properties and find out additional information about their data seamlessly while working in VR space. It also allows users to globally toggle visibility on and off for clipping planes, measurements, markers, overlays and segments and enables to skip along the time dimension of their data.
The main menu provides all the necessary controls to work in VR space once a user has loaded a dataset. The arivis InViewR main menu provides full control over the virtual room, rendering settings, tools, and the properties of all measurements, markers, overlays, and segments. It is also possible to change menu themes her or switch between light and dark modes.
Data sets can be imported from imaging techniques such as:
- Light Sheet Microscopy
- Confocal Microscopy
- Electron Microscopy
- Computer Tomography (CT / X-Ray)
- Magnetic Resonance Imaging (MRI)
Utilize arivis InViewR
A multitude of scientific disciplines benefit from arivis InViewR including:
- Developmental Biology
- Cell Biology
- Cancer Research
Image Courtesy: Carl ZEISS Microscopy