Case Study: Volum + Newmark Collaboration


An exploration by Volum and the Newmark J-School, with support from the New York Times and the New York City Media Lab.
Objective
Test, refine, and document Volum's volumetric video capture process for use in the field, and make short volumetric videos that readers can interact with in AR.
Toolkit
The Volum Box depth camera, built around the Intel RealSense D415 depth + RGB camera, provides the volumetric video capture. The photogrammetry process is used to capture and produce a static reconstruction of the location in 3D. Using the Volum workflow devleoped by Ben Kreimer, the volumetric capture and photogrammetry scene scan are then merged together in Unity, recreating the captured scene for viewing in augmented reality.
The Challenges
There is no field-ready hardware available.
There is no established post-production workflow for field capture.
The Context
For 12 weeks in the spring of 2020 (including a month of the COVID-19 pandemic), we looked at the existing Volum workflow.
Team
Screen Shot 2020-04-03 at 3.47.49 PM.png
() | Matt MacVey | Keishel Williams |

What we did
A look behind the scenes at our first live capture efforts
Watch a screen capture of our first prototype
What we learned

Next Steps
Finish Azure Kinect version of the Volum Box
Refine multiple 3D camera workflow
Optimize photogrammetry and volumetric video content for better performance on mobile and non-mobile platforms
Final Presentation
_____________________________________________
Want to print your doc?
This is not the way.
Try clicking the ⋯ next to your doc name or using a keyboard shortcut (
CtrlP
) instead.