Mar 5, 2024: FarfetchFusion
Challenges:
Proposed Design: Disentangled Fusion
Separate static (ear, nose, forehead) and dynamic (mouth, eye) facial information
Combine static info from multiple frames while fusing only the dynamic parts from recent frames
⇒ leverage spatio-temporal redundancy in multi-view video streams
⇒ reduce processing time of static information
TSDF (distance measurement from voxel to surface) - stored in hash map
Volumetric Fusion
Alignment (feature extraction + registration)
Fusion
Rasterization
→ highly optimized but need TSDF voxel to point cloud conversion
→ Marching Cubs Algorithm (to convert TSDF voxels to point cloud)
Raycasting: generate rendered images without converting TSDF voxels to 3D data (point cloud)
→ find surface voxels (TSDF values closest to zero) by casting ray from viewer POV
→ color information used to form 2D pixels
System Design: