SGN-34006 3D and Virtual Reality, 5 cr
Additional information
Suitable for postgraduate studies.
Person responsible
Atanas Gotchev
Lessons
Implementation | Period | Person responsible | Requirements |
SGN-34006 2016-01 | 2 |
Atanas Gotchev Olli Suominen |
Passing three lab exercises and one project assignment. |
Learning Outcomes
The course targets new and emerging technologies for 3D visual scene capture, processing and visualization as well as their use in applications such as 3D video and virtual reality. The course will provide in-depth knowledge about the creation, processing, delivery and visualization of 3D moving scenes. After the course, the students will know the basics of binocular vision, and how those basics are taken into account while capturing and processing 3D video. The students will be able to design application-specific multi-camera and multi-sensor capture systems. They will know the principles and the state-of-the-art techniques for compression and transmission of multi-modal signals used for conveying 3D information. The students will know the state-of-the-art and emerging visualization techniques, such as auto-stereoscopic, light-field, holographic, head-mounted and haptic-augmented displays. The students will have knowledge about typical visual artifacts appearing during the stage of 3D content creation, compression and transmission and about the advanced signal processing methods which can tackle those artifacts. 3D tracking sensors: mechanical, optical, magnetic, audio, etc. will be overviewed and basics of augmented and virtual reality will be presented.
Content
Content | Core content | Complementary knowledge | Specialist knowledge |
1. | 1. Binocular vision; perception of depth | Vergence and accommodation | Visual processing paths |
2. | 2. 3D scene capture: multi-camera, multi-sensor approaches. | Calibration and rectification of multi-camera setting with different topologies. Depth estimation. | Capture-specific 3D visual artifacts |
3. | 3. 3D scene representation formats: multi-view + multi-depth; layered depth; epipolar image; point-clouds; dynamic meshes | Format conversion methods | Representation-specific 3D visual artifacts |
4. | 4. Copression of 3D imagery. MVMD compression. Point-cloud and mesh compression. Emerging standards. | Light-field compression. Multi-modal 3D scene compression. | Compression-specific 3D visual artifacts |
5. | 5. 3D and VR displays. | Display-specific artifacts | Quality measurements for 3D displays. |
6. | 3D tracking sensors: mechanical, optical, magnetic, audio. Motion capture | Augmented reality | Virtual reality |
Instructions for students on how to achieve the learning outcomes
Three lab works and one project assignment should be completed. Course grade is computed as follows: firt lab work - 20%; second lab work - 20%; third lab work - 20%; completed project - 40%. Total grade=0.2L1+0.2L2+0.2L3+0.4P. All four units must be passed otherwise final grade is not given. There is no exam.
Assessment scale:
Numerical evaluation scale (0-5)
Prerequisites
Course | Mandatory/Advisable | Description |
SGN-12006 Basic Course in Image and Video Processing | Advisable | |
SGN-51506 Human Visual System | Advisable |
Correspondence of content
Course | Corresponds course | Description |
SGN-34006 3D and Virtual Reality, 5 cr | SGN-5456 3D Media Technology, 4 cr |