These days 3D tours are widely used for visualizing and marketing indoor spaces. They are also used to provide detailed space documentation such as floor plans and measurements. Anyone looking at getting into making 3D tours is likely overwhelmed by the number of technological offerings from many different vendors and may have a hard time choosing which one to use.
This review provides a comparison of the most popular methods and products for capturing property data and creating 3D tours. It discusses the advantages and weaknesses of different types of cameras and equipment. I will be drawing on my previous experience developing optical metrology systems and more recently designing one of the discussed technologies (iGUIDE).
For the most part, this discussion will focus on 3D tours of real physical spaces and not computer-generated 3D environments.
Most often, a 3D tour is a series of 360° images, also called photospheres, panoramas, or panos, where a user can navigate from one 360° image to another. There is an opinion that simply using 360° images without an underlying 3D mesh, or 3D point cloud collected by a camera, is not enough for such a presentation to be called a 3D tour, so an explanation is needed.
The term 3D, shorthand for three dimensional, means having three degrees of freedom in which a user can move. In each 360° image a user can look left-right and up-down and these degrees of freedom add up to two dimensions. In addition, a user can move between 360° images, zooming in and out of the images during such transitions. This degree of freedom adds yet another dimension, resulting in the experience of moving through 3D space. Because of that, the term “3D tour” is well justified for tours made using connected 360° images.
Another way to show a 3D tour is to use a 3D mesh with image textures applied to it. This approach is generally not used in a web browser for rendering real physical spaces captured with a camera, due to the high complexity of 3D mesh for the limited resources of a web browser. However, computer-generated 3D environments are often rendered in web browsers using the textured 3D mesh approach because in such cases the 3D mesh is a lot simpler.
When viewing a 360° image, a user can look all the way around and up and down. They can also zoom in and out. When it comes to navigating to another 360° image, there are a few different options available.
Nearly all 3D tours allow navigation through 3D space by clicking in 360° images on some kind of a hotspot or simply in a desired direction of movement. The 3D tour then renders a transition to the next location simulating movement in the 3D space. This type of simulated movement doesn’t render in an entirely realistic way and even in cases where an underlying 3D mesh is available, the transitions often exhibit noticeable unrealistic artifacts. Nevertheless, such imperfections have a relatively small effect on a user, because during the transition the human brain processes data mostly to assess speed and direction of motion, rather than details of the surrounding 3D space.
Without some kind of a map as a reference, it can be very difficult to make sense of a floor layout and after a couple of 360° image transitions, the user can find themselves lost in space and unsure of where they are going or have been.
This is where showing a floor plan or a map alongside the 360° image is very helpful and improves the user experience, as the user’s current position and viewing direction can be displayed on the floor plan, providing constant location tracking.
Simply clicking on a floor plan can allow instant teleportation to another location across the floor, providing for very efficient and intuitive navigation which is possible only with a computing device. This is opposite to the 360° image navigation that attempts to simulate a real-life walking experience and requires a lot of mouse clicks or “steps” to move across the floor.
Floor plan or map navigation is used in very few 3D tours, notably in iGUIDE, Cupix, InsideMaps, and Google Street View. In those tours it is used in addition to 360° image navigation, giving users the option to choose how to explore the space.
Some 3D tours feature a 3D dollhouse, such as those produced by Matterport. The dollhouse can be rotated in space to get an idea of property layout for all floors at once. It is possible to navigate using the dollhouse but is not very practical given the number of clicks required to enter the dollhouse mode, rotate the dollhouse, zoom in or out, and finally select a new location. It is definitely quicker than using the 360° image walking-type navigation, but certainly not as efficient as using floor plan navigation.
Since it is impossible to capture a 360° image with one lens and in one shot, either several lenses or several shots with a single lens, rotated each time, or a combination of both are used. Several resulting source images are then stitched together to produce a 360° image.
When using several lenses, the 360° images can exhibit stitching artifacts for objects close to the camera due to parallax errors, because the same objects are viewed from different locations by different lenses. When shooting indoors, more often than not, there will be objects in a scene that are close to the camera, so stitching artifacts most often happen when inside buildings.
To see what a parallax error looks like, hold one pen about a foot or 30 cm in front of you and a second pen two feet in front of you and close one eye at a time. You will see that the distant pen will be either to the left or to the right of the closest pen, depending on which eye is closed.
When using a single lens and rotating it several times, parallax errors can be practically avoided, if the lens is rotated around a special point called the entrance pupil, with the exact location depending on the lens design.
Besides the number of lenses and other design details, 3D tour capture technologies can also differ in whether they take any spatial measurements and which measurement principle is used. Several popular 3D tour capture technologies are discussed below.
Among the many different 360° cameras, the most popular and supported ones are the Ricoh Theta and the Insta360. These cameras feature a slim two-lens design which reduces, but does not completely eliminate parallax errors, while cameras with more than two lenses will have higher parallax errors and more stitching artifacts.
Because 360° images are captured in one shot, the onsite capture speed is rather fast with these cameras. However, there is a need to run and hide in another room when taking a 360° image, which does affect shooting time. Some 360° cameras support a two-shot mode, where a photographer needs to step around the camera and take two shots for one 360° image.
The 360° cameras do not include any measurement technology. Nevertheless, images can be used for inferring 3D space structure and extracting floor plans, though with limited accuracy.
The 360° images from 360° cameras are softer than from DSLR cameras, especially around the edge of each of the two lenses. Nevertheless, since more than 50% of 3D tours are viewed on mobile devices and 3D tours on the desktop aren’t often viewed full screen and on a large monitor, limited image resolution is not a big issue.
Small image sensors used in most 360° cameras do cause extra noise in images under low light conditions typical for indoor shooting, which can make the resulting 360° images appear flat and lack depth. The Ricoh Theta Z1 was able to address this issue by using larger image sensors.
DSLR camera setups feature a wide-angle or fisheye lens that is typically mounted in a way to avoid parallax errors when the camera is rotated into multiple positions to produce a 360° image. The number of positions required to produce a 360° image depends on a field of view of the lens and typically is four to twelve.
There are several software packages that allow for stitching source images with varying degrees of speed and ease of use. Once the 360° images are produced, platforms like Cupix can be used to turn them into 3D tours.
The capture speed with DSLRs is slower than with 360° cameras due to the need for multiple rotations. However, the photographer is always behind the camera and there is no need to hide in another room. Low light performance indoors is improved due to large image sensors used in DSLR cameras.
Just as with 360° cameras, DSLR setups do not include any measurement technology, but the images can be used for inferring 3D space structure and extracting floor plans.
Mobile phones can be used with automated rotators to take several shots that are stitched into a 360° image. If the rotator and phone are compatible and with the correct setup, the phone can be rotated in a way to minimize parallax errors.
Due to phone lenses having a smaller field of view than lenses in 360° cameras, 12 to 18 shots are required for a single 360° image, but the resulting 360° image resolution is higher than with 360° cameras.
Usually, the phones are not rotated all the way up, so the resulting 360° image does not cover a complete photosphere and part of the ceiling area is missing.
Because of the required phone rotation, the onsite capture speed is much slower with phones as compared to 360° cameras, even though it takes a comparable number of camera positions on the floor.
The source images are processed into a 3D tour through software platforms, such as InsideMaps and these vendors also provide their custom motorized rotators that typically cost under $200.
The Zillow 3D Home app and Matterport’s recently announced handheld smartphone capture do not need a rotator to produce 3D tours. Due to their handheld workflow, parallax errors are likely. The capture speed with the handheld process is slower than with phones on rotators.
Much like 360° cameras, phones normally do not include any measurement technology, but the images can be used for inferring 3D space structure and extracting floor plans.
Matterport’s 3D camera uses three lenses and is rotated into six positions with the resulting 18 source images stitched into one 360° image. Because of that optical setup, stitching artifacts can be visible for close-up objects in the image.
The Matterport camera also incorporates three 3D scanners. These scanners use a structured light measurement principle, where a pattern of infrared light is projected outwards and viewed by an infrared camera to compute a 3D point cloud. The camera data is processed through the Matterport platform to produce a 3D tour which includes 3D mesh with the dollhouse and optional floor plans.
The need to collect overlapping 3D point clouds and produce a 3D mesh may require two to three times more camera positions compared to 360° cameras and the overall capture speed will, respectively, be slower.
The Leica BLK360 Imaging Scanner includes several panoramic lenses and the resulting 360° images can exhibit stitching artifacts.
The BLK360 employs a time-of-flight 3D laser scanner (lidar) with the highest accuracy and longest range of all the systems discussed in this article. The system was designed for applications in construction and architecture and costs quite a bit more than the other discussed equipment.
Leica’s BLK360 software does not produce 3D tours, but the data can be processed through the Matterport Cloud, resulting in a 3D tour with the standard Matterport features.
The BLK360 can take up to three minutes to collect each 360° image. However, due to the much longer range of the 3D scanner, not as many capture positions are required and the overall capture speed is generally comparable to that of the Matterport camera.
The iGUIDE IMS-5 Camera System by Planitar features a DSLR body with a fisheye lens that is mounted in a way to avoid parallax errors when the camera is rotated into three positions to produce a 360° image.
The camera includes a time-of-flight 2D laser scanner with high accuracy and long-range. The camera’s fisheye lens is calibrated for measurements and the combined 360° imagery and 2D laser point cloud data allow for not only constructing accurate floor plans but also making measurements in 3D space without having a 3D mesh.
The camera data is processed through the iGUIDE platform to produce a 3D tour with floor plans and property measurements.
Due to the long-range of the lidar, the iGUIDE camera needs fewer camera positions to collect the necessary data and its onsite capture speed is about the same as with 360° cameras. A 3000 sqft home can be captured in 20 minutes.
There is a broad range of technologies that can be used for creating 3D tours all with varying degrees of accuracy, ease of use, and cost. We looked at the most popular ones in the North American marketplace and discussed their most important features.
When choosing a 3D tour technology, each particular use case may have different requirements and constraints. In the most common scenarios, speed of capture and availability of floor plan navigation would seem to be the most desirable features for a tour producer with all else being equal. In addition, according to both the National Association of Realtors® (NAR) and Zillow studies, over 80% of home buyers consider floor plans important or very important.
Understanding the public expectation that presented floor plans are a result of precise measurement and with property measurement standards requirements becoming more common in the industry, the measurement accuracy of a 3D tour technology becomes just as important as capture speed.
Alex Likholyot is co-founder and CEO of Planitar, the company behind the iGUIDE technology. Alex holds a PhD degree in physics and spent many years developing optical measurement systems in industrial and medical metrology fields before applying that expertise to the proptech industry.
REAL Trends has been The Trusted Source of news, analysis, and information on the residential brokerage industry since 1987. We are a privately-held publishing, consulting and communications company based in Castle Rock, Colorado.
Accessibility: We are making efforts to be ADA Compliant. Should you have any challenges or questions please contact us at (303) 741-1000.