The detector's multi-bounce reflection data is used to reconstruct the shape of objects that are visible from the position of the wall, but not from the position of the camera. The duration of each laser burst is so short that the system can gauge how far the bursts have traveled by measuring the time it takes them to reach the detector. The Conar system compares the different times at which the returning light arrives at different parts of the detector, uses multi-path analysis to crunch the data, and constructs an image of the room’s geometry. (You can watch a video lecture discussing the technology in more detail here.)
Conar's technology has initially produced recognizable 3D images of a wooden figurine and images of foam cutouts that are placed outside the femtosecond laser camera's line of sight.
Eventually, the researchers expect the technology could lead to imaging systems that help emergency responders evaluate dangerous or unknown environments, or collision avoidance systems for vehicles and vehicle navigation systems that can negotiate blind turns. Other possible applications include robot navigation systems for industrial environments and instruments that can investigate conditions in cramped spaces inside machinery with moving parts.
Yes, this reminded me of the Lytro camera as well. The Lytro camera allows setting of the range of focus with the picture data, through software, after the picture is taken. This concept allows reconstruction of laser topology reflected back into the scene. Also reminds me of the laser-based window listening devices, what will they think of next?
bob from maine, the article gives links to two videos worth watching, one short and one more detailed, on how this technology works. Military surveillance and security are definitely some apps this could be used in, and probably lots more we haven't thought of yet that this will make possible.
New technology solves old problem with better resolution! Seeing beyond the electromagnetic horizon. Reminds me of over the horizon Radar popular during the Cold War era. Different wavelengths of electromagnetic spectrum but somewhat similar idea. Reflect off of the ionisphere and listen for back scatter to provide a target echo.
Do I understand correctly? The laser emits a light beam which scatters, then the camera 'averages' the returned photons to memorize what is there. When something new is added, the returning photons take a different time of flight and thus the shape and position of the 'new' thing is derived and displayed. Depending on the acquisition speed of the camera/computer, this would have a great usefullness in any number of internal/external building security, military field surveylance, etc.. A really neat concept.
Laser scanning is used everywhere to define the real-world and convert it to 3D point cloud data that can be used in plant design, architectural remodeling, accident reconstruction, crime scene reconstruction, etc.
The way the technology works today, you have to set up the tripod in a few different places with targets that can be matched by the software to get a complete picture of the area. It works pretty good, but consider one setup, one scan and you're done. Amazing.
These new 3D-printing technologies and printers include some that are truly boundary-breaking: a sophisticated new sub-$10,000, 10-plus materials bioprinter, the first industrial-strength silicone 3D-printing service, and a clever twist on 3D printing and thermoforming for making high-quality realistic models.
Using simulation to guide the drafting process can speed up the design and production of 3D-printed nanostructures, reduce errors, and even make it possible to scale up the structures. Oak Ridge National Laboratory has developed a model that does this.
Focus on Fundamentals consists of 45-minute on-line classes that cover a host of technologies.
You learn without leaving the comfort of your desk. All classes are taught by subject-matter experts and all are archived.
So if you can't attend live, attend at your convenience.