That's a valid question. There's more than one way to create 3D in machine vision. Chuck's upcoming February feature, already out in the print edition of DN, discusses this subject. The simplest, easiest, cheapest method is by using two 2D cameras in stereo, as does this QuantumVision product. This roughly emulates the stereoscopic vision of humans (and other primates), in that both of our eyes used together creates 3D stereo images. Others use more complex math and/or special image sensors, and/or special image processing.
Basically, this is smaller than other stereo 3D cameras, and it's way smaller than other 2-camera 3D stereo cameras. Since it's a stereo 3D system, it's created with two 2D cameras, so there's really no new paradigm in that sense; you are still processing 2D data. You can process that data faster if you use the cameras independently. Another thing about this system is its rugged enclosure, which is why it's shown with water drops.
Ann, what's the use case for this type of system compared with a traditional 2D vision system or any of the stereo 3D systems? Am I saving money by going this route or is it purely a matter of increasing visibility without having to move to a totally new 3D paradigm?
Determining the quantities and location of sensors in an Internet of Things application requires a thorough problem statement and a clear vision of success, an expert will tell engineers at the upcoming Design & Manufacturing Show in Minneapolis.
Focus on Fundamentals consists of 45-minute on-line classes that cover a host of technologies.
You learn without leaving the comfort of your desk. All classes are taught by subject-matter experts and all are archived.
So if you can't attend live, attend at your convenience.