Most approaches to capturing 3D models of real-world objects involve multiple cameras that are rarely cheap, and are sometimes tricky to calibrate. The University of Glasgow has developed a method that ditches those cameras altogether. Its system has four single-pixel sensors stitching together a 3D image based on the reflected intensity of light patterns cast by a projector. Reducing the pixel count lowers the cost per sensor to just a few dollars, and extends the sensitivity as far as terahertz wavelengths. Real-world products are still a long way off, but the university sees its invention as useful for cancer detection and other noble pursuits. Us? We'd probably just waste it on creating uncanny facsimiles of ourselves.