RIEGL proudly announces its new iPad point cloud viewer: RiALITY, now available for free in the iTunes App Store.
This new, innovative App, the first of its kind, allows users to experience LiDAR data in a completely new environment. It also allows easier LiDAR data demonstrations through the use of an iPad.
RIEGL’s RiALITY App enables users to visualize and navigate through point clouds acquired with RIEGL laser scanners. As an example, users are able to explore a dataset of the beautiful Rosenburg Castle in Austria. RIEGL scans can also be imported from RIEGL’s RiSCAN PRO software into the App, as well.
“We’re pleased to present a new way of visualizing point clouds. RiALITY delivers this new technology by providing Augmented Reality technology in an easy-to-use app. Now you can easily send your client a 3D point cloud that they can visualize on their iPad, for free.” said Ananda Fowler, RIEGL’s manager of terrestrial laser scanning software.
RiALITY features true color point clouds and 3D navigation. In a breakthrough technological development, the app features an Augmented Reality Mode. The Augmented Reality Mode allows point clouds to be virtually projected into the real world.
https://scanable.com/wp-content/uploads/2014/02/RIEGL_Software_RiALITY_Screen-5d.jpg15362048Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2014-02-27 15:26:552014-02-27 15:26:55Introducing the World’s First App for LiDAR data visualization on the iPad: RiALITY
Film director and cinematographer Patryk Kizny – along with his talented team at LookyCreative – put together the 2010 short film “The Chapel” using motion controlled HDR time-lapse to achieve an interesting, hyper-real aesthetic. Enthusiastically received when released online, the three-minute piece pays tribute to a beautifully decaying church in a small Polish village built in the late 1700s. Though widely lauded, “The Chapel” felt incomplete to Kizny, so in fall of 2011, he began production on “Rebirth” to refine and add dimension to his initial story.
Exploring the same church, “Rebirth” comprises three separate scenes created using different visual techniques. Contemplative, philosophical narration and a custom orchestral soundtrack composed by Kizny’s collaborator, Mateusz Zdziebko, help guide the flow and overall aspirational tone of the film, which runs approximately 12 minutes. The first scene features a point cloud representation of the chapel with various pieces and cross-sections of the building appearing, changing and shifting to the music. Based on LIDAR scans taken of the chapel for this project, Kizny generated the point clouds with Thinkbox Software’s volumetric particle renderer, Krakatoa, in Autodesk 3ds Max.
“About a year after I shot ”The Chapel,” I returned to the location and happened to get involved in heritage preservation efforts,” Kizny explained. “At the time, laser scanning was used for things like archiving, set modeling and support for integrating VFX in post production, but I hadn’t seen any films visualizing point clouds themselves, so that’s what I decided to do.”
EKG Baukultur an Austrian/German company that specializes in digital heritage documentation and laser scanning, scanned the entire building in about a day from 25 different scanning positions. The collected data was then registered and processed – creating a dataset of about 500 million points. Roughly half of the collected data was used to create the visualizations.
Data processing was done in multiple stages using various software packages. Initially, the EKG Baukultur team registered the separate scans together in a common coordinates space using FARO Scene software. Using .PTS format, the data was then re-imported into Alice Labs Studio Clouds (acquired by Autodesk in 2011) for clean up. Kizny manually removed any tripods with cameras, people, checkerboards and balls that had been used to reference scans. Then, the data was processed in Geomagic Studio to reduce noise, fill holes and uniformly downsample selected areas of the dataset. Later, the data was exported back to the .PTS ASCII format with the help of MeshLab and processed using custom Python scripting so that it could be ingested using the Krakatoa importer. Lacking a visual effects background, Kizny initially tested a number of tools to find the best way to visualize point cloud data in a cinematic way with varying and largely disappointing results. Six months of extensive R&D led Kizny to Krakatoa, a tool that was astonishingly fast and a fraction of the price of similar software specifically designed for CAD/CAM applications.
“I had a very basic understanding of 3ds Max, and the Krakatoa environment was new to me. Once I began to figure out Krakatoa, it all clicked and the software proved amazing throughout each step of the process,” he said.
Even with mixing the depth of field and motion blur functions in Krakatoa, Kizny was able to keep his render time to roughly five to ten minutes per frame, even while rendering 200 million points in 2K, by using smaller apertures and camera passes from a higher distance.
“Krakatoa is an amazing manipulation toolkit for processing point cloud data, not only for what I’m doing here but also for recoloring, increasing density, projecting textures and relighting point clouds. I have tried virtually all major point cloud processing software, but Krakatoa saved my life on this project,” Kizny noted.
In addition to using Krakatoa to visualize all the CG components of “Rebirth” as well as render point clouds, Kizny also employed the software for advanced color manipulation. With two subsets of data – a master with good color representation and a target that lacked color information – Kizny used a Magma flow modifier and a comprehensive set of nodes to cast and spatially interpolate the color data from the master subset onto the target subset so that they blended seamlessly in the final dataset. Magma modifiers were also used for the color correction of the entire dataset prior to rendering, which allowed Kizny greater flexibility compared to trying to color correct the rendering itself. Using Krakatoa with Magma modifiers also provided Kizny with a comprehensive set of built-in nodes and scripting access.
The second scene of “Rebirth” is a time-lapse reminiscent of “The Chapel,” while the final scene shows live action footage of a dancer. Footage for each scene was captured using Canon DSLR cameras, a RED ONE camera and DitoGear motion control equipment. Between the second and third scene, a short transition visualizes the church collapsing, which was created using 3ds Max Particle Flow with help of Thinkbox Ember, a field manipulation toolkit, and Thinkbox Stoke, a particle reflow tool.
“In the transition, I’m trying to collapse a 200 million-point data cloud into smoke, then create the silhouette of a dancer as a light point from the ashes,” shared Kizny. “Even though it’s a short scene, I’m making use of a lot of technology. It’s not only rendering this point cloud data set again; it’s also collapsing it. I’m using the software in an atypical way, and Thinkbox has been incredibly helpful in troubleshooting the workflow so I could establish a solid pipeline.”
Collapsing the church proved to be a challenge for Kizny. Traditionally, when creating digital explosions, VFX artists are blowing up a solid, rigid object. Not only did Kizny need to collapse a point cloud – a daunting task in of itself – but he also had to do so in the hyper-realistic aesthetic he’d established, and in a way that would be both ethereal and physically believable. Using 3ds Max Particle Flow as a simulation environment, Kizny was able to generate a comprehensive vector field of high resolution that was more efficient and precisely controlled with Ember. Ember was also used to animate two angels appearing from the dust and smoke along with the dancer silhouette. The initial dataset of each of angels was pushed through a specific vector noise field that produced a smoke-like dissolve and then reversed thanks to retiming features in Krakatoa, Ember and Stoke, which was also used to add density.
“To create the smoke on the floor, I decided to go all the way with Thinkbox tools,” Kizny said. “All the smoke you see was created using Ember vector fields and simulated with Stoke. It was good and damn fast.”
Another obstacle was figuring out how to animate the dancer in the point clouds. Six cameras recorded a live performer with markerless motion capture tracking done using iPi Motion Capture Studio package. The data obtained from the dancer was then ported onto a virtual, rigged model in 3ds Max and used to emit particles for a Particle Flow simulation. Ember vector fields were used for all the smoke-like circulations and then everything was integrated and rendered using Thinkbox’s Deadline, a render management system, and Krakatoa – almost 900 frames and 3 TB of data caches only for particles. Deadline was also used to distribute high volume renders and allocate resources across Kizny’s render farm.
Though an innovative display of digitally artistry, “Rebirth” is also a preservation tool. Interest generated from “The Chapel” and continued with “Rebirth” has enticed a Polish foundation to begin restoration efforts on the run-down building. Additionally, the LIDAR scans of the chapel will be donated to CyArk, a non-profit dedicated to the digital preservation of cultural heritage sites, and made widely available online.
The film is currently securing funding to complete postproduction. Support the campaign and learn more about the project at the IndieGoGo campaign homepage at http://bit.ly/support-rebirth. For updates on the film’s progress, visit http://rebirth-film.com/.
About Thinkbox Software Thinkbox Software provides creative solutions for visual artists in entertainment, engineering and design. Developer of high-volume particle renderer Krakatoa and render farm management software Deadline, the team of Thinkbox Software solves difficult production problems with intuitive, well-designed solutions and remarkable support. We create tools that help artists manage their jobs and empower them to create worlds and imagine new realities. Thinkbox was founded in 2010 by Chris Bond, founder of Frantic Films. http://www.thinkboxsoftware.com
https://scanable.com/wp-content/uploads/2014/01/LiDAR-for-Visual-Effects-Rebirth.jpg378756Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2014-01-25 05:09:482014-01-25 05:09:48Krakatoa Creates CG Visual Effects from LIDAR Scans for Short Film “Rebirth”
The Only Portable Mobile Mapping & GIS Data Collection Solution is now available for daily, weekly and monthly rental.
Mobile Surveying Systems or Mobile Mapping Systems, as they are becoming known, are not new. Several laser manufacturers have introduced and sold Mobile Mapping Systems over the last few years. The difference is that these systems (with one exception) are adaptations of tripod mounted (static) laser scanners. They are all large, very heavy, complicated and very, very expensive.
Measurement Devices Ltd. announces the world-wide availability of Dynascan equipment hire/rental giving even the smallest survey organizations the ability to add mobile mapping to its service offering, enabling them to undertake previously undreamed of ‘detailed 3D, large scale mapping projects’, formerly the domain of large companies or government organizations. Productivity of Laser Surveying Systems will no longer be measured in terms of “Points per Second” but in terms of “Square Miles per Hour” and available for the same cost-effective daily hire rate as static terrestrial 3D laser scanning systems (i.e. Leica Geosystems ScanStation C10, HDS6200, HDS7000).
Furthermore, MDL has assembled a global team of LiDAR industry experts to provide field and office support services and custom-tailored training on “real-world” workflows and procedures. Travis Reinke, Chief Operating Officer of Measurement Devices US, LLC stated, “By providing easy access to the latest technology and expertise, we are able to change the decision-making process. Project managers and the like now have the ability to integrate new technologies into their workflows without the huge risk of taking on large capital expenditures. We’re disrupting the market by introducing mobile mapping to the masses.”
MDL’s new business model of providing ‘complete measurement solutions’ gives greater support to customers with irregular or infrequent resource needs or project specific requirements. Reinke went on to add, “Our highly-skilled team of professionals are available to ensure that first-time users are successful on their first project. We accomplish this by offering a discounted bundled rate for an experienced technician and the mobile system when needed.”
MDL’s mission is:
To assist and enable more end-users in the real-world application of various rapid 3D data capture technology by providing easy access to a comprehensive range of equipment rental, consultancy and education.
To help end-users close the information gap between the Hardware/Software Vendors and the real world.
To empower and educate our client base in order to enhance industry growth and spawn further advancements.
To support customers with irregular or infrequent resource needs or project specific requirements.
With decades of combined experience gained from a real world application of laser measurement systems, MDL’s team of industry professionals has obtained a highly qualified and proven skill set focused on Quarry and Mining Surveying, Dynamic Land Surveying, High Definition Surveying, and Hydrographic Surveying.
Eliminate the Investment Risk
Rent equipment until you’re ready to buy – expense on per project basis.
Take advantage of knowledgeable experts with real world experience to support your project team.
Ramp up or down as needed to keep your capital expenditures minimal.
Rely upon our experience to provide you with practical user information and get more than just a demo.
The Dynascan ‘Plug and Play’ LIDAR system is a fully integrated high speed laser scanner, high accuracy GPS positioning system and Inertial Measurement Unit. The system is light weight, highly portable and may be used on land vehicles or marine vessels to acquire 3D survey quality data of topography, urban developments, industrial plants, including overhead utility cables, bridges, dams, harbors, beaches, rivers and canal banks, and much more.
By fully integrating all the sensors in one package, MDL has eliminated the need for field calibrations as all sensor offset measurements are fixed during the factory calibration process and pre-configured in the data acquisition software. The LIDAR system is shipped ready for operation and can be mobilized in a matter of minutes.
Dynascan system comes complete with data acquisition and post processing software suite which is compatible with all known brand multi-beam echo sounders, swathe sonar systems, positioning systems and Inertial Measurement Units. Data from all available sensors is synchronized, “Time Tagged” and Recorded. Post processed and raw data may be exported to most 3D data base and CAD software packages. Dynascan is highly affordable and represents an unrivalled ‘Price-Performance’ advantage, opening up the benefits of 3D LIDAR to numerous applications and market opportunities.
https://scanable.com/wp-content/uploads/2012/06/198361_194154690711110_211518139_n.jpg356960Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2012-07-13 21:09:562015-05-22 19:48:27Mobile Scanners Enter the Rental Market
Brian Ussery is reporting that Google is back in Atlanta, GA making Street View images for Google Maps but, this time they brought in the big guns. Beu Blog reported on April 28, 2010, “The cars here today are equipped with GPS, high resolution panoramic cameras and multiple SICK sensors. These sensors collect LiDAR data that can be used for 3D imaging and visualizations like that seen in Radiohead’s recent “House of Cards” music video. Google Earth and SketchUp, Google’s 3D virtual building maker for Maps also use this type of data.
Last week Google announced the release of a plugin which allows users access to Google Earth imagery via Maps. As a result it’s now possible to view 3d images in Google Maps. The problem here is fairly obvious, Google Earth’s aerial imagery is taken from above and as a result not from the same perspective as users interacting with the data. Not to worry though, the StreetView team has been working on these kinds of problems for sometime. When it comes to Navigation, Maps or StreetView, earthbound LiDAR enhanced imagery processed via Sketchup seems like a perfect complement to Google’s existing view from above. Combining high resolution imagery taken from the user’s perspective with advanced 3D image technology, presents some new possibilities to say the least. Factor in new releases like business ads in Maps, now being available in 3D on your mobile device and it’s pretty clear how Sketchup will be monetized.”
It is expected that Google’s incorporation of LiDAR into their mapping efforts will lead to some significant changes to our industry. If you have not previously seen the “House of Cards” video, be sure to check out the interactive music video code to see how Google made the point cloud data readily available for manipulation in a standard web browser. Point clouds are finally becoming more natively accepted in most CAD platforms and with Google getting involved in the industry, who knows where we will be in the near future.
https://scanable.com/wp-content/uploads/2010/05/GoogleEarthLiDAR.jpg234400Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2010-05-01 18:00:222010-05-01 18:00:22Google Maps Street View 2.0 [LiDAR]
From Tiltan’s website: TLiD is Tiltan’s innovative solution for fast, automated creation of 3D maps and GIS information from LiDAR point clouds.
TLiD Main Features:
– Automatic extraction of DTM (bare earth) and DSM
– Automatic features extraction (houses, trees, power lines)
– Automatic full scene 3D reconstruction
– LAS or free ASCII txt input
– LAS, SHP, DTM and other output file formats
– Multiple input/output coordinate systems
– Integrated with a 3D Viewer
TLiD Advantages:
– Fast parallel processing for cost reduction
– No limitation on input file size
– Standalone product
– Special Applications
– Trees counting – height and size
– Power line mapping and clearance
– Line of sight
– Other applications – available on request
https://scanable.com/wp-content/uploads/2010/04/Tiltan-e1271777040959.jpg165716Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2010-04-20 08:24:232010-04-20 08:24:23Tiltan TLiD Transform LiDAR Point Clouds to 3D Models in One Keystroke
April 12, 2010 – Houston – Increasing its cadre of laser mapping researchers, the University of Houston will expand its pioneering work in such areas as homeland security, disaster recovery, oil and gas exploration, wind farm site planning and environmental studies.
The NSF National Center for Airborne Laser Mapping (NCALM) and the groundbreaking researcher leading it recently moved operations to the University of Houston. Based upon historical information, revenues generated by the center’s operation are anticipated to be $1 million per year and will be reinvested in the program.
NCALM is UH’s first and only NSF-supported center, established and sustained by funding from the National Science Foundation. This differs from the way the university typically sets up centers, using university funds or grants from multiple sources for multiple projects. These types of centers support NSF’s focus on interdisciplinary research, spanning several institutions and departments.
Ramesh Shrestha, Hugh Roy and Lillie Cranz Cullen Distinguished Professor of Civil and Environmental Engineering, brought NCALM to UH from the University of Florida. He has been director of the center, focused on ground-based scanning laser technology and airborne laser swath mapping research, since it was established in 2003. Shrestha brought much of his Florida team with him to Houston, where they now operate NCALM jointly with the University of California-Berkley.
“With the center, we have brought laser mapping’s uses to the forefront and expect to continue to have this impact in our new Houston home,” Shrestha said. “We plan to establish curriculum catered to this specialty and eventually add a graduate degree in geosensing systems engineering. This is in addition to carrying out research far surpassing what is capable in laser mapping to date.”
Shrestha’s work with laser mapping goes back to the 1990s, when this once niche research area was just making its debut. Bill Carter, now a research professor at UH, worked with him early on and helped establish NCALM.
“Together, we saw its potential to far exceed what was possible with many traditional methods, such as airborne photogrammetric mapping that uses cameras to detail terrain,” Carter said. “Laser mapping has the ability to work day or night, as well as generally map areas even though they were covered by forests and other vegetation where photogrammetric methods couldn’t.”
It wasn’t long before other scientists would see its same benefits, especially as the two developed techniques to remove and minimize some of the errors seen in the early years. Their equipment became fine-tuned to collect even more data, now mapping as many as 167,000 points per second compared to the 3,000 they were able to achieve when they first started.
Their work has changed the way the state of Florida monitors erosion on its coastline, produced the highest resolution 3-D images in existence of the San Andreas Fault and taken them across the globe to map Mayan Ruins in Belize and volcanoes in Hawaii. While the impact of their work is already far reaching, their plan for the coming years indicates they are not close to completion. The value of this work is evident in evaluating the before and after of hurricanes and earthquakes in terms of improving building design and other mitigation efforts, as well as offering predictive tools for subsequent powerful events.
Aided by NSF, future NCALM efforts explore the possibility of using Light Detection and Ranging (LiDAR) to map everything from glacial movements to the migration of penguin colonies in Antarctica. Using LiDAR, researchers take measurements of the ground’s surface from their Cessna 337 Skymaster airplane.
From roughly 2,000 feet, this remote technology measures properties of scattered light through the use of laser pulses. Thousands of small cone-shaped pulses travel through a hole in the bottom of the plane to the ground below, and a unique detector picks up rays reflected from the ground. Then, each point’s distance is determined by measuring the time delay between the transmission of a pulse and the detection of reflected signals. The plane’s location and movement in the air are tracked by an inertial measurement unit fixed inside the laser system with a GPS receiver mounted to the plane and others on the ground. Both are used, along with the laser data, to produce detailed 3-D topographical images of the terrain.
“In coming years, our group plans to develop a next-generation LiDAR system. The unit would be less expensive than commercially available systems and allow for some of the most accurate, highest-resolution observations possible in laser mapping,” Shretha said. “We want to develop a system like no one else has developed. It would really change what could be done with this technology. It would have new features, be faster, smaller and capture more during each flight than we can today.”
According to Shrestha, this system would use a much shorter pulse-length laser, increasing the number of points that could be mapped per second to 800,000. This would add to data accuracy and reduce the amount of time needed in the air to collect the information. Additionally, it would be able, for the first time, to penetrate shallow water depths.
###
NOTE TO JOURNALISTS: High-resolution photos of Ramesh Shretha and the Cessna 337 Skymaster airplane are available to media by contacting Lisa Merkl.
About the University of Houston
The University of Houston, Texas’ premier metropolitan research and teaching institution, is home to more than 40 research centers and institutes and sponsors more than 300 partnerships with corporate, civic and governmental entities. UH, the most diverse research university in the country, stands at the forefront of education, research and service with more than 37,000 students.
About the Cullen College of Engineering
The Cullen College of Engineering at UH has played a vitally important role in educating engineers in Texas. Taught by innovative faculty, eight of whom are in the National Academy of Engineering, the college offers degree programs in biomedical, chemical, civil, computer, electrical, environmental, industrial, mechanical and petroleum engineering, as well as specialty programs in materials, aerospace, and computer and systems engineering.
For additional news alerts about UH, follow us on Facebook and Twitter.
00Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2010-04-14 12:14:212010-04-14 12:14:21National Center for Airborne Laser Mapping Comes to Houston [LiDAR]
HOUSTON, TX March 18, 2010 Coign Asset Metrics & Technologies (CoignAMT), at the direction of Zoo Film Productions of Hollywood, CA, has helped produce IBM’s first globally released television commercial created entirely from Light Detection and Ranging (LiDAR) 3D point cloud data.
CoignAMT used the HDS6100 phase-based laser scanner from Leica Geosystems to create scenes of cars on a freeway, patients in a hospital, electricity grids and much more. The 30-second LiDAR-based commercial is a key part of IBM’s Smarter Planet initiative to portray that data is all around; and that by changing the way the world thinks, companies can maximize the use of data to lower their costs and reduce environmental impact.
Travis Reinke, business sector manager for CoignAMT, says, “Coincidentally, IBM’s perspective is a core part of CoignAMT’s business practice. We help our clients see the long-term value of the data they currently have by using the latest technology, such as 3D laser scanning, to quickly gather an immense amount of data to support their existing “intelligent” systems.”
The Zoo Film Productions crew spent a week in Houston, TX with CoignAMT personnel capturing laser scan data of transmission lines and over 16 blocks of downtown Houston streetscape and surounding buildings. CoignAMT then merged the point cloud data gathered in downtown with 6 square miles of low-altitude helicopter-based LiDAR provided by Aerotec LLC out of Birmingham, AL. Zoo Film Productions crews also created numerous scenes of hospital activities as well as automobiles with and without drivers that CoignAMT scanned individually for use in the commercial.
Reinke continues, “We were honored to be part of this innovative project given the direct correlation between the services CoignAMT provides and IBM’s Smarter Planet initiatives. Using the latest laser scanning technology to visually portray the importance of the data surrounding us was an unforeseen irony. I would never have imagined that we would be using this technology to scan people and cars, objects that are often considered ‘noise’ on a typical inventory project.”
About Coign Asset Metrics & Technologies LLC:
Coign Asset Metrics & Technologies, LLC (CoignAMT) is a HUBZone qualified, SBA certified, small business that provides a full range of asset management life cycle services and specialized technologies, including high definition laser scanning (HDLS). Its mission is to strengthen public and private sector organizations by aligning their assets and operational practices with their strategic initiatives. CoignAMT is headquartered in the Pittsburgh, PA area, with regional offices located in Colorado Springs, CO, and Houston, TX. Clients include federal, state, and local governments, as well as private sector customers in the construction, energy, transportation, manufacturing, and security industries.
https://scanable.com/wp-content/uploads/2010/03/IBM_Data_Anthem_02_web1-e1269575231374.jpg539958Travis Reinkehttps://scanable.com/wp-content/uploads/2025/01/SCANable_logo_emblemSimple-180x180.pngTravis Reinke2010-03-18 10:58:402010-03-18 10:58:40IBM Visualizes a “Smarter Planet” using 3D Laser Scanning Technology [LiDAR]
By combining the best of two different distance measurement approaches with a super-accurate technology called an optical frequency comb, researchers at the National Institute of Standards and Technology (NIST) have built a laser ranging system that can pinpoint multiple objects with nanometer precision over distances up to 100 kilometers. The novel LIDAR (“light detection and ranging”) system could have applications from precision manufacturing lines on Earth to maintaining networks of satellites in perfect formation, creating a giant space-based platform to search for new planets.
LIDAR transmits light through the air and analyzes the weak reflected signal to measure the distance, or range, to the target. NIST’s new LIDAR, described in Nature Photonics,* has a unique combination of capabilities, including precision, rapid updates from multiple reference points at the same time, and minimal “measurement ambiguity.” The system can update measurements to multiple targets simultaneously every 200 microseconds. Measurement ambiguity in a LIDAR system is due to the fact that, if the target is at long range from the instrument, the system can’t distinguish between two different distances that are multiples of its “ambiguity range.” The new NIST LIDAR has a comfortably large ambiguity range of at least 1.5 meters—large enough to check the coarse distance with widely available technologies such as GPS.
No other ranging system offers this combination of features, according to the new paper. NIST’s LIDAR could enable multiple satellites to maintain tight spacing and pointing while flying in precision formations, acting as a single research instrument in space, the paper states. Formation flying has been proposed as a means to enhance searches for extraterrestrial planets, enable imaging of black holes with multiple X-ray telescopes on different satellites, and support tests of general relativity through measurements of satellite spacing in a gravitational field. The new LIDAR could enable continuous comparisons and feedback of distances to multiple reference points on multiple satellites. There also may be applications in automated manufacturing, where many parts need to fit together with tight tolerances, according to Nate Newbury, the principal investigator.
NIST’s LIDAR design derives its power from combining the best of two different approaches to absolute distance measurements: the time-of-flight method, which offers a large ambiguity range, and interferometry, which is ultraprecise. The LIDAR relies on a pair of optical frequency combs, tools for precisely measuring different colors (or frequencies) of light. The frequency combs used in the LIDAR are based on ultrafast-pulsed fiber lasers, which are potentially smaller and more portable than typical combs that generate laser light from crystals. The two combs operate at slightly different numbers of pulses per second. Pulses from one comb are reflected from a moving target and a stationary reference plane. The second comb serves as precise timer to measure the delay between the reflections returning from the target and from the reference plane. A computer calculates the distance between the target and the reference plane by multiplying the time delay by the speed of light.
* I. Coddington, W.C. Swann, L. Nenadovic and N.R. Newbury. Rapid, precise absolute distance measurements at long range. Nature Photonics. Published online May 24, 2009.