How Ultra-Intelligent Point Clouds Could Make Modeling Obsolete


Paul Tice, 3D-scanning consultant and CEO of ToPa 3D, posted a LinkedIn article last month about current technologies for processing point cloud data. Notably, he takes a little time to cover Euclideon, and that’s where it gets interesting: Tice calls the company’s much-vaunted Unlimited Detail technology “such a revolutionary concept” that he believes the following statement to be true:

“…it should be possible to assign metadata to each point with said data hosted perhaps as [IP addresses] in the IoT. And, if that’s possible, then we at ToPa 3D believe that one day, component and mesh modeling will be antiquated because intelligent models of the future will be composed of pure point cloud data.”

It reads like a provocation and a statement of vision. So I caught up with Tice via email to unpack this bomb, and get an idea for what the future holds for the humble point cloud.

What He Means by That

First off, Tice says the basic idea is that clever programming should enable us to “do a lot more with point clouds than ever before.”

He points out that Euclideon’s method performs a kind of “pre-rendering” of point clouds. This shows that programming can save users the trouble of rendering the point cloud as they view it. It also shows that we can program “particles in a point cloud to perform in specific ways.”

Tice’s big idea follows from this: Clever programming should enable us to “attach metadata to each individual point” in the cloud.

The Internet Protocol defines the system of unique addresses that we use to locate devices on the internet (better known as IP addresses). Tice tells me that it has recently upgraded to its sixth version, which has expanded the number of possible addresses by a quantum leap, generating enough unique device identifiers for every grain of sand—roughly 670 quadrillion addresses per square millimeter of the Earth’s surface.

tice-portland-state-university-point-cloudTice says this means we could easily assign each point with its own address, allowing us to treat it as a device on the internet of things, and a unique Internet location for information-storage purposes.

This raises a lot of big possibilities. “What if each and every point knew where it belonged in 3D space?” he asks. “What if one point was assigned to a wall, another to a doorknob? If each point cloud could have this sort of information, maybe we could manipulate a point cloud just like a vector model.”

In other words, Tice argues that programming can turn the much maligned “dumb” point cloud into a replacement for the “smart” models that we use to index and access our data.

Point Data vs. Vector Data

But, why would we want to attach metadata directly to the point cloud?

We usually store metadata (like BIM information) in meshes or components we define using auto-extraction software. These are otherwise known as vector models. The problem with vector models, according to Tice, is that we have no efficient way of rendering them. “A simple example is a detailed Revit model,” he says. “These are difficult to navigate in 3D space.”

Point clouds are much faster to render, especially if we “bake” them like Euclideon does. This holds true despite the point clouds having “exponentially more visual detail.” If we use the point cloud as a means of storing our metadata, Tice argues, that would be a lot easier to navigate and much more efficient to manage.


It would also allow for more precise storage of information. “If we have point clouds that are precisely classified with unique identifiers, we would have greater precision with the metadata as it relates to the acquired geometry. With component models, for example, users tag polygons or components with metadata—but there is a limitation on how many polygons users can easily navigate at one time with current BIM platforms.”

Tice’s vision for the point cloud sees the point cloud itself becoming a smart model. This would mean that the processes we use to make models from point clouds “no longer relevant.” To use the words from Tice’s original statement, “intelligent models of the future will be composed of pure point cloud data.”

Scan and Deliver

To understand the value this would offer, imagine a construction project where you have installed sensors to capture a full, real-time scan of the whole site. If each point in a scan holds a wealth of relevant metadata, that scan is already a smart model. If you capture multiple scans, you’re essentially creating an archive where each entry is a smart model for that point in time.

tice-st-johns-bridge_perspective-1Why Aren’t We Doing it Yet?

The biggest challenge is a fairly obvious one, and a tall order: We need to figure out how to assign the metadata to millions of points simultaneously.

Tice suggests the answer is digging deeper into the data that the point cloud already offers, and finding creative ways to exploit it with programming. “I think that, with anything, it’s always good to start small,” he explains. “Working on better classification methods with a goal of making this an autonomous process is a start for testing out this concept for big data retrieval. We will need intelligent classification and grouping.”

Though there are details left to be worked out, and future challenges still unknown, the need is clear to Tice. “We know that polygons are computationally cumbersome with large numbers,” he says. “As our data sets become large for applications such as 3D city modeling, we may run up against severe limitations in the navigation and rapid access to the data we need.”

In the world of the large-scale smart model, the “dumb” point cloud might be our only hope—we just have to get to work and make it smarter.

Be sure to see Paul Tice presenting at SPAR 3D 2017, in Houston Texas from April 3-5. We have an exciting program covering all tools for 3D capture, processing, and delivery. Register now!


About Author

SPAR 3D Editor Sean Higgins produces SPAR 3D's weekly newsletters for 3D-scanning professionals, and Sean has previously worked as a technical writer, a researcher, a freelance technology writer, and an editor for various arts publications. He has degrees from Hampshire College in Amherst, Massachusetts and the University of Aberdeen in Scotland, where he studied the history of sound-recording technologies. Sean is a native of Maine and lives in Portland.


  1. Nice article ! Always good to see that our industry moves in a shared vision. Well Smart Point Clouds are coming sooner than expected, this is already on the shelf within my research ! Be sure to check a Smart Point Cloud published introduction here :
    It covers the base principles of what this article envision.
    The launch toward industrial applications is what is left to do now, the trick being efficient data managment to directly reason on information extraction rather than interpolated meshes, models, …
    Also, being a representative of Euclideon in France, pre-processed data is already queryable per-point in real-time in the latest version (which is a silent breakthrought ^^). Very efficient indexing method.

  2. 3D modeling with point clouds has already been accomplished. The points contain not only attributes that can be queried but coordinate data as well. As long as you have one common or known point and assign its location, the rest of the points can simply associate to it and fall in place where the belong.

    A team I worked with previously used a quadcopter with a geospatial capable camera to take pictures of a beach. Then we used freeware to convert the geospatial imagery to a massive point cloud where it was then placed into AutoCAD Civil 3D. In our case, the points already had the geospatial data assigned to them and we simply set the coordinates for our drawing in Civil 3D.

Leave A Reply

© Diversified Communications. All rights reserved.