Efficient data acquisition and reconstruction for air-coupled ultrasonic robotic NDE

Non-destructive evaluation of complex parts using surface scanning techniques, such as ultrasonic testing and eddy current testing, requires complex manipulation of such sensors to ensure quantitative results. A robotic arm may function as a complex manipulator for surface scanning, controlling the position and tilt between the probe and specimen’s surface. To ensure accuracy in probe manipulation, accurate geometric information of the specimen is required. This article explores a methodology that uses structured light for physical-to-virtual reconstruction, providing submillimeter scale and accurate surface geometries. Reconstruction aids in path planning through a novel ray-triangle intersection array algorithm, establishing movements for the NDE probe to orient itself on the specimen at a constant probe to specimen surface distance, or lift-off. The proposed technique is demonstrated and validated through experimental air-coupled ultrasonic inspection of automotive CFRP composite samples with simulated flaws such as interlaminar delamination. The proposed method employs guided waves and a pitch-catch configuration of air-coupled ultrasonic probes, enabling single-side access scans. A Fanuc 100ib robot arm was used to manipulate the ultrasonic probes along a sample reconstructed with a CR-Scan 01 structured light sensor. The probes were excited at 200khz from a SonoAir system, while also recovering defect vs background information synchronized with the probe’s orientation. Additionally, a framework for potential automation is proposed, with further details to be explored in future works.

To overcome the limitations associated with predefined CAD models, we propose a framework employing surface reconstruction for NDE inspection with robotic arms.This approach offers insights into the scanning environment during the component's examination, eliminating previously mentioned disparities.The reconstructed environment serves a dual purpose, allowing scrutiny of background components for advanced NDE path planning and collision avoidance.The basic outline of the framework is shown in Fig. 1.Reconstruction will obtain a surface geometric profile.Registration aligns this geometric information to the robot's workspace.To create the movements for the probe, toolpath generation employs a novel approach, using a ray-triangle intersection array algorithm to parse movements on the surface geometry.The physical movements are then generated and sent to the robot, which will synchronize the probe's information with its orientation.

Surface reconstruction for autonomous robotic NDE
Surface reconstruction may be used to autonomously obtain the surface details of an "inspection space, " including the geometry of the specimen which the robot may reach, and background components to avoid collision with.This "inspection space" can be considered as a 3D virtual environment, which are meshes or CADs containing important components for a robot to consider.Specifically, surface information of the specimen may be used to parse toolpaths to orient a probe, including its position and tilt.For reconstructed meshes, a device using methods such as structured light or blue light can be used to obtain a sub-millimeter accurate virtual model of physical objects.Meshes obtained this way require alignment to the robot's base frame, else the frame will be aligned towards the devices local frame.
Ensuring higher accuracy in reconstruction and alignment is crucial for precise sensor positioning in NDE scanning and sensory data collection, particularly in techniques like air-coupled UT or eddy current scanning with low lift-off distances requirement for better energy coupling.The accuracy between virtual and physical geometries is beneficial for both complex and non-complex components, such as flat coupon samples.Inconsistent or large lift-off issues can lead to misleading or poor quantitative information in the output scan.Lift-off issues may be countered by advanced techniques such as with deep learning 11 or transfer learning 12 for different NDE surface scanning techniques.There is still, however, risk that no valuable information is picked up from poor positioning.Robotic systems, prone to errors, especially in stability, may require thorough examination of the impact of these errors on probe accuracy.Errors may occur due to imperfections in mastering or mechanical issues, including errors in the probe's movement trajectory.

State-of-the-art autonomous scans and surface reconstruction
Solutions addressing the inverse kinematic problem, achieving 6-DOF transformations for sensors on robot arm platforms, have been developed 13 .Dual ultrasonic testing (UT) robotic arm systems have demonstrated applications in scanning cylindrical composite objects using water-jet transmission 14 , as well as quality control of helicopter components with carbon-fiber reinforced polymer (CFRP) laminates 15,16 .Challenges arise in managing the complexity of helicopter components in such approaches.Synchronization of twin robot systems 17 and the development of inspection and repair robot crawlers for power plant boiler inspections 6 showcase progress in mobile platform systems.An in-line inspection system for cylindrical pipes, integrating endoscopic laser profiling and machine learning, addresses deformations within pipes 9 .Various imaging and reconstruction solutions include hyperspectral cameras for CFRP panel imaging using robot arms 18 , x-ray computed tomography for reference-free 3D scans to localize damages 19,20 , and the use of Structure from Motion techniques in robot arm www.nature.com/scientificreports/environments 21 .Time-of-flight cameras find applications in environment reconstruction for both mobile robot platforms and robot arms 22 .Stereo cameras, explored for NDE and welding applications 23,24 , offer portable and fast computation, particularly when compared to methods like structured light.A crawler robot for NDE was designed for plate inspection using stereovision 25 .Thermography, suitable for inspecting complex parts such as wind turbines 26 , has been examined in conjunction with robotics 27 .An advanced state-of-the-art method for UT robotic inspection is to utilize UT probes as an "in-process" path planner and NDE inspection device.One work has been done operating a UT controlled robot, which the UT probe dictates robot path planning via real-time processing.This system can determine the trajectory of the UT probe for surface mapping from a single probe start position alone, while also obtaining volumetric evaluation of a curved steel sample 28 .

Proposed framework for robotic NDE 4.0 surface reconstruction
A new framework for conducting autonomous scanning on complex parts from surface scanning NDE techniques is proposed.The first process of this framework starts with the initial environment reconstruction.This is followed by localizing the robot by registration of the environment to the scanning workspace.The next step is toolpath generation, in which a novel method is proposed: ray-triangle intersection arrays.
This process involves projecting rays onto the initial mesh and checking for intersections, resulting in a zig-zag pattern of "waypoints" holding position and tilt information for the probe on the toolpath.This provides robust scan profiles for any input mesh, albeit with smoothing limitations with respect to tool movements compared to b-splines.Figure 1 illustrates the considered framework depicting the reconstruction-to-NDE scanning process.Following path computation, the generated path is transmitted to the robot, initiating the NDE scanning process.The NDE scan synchronizes the recorded position of the robot's end effector with the NDE data, generating NDE point data within 3D space respective to the robot's workspace.This resultant data undergoes post-processing, involving actions such as removing trend effects or converting the 3D NDE point data into a mesh.

Geometric point cloud from physical environment
The goal of this section is to define an output to a cyber environment E cy based on a physical environment that will be used for path planning.Starting with the basics, the raw output from a reconstruction device is a point cloud labeled as a set of points P.This process is known as point cloud reconstruction.Eventually, P needs conversion into a mesh, forming the reconstructed environment E cy .A point cloud P contains a local set of points with indices p i , holding vertex information t xyz and color information rgb.Therefore, p i = [t xyz i , rgb i ] , and P = [p 1 , p 2 , p 3 , . . ., p LP ] , where LP is the number or "length" of points in the point cloud.Point clouds are obtained by an "observer, " which in this case would be the reconstruction device of choice.
From point clouds obtained from stereo cameras, the points are typically organized in a grid-like fashion along local coordinates x and y, with variations along z to represent depth.However, the output is unraveled into a 1D vector form.For example, if a stereo camera has a resolution of 1920×1080 points, then LP = 2073600 in the 1D vector P cy .Each point falls within width w and height h boundaries local to the point cloud's frame, creating a grid-like formation of points.Parameters w and h depend on the camera's field of view and the distance between a physical component inside the workspace.An example is shown in Fig. 2. In this example, a 4 × 4 grid-like point-cloud is shown, with dimensions w × h × d for width, height, and depth.From the point cloud definition, P cy = [p 1 , p 2 , p 3 . . ., p 16 ] with LP = 16 and the orange points are closer to the observer than the blue points.The clear points are just placeholders to show orthogonality between the depths, as in, the arrays will hold the same x and y positions with depth.From a reconstruction device, the depth of these points should move along the

Mesh generation from point cloud
From the point cloud generated, a mesh is generated by connecting vertices together.This process is known as mesh reconstruction.This will obtain a list of faces that will define the surface profile of the physical environment within virtual space.Covered are the basic principles of faces generated from reconstruction and its relation to path planning, with a breakdown shown in Fig. 4. Two popular algorithms are Poisson reconstruction 29,30 and Delaunay triangulation 31 .A mesh comprises a set of faces F with face indices fi, where and LF is the number of faces.Faces are constructed from vertices, with where LP F is the number of vertices in a face with LP F ≥ 3 and LP F ≥ LF + 2 assuming LF ≥ 1 .Vertices function as the foundation of meshes and are defined from P through reconstruction.An example of faces defined by vertices is shown in Fig. 3.The vertex list V differs subtly from P since reconstruction may reorganize V. Mesh processing may further reorganize V.Although the organization of faces through vertex indexing is abstracted, it is mentioned for contextual purposes.Output data sets typically use this organization to significantly reduce the redundancy of overstated vertices.
Faces are essential in the ray-triangle intersection array algorithm later discussed.Faces contain a flat area between its vertices.A point of intersection will lay on this area, determining probe positioning.Faces also contain a single normal throughout their area, determining probe tilt.A normal from a triangular face, or LP F = 3 , may be calculated from the following equation: Faces may be triangulated using face culling algorithms 32 .
The order of vertices defined by a face is important for normal orientation.The position of vertices per face determines facial orientation, including the normal or the direction the face is pointing towards.For example, f = [v a , v b , v c ] in a counterclockwise formation will face the normal in a positive orientation, while f = [v a , v c , v b ] in a clockwise formation will flip the normal in the negative direction.In other words, an incor- rect vertex order will flip the probe in the opposite direction, inevitably leading to a collision between the probe and sample!

Scanning environment post-processing
Post-processing on the mesh is recommended for effective path planning.Cropping out background components is recommended to prevent manipulation in areas outside the scanning region.This step may also aid in eliminating curved beveling on edges of a sample that merges with background components, a common occurrence during reconstruction.Additionally, smoothing the scan surface is recommended to avoid issues with coarse  1.
Table 1.Properties of different face formations from a set of vertices, shown in Fig. 3. Table (a) shows the configuration for each set of faces F, while table (b) shows the formation of each point per face, which direction of points is relevant for normal calculations per face on the mesh.
Vol.:(0123456789) www.nature.com/scientificreports/surfaces that might lead to overcompensation in the rotation of the actuation system.While this article does not delve into the optimization and automation of post-processing procedures, techniques such as manual face cropping, face simplification or decimation, and Laplacian smoothing 33 are commonly employed to enhance path planning and scanning results.

Background removal
In a workspace, two classifications of physical objects are considered within a point cloud snapshot: the sample to reconstruct and the background.The toolpath generation algorithm aims to retain only the sample under test, excluding other components.It is crucial to identify what qualifies as background in the workspace.This may include holders like vices, which may appear attached to the sample and need removal.If the sample is on a table, the table itself is considered part of the background.Components outside the robot's workspace might also appear in the snapshot as background, including the robot and any related wires to the NDE probe.It is essential to note that background components may be important for collision detection.Thus, it is recommended to keep one mesh with the necessary background for collision detection and another for path planning.
To eliminate backgrounds, two methods were employed: manual removal and statistical outlier removal.Statistical outlier removal was a default technique applied to stray point clouds to prevent them from affecting the mesh reconstruction process.Manual cropping was also employed, typically if a background component, such as vices or tables, merged with the sample.For future works, an autonomous solution for background removal will be considered.

Path Planning: Ray-triangle intersection array algorithm
The path planning process is expected to produce a linear path of waypoints, W, derived from the processed reconstructed environment mesh, E cy .Each waypoint w within W includes translation and rotation properties.W serves as input for the inverse kinematics algorithm, generating the joint set as the toolpath where the NDE probe actuates.
A novel approach was employed to derive W from E cy , using ray-triangle intersection arrays.This neces- sitated the development of a customized 3D engine capable of calculating ray-to-triangle intersections along an input mesh.E cy requires processing to eliminate obstructive environmental components and smooth the surface.
A ray is simply two vertices forming a line in space.Ray-triangle intersection, often implemented through the Möller Trumbore algorithm 34 , allows a ray to intersect with a triangular face.This intersection will provide the location of the waypoint, and the normal of the intersected face for the rotation of the waypoint.Euler rotations can be derived from n f .To ensure the correct direction, the normal direction is flipped to align the probe with the surface.Otherwise, the probe will have a flipped orientation inside the sample which will inevitably cause collision and improper scanning.If a ray does not intersect with the mesh, then it is ignored.Figure 5 shows a zig-zag patern on a low-polygon sphere with intersections and approximate rotations per face.
The objective is to create a grid of rays, each checking the mesh for position and rotation properties of a waypoint.This grid may be envisioned as a formation of rays in a rectangular prism, oriented on the spot to examine on the mesh.The volume of the prism is defined by the scan zone area and the ray length.The ray length should be defined finitely, though may be technically infinite or a high number without affecting the algorithm.The number of rays per row and column represents the resolution of the toolpath.For simple raster scans, rows indicate the number of "swipes" the NDE sensor will move along, and columns represent the number of movements Figure 4. Breakdown of process of mesh reconstruction and properties.The process involves obtaining a point cloud P cy from a reconstruction device, then obtaining surface information through mesh reconstruction.The results are placed for further processing.per swipe.Larger numbers of columns enhance fidelity over curved surfaces, though in practice, the robot may "stutter" at each of these points, increasing scan time due to deacceleration and reacceleration.
Once the waypoint path, W, is determined, the joints to actuate the NDE probe must be resolved through inverse kinematics.This process is executed by inputting a finite transformation of the probe, and the joint solution is output.This can be done through simulation, then submitted to the physical robot.As inverse kinematics presents an infinite number of joint configurations, there may be a resolution for an optimal set of joint movements to optimize scan times.This can be done by reducing the movement time of the joints.Inverse kinematics needs to also avoid singularities or impossible movements.Inverse kinematics is solved with RoboDK, a commercial software tremendously helpful for robot arm integration.
It is advantageous to organize points to optimize scan times, often using zig-zag raster scanning.For example, because rays may not intersect, a simple raster algorithm may increase scan time or obtain unwanted data.For the implementation of this algorithm, including zig-zag and line-by-line instances and optimization using a greedy traveling salesman problem (TSP) approach, refer to 35 .TSP approaches create more advanced paths to increase scan speeds at the cost of computation time.More advanced algorithms may also incorporate collision avoidance with background and sample, if the sample contains geometries that may cause self-collision.

Results and validation of the proposed framework for complex surface reconstruction
Experiments were conducted that demonstrated reconstruction relevant for later path planning.The first shown are the capabilities of reconstruction devices for path planning, and the second are show the capabilities of NDE scanning.Stereovision and Structured light methods were used to obtain a cyber environment from the sample under test.Initially, an Intel RealSense D425i camera used for stereovision was considered.This camera has a depth accuracy of 2% or 40 mm at 2 m lift-off, and output field-of-vision (FOV) of 87 • × 58 • , and an ideal range for acquisition between 0.3 to 3 m.However, as shown later, it had poorer performance than the Creality CR-Scan 01 structured light sensor.This sensor has a depth accuracy at 0.1mm, a reconstruction area 53.6 × 37.8cm 2 for FOV, and a required lift-off range of 1 m.
Five samples were selected with various geometries, shown in Fig. 6 and general dimensions shown in Table 2. Car samples S1 through S3 have variances in thickness, due to regions having honey-comb structures and other regions only containing the CFRP surface.These samples have hills and valleys that need to be reconstructed as well.

Mesh processing
The same post processing process was used for each sample: crop out the background, decimate mesh to 1000 faces, and Laplacian smoothing with 10 iterations.Alignment procedures are not covered in this article, though it should be noted that point-pairs picking from CloudCompare was used for registration of meshes back to the robot's base frame.For environment post processing, CloudCompare for statistical outlier filtering and Poisson

Results
The results for each car piece for samples S1-S3 are shown in Figs. 7, 8, and 9, for both stereovision and structured light.Each colormap is set to depth (white is closer).The first experiment was to use stereovision to reconstruct    the three car pieces.Each piece was placed on a vice so that the piece would be orthogonal to the ground while also minimizing any background point cloud data.Otherwise, the background data would merge with the sample, increasing the difficulty of post-processing.The camera was positioned 450mm away from each sample.Each was a direct mesh from the Intel Realsense SDK.There are holes that are seen in the snapshots for sample S2 and S3.Specifically, on sample S2, there is a large gap on the left side where the ramp is.The general depths of each are shown, however the resolution is rather poor.Because of this, structured light was considered for further testing.
Stereovision may be considered for other applications and its benefits for wider range and portability would be used for NDE 4.0 pathing for unmanned vehicles or drones.Structured light was used to reconstruct all car samples, the aluminum sample, and the x-brace sample.The position of the camera was dynamic, as the Creality software allows for fast stitching of multiple snapshots while the camera is moving above each sample.This significantly helps increase the size of the area to apply path planning on.It also requires tracking, which has been found difficult to do through the air for certain samples.Because of this, the car samples were placed stationary on a table.Figures 7, 8, and 9 show the results after processing for the first three samples.A glass platform was used to differentiate the sample from the table background.Each sample is shown with higher resolution and depth clarity than the stereo camera.For samples S2 and S3, areas of the mesh are missing, which can be helped by stitching more snapshots together at different viewing perspectives of reconstruction device.

NDE 4.0 surface scanning on reconstructed specimens using ACUT
Once validation was done for using structured light as the main reconstruction method, experiments for UT scanning were next conducted.Anomalies were found using the first order reflection of the air-coupled UT (ACUT) transducers set in pitch-catch mode, with one transmitter sending ultrasonic guided waves and one receiver placed at an angle between each other.Figure 11 illustrates the setup for air-coupled inspection of CFRP test samples using Sonoscan CF200 probes from Sonotec.A SonoAir system was used to excite the probes with a 4-cycle square pulse at 200kHz bursts of four at 200V and 70dB preamp gain.A band pass signal was placed onto the time signal between 180 and 220 kHz.To prevent direct wave propagation through the air between the probes, scattering foam is employed.Beneath the foam, waves penetrate the test sample, emerging on the opposite side where they are detected by the receiver.The angle of wave incidence is carefully chosen to activate the fundamental A0 mode within the specimen.The received signal is then amplified using a preamplifier positioned on the robot's end effector.To streamline the process, measured ultrasonic (UT) signals are automatically time windowed.The energy of the received signal serves as a feature for flaw detection.Figure 10 shows the ACUT sensors and laser calibration tool.A Banner LM150KUQP depth laser was attached to the end effector of the robot to obtain calibration information which ties known physical spots to virtual space.This is done through point-pairs picking in CloudCompare.A Fanuc ARCMate 100iB robot arm capable of 6-DOF movements was used to manipulate the ACUT probes and laser calibration tools.
The reconstructed environments were obtained by the Creality CR-Scan 01 structured light sensor.After mesh processing, like the car pieces from samples S1-S3, the meshes were placed into the ray-triangle path generator.The generated toolpath on the mesh is input into RoboDK to create a set of instructions for simulated and physical robots to manipulate the ACUT probe.This simulation sets the ACUT probes lift off to 5 cm.The instructions are then run on the physical robot, in which the robot's tool orientation is synchronized with the UT signal within the scan duration (Fig. 12).
Samples S4 and S5 shown in Fig. 6 and details in Fig. 2 were used.A flat aluminum block as sample S4 was used as a base reference scan for a simple geometry.This sample includes milled holes and surface calibration stickers for detection.The ACUT surface detection scan and its process are shown in Fig. 13.For the aluminum sample, the nine calibration stickers are clearly visible on this sample.Each sticker was 0.25 in 2 in area and approximately 0.07 mm in thickness, however because of the small area size and convolution, the stickers are rounded into a circular formation.The milled holes were cropped from these results, but they are visible on the top and right corners.There also is a slight gradient indicating that the bottom left of the sample is slightly closer to the robot's base than the top right, which is a concern for alignment.Despite this, the surface features are clearly picked up.
Sample S5, which will be referred to as an x-brace, contains a complex geometry and is more difficult to detect defects.This sample is 7.5 mm in thickness and contains three subsurface defects with various intensities, with the largest on the bottom left and smallest on the top right.As these scans were taken at separate times, the x-brace was reconstructed twice.To help with this new reconstruction, tape was placed to be used as a reference for tracking using structured light.Two different orientations were evaluated with similar scanning procedures, with the only major differences being the sample's physical location.
Post processing involves processing each 3D point cloud using known information about the raster pattern of the scan.The process is shown in Fig. 14 and outputs at Fig. 15.From even the raw data, each defect is visible, albeit the smallest defect found top left is difficult to see.First, signals recorded in the air are removed.This is a straightforward process, as signals in the air will return a lower reflected signal than in the sample.Values under 0.02V are removed from the scan.The next step is to apply detrending to alleviate probe orientation errors.This is done throw median subtraction per scan line.As each line is a set of 3D vertices unorganized in space, these needed to be calculated using the initial scan parameters to parse and index each line.Once each scan line is found, simply the median of that line is subtracted.The results for both scans are shown in Fig. 15.Note that after median filtering, then the voltage will be centered around zero rather, hence why further voltage values are considered as arbitrary.
After processing, the defects may be seen around ±0.03 for the large defect 1, ±0.01 for the middle defect 2, and ±0.005 for the smaller defect 3, with the background being close to 0 through the median subtraction detrend, all values in arbitrary voltage units.The edges of the sample also appear dark as well.This is due to the structure of the x-brace having curved edges that may affect the reflected signal, despite attempting to scan along these regions at the proper orientation.The defects also are parallel with the probe's orientation.For example, if the probe's local z-axis were rotated 90 • , the defects would also rotate 90 • .

Conclusions and future work
Through the experimental results, it is shown that using structured light environmental reconstruction along with point-pairs picking on calibrated regions for registration is an effective technique to parse surface and subsurface defects with air-coupled UT probes.Importantly, the path planning through the ray-triangle intersection algorithm enables toolpath generation by parsing a position through intersection on the mesh, and rotation through the face from intersection.In turn this also enables autonomy for scanning with only basic parameters such as area, resolution, and location of the inspection area needed to be known.This assumes the processes in the framework, which are environment reconstruction and alignment, are also automated.Reconstruction may be automated by either placing the reconstruction device on the robot arm for mesh reconstruction scanning or placing the device statically beside the area of interest.Registration may be automated by various means, such as April tags or iterative closest point algorithms 36,37 .
Figure 12.Setup diagram.The robot is controlled by operations sent from the robot controller and PC.The robot holds a tool containing the UT probes and a laser for mesh registration purposes, both connected to the PC for synchronization with the robot's tool frame.The UT probes will also be excited from a SonoAir system.The sensors on the robot tool, along with the reconstruction device, interact with the scanning environment.As seen in Figs. 13 and 15, the areas of interest are seen versus the sample's background, either as surface stickers or subsurface defects.Such areas of interest may be found by the robotic system mostly independent of orientation and be localized with respect to the robot for further analysis.The importance of the robot's awareness of each defect may be for secondary operations after the scan, more-so than just giving acknowledgment for a human operator.For example, if the robot were to have an operation to repair by composite scarfing, then this gives insight on where the operations should take place 38 .Examples exist for this repair on robotic arm systems 39 and mobile systems 40 .For these secondary operations to be applicable, more information such as the precise depth and the amount of scarfing required would be up to a future endeavor.
There may be more room for different environment sensing research.For the examined stereovision results, it is shown less effective reconstruction than structured light results, which provides significant issues in determining the option for path planning.These issues do not disqualify the practicality of stereovision for more portable systems, such as unmanned vehicles or drones, and these results are down to the sensors used in the experiments.As mentioned previously, there has been research done with stereovision with robotics [23][24][25] .
The system itself may incorporate more autonomous procedures or otherwise improve robotic accuracy.For example, autonomous movement of the structured light sensor during environment reconstruction may be implemented, by simply conducting a raster-scan above the surface.There have been two issues with this procedure.One is the difficulty parsing an orientation proper to the robot end-effector, which seems to change with an error up to 2cm.If using autonomous registration methods such as center-point transformation, this error is not acceptable.Therefore point-pairs picking was used, which used human interaction.This human interaction contradicts the autonomous nature of the setup.The second issue is tracking to expand the workspace of the digital reconstruction by parsing similar point clouds together.This is a built-in feature with Creality

Figure 1 .
Figure 1.Framework and system layout.(a) Proposed 6-DOF NDE 4.0 framework, showing reconstruction, alignment, generation, scanning, and post processing.(b) System layout showing the robot, sensors, and flow of data.

Figure 2 .
Figure 2. Point cloud example from different perspectives, with a 4 × 4 grid-like point cloud with area w × h two sheets of points at different depths defined by d.(a) Bird's eye perspective.(b) Isometric perspective.

Figure 3 .
Figure 3. Example showing face formations from vertices, oriented to the reader with LP = 4 .The properties of each formation is shown in Table1.

Figure 5 .
Figure 5. Path planning example using a custom 3D engine, showing the grid of arrays intersecting with a low polygon sphere.Each waypoint is seen, organized in a zig-zag pattern starting from red to blue, where position and rotation is demonstrated.Per each face from respectively combined triangles, the rotation remains constant, but changes for every other face to match the normal project of associated triangles.

Figure 6 .
Figure 6.Samples for testing and their respective labels.

Figure 7 .
Figure 7. Sample S1 snapshots between the two reconstruction methods.Colormap is set to depth.The following figures will use similar configurations.As seen for the next figures, the stereovision results have a spuratic formation, while the structured light results are more detailed to the physical object.(a) Stereovision.(b) Structured light.

Figure 10 .
Figure 10.Robot and ACUT probe in physical environment.(a) ACUT probe and calibration tool as endeffectors.The tool is held and controlled by the robot arm with its connections shown in Fig. 12.(b) ACUT probe conducting a scan on sample S5 (CFRP x-brace).To help reconstruction, foam was placed behind the sample and tape surrounded the area of inspection desired..

Figure 11 .
Figure11.ACUT transducer setup, showing a transmitter and receiver against a sample with an embedded defect.The defect is found through the reflections within the sample, excited from the transmitter which are picked up from the receiver in the air.Foam is placed in between to prevent transmission directly to the receiver.

Figure 13 .
Figure 13.Aluminum sample reconstruction to ACUT results.(a) Point cloud from structured light.(b) Toolpath generation after mesh processing, with an area of 100 mm × 100 mm and resolution of 100 × 2 waypoints.(c) Simulated movements in RDK with a global lift-off of 50 mm along the z-axis.(d) ACUT surface results, showing surface profile including the calibration stickers.

Figure 14 .
Figure 14.UT subsurface scan air removal, with the air removed results shown next in Fig. 15 for scan 1.The same method was used for scan 2. (a) Raw ACUT point cloud containing data within ai, which is notified as black as there is no reflection returned.(b) Histogram of raw data used for filtering, showing a large amount of air signal for lower voltages, as well as the peak of voltages indicating the sample background.

Figure 15 .
Figure 15.Post-processed ACUT results for both orientations.All three defects are seen in both images, regardless of orientation.(a) Scan 1.(b) Scan 1.

Table 2 .
Properties of each sample to reconstruct.