Author Notes
Event cameras differ from conventional frame cameras: Rather than capturing images at a fixed rate, they asynchronously measure pixelwise brightness changes and stream an event sequence that includes time, location, and sign of the brightness changes. Event cameras provide attractive properties such as high temporal resolution and latency (both in the order of ms), very high dynamic range (120 dB), and low power consumptions because no routine operation needs to be recorded. Hence, event cameras have been successfully used, for instance, for object tracking, pose estimation, or image deblurring. Common in-process sensors for process monitoring in laser welding are photodiodes or high-speed frame cameras. Due to the described advantages of event cameras compared to these sensors, this paper for the first time investigates the potential of event cameras in laser welding. We present different promising ways to process the unconventional asynchronous output of event cameras for process monitoring, including time-series representations, as well as image representations. Our laser welding experiments show clear differences for different production scenarios in those event representations. Additionally, we propose a machine learning pipeline for the automatic detection of production errors with event camera.
I. INTRODUCTION
In production, laser welding processes—if not executed properly—are posing a risk of component failure. Consequently, to ensure the reliability of the components, stringent monitoring of the weld quality is needed. One promising option is the in-process monitoring of the welding keyhole, molten pool, or spatters. Monitoring sensors are acoustic sensors,1 photodiodes,2,3 x-ray sensors,4 charge-coupled device (CCD) cameras,5 and high-speed cameras.6 The analysis of the sensor signals includes approaches based on machine learning, such as support vector machines,7 decision trees (DTs),8 random forests,9 or convolutional neural networks (CNNs).10–12
Event cameras (ECs), which are inspired by biological mechanisms, measure pixelwise changes in brightness, providing advantages such as high temporal resolution, high dynamic range, low power consumption because no routine operation needs to be recorded, and low latency13 compared to standard cameras. These properties have led to promising applications in areas such as vibration measurement,14 drowsiness detection systems,15 counting of fast moving objects,16 or image deblurring.17 Depending on the further evaluation algorithm of an application, the asynchronous event data are processed and often transformed into alternative representation:13 For instance, individual events can be directly handled by event-by-event methods like spiking neural networks.18,19 However, those networks require special hardware and, therefore, are not considered further. Moreover, events can be represented as event packets, whereby a predefined amount of events is collectively processed.20 Additionally, events can be accumulated into images that can be easily processed using computer vision algorithms.21
As the EC facilitates a more comprehensive understanding of exceptionally swift processes, the advantages of the EC compared to standard cameras for process monitoring in laser welding might be interesting. The adaptability of the EC in different domains emphasizes the potential to provide insights into the dynamic processes of laser welding processes. To the best of our knowledge, no prior publications have investigated the potential of EC for quality monitoring of laser welding processes. Therefore, the present paper includes the following contributions:
•Transformation of the asynchronous events of the EC data to different promising event presentations for in-process quality monitoring of laser welding defects.
•Proposal of different sampling strategies of the event camera data for quality monitoring of laser welding processes.
•Analysis and interpretation of the introduced event representations for different in production occurring laser welding defects.
•Usage of machine learning algorithms including decision trees and convolution neural networks to distinguish the introduced defects.
II. EVENT CAMERAS
The explained EC encoding is inspired by the spiking nature of biological visual pathways.22 Compared to standard cameras, the EC offers numerous potential advantages: ECs have an extremely high temporal resolution and low latency13 (both in the order of microseconds), which make these sensors react quickly to visual stimuli. Moreover, ECs have a very high dynamic range23 (120 versus 60 dB of standard cameras) and low power consumption.24 Therefore, ECs are currently used for feature detection and tracking like corners,25 fastmoving objects,16 optical flow estimation,26 3D reconstruction,27 pose estimation,28 visual-inertial odometry,29 image reconstruction,20 motion segmentation,30 neuromorphic control,31 or image deblurring.17 However, the spatial resolution of the EC is generally lower than that of standard cameras.
After having seen the advantages and applications of EC in diverse fields, we now look at how the EC can be used for the quality monitoring of laser welding processes.
III. EXPERIMENTAL SETUP
Based on an industrial welding process, two metal plates were welded together. Therefore, in production occurring scenarios were provoked. The process is observed with the EC, and the aim is to recognize differences in its data. Figure 1 shows the experimental setup. Therefore, a laser beam was directed onto two thin metal plates by a 2D galvanometer scanner. The two metal plates have a dimension of 80 × 80 mm2 and a thickness of 75 μm. As a laser, an IPG YLR-2000-WC fiber laser with an infrared wavelength of 1070 nm at a power of 300 W (max. power 2000 W) and a speed of 0.5 m/s was used. A scanner, a Raylase AS Fiber 30, focused the beam to a diameter of about 50 μm on the work piece surface. An event camera (EC) observed the laser welding process coaxially and in-process. As the EC, a Prophesee Metavision Evaluation Kit 4HD (EVK4) was used. The EC data sequence then is transferred and processed on a computer.
Experimental setup of the laser welding process of two thin metal plates. An event camera (EC) was used to observe the process in situ. The EC data sequence is transferred and processed on a computer.
Experimental setup of the laser welding process of two thin metal plates. An event camera (EC) was used to observe the process in situ. The EC data sequence is transferred and processed on a computer.
Anticipating scenarios that may arise during production, experiments were conducted. Figure 2 schematically shows the front view of the metal plates of five scenarios. In the reference scenario (a), the metal plates were positioned on top of each other. In (b), a thin film of oil was applied between the metal plates, which led to a spattering process. In (c), only one plate was welded. In (d), there is a defocus effect of the laser beam, and in (e) a gap was inserted between the two plates.
Five different production scenarios have been considered: (a) reference, (b) oil, (c) one plate, (d) defocus, and (e) gap.
Five different production scenarios have been considered: (a) reference, (b) oil, (c) one plate, (d) defocus, and (e) gap.
Having seen the experimental setup, we look at the resulting EC data. Figure 3 visualizes a sequence of events E [see Eqs. (1) and (2)] captured during the observation of the laser welding process. The events are shown in the three-dimensional space with the -location and the time t as axis. The individual events are marked in blue. The position of the laser is given by the gray line. While at the beginning less events occurred, more events were captured afterwards as the process spattered.
Visualization of events (blue) during the observation of a laser welding process in the three-dimensional space, with the -location and the time t as the axis. The laser position is given by the gray line. The polarity p of the events is neglected.
Visualization of events (blue) during the observation of a laser welding process in the three-dimensional space, with the -location and the time t as the axis. The laser position is given by the gray line. The polarity p of the events is neglected.
IV. EVENT REPRESENTATIONS
One of the key questions of event cameras is how to extract meaningful information from the event data. Therefore, the events are transferred to alternative representations that make it easier to observe changes in the laser welding process. In the following, we present three different promising event representations to monitor laser welding processes, namely, event rates (Sec. IV A), projection on the x, -plane (Sec. IV B), and projection along scanner data (Sec. IV C). Afterward, the resulting representations can be further analyzed, e.g., by classification algorithms.
A. Event rate
An event rate transfers a data sequence E into a time series. More precisely, an event rate gives the number of events that occur in equidistant time intervals per length of the intervals. Mathematically speaking, this means that assume having observed a laser process during the time interval , which results in an EC data sequence E. Let us divide T in L equally long subintervals of length , so .
B. Projection on x, y-plane
1. Binary
So P has the entry if at least one event occurs at position and if there is no event occurrence.
2. Density
3. Temporal
C. Projection along scanner data
V. SAMPLING STRATEGIES
After having seen different event representations, the question arises as to when and how often the representations should be generated during a welding process for quality monitoring purposes. Therefore, in the following, we present two different sampling strategies, namely, constant time period (Sec. V A) and constant event rate (Sec. V B).
A. Constant time period
Assume having observed a laser process during the time interval T, which results in an EC data sequence E. Let us divide T in equally long subintervals of length . The constant time period sampling strategy then generates the event representations for each subinterval of length .
Figure 4(a) visualizes this sampling strategy. As in Fig. 3, the events (blue) of a sequence of events E are shown in the three-dimensional space with the -location and the time t as axis. When the time has passed, a binary projection P is generated. In the figure, this results in three binary images. A drawback of considering a constant time period as sampling strategy is that even if few or no events occur—meaning that no or few changes are in the laser process—a representation is created. This can be seen in the figure: In the first of the three intervals, there are almost no changes in the process, yet an almost completely black image is created. This drawback is overcome in the next sampling strategy.
Sampling strategies: A visualizes the constant time period sampling strategy. So, when the time has passed, an event representation is created resulting in three binary images. B visualizes the constant events sampling strategy. So, when an amount of events have occurred, an event representation is created also resulting in three binary images. It is and .
Sampling strategies: A visualizes the constant time period sampling strategy. So, when the time has passed, an event representation is created resulting in three binary images. B visualizes the constant events sampling strategy. So, when an amount of events have occurred, an event representation is created also resulting in three binary images. It is and .
B. Constant events
Assume having observed a laser process during the time interval T, which results in an EC data sequence E. Let us divide E in subsets with the same amount of events . So each entry of the event rate has the value . This constant event sampling strategy then generates the event representations for each subset with events.
Figure 4(b) visualizes this sampling strategy. When events have occurred, a binary projection P is generated. In the figure, three binary images are shown. Thereby, projections are created after differently long time intervals. Compared to Fig. 4(a), the first of the three projections is created later, as fewer events occur at the beginning. To determine the time of an anomaly during quality monitoring, the time after each has to be remembered.
VI. RESULTS AND DISCUSSION
After gaining understanding about different sampling strategies and event representations, we now look at which differences can be seen in the different event representations for our in production occurring scenarios according to Fig. 1 like gap or spatter (Sec. VI A). Subsequently, to further show the potential of EC data for process monitoring of laser processes, different configurations in which the EC parameters, the process parameters or the hardware was changed were realized (Sec. VI B). Afterward, we present a machine learning pipeline for quality monitoring including DTs and CNNs (Sec. VI C).
A. Event representations of different scenarios
The first column of Fig. 5 shows the event rate ε determined during the welding process of the introduced scenarios according to Fig. 2. The event rate with is shown over the time in ms. The graph line is shown in rainbow colors. The color changes constantly over time. In the first row, the event rate of reference shows minor changes over time. The next two rows show oily processes. Compared to the reference, the event rate shows high peaks due to the spatters. The event rate of one plate shows nearly no events, the event rate of defocus shows tiny rises and falls, and the event rate of the gap shows few small peaks.
Comparison of different event representations: The first column gives the event rate ε with over the time t in ms. The second column gives the projections along scanner data V, and the last column gives the time projections . For the event representations, the different scenarios (a)–(e) according to Fig. 2 are shown. So every row shows the different event representations of one scenario. Every process lasts 100 ms, and the rainbow coloring gives the time. So the coloring starts with red at 0 ms and changes to green at about 30 ms, to blue at about 60 ms, and finally back to red at the end of the process at 100 ms. Therefore, corresponding points between the event representations can be seen by matching colors.
Comparison of different event representations: The first column gives the event rate ε with over the time t in ms. The second column gives the projections along scanner data V, and the last column gives the time projections . For the event representations, the different scenarios (a)–(e) according to Fig. 2 are shown. So every row shows the different event representations of one scenario. Every process lasts 100 ms, and the rainbow coloring gives the time. So the coloring starts with red at 0 ms and changes to green at about 30 ms, to blue at about 60 ms, and finally back to red at the end of the process at 100 ms. Therefore, corresponding points between the event representations can be seen by matching colors.
Figure 6 shows the projections on the -plane determined during the welding process of the introduced scenarios. The first column of Fig. 6 shows the binary projection P, the second column the density projection Q, the third column the time projection , and the last column the time projection . The images include the whole image field of the EC of pixels. In the binary projection, differences in the images between the scenarios are visible. While in reference, a regular spatter pattern with most spatters being completely visible in the image, in oil, far more spatters are visible. In one plate, nearly no events occur. The fifth row results from defocusing smaller than 1.5 mm and the sixth row from defocusing greater than 1.5mm. Therefore, the fifth row shows many thin spatters, whereas the sixth only shows few spatters. In the last row, a gap was introduced on one side of the rectangular welding geometry (see Fig. 1). As seen in the event rate (see Fig. 5 first column and last row), few events occur in that scenario. On closer inspection, one can see that few spatters occur on the left-hand side of the image, which is due to the introduced gap. In addition to the information whether or not an event occurred at a certain position in the -plane, no further information like how many events or when the events occurred is available. This information is additionally shown in the next columns. The images of the density and temporal projections are shown in rainbow colors. So, for instance, for the density projection, the color map starts with red (no events), goes to yellow (medium number of events), and ends with green (most events). In the temporal projection, the rainbow colors change constantly over time as in Fig. 5. With this additional temporal information, it can be seen that the spatters in the one plate scenario occur at the beginning of the process when the laser beam penetrates. Moreover, the other events in the one plate scenario occur at four time intervals (pink, orange, green, light blue). Those events are due to reflections on a clamping device in the four corners of the welding geometry. At closer look, they are visible in all scenarios.
Projections on the -plane: The first column gives the binary projection P, the second column the density projection Q, the third column the time projection , and the last column the time projection . For those projections, the different categories (a)–(e) according to Fig. 2 are shown. Thereby, the fifth row results from defocusing smaller than 1.5 mm and the sixth row from defocusing greater than 1.5 mm. So every row shows the different projections on the -plane of a category. The projections are calculated over the events occurring during the whole laser process of the two metal plates (see Fig. 1). The projections of the second, third, and last column have a rainbow coloring. So, for instance, in the density projections red coloring means few events per position, yellow means more, and green means even more events per position. In the temporal projections, the rainbow colors change constantly over time.
Projections on the -plane: The first column gives the binary projection P, the second column the density projection Q, the third column the time projection , and the last column the time projection . For those projections, the different categories (a)–(e) according to Fig. 2 are shown. Thereby, the fifth row results from defocusing smaller than 1.5 mm and the sixth row from defocusing greater than 1.5 mm. So every row shows the different projections on the -plane of a category. The projections are calculated over the events occurring during the whole laser process of the two metal plates (see Fig. 1). The projections of the second, third, and last column have a rainbow coloring. So, for instance, in the density projections red coloring means few events per position, yellow means more, and green means even more events per position. In the temporal projections, the rainbow colors change constantly over time.
The second column of Fig. 5 shows the projections along scanner data V determined during the welding process of the introduced scenarios. The rainbow colors refer to the time, as in the event rate of Fig. 5. The horizontal, rainbow-colored lines give the position of the laser. While the event rate of A gives the amount of occurred events, B shows what happened spatially. For example, the event rate of reference shows small rises and falls. In the projection along the scanner data, it is visible that these events are caused by the regular occurrence of small spatters. Moreover, the high peaks in the event rate in the oil scenario are caused by strong ejections with several spatters that fly both forward and backward. So the projections along the scanner data V show the direction of the spatter relative to the global work piece position.
The third column of Fig. 5 shows the time projections determined during the welding process of the introduced scenarios. The rainbow colors give the time as before. Corresponding points among the representations can be seen. For example, the light green serpentine flying spatter in the reference scenario in the projection along the scanner data is also visible in the temporal projection. However, the spatter path appears longer in the -direction in the temporal projection. This is due to the fact that in the temporal projection, the events are shown relative to the laser position, whereas in the projection along the scanner data, the events are shown relative to the work piece. As the laser moves from left to right, the path of the backward flying light green spatter appears longer in the temporal projection than in the projection along the scanner data. If a spatter flies forward relative to the work piece, like the light blue spatter in the gap scenario in the projection along the scanner data, then its path appears shorter in the temporal projection. Moreover, it can be seen in the temporal projections of the oil scenario, that not all spatter paths are completely visible in the EC observation field. This means that some spatter path in the projection along the scanner data are cut off.
B. Event representations of different configurations
To further show the potential of EC data for process monitoring of laser processes, four different configurations I–IV (see Table I) have been realized. Thereby, configuration I describes the previously used configuration according to Sec. III. Compared to configuration I, in configuration II parameters of the EC have been changed: The threshold for triggering an event has been reduced. This means that events are triggered for smaller intensity changes. This is indicated in the table as the sensitivity . In configuration III, process parameters, namely, the weld speed ν and the weld power , have been increased. In configuration IV, experiments were carried out on a different system with a different EC and scanner as well as process parameters.
Configurations. As EC a Prophesee Metavision Evaluation Kit 4HD (EVK4) and a Prophesee Metavision evaluation kit 3HD (EVK3) were used. The sensitivity σ (Ref. 13) indicates how sensitive the EC parameters were set. As a laser, an IPG YLR-2000-WC (IPG 2000) and an IPG YLR-1000-WC-Y14 (IPG 1000) laser were used. As a scanner, a Raylase AS Fiber 30 (Raylase) and a Scanlab intelliWELD PR (Scanlab) were used. Moreover, the weld speed ν in m/s and the weld power η in W are given.
Configuration . | EC . | σ . | Laser . | Scanner . | v (m/s) . | η (W) . |
---|---|---|---|---|---|---|
I | EVK4 | 0 | IPG 2000 | Raylase | 0.5 | 300 |
II | EVK4 | 1 | IPG 2000 | Raylase | 0.5 | 300 |
III | EVK4 | 1 | IPG 2000 | Raylase | 3.0 | 900 |
III | EVK3 | 0 | IPG 1000 | Scanlab | 0.5 | 250 |
Configuration . | EC . | σ . | Laser . | Scanner . | v (m/s) . | η (W) . |
---|---|---|---|---|---|---|
I | EVK4 | 0 | IPG 2000 | Raylase | 0.5 | 300 |
II | EVK4 | 1 | IPG 2000 | Raylase | 0.5 | 300 |
III | EVK4 | 1 | IPG 2000 | Raylase | 3.0 | 900 |
III | EVK3 | 0 | IPG 1000 | Scanlab | 0.5 | 250 |
Figure 7 shows different event representations of the different scenarios for the four configurations. The first column shows the projections along the scanner data V of the reference scenario. A cut-out in which the welding path first is a curve and then continues in a straight line is shown. The other columns show the time projection of the different categories. The first line shows configuration I. Due to the more sensitive EC parameters in configuration II compared to configuration I in each image, more events are visible. In addition to the visibility of spatters in configuration II, more events are generated along the welding, which give information about the keyhole. In contrast, more events are generated in uninteresting areas as well: Irrelevant events due to reflections on the clamping device in the curve but also on the straight line occur. In configuration III, more events occur, especially around the curve. In configuration IV, similar patterns like in configuration I are visible. In summary, configurations I and IV prove to be particularly suitable because they give the spatter behavior of the process and do not generate any additional irrelevant events.
Event representations of different configurations. I–IV give the configuration according to Table I. The first column gives the projections along the scanner data V and the second column the temporal projections . Different scenarios according to Fig. 2 are given. Therefore, the second last column results from defocusing smaller than 1.5 mm and the last column from defocusing greater than 1.5 mm. The coloring between the projection along the scanner data and temporal are not correlated and are chosen due to illustration purposes: While in the projection along the scanner data V, the rainbow pattern is traversed twice, in the temporal projection the rainbow color is traversed once.
Event representations of different configurations. I–IV give the configuration according to Table I. The first column gives the projections along the scanner data V and the second column the temporal projections . Different scenarios according to Fig. 2 are given. Therefore, the second last column results from defocusing smaller than 1.5 mm and the last column from defocusing greater than 1.5 mm. The coloring between the projection along the scanner data and temporal are not correlated and are chosen due to illustration purposes: While in the projection along the scanner data V, the rainbow pattern is traversed twice, in the temporal projection the rainbow color is traversed once.
C. Event camera process monitoring with machine learning
Figure 8 shows our proposed machine learning pipeline. After performing the experiments according to Sec. III, the received events are transformed either to event rates or to projections on the -plane. A subset of transformed events is then used for the supervised training of a decision tree (DT) or a convolution neural network (CNN). DTs are well-suited due to their interpretable decision rules, while CNNs are explored due to their state-of-the-art performance in computer vision and pattern recognition tasks. Therefore, the label is given by the scenarios introduced in Fig. 2.
Machine learning pipeline. During welding, the received events are transformed either to event rates or projections on the -plane. The transformations are then used for the supervised training of a decision tree (DT) or a convolution neural network (CNN). Therefore, the label is given by the scenarios according in Fig. 2.
Machine learning pipeline. During welding, the received events are transformed either to event rates or projections on the -plane. The transformations are then used for the supervised training of a decision tree (DT) or a convolution neural network (CNN). Therefore, the label is given by the scenarios according in Fig. 2.
With T being the time interval in which a complete rectangular geometry according to Fig. 1 is welded, we select . This means that for the welding of two metal plates, the event rate ε contains one entry, namely, the number of events that occurred during T. The same applies for and . For and , visually a radius pixels was chosen. The training data instances of the DT, therefore, consist of three values, namely, , , and . Those training data instances are visualized by the dark boxes in Fig. 8. Instead of an event rate, for the CNN, one projection on the -plane is created per time interval T. Therefore, CNNs were trained for the binary projection P, the density projection Q, and the time projection as grayscale images.
Table I gives the number of metal plate pairs of the dataset. The dataset consists of 249 welded metal plate pairs. The number of metal plate pairs for training, validation, and testing is given by (60%), (20%), and (20%). As given in Table II, the data are imbalanced between the scenarios. To compensate this, DT used Synthetic Minority Oversampling Technique (SMOTE). For the CNN, a data augmentation with balancing factors was incorporated.
Number of metal plate pairs of the dataset.
Category . | ntrain . | nval . | ntest . | n . |
---|---|---|---|---|
a) Reference | 32 | 10 | 10 | 52 |
b) Oil | 41 | 14 | 14 | 69 |
c) One plate | 11 | 4 | 4 | 19 |
d) Defocus | 22 | 7 | 7 | 36 |
e) Gap | 43 | 15 | 15 | 73 |
Total | 149 | 50 | 50 | 249 |
Category . | ntrain . | nval . | ntest . | n . |
---|---|---|---|---|
a) Reference | 32 | 10 | 10 | 52 |
b) Oil | 41 | 14 | 14 | 69 |
c) One plate | 11 | 4 | 4 | 19 |
d) Defocus | 22 | 7 | 7 | 36 |
e) Gap | 43 | 15 | 15 | 73 |
Total | 149 | 50 | 50 | 249 |
The DTs were implemented with the library Scikit-learn.32 The library uses an optimized version of the CART algorithm. To measure the quality of a split, the Gini impurity was used as criterion. A grid search to find good hyperparameters, namely, the maximum depth of the tree, and the minimum number of chunks required to split an internal node was performed.
To teach the CNN invariances and robustness properties, a data augmentation including rotation, flipping, shifting, and shearing was used. The image size was resized from 720 × 1280 pixels to 100 × 177 pixels to accelerate the training process. The models were developed and trained using TensorFlow.33 A wide range of CNNs were tried: Initially, the networks were kept shallow extended later to deeper networks. The filter size was changed between 3 × 3 to 7 × 7. In addition to convolution layers, max pooling, global average pooling, batch normalization, or dropout have been used. Different activation functions such as ReLU, LeakyReLU, or tanh were tried. The final layer used a softmax activation. During the training of the CNN, the categorical cross-entropy was used as loss function. Moreover, an Adam optimizer with a learning rate was used.
Table III gives the weighted classification accuracies of the different event representations as input. Given the class imbalance in the data, standard accuracy can be misleading due to the underrepresentation of some classes. To overcome this issue, the weighted accuracy was use, which adjusts the overall accuracy by accounting for the number of instances in each class. The DT delivers an accuracy of 64.29% for the multiclass classification problem with five classes. In comparison, the CNN with binary projection P provides a higher accuracy of 72.96%. This makes sense, as the CNN processes a complete picture with information about the positions of the events, rather than relying on just three simple features as in the DT. Nevertheless, our results show that a density projection Q does not improve the classification accuracy and a time projection significantly degrades the classification.
Accuracies for training (Atrain) and testing (Atest) for DT and CNN for the different event representations of configuration I. The CNN had the following architecture: for the binary projection P, the CNN consists of one convolution layer with a filter size of 7 × 7 followed by max pooling and one dense layer. The time projection R1 results in the same CNN but with a filter size of 3 × 3. The density projection Q consists of three convolution layers with a filter size of 3 × 3 with max pooling followed by two dense layers. The CNNs have ReLU as the activation function.
Method . | Event representation . | Atrain (%) . | Atest (%) . |
---|---|---|---|
DT | , with r = 60 | 65.97 | 64.29 |
CNN | Binary projection P | 79.47 | 72.96 |
CNN | Density projection Q | 75.62 | 72.50 |
CNN | Time projection R1 | 46.04 | 42.50 |
Method . | Event representation . | Atrain (%) . | Atest (%) . |
---|---|---|---|
DT | , with r = 60 | 65.97 | 64.29 |
CNN | Binary projection P | 79.47 | 72.96 |
CNN | Density projection Q | 75.62 | 72.50 |
CNN | Time projection R1 | 46.04 | 42.50 |
VII. SUMMARY AND OUTLOOK
This paper, for the first time, investigates the potential of event cameras in laser welding. Event cameras (EC) differ from standard frame cameras: Rather than capturing images at a fixed rate, they asynchronously measure pixelwise brightness changes and output an event sequence that includes time, location, and sign of the brightness changes. To extract meaningful information from the asynchronous events that occur during a laser welding process, we present three promising event representations for in-process quality monitoring of laser welding processes: First, we propose the use of event rates that count the number of events occurring in a predefined time interval. Therefore, spattering processes can be quantified. Second, to also capture spatial information, we propose to transform the events into projections in the EC observation field. Three such projections, namely, binary, density, and temporal, have been analyzed. Third, to also incorporate scanner information, we propose to project the events along the scanner data. In our experiments, we observe the laser process of two metal plates coaxially and in process with an event camera and introduced different in production occurring scenarios. Those scenarios include contamination, leading to a spattering process or a gap between the two plates. Our laser welding experiments show clear differences for the scenarios in the event representations. To further encourage the potential of event cameras for process monitoring in laser welding, different configurations in which the event camera parameters, the process parameters, or the hardware parameters were changed are realized. After having analyzed the differences in the event representation when changing the laser welding process, we then present a machine learning pipeline for quality monitoring. Therefore, event rates were used to train a decision tree (DT) and projections of the events in the event camera observation field—so images—were used to train convolutional neural networks (CNNs). Therefore, the CNN outperformed the DT. Moreover, binary and density representations outperformed time representations across the CNN architectures.
In the context of future research, to improve the classification accuracy, additional features such as spatter count, spatter angle, and spatter speed may be analyzed. In addition to additional features, further classification algorithms could be applied. Moreover, while our analysis classified the laser welding process after the weld of one experiment, other sampling strategies like analyzing the event data after a predefined amount of events has occurred could be applied.
AUTHOR DECLARATIONS
Conflict of Interest
The authors have no conflicts to disclose.
Author Contributions
Patricia M. Dold: Conceptualization (equal); Data curation (lead); Formal analysis (lead); Investigation (equal); Methodology (lead); Software (equal); Visualization (lead); Writing – original draft (lead); Writing – review & editing (equal). Praveen Nadkarni: Investigation (equal); Methodology (supporting); Software (equal); Visualization (supporting); Writing – original draft (supporting); Writing – review & editing (equal). Meiko Boley: Investigation (supporting); Project administration (equal); Supervision (equal); Writing – review & editing (equal). Valentin Schorb: Investigation (equal); Writing – review & editing (equal). Lili Wu: Supervision (supporting); Writing – review & editing (equal). Fabian Steinberg: Supervision (supporting); Writing – review & editing (equal). Peter Burggräf: Supervision (supporting); Writing – review & editing (equal). Ralf Mikut: Conceptualization (equal); Project administration (equal); Supervision (equal); Writing – review & editing (equal).