1 Introduction

Next-generation Reconfigurable Manufacturing Systems (RMS), namely, adaptive production systems designed to quickly adjust their configurations to accommodate changes in production requirements, processes, or product designs, are being developed to achieve flexible and intelligent industrial automation [1]. According to the Fourth Industrial Revolution, significant improvements have been made in both hardware (HW), which comprises physical components such as robotic arms, conveyors, sensors, and computing devices used for product handling, data acquisition, processing, and communication; and software (SW), which includes computer programs, algorithms, and applications that control the system operation. However, despite Industry 4.0 (I4.0) envisioning the use of modern smart technologies to automate traditional manufacturing and industrial practices, automation is still not capable of tackling challenging tasks in variable manufacturing environments, primarily executing pre-programmed jobs. Therefore, adaptability to target changes and autonomy are the next trend in the transition towards flexible industrial automation [2, 3].

The virtual programming of flexible physical systems can be framed in a broader field of research that is particularly relevant for the next generation of reconfigurable industrial robots that exhibit adaptable behavior in cyber-physical manufacturing environments [3, 4]. Cyber-Physical Systems (CPSs), which integrate heterogeneous software and hardware components, play a crucial role in reconfigurable systems, attracting researchers in various areas, particularly in manufacturing. This work focuses on a subcategory of CPS that relies on digital models in the field of industrial vision systems for inspection purposes. It includes the set of digital Computer-Aided (CA) information, assisting in reconfiguring a physical copy from the virtual version. Analogies to other flexible machines and systems are also presented [5].

Vision systems belong to the interdisciplinary fields of system engineering or, more generally, automation [6], and incrementally benefit from the fast progress in electronics (e.g., sensors, processors, specialized cards) and computer science (e.g., algorithms) [7]. Despite the considerable level of automation introduced by these solutions, and the integration with I4.0 cutting-edge technologies such as the Internet of Things (IoT), enabling device connectivity [8]; autonomous robots, capable of self-operation [2]; big data and analytics, for data-driven insights [9]; Additive Manufacturing (AM), using 3D printing [10, 11]; and Augmented Reality (AR), enhancing real-world experience [12], the reconfiguration activities (i.e., hardware setup and software parametrization) are still human-intensive based [2, 13, 14].

Among the main limitations in applying Vision Inspection Systems (VIS, i.e., automated systems using HW and SW for contactless inspecting products through electromagnetic sensing) in industrial environments as an alternative to manual visual inspection is the high degree of HW adaptability required to cover different products (e.g., the wide variability of products’ characteristics and defect types). Concerning the hardware side, most VIS are designed ad hoc for a given application, with low degree of flexibility, often presenting rigid architectures (i.e., not mobile components) or significant setup time. On the software side, a major limitation is the manual low-coding activities required from skilled staff to adapt specific algorithms and thresholds via reparameterization [15]. However, owing to a lack of expertise in programming, traditional programming methods are challenging for operators. As a result, there is a growing demand to make SW programming easier for the new generation of reconfigurable VIS that require flexible HW/SW architecture as an enabling factor to pursue the aimed reconfigurability.

The novelty of the current work and main contribution consists in the following:

  1. i)

    proposing a framework for designing a Flexible Visual Inspection System (FVIS—an adaptable inspection system) that can be easily reconfigured via a Reconfiguration Support System (RSS—a system supporting reconfiguration) according to a variant approach inspired by the variant process planning, originally defined for Computer-Aided Process or Inspection Planning (CAPP/CAIP) [16].

  2. ii)

    validating the proposed framework in a real industrial setting with a high degree of complexity (defect types and products variance) to stress-test its actual flexibility and reconfigurability capabilities.

The two contributions of the paper are illustrated in Fig. 1 along with some promising future developments. Specifically, the theoretical components of the proposed framework are outlined on the top side of Fig. 1, while the case study validation setup for testing the practical implementation of the proposed framework and measuring some indicators to showcase the achieved results is shown in the bottom side of Fig. 1.

Fig. 1
figure 1

The two contributions of the current work. Proposed framework definition (Section 2: the proposed framework: FVIS enabling RSS) and proposed framework validation (Section 3: experiments). Future directions are highlighted in dotted box for both the contributions

With a particular focus on the proposed framework, it is built on RSS, which is in turn based on the FVIS architecture. RSS is rooted in the computer-aided extraction of (digital) information from different sources of data (i.e., actual images, Computer-Aided Design (CAD) model information, and 3D model rendered images) by Automatic Feature Recognition (AFR), which is a process for recognizing geometrical and topological features automatically. The aim is to allow non-skilled operators to easily reconfigure the FVIS by extracting parameters and generating a Reconfiguration (ReCo) file (i.e., file containing both HW parameters, for example, sensors, and SW parameters of the vision inspection algorithms), through the Graphical User Interface (GUI) of the RSS. On the other hand, FVIS conceptual design concerns the census of all possible defect types to be inspected, the HW and SW initial selection and configuration according to the working context.

The effectiveness of this proposed framework is demonstrated through validation on a real inspection case study where the FVIS and RSS have been designed and implemented in accordance with the proposed framework. The validation process starts by accounting for all digital parts (220) and defect types (60) from 12 families. The first step concerns the FVIS virtual simulation, then the FVIS lab prototype is developed. Secondly, the FVIS industrial system is implemented in the actual plant manufacturing setting and run over 4 years. The implemented solution is evaluated via a quantitative index on the rate of true positive and false positive detection.

The proposed framework draws inspiration from other well-known industrial automation frameworks, which are reviewed in Table 1. A comparative overview of characteristics (rows) among different classes of manufacturing systems (columns) showcases representative differences and similarities of the proposed framework for visual inspection (in the last column).

Table 1 Synopsis of manufacturing systems’ (columns) attributes (rows). The last column refers to the general attributes of the proposed framework in the current work

Particular attention is given to flexibility enablers, which, as defined in the recent literature, encompass four critical properties of manufacturing systems with low ramp-up/reconfigurability cost, time, and effort: (1) modularity—major components are modular, (2) integrability—hardware and software modules provide interfaces for easy integration, (3) customization—equipment is designed to meet the specifications for processing a specific product family, and (4) convertibility—hardware or software changes can be made quickly [3]. The percentages reported in the table are based on the level of satisfaction of these four enablers for each system.

Another important attribute concerning flexibility is the type of machine or system adopted. Reconfigurable Machines (RMs—machines that can adapt themselves for a set of operations to produce within a specific product family) form the foundation for RMSs. According to [17], machines used in a Dedicated Manufacturing System (DMS—rigid automation) are typically specialized for a specific, mass-produced part. They are designed to carry out a single operation with high precision, consistency, and efficiency making them relatively simple and cost-effective. In contrast, machines used in a Flexible Manufacturing System (FMS, e.g., flexible manufacturing cell) are designed for versatility and can produce a variety of parts by altering their computer programs and flexible HW. These Numerically Controlled (NC—computer programmable) machines are typically more expensive due to the need for high production volumes [17]. RMs falls between DMS and FMS machines, providing limited, specialized flexibility that allows for cost savings while enabling quick adaptation to changing product designs [17]. Unlike a standard VIS and similar to an RMS/FMS, the FVIS includes hard automation NC motorized components, such as cameras, lighting sources, grippers, and tailstocks, which can be reconfigured via software using a specific ReCo file generated through the AFR methods implemented in the RSS.

Other attributes, such as design principles (i.e., requirements the systems need to address), output (physical product/product's information resulting from the system), setup (approach used to setup the system), reprogramming/process planning (techniques adopted to reprogram the system), or output of the reprogramming (hardware/tool adopted to reprogram the system) are also considered in Table 1.

Table 2 reports the acronyms, full names, and definitions used throughout the paper, while the logical structure of the paper is outlined as follows: Section 1.1 defines the VIS background and core concepts; Section 1.2 introduces the main works related to Computer-Aided Vision Inspection (CAVI—adoption of computer-based tools, algorithms, and image processing techniques to automate and optimize visual inspection tasks), highlighting the need for a FVIS to achieve reconfigurability. In Section 1.3, a detailed review of the main techniques adopted in the literature for AFR via actual images, 3D CAD data, and synthetic images is provided. Section 2 details the proposed framework for the design of FVIS and RSS. Section 3 presents the industrial case study where the proposed framework has been applied. Section 4 reports the results both for the general framework (top-down) and for the case study (bottom-up), and Section 5 concludes the present work.

Table 2 Glossary of acronyms in alphabetical order and their extended definitions

1.1 VIS background

Production flexibility and efficiency have become the new challenges in established manufacturing paradigms [18, 19], which have undergone a transition from mass production, based on a push approach, to mass customization, rooted in customer demand [10, 20, 21]. Growing customer expectations have lowered the maximum level of acceptable defects to drastic values in the order of parts per million (ppm). Therefore, for many manufacturing industries, such as automotive production, the role of quality assurance—e.g., Total Quality Management (TQM), Six Sigma, zero-defect manufacturing—has gained a strategic relevance [22].

Although quality management standards rely on the capability of the process to meet certain specifications and statistically ensure the related product conformance, final product inspection remains a central activity for manufacturing processes [23]. According to a definition from the International Organization for Standardization, in ISO 9000, inspection is the process of determination of conformity, which occurs when a product meets specified requirements [24]. In this context, human visual inspection is one of the most implemented approaches due to its flexibility (i.e., versatility in inspecting variable products and variable defects) [25]. On the other hand, it is a source of inefficiencies, such as time-consuming, inaccurate, and inconsistent activities due to unavoidable human variability, as well as other human-related physical limitations [23, 26].

Vision systems (also known as artificial or machine vision), which are widely used today to capture images and perform various tasks such as surveillance, guidance, tracking, location, sorting, and 3D reconstruction, in a wide range of application domains, extending from e.g., food processing [27] and environment monitoring [28] to civil construction [29], assisted medicine [30], and pharmaceutical inspection [31], have become a valuable alternative for automatic contactless inspection (i.e., VIS or automated visual inspection) [13, 32, 33].

VIS can be defined as a set of existing off-the-shelf, commercial, or prototypical electromechanical (i.e., hardware) and information technology (i.e., software) components integrated in a novel way to solve inspection problems. The embedded technology includes contactless sensors in a broad electromagnetic wavelength band (e.g., X-ray, ultraviolet, visible, infrared, microwave, and multispectral).

From a general perspective, VIS essentially relies on the extraction and interpretation of information acquired from digital images and making decisions accordingly. The work proposed by [34] classified automatic visual inspection as a three-step process that can be collapsed into the following two-step process:

  • Image acquisition: this step involves the hardware and software components required for image acquisition, storage, and processing. As shown in Fig. 2 (left side), the vision hardware mainly includes camera sensors [14], focal lenses, optical filters, port connections, lighting, and handling sub-systems. Software components include embedded computing integrated with vision data management tools and processing algorithms to perform specific visual inspection activities (Fig. 2, right side) [7].

  • Image (pre-/post-) processing and analysis: once the image is acquired, a variety of visual inspection activities (e.g., surface control, pattern matching, dimension gauging, inner defects check, shape control, and general identification) are performed to extract the desired characteristics from the inspected object [35]. The characteristics are analyzed via context-related algorithms parametrized for the specific case. Among several inspection algorithms, two main categories can be defined, namely, rule-driven and Artificial Intelligence (AI)-driven [13]. The rule-driven category includes classic computer vision algorithms such as color analysis, blob detection, and edge detection, while AI-driven is based on Machine Learning (ML) algorithms able to learn from data, recognize patterns, and make predictions without explicit programming (e.g., Convolutional Neural Network (CNN)) [13].

Fig. 2
figure 2

Vision Inspection Systems (VIS) are vision systems made by conventional or customized HW/SW components and sub-systems (left) integration but specifically designed for inspection activities (right)

1.2 Computer-Aided Vision Inspection (CAVI)

Digital tools have brought significant value to the field of CAVI, and the integration of CAD technologies and data management systems such as Product Data Management (PDM—data base containing product-specific information) for offline programming and reconfiguration of vision inspection assets has been a promising research topic since the late 1980s [36]. Over the past decades, research into the reconfiguration of VIS has led to the development of visionary frameworks like the Computer-Aided Vision Inspection System (CAVIS) [37], where the authors propose a solution to overcome the need for specialized skills in the vision system reconfiguration, algorithm definition, and programming phases in assembly contexts. Despite the merits of these theoretical models in conceptualization and initiating CAVI reconfiguration, there have been only a few practical applications on flexible architectures enabling this approach in industry and scientific literature in recent years.

With the advent of I4.0, more powerful, flexible, and affordable hardware and software solutions have become available on the market. As a result, many Small and Medium Enterprises (SMEs) and companies facing high technological and automation barriers are now transitioning to automated systems. In this rapidly evolving automation environment, there is a growing interest in introducing and implementing new technological concepts to fully interconnect production elements by bridging the physical and digital worlds [38], creating new opportunities for CAVI reconfigurable systems [32]. As noted by [13], the global machine vision system market is expected to grow rapidly from $8.6 billion in 2020 to $17.7 billion by 2027, driving the demand for improved performance and new capabilities such as reconfigurability.

Consequently, there is a growing demand for flexible and reconfigurable systems capable of eliminating the laborious and time-consuming hardware setup and software programming work [39]. In this context, parametric offline reprogramming is gaining increasing attention in various robot-based activities [15, 40,41,42]. Among robot-supported visual inspections, a new approach for specifying the spatial and temporal trajectory of visual inspection based on CAD models of workpieces has been proposed [43]. In this work, the authors claim the ability to reduce the effort required from the engineer to setup vision hardware, a task that remains a non-value added and tedious for production line engineers [43]. Similar to [43], our work considers CAD data for the system setup, offering a higher degree of flexibility enabled by NC programmable hardware components.

Another related work based on an adaptable CAVI system can be found in [44]. In their work, the authors propose a flexible hardware solution comprising standardized (e.g., lights, robots) and reconfigurable (i.e., handling system) components, integrated with flexible image processing algorithms using CNN and Explainable Artificial Intelligence (XAI), along with a knowledge transfer system for software parametrization support.

Despite the close relationship with the general framework proposed in [44], we aim to provide a more general model not tied to a specific HW/SW architecture. To validate the proposed framework and present an industrial case study, a successful real-world implementation in the automotive industry is showcased.

1.3 Automated Feature Recognition (AFR) methods

The ReCo file serves as the link between the digital and physical worlds. It is used to reconfigure the NC hardware components of the VIS and contains the parameters for the inspection algorithms. To expedite the parametrization of the ReCo file, a virtual environment based on AFR is adopted to minimize the user’s programming effort and facilitate the process of CAVI reprogramming.

As shown in Fig. 3, a trade-off emerges when embracing an increasing virtual AFR approach in the effort to calibrate the cyber-twinned entity to mimic the physical environment. As a matter of fact, creating an almost identical copy of a physical counterpart (i.e., with a very high degree of accuracy) is a demanding task. The whole process of developing highly accurate computer models and simulation of processes is time-consuming and demands significant investments [45]. This underscores the importance for industries to conduct a cost-benefit analysis before implementing a digitally twinned approach over another.

Fig. 3
figure 3

Automatic Features Recognition (AFR) techniques at different levels of cyber-physical configuration effort and production line setup impact. A trade-off is needed between cyber-physical configuration effort and production line setup impact

In the following Sections (1.3.11.3.3), we will review three main categories of possible approaches for extracting digitally twinned vision features (to provide input for the ReCo file parameters) for the reconfiguration of the FVIS. As illustrated in Fig. 3, the first AFR approach, examined in Section 1.3.1, pertains to extracting digital information directly from actual images of the authentic product, captured on-site by the actual camera under operational lighting conditions. The second approach, treated in Section 1.3.2, is related to extracting geometrical information using a three-dimensional (3D) CAD virtual product model. The third approach, reviewed in Section 1.3.3, utilizes synthetic rendered images of the virtual product as source input. All these categories can be implemented in the proposed framework.

1.3.1 AFR via actual images data

In the field of industrial vision inspection [7], employing authentic images to parametrize the vision algorithms is a conventional approach facilitated by camera manufacturers, who embed their products with dedicated image analysis toolboxes or SW extensions [46]. Parametrization can be accomplished through user-friendly interfaces that enable the selection of Regions Of Interest (ROI) using drag and drop functionality, among other features.

On the other hand, more advanced approaches (e.g., deep learning-based) commonly employed for surface defects detection can be adopted to partially or fully automate the classification of features and extract information for parameterizing the ReCo file [47].

Undoubtedly, these approaches reduce the complexity of transitioning from the virtual environment to the real production environment. As a result, the reference system remains the same, and there is no need to consider additional lens distortion or calibration [48]. However, a significant issue persists: the impact on the production line flow.

Due to the time required for setting up the actual vision system, offline parametrization is not entirely offline since real images must be collected using the real machine, which interrupts the production flow. Therefore, the literature has explored more advanced solutions with zero impact on shop floor productivity and fully offline parameterization. Existing approaches are reviewed in the subsequent two sections.

1.3.2 AFR via 3D CAD data

CAD-based AFR methods are aimed to extract geometric features (e.g., holes, slots, pockets, bushings) and tolerance specifications (e.g., dimensions, parallelism, concentricity) from 3D models without the necessity of the real product [49, 50]. These methods are typically adopted for extracting specific features for manufacturing purposes but can also be used to extract inspection-related or vision-specific features.

AFR techniques for CAD models are well-established, and research has been ongoing in this field for decades [51,52,53,54,55]. CAD has been instrumental in streamlining the design process for new products, using recurring or scalable elements [56, 57] and group technology [58]. The use of CAD, CAPP, Computer-Aided Manufacturing (CAM), and CAIP has significantly increased over the years, intensifying the already existing link between design and production fostered by the AFR process [54, 59, 60]. CAPP, CAM, and CAIP tools often involve the “decomposition” or “extraction” of distinct sets of faces to be machined or inspected, defining processing information such as the selection of cutting/inspection tools, machining/inspection sequence and strategy [61, 62]. AFR techniques are the first and most crucial step in transforming the 3D CAD data into processable information for manufacturing (i.e., CAPP, CAM) or inspection (i.e., CAIP) and can be readily adapted for CAVI.

Table 3 reports the main AFR techniques for 3D CAD data retrieved from recent state-of-the-art reviews [53,54,55].

Table 3 Main AFR techniques for 3D CAD data summarized from [53,54,55] along with a synthetic description

Although many established CAD software programs now incorporate generic functions for AFR [63, 64], evidence of their industrial implementation is lacking [55]. Furthermore, with the increasing complexity of products and processes, the need to develop new, efficient strategies for feature extraction remains an open topic [53].

As previously highlighted, the main advantage of adopting AFR based on 3D CAD data lies in the extraction of geometric features from a digital model. This enables the parametrization of the ReCo file without disrupting the production line. Additionally, this approach is significantly faster than the one outlined in Section 1.3.1, as it automates numerous activities associated with manually setting the desired tolerance values embedded in the 3D CAD model, which are automatically extracted.

However, a fully digital environment also introduces challenges in calibrating the real vision system. The extracted information must be converted to account for real-world lens distortion and other calibration factors. This requires additional calibration steps to migrate the parameters from the digital to the actual environment. Another obstacle to the widespread application of CAD-based automation in the industry is the compatibility among different commercial CAD file formats and the need for a neutral file format [65]. To address this, file formats such as the Standard Triangulation Language To Layer (STL), STandard for the Exchange of Product model data (STEP), or the Initial Graphics Exchange Specification (IGES) have been defined to unify and provide a common ground for developing new methods and algorithms [66,67,68].

1.3.3 AFR via synthetic images

Photorealistic rendering is a technique that allows the generation of images, starting from a digital 3D model, which are challenging to distinguish from actual photographs [69]. A significant segment of the realistic rendering market (e.g., architectural, automotive, and product visualization in general) has relied on Physically Based Rendering (PBR) technology [70].

The utilization of synthetic rendered images for the extraction of valuable information for VIS reconfiguration was first proposed within the broader CAVIS framework by Lanzetta et al. [37]. As an application, CAD-rendered images of the product have been successfully employed to extract the ROIs and set other vision system parameters [71], or to train supervised vision inspection algorithms in the field of Artificial Intelligence (AI) [72].

The use of synthetic rendered images as a hybrid approach allows for combining the benefits of virtual CAD data and real digital image data (Section 1.3.1 and 1.3.2). This advantage can be reached by simulating the camera parameters and settings within a fully digitally twinned reality [73, 74], reducing the gap between the digital model and the actual visual environment. In this case, no extra calibrations are needed to migrate the information retrieved from the rendered image to the real VIS. Nonetheless, despite the significant advancement of CAD systems over the past two decades, the use of synthetic images for feature extraction remains restricted due to the considerable number of parameters required in the rendering setup process (e.g., material, texture, lights, environment, camera specifications), as well as the computational power demands.

2 The proposed framework: FVIS enabling RSS

The proposed framework is illustrated in Fig. 4. At the bottom, the product to be inspected is the source for three different types of data, which allow for the three different AFR techniques presented earlier (Section 1.3.11.3.3). Depending on the selected approach, actual images, CAD or rendered images files are the input for the RSS, which enables offline reconfiguration via ReCo file using three main modules: the Graphical User Interface (GUI), hardware setup, and software reconfiguration.

Fig. 4
figure 4

Flexible Vision Inspection System (FVIS) and Reconfigurable Support System (RSS) framework

During the AFR step, a ReCo file (e.g., in Extensible Markup Language (XML) format) is generated, and the VIS hardware setup and algorithms parameters are updated accordingly.

The online inspection begins taking images via the vision hardware, pre-processing, and analyzing them via the inspection software. The ReCo file can be stored in a shared database among plants, expediting the reconfiguration process for various products, following a variant approach inspired by the variant process planning defined for CAPP. The premise for this feature is that similar parts require a similar inspection task and, consequently, a similar ReCo file (i.e., similar HW/SW configuration).

The main two steps of the system design are presented as follows: Section 2.1 recalls the preliminary activities that should be carried out in the conceptual design of the FVIS. Section 2.2 delves into the offline reconfiguration process enhanced by the AFR approach implemented into the proposed RSS for the generation of ReCo files.

2.1 Flexible Vision Inspection System (FVIS) conceptual design

The FVIS conceptual design focuses on designing a flexible hardware and software configuration (right-hand side of Fig. 4, central-top side of Fig. 1) tailored to the product’s defects to be inspected, physical and geometric characteristics of products and families, as well as environmental constraints (e.g., inspection context and production rate). It is important to note that this section solely serves to define the conceptual design methodology, which includes the two steps as follows:

  • Step 1 (defect types census and working context): to establish a comprehensive list of defects, a combination of top-down and bottom-up approaches should be employed.

    Top-down generation (i.e., product families database census): PDM exploration and products/families census can be easily achieved by retrieving the information stored into the CAD files. The main characteristics and components of each type, for product families, can be analyzed by checking the Bill Of Materials (BOMs). This procedure facilitates the generation of a list of possible related defects.

    Bottom-up generation (i.e., historical data on defects): quality assurance records, derived from the experience with similar products, explicit requests from customers, inspection policies of suppliers, and raw material can be used as valuable historical data to define already registered defects.

    The culmination of this process involves aggregating all the top-down and bottom-up generated defects. The list is further refined through cleaning (i.e., duplicate removal), and summarizing activities. Additionally, the working environment (e.g., physical variables such as ambient light, vibrations and atmospheric agents like dust, humidity) and production constraints (e.g., productivity, accuracy) must be considered. This Step 1 is inspired from the first design principle for a reconfigurable machine proposed by [17], in which “a reconfigurable machine is designed around a specific part family of products”. It also aligns with principle 2 for customized flexibility, as well as principle 3 for easy and rapid convertibility.

  • Step 2 (HW and SW selection and initial configuration): the defects list defined in Step 1 must be processed to establish a set of defect clusters that will be adopted for hardware and software components selection.

    The hardware and software must be selected based on the principle of maximum flexibility, including the use of NC motorized axes and PLC for hardware setup and control (e.g., robotic arm, moving cameras, and adjustable lighting conditions). Vision commercial libraries or custom-developed inspection algorithms can be implemented for software inspection. The flexibility can be associated with the coverage of each defect to be inspected by using different configurations of hardware and software components. Once the initial HW/SW configuration is selected and integrated into a flexible architecture (i.e., FVIS), individual features and attributes to be inspected will have its corresponding inspection hardware and software algorithm defined.

2.2 Reconfiguration Support System (RSS)

The frequent turnover of products resulting from the mix of components amplifies the need for reconfiguration. In common practice, even a straightforward reconfiguration for changeover of VIS requires specialized staff. Minor software modifications to ROIs or threshold algorithms are generally considered a complex task to outsource, except in the rare case of specialized internal staff possessing programming skills. Similarly, concerning the hardware setup, tasks like repositioning a camera or adjusting a light source and subsequent calibration require specialized staff if the system lacks proper flexibility features. This deficiency often leads to the necessity of redesigning the system and incurring substantial adaptation costs.

According to the FVIS architecture defined in Section 2.1, a user-friendly RSS is proposed as the second ingredient to enable a fully flexible framework. The graphical representation of the offline flexible reconfiguration process facilitated through the proposed RSS is presented on the left-hand side of Fig. 4 and central-top side of Fig. 1. The flexibility aspect is the main innovation of the proposed solution because it drastically reduces the FVIS programming time (Section 1.1). In pursuit of this objective, the RSS encompasses three primary modules:

  • SW reconfiguration module: this module is responsible for acquiring specific problem knowledge for ReCo file parametrization. It accomplishes this by extracting pertinent information, including inspection algorithm-related parameters (e.g., ROIs and thresholds) and NC hardware component-related parameters (such as actuator and NC motor settings), leveraging the AFR techniques implemented into the GUI. This information is then saved in the ReCo file.

  • HW setup module: this module focuses on integrating the RSS reconfiguration commands stored in the ReCo file into the controller employed for NC programming. This allows for automatic changes in screw lengths through actuators for cameras, lighting, grippers, and tailstocks positions, as well as moving the robotic arm or other flexible mechatronics hardware via the SW interface.

  • Graphical User Interface (GUI): it serves as the interface that assists users, providing information and describing the AFR process through digital support. The GUI also enables operators to add, update, and verify parameters stored in the ReCo file to optimize the visual inspection process. Additionally,the GUI is utilized during the inspection process to display defect reports  to the operators on the production line. These functionalities can include explainable AI (XAI) models, which offer insight into how the inspection system reached its conclusions, as well as visualization tools to help the operator understand the inspection results. The XAI functionalities assist operators in understanding the detection rationale for a specific defect, its location, and how it can be managed according to the quality policies. The requirement for XAI functionalities can vary depending on the selected inspection algorithm (i.e., rule-driven versus AI-driven).

The main output from the reconfiguration process is a ReCo file, which is generated by the RSS in an interoperable format, such as XML, and is easily manageable by conventional operating system libraries. This file is used to update the FVIS configuration, effectively acting as an integrated AFR tool. Figure 5 offers a simplified visual representation of the workflow involving GUI interaction during AFR tasks. The three AFR approaches that are considered in the general framework are highlighted. Regardless of the chosen AFR approach, the part number information (such as assembly, schemes, and actual images) must be previously stored in PDM, even for new (i.e., never inspected) part numbers and related components.

Fig. 5
figure 5

Procedural workflow of the RSS GUI. Information flow is depicted with dotted lines. Solid lines indicates the logic flow of the activities (represented by squares). Decision steps are represented via diamond shapes. The central section of the figure depicts common activities. After the AFR technique branch, the upper side represent CAD-based approach, while the lower side represents the actual/rendered image-based implementation

The image-based (actual and synthetic) approach of the RSS AFR (as shown in the central and bottom side of Fig. 5) consists of a wizard interface to define the ROI and the operational parameters by selecting predefined points, lines, and areas on an image taken by the actual camera or from rendered images. As reviewed in Section 1.3, the strength of the actual images approach lies in operating on real images, which guarantees compensation for optical distortions. Additionally, it offers the possibility of calibrating the system by only computing the mm/pixel ratio for each component. In contrast, synthetic rendered images require more effort in the photorealistic rendering process to approximate actual images directly from CAD data; however, they ensure an offline process solution. Once the image (actual or rendered) is obtained, a customized interface containing a guided wizard procedure is utilized to parametrize the ReCo file. This method requires several steps, including searching for images stored in the PDM (if any), or imposing an online manual activity to take images of the part/generate the rendered images.

An alternative approach for offline reconfiguration is based on AFR directly from CAD files (central and upper side of Fig. 5). Currently, some commercial CAD software include the capability to develop macros using a Visual Basic for Applications (VBA) engine. This enables the development of a single macro that can be adapted with minimal modifications for use in other CAD environments, particularly when standardized CAD data format (such as STEP and IGES) are employed.

Regardless of the adopted AFR approach, the generated ReCo file contains information about the product’s inspection features (e.g., tab slot, dust shield, O-ring) and their attributes (such as tab slot’s width and length). Each feature and attribute correspond to an inspection algorithm defined during the FVIS design phase. Alongside the extracted vision features for hardware and software setup, the ReCo file also contains details about the hardware components (e.g., size, resolution of the sensor, focal length of the optics, distance of the optics from the axis, distortion factors, actuators to be activated). This information is used when adopting the CAD-based AFR approach to convert the CAD model dimensions in millimeters into the respective dimensions in pixels required to parametrize the inspection algorithm that processes real images.

Once the new ReCo file is generated, it is stored in the program database, together with all the generated ReCo file (e.g., shared cloud among different production plants), aligning with the I4.0 approach. If a different plant in another location needs to inspect a product with an existing ReCo file, an updated version can be conveniently accessed for direct download and variant adaptation (as shown in the right side of Fig. 4).

Notably, mechanical parts with different functions, such as splines or threads, might exhibit similar appearance and share the same inspection algorithm, given the close relationship between vision features and geometrical features. As a result, it is possible to develop an application that selects an already existing ReCo file from a database (associated with a previously processed part number) with similar vision inspection features. This ReCo file contains configuration parameters closest to those required for industrializing the new code, and it can be adjusted as necessary using a variant approach.

3 Experiment

This section presents the implementation and testing of the proposed framework in a real-world industrial case. The specific case of automotive half shaft production is detailed in Section 3.1. The conceptual design of the FVIS for this case is outlined in Section 3.2, while the implementation of the RSS is presented in Section 3.3.

3.1 Case study background

The development of an FVIS and RSS arose from demanding customer requirements for a global leader in the half shaft production sector, requiring less than 10 ppm of defective products. A half shaft is a component that transfers torque to a vehicle's wheel using two Constant Velocities joints (CVJ) and splines on each end [75, 76]. Figure 6(a) offers an overview of a typical half shaft product.

Fig. 6
figure 6

a Half shaft is the dynamic connection between the differential and the driving wheel of a motor vehicle. It transmits power and facilitates steering, adapting to suspension movements, and isolating vibrations. Each half shaft consists of two CVJs. A CVJ is a mechanical coupling ensuring that the output shaft's rotation speed matches the input shaft at any working angle. A bar shaft connects these two constant velocity joints. b Defects identification phase. Starting right to left the labels indicate the following defects: circlip, spline, oil drain, bearing, differential side boot clamp, boot, shaft side boot clamp, shaft side boot clamp, boot, wheel side boot clamp, dust shield, wheel side back face, stem, spline, thread, and tab slot

Half shafts are safety-critical mechanical components in vehicles, enduring mechanical stresses (e.g., torsion and bending), leading to wear, fatigue, and even fractures. Given their safety-critical nature and the persistent mechanical stresses, stringent quality control measures during manufacturing are essential to prevent defects and failures.

In the specific case, the manufacturer monitored 150 machining and assembly process parameters during half shaft production. However, human errors and defects in supplied parts could only be detected through 100% final product inspections before packaging and shipping (e.g., some hidden defects of the assembled half shaft require product manipulation to achieve a complete inspection).

Manual inspection remains common practice in many robotized assembly lines [77, 78]. Similar to the current case, manual operation is often preferred due to the complexity and variability (production mix) that complicates automating the inspection process. However, due to human-related physical limitations, operators could only target a few critical defects during the visual inspections. Therefore, a plausible introduction of a FVIS lies in some of the intrinsically lower operators’ perception capability and reliability compared to machine vision systems.

This work began in 2016 as an industrial-university cooperation to develop a general-purpose FVIS for quality inspection, reconfigurable via an RSS. Various motivations have pushed this work to the maximum potential. However, a primary driver has been achieving the complete automation of assembly.

The first dilemma was deciding between distributed inspection in the individual assembly stations or a single final inspection, providing feedback on all the previous operations. The initial outcome of this cooperation is an inspection station—online since 2017—at the end of an assembly cell, which can detect assembly or machining defects.

This system has been cloned and operating across various automated and semi-automated international assembly cells and can detect over 60 defects in less than 30 s, processing 2.500 parts/day. This achievement is enabled by reconfigurable hardware components such as eight NC cameras, grippers, tailstocks, and adjustable lights, as well as parametrizable inspection algorithms tailored to specific product types. Such a complex system is a significant challenge and represents a research achievement for its complexity. However, its detailed description is outside the scope of this paper, as the main focus centers on the general proposed FVIS and RSS framework. Specifically, Section 3.2 delineates the conceptual design of the FVIS, while Section 3.3 presents the RSS development rooted on AFR techniques.

3.2 FVIS conceptual design

According to the first step outlined in Section 2.1, Table 4 provides valuable information into the preliminary conceptual design of the flexible hardware components within the machine vision system. This design takes into account the identified defects labelled in Fig. 6(b). The full list of defects was generated using a hybrid (i.e., top-down + bottom-up) approach according to Section 2.1.

Table 4 FVIS design options (rows) and defects (columns). For each combination, a score is given as follows: 0 = impossible, 1 = difficult, 2 = good, 3 = optimal. For confidentiality, a sub-set of the 60 defect types investigated is shown. The last column summarizes the average score for each design option (green scores are the selected) for the selection and clustering of solutions suitable for multiple needs

Each column (i.e., defect type) in Table 4, is associated with a score that reflects the compatibility of hardware options (e.g., camera view, background, lighting, camera type, camera distance) with the specific defect. The final column shows the average score obtained for each solution; the green cells highlight the optimal score for VIS hardware components. The final row, labeled “defect location”, shows the average positions and ranges spatial coordinates of defects (in this instance, only y is considered due to the longitudinal product extension). Such measurements are used to define how many cameras should be placed and where (if moving the camera via NC or not, or move the part, e.g., rotation), according to the Field of View (FoV) to ensure the coverage of all potential defects. For the current scope and objectives of the preliminary conceptual design of the vision system, the centimeter level of accuracy is satisfactory, and any minor uncertainties can be effectively managed through practical approaches such as overlapping FoV. Occasionally, a suboptimal solution may be selected if it accommodates various defects.

The mean and variability of half shaft defects’ positions are derived from analyzing the part number families in output from the identification stage. Once the defects’ position within the FoV are defined, spatial clustering is carried out according to the camera specifications (e.g., FoV, focal length, resolution). A similar procedure is followed for the software components selection (e.g., inspection algorithm type, sequence, parameters).

Additional factors and technicalities, including the working environment and production constraints, are also considered into the final FVIS architecture design.

In order to achieve a fully flexible solution from a reconfiguration perspective, eight NC motorized axes cameras have been selected to cover the identified ROI clusters. By considering the longitudinal extension of each cluster, the maximum distance at which cameras are expected to be positioned, as well as the focal length of the optics, and FoV, it is possible to control which camera should be activated and its specific position. In addition to the previously highlighted aspects of the hardware design, mapping defects’ positions aids in defining the maximum working range of the adaptable (i.e., universal for all the families) gripper and tailstock system for handling the parts. This system is also NC controlled (Fig. 7(a)). A similar approach was adopted for NC-controlled and -adjustable lights.

Fig. 7
figure 7

a Different product family HW configuration (e.g., cameras activation and axial movement based on the parametrized ReCo file). Red cameras are deactivated. Camera, lights, grippers, and tailstocks can be repositioned using the NC axis. b Product families variability examples

In Fig. 6(b), defects are clustered based on their positions, with more detailed sub-groups illustrated in Fig. 8. In the example shown, the 16 different sub-groups of defects checked in the area inspected by NC camera#8 are presented. This approach of defect clustering is then extended to each variant within the different product families, as exemplified in Fig. 7(b). The associated ROIs for each defect types (about 60) for the current half shaft models (about 220) are then exhaustively checked as shown in Fig. 7(a). This process aids in determining the moving range of each camera and light.

Fig. 8
figure 8

Wheel joint side cluster consisting of 16 (out of 60) defect types categorized into 6 defect groups within the wheel joint side, highlighted by the yellow area. The identified defects are used to select the appropriate vision inspection algorithm and tool for quality check

The examined set includes all the identified half shaft models in order to be representative of the historical, current, and forthcoming production scenarios. In other words, for the system to be flexible enough, all the future models must be combinations of previously examined features, like in the present case (i.e., variant approach, Fig. 1).

As for the inspection software side, given the diversity across product families (Fig. 7(b)), different algorithms must be defined for inspection purposes. In the interest of brevity, Table 5 provides a condensed selection of examples, related algorithms, and some of the parameters identified in the FVIS design stage. Given the requirements of intensive industrial application and the nature of defects, commercial software libraries have been selected and a specific commercial SW has been used (in alignment to the company’s prior experience [46]). For each quality check, the chosen inspection algorithm, the corresponding software tool, key configuration parameters, output details, and the applied threshold to determine conformity are listed. Additionally, more advanced techniques for surface defect identification and classification can be adopted [47, 79, 80].

Table 5 Examples of quality checks carried out by the developed FVIS on a sub-set of 8 out of 60 currently implemented and operating. The inspection algorithm and its output are generalized to all the product families considered for the FVIS design. The parameters and thresholds are automatically extracted for each new reprogrammed part to be dynamically updated in the RSS AFR in Section 3.3. Inspection algorithms are first defined by the system developer and selected by the user for new part numbers through the wizard interface in Fig. 9

3.3 Implemented RSS for AFR

As anticipated and according to the approach proposed in Section 2.2, the RSS eliminates the necessity for an expert programmer and speeds up the ReCo file parametrization within an entirely virtual as well as offline environment for AFR, based on CAVI. The user's proficiency requirement pertains to the product rather than the FVIS. The sole task required for the user is to select the unified defect tags recognizable by the AFR software via the guided procedure implemented in the RSS GUI and presented in Fig 9. In the procedure, a variant approach is adopted, enabling the user to select from a list of predefined defects organized according to tree logic product families (e.g., group A, B, C). The user is then able to assign tags for each identified defect outlined in Section 3.2 (Table 4).

Fig. 9
figure 9

Three screenshots of the developed user-friendly GUI part of the implemented Reconfiguration Support System (RSS) for the half shaft industrial case study

The implemented RSS for the current case presents a CAD-based AFR approach. Once the product family half shafts have been selected and the CAD information retrieved from the PDM system, the AFR macro can discern the location and dimensions of the elements labeled by the user via an interactive procedure. According to the implemented AFR technique, a dedicated macro provides a digital interface to detect the type and the relevant dimensions (or presence) of the labeled 3D element. In Fig. 10, an example of the RSS GUI tagger developed for CAD-based AFR is depicted.

Fig. 10
figure 10

Examples of assigning labels in Autodesk Inventor® via macros, which integrate the RSS with the CAD system and specific 3D model of the product under consideration to extract the specific features and attributes. The blue highlighted area supports the user in identifying the part of the product under selection. The right pop-up shows the extracted features

Once this step has been completed, the ReCo file in XML format is automatically generated and ready to reconfigure the FVIS hardware and parametrize the inspection software. Figure 11 provides an example of some lines for the generated ReCo file in XML format; encompassing various labels, including:

  • <job>: This element represents the overall jobs and includes attributes such as “name”, “cameraName”, and “runCount”, which indicate the job’s name, the camera used, and the number of times the job is executed (36 times).

  • <tool>: This element represents a tool utilized in the job. There are two tools shown in Fig. 9, namely “WSB_JointClamp” and “WSB_ShaftClamp”. Each tool has specific parameters enclosed within the <params> label.

  • <params>: This section contains various parameters related to the tools. These parameters specify the regions of interest, detection points, and compensation points used in each tool.

  • <checkList>: List of different checks performed using the tools. It includes elements such as “WSB JointClamp_diameterup”, “WSB JointClamp diameterdown”, “WSB JointClamp position”, and “WSB_JointSide_overflow”, each with respective attributes (e.g., nominal values, tolerances, and constants).

Fig. 11
figure 11

Example (extract) of a ReCo file in XML format for a new part to be inspected, generated by the developed RSS. It contains ROIs, conversion parameters, nominal values of dimensions, and relative tolerances. Each set of images taken by a single camera is processed by a job containing a series of tools specifically developed to carry out a certain number of checks on a specific area of the half shaft. Each tool operates calling the parameters. The additional required variables for the vision software (e.g., thresholds) are universal for all past and new parts

4 Results

The primary object of this work has been to develop a framework for a variant inspection vision system, called FVIS, and based on RSS, utilizing AFR applied to actual or rendered images, as well as CAD 3D models. The outcomes associated with this endeavor are presented in Section 4.1.

The second contribution of this work is the validation of the proposed framework through a challenging case study conducted in the automotive industry (Section 4.2). The purpose of presenting this case study is to validate the effectiveness of the proposed framework in a real industrial environment. The selection of a particularly challenging case, which involves more than 60 different defects and 220 products spanning 12 product families, serves as high-quality benchmark for evaluating the applicability of the proposed framework. The experimental demonstration of the proposed approach showcases its capability to inspect a wide range of defects and products, making it well-suited for handling challenging scenarios involving multiple product families and diverse inspection tasks. The successful reprogramming capabilities for new products and the flexibility features of the hardware have been proven in this real case study. However, to fully establish the effectiveness and to expand the boundaries of validation, further testing across various experimental cases and real-world validation through multiple applications and years are necessary, as indicated in Fig. 12.

Fig. 12
figure 12

Proposed framework validation and improvement. The dark blue area, representing current case study, highlights the initial step in validating the proposed framework. Its wide area indicates the adoption of a significant and complex case, encompassing both lab and industrial setups, which ensures a crucial first validation. The dotted lines and arrows depict the expansion of validation boundaries and possible improvement after implementing the framework in additional new case studies

4.1 Framework results

The proposed framework enables the retrieval of information from the CAD-based AFR approach, generating Reconfiguration file (ReCo file) accordingly, and parametrically reconfiguring the flexible VIS. NC hardware components such as cameras, lighting systems, grippers, and tailstocks can be offline parametrized based on the specific ReCo file parameters. Inspection software algorithms can also be reparametrized using the same source of information stored in the ReCo file.

The proposed framework has several merits including the ability to improve knowledge management by embedding a RSS, which makes knowledge explicit and easy to access, manage, and transfer without spillover with staff turnover. Additionally, the system increases the amount of standard/compliance information retrieved during inspection, such as image databases. The system also aims to increase the efficiency of quality control by reducing programming time and increasing effectiveness offering the ability to carry out CAVI offline, which reduces the impact on shop floor productivity. Finally, the proposed framework also allows for hardware reconfiguration and software parametrization for new product families with similar vision features reducing the need for product-specific staff skills [81].

4.2 Case study results

A virtual prototype has been successfully implemented, and different hardware setup as well as inspection algorithms have been tested. The final FVIS architecture has been tested both virtually (Fig. 13(a)) and experimentally (Fig. 13(b)). Figure 13(a) depicts the 3D model of the final vertical layout configuration, showing four of the eight labelled NC cameras, NC ad hoc grippers and tailstocks for centering and revolution, NC axial diffuse backlight system, and a half shaft under inspection. This prototype validated the FVIS designed in Section 3.2.

Fig. 13
figure 13

a Autodesk Inventor® CAD virtual prototype design of the developed FVIS. Four NC camera out of 8. The half shaft is vertically placed and axially aligned by 360 rotational motorized special purpose grippers and tailstocks. b Lab’s real prototype tests

In Fig. 13(b), a prototype of the FVIS represented in the 3D model has been tested in laboratory environment before achieving its final state on a production line.

The developed station inspects 100% of the assembled half shafts, following the main steps summarized as follows:

  1. 1.

    The half shaft is loaded into the inspection station by an anthropomorph robot.

  2. 2.

    Half shaft rotation starts via special NC grippers and tailstocks, all the lights are activated, and the door is closed.

  3. 3.

    The PLC controlling the inspection station sends a signal to the vision software, initiating the automatic inspection of the half shaft. Each camera captures 36 frames during a full rotation.

  4. 4.

    Captured images are transferred to the Personal Computer (PC) for processing via software.

  5. 5.

    Upon completion of the analysis, the results are sent back to the handling robot via the PLC deciding whether the half shaft is accepted and palletized for shipment, or if it necessitates manual reinspection for potential rework or disposal.

  6. 6.

    The GUI report is updated accordingly, displaying a red instead of a green box on the monitor of each camera if a defect has been detected (e.g., Figs. 14, 15)

Fig. 14
figure 14

Cell operator’s GUI on a touch screen on the inspection station at the end of the robotized assembly cell. Inspection of clamp differential side closure, positioning, lubricant for the part number 10302233. Green boxes (camera1, camera2, camera4DX, camera5) showcase conformity, while red (camera 3) highlights an identified defect

Fig. 15
figure 15

A screenshot of the commercial vision software adopted. The block structure of the developed inspection software is shown on the left side. Right side provides an edge detected in red for the clamp under inspection

As for the quantitative results, the system’s efficiency is evaluated at the rate of True Positive (TPR) and False Positive (FPR) via the following formulas [82]:

$$TPR=\frac{\textrm{TP}}{\textrm{TP}+\textrm{FN}}\kern0.5em FPR=\frac{\textrm{FP}}{\textrm{FP}+\textrm{TN}}$$

where

TP (True Positive): the number of defective products that have been correctly classified as defective.

FN (False Negative): the number of defective products that have been incorrectly not classified as defective.

FP (False Positive): the number of non-defective products that have been incorrectly classified as defective (false alarm).

TN (True Negative): the number of non-defective products that have been correctly classified as non-defective.

TPR should tend to 100%, while FPR should tend to 0%. Note that a value over 0% for FPR can be chosen for conservative purposes, e.g., during the system development.

In the initial testing on the line, a sample of 60 products has been used, obtaining the following results: TP = 15, FP = 4, TN = 41, FN = 0. Therefore, TPR = 100%, while FPR = 9%. After fine tuning via a large experimental set from production and the optimization of threshold values for vision algorithms  (e.g., binarization and pattern matching), performance reached 2% FPR and 100% TPR, also for new parts.

The system can assess more defects concurrently, but the occurrence of multiple defects in this type of production is considered a rare and exceptional event. Additionally, only certain defective parts can be reworked because most components cannot be disassembled without permanent damage (e.g., spline bumps cannot be recovered). As a result, after a defect is detected, the faulty part is segregated from the good parts and manually inspected to determine the appropriate actions for managing the non-conformity. Given the relative low frequency of defects, in most cases it is simply discarded. The re-programming time and cost for new parts inspection were also reduced by 40%, and low-level coding is not requested anymore. This transition has garnered unanimous operators’ satisfaction (100% of the 10 interviewed operators dedicated to this task report an “easy-to-use” and “functional” system).

Not only the system implemented and presented is complex and challenging, it also has innovative flexibility features answering the main research questions: how to virtualize the reconfiguration of a flexible vision inspection system and validate the proposed framework?

Table 6 provides a summary overview of the general proposed framework and the implemented solution for the developed inspection station #R100GKN. Significantly, Table 6 highlights the adherence of the implemented system to the four key flexibility enablers presented in Section 1 (Table 1), namely, modularity, integrability, customization, and convertibility.

Table 6 Technical solutions and main aspects of the general implemented framework and the specific solution for the inspection station #R100GKN. 

Modularity: the system’s vertical layout enables flexible arrangement and easy integration of NC hardware components, including cameras and lighting systems. An automated loading system with a robotic arm enhances modularity, and NC special conformed grippers allow quick adjustments for different product sizes and shapes.

Integrability: the system focuses on seamless interaction between hardware and software components. A PLC serves as a centralized control unit, facilitating communication and coordination with motorized components. The system’s communication with the PLC exemplifies its integrability.

Customization: the FVIS is highly customizable, catering to specific processing needs of diverse product families. An automated loading system with specialized grippers allows precise handling, adapting to various product configurations. Carefully selected NC hardware components can be easily customized for different inspection tasks.

Convertibility: the FVIS is remarkably convertible, allowing swift hardware and software changes for varying inspection demands. Vision SW algorithms controlled by a PC enable quick software modifications. The RSS with hardware setup and software reconfiguration modules facilitates rapid changes through a GUI developed in Python and a CAD-based AFR in VB.net.

As for main limitations, the current system also has challenges that may represent new research opportunities. These include a lack of testing on different product types and a lack of a new AFR technique that relies on synthetic rendered images in digitally twinned environments. Additionally, the system has no specific focus on scalability and applying modularity, but these factors can be selected as driving factors in designing other FVIS.

5 Conclusion

The presented article introduces a comprehensive framework for the Flexible Vision Inspection System (FVIS), which is seamlessly supported by a Reconfiguration Support System (RSS). The core concept is based on the RSS generating Reconfiguration files (ReCo) through a user-friendly GUI and Automatic Feature Recognition (AFR) techniques. These ReCo files encompass both hardware parameters, such as sensors and light sources, and software parameters for the vision inspection software, such as position of ROIs and thresholds. This framework facilitates the creation of a unified parameter database that can be shared across multiple production lines and production plants. Extensive experimental testing has been conducted within a laboratory setup, focusing on an automotive half shaft inspection case, aimed at validating the framework’s effectiveness. Subsequently, the framework has undergone further development and transformation into a fully operational system deployed at the factory, where it now serves as an integral component of the daily production processes successfully performing for 4 years. The complexity of the presented case study underscores the promising potential for the broader adoption of the proposed FVIS framework across a wide range of inspection settings. The potential for future research on a fully flexible (generative) approach has been addressed within the proposed flexible and reconfigurable framework [68].