Hostname: page-component-7479d7b7d-qlrfm Total loading time: 0 Render date: 2024-07-12T03:17:46.411Z Has data issue: false hasContentIssue false

Finite-element analysis case retrieval based on an ontology semantic tree

Published online by Cambridge University Press:  14 May 2024

Xuesong Xu
Affiliation:
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou, Zhejiang, China College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, Zhejiang, China
Zhenbo Cheng*
Affiliation:
College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, Zhejiang, China
Gang Xiao
Affiliation:
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou, Zhejiang, China College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, Zhejiang, China
Yuanming Zhang
Affiliation:
College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, Zhejiang, China
Haoxin Zhang
Affiliation:
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou, Zhejiang, China
Hangcheng Meng
Affiliation:
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou, Zhejiang, China
*
Corresponding author: Zhenbo Cheng; Email: czb@zjut.edu.cn
Rights & Permissions [Opens in a new window]

Abstract

The widespread use of finite-element analysis (FEA) in industry has led to a large accumulation of cases. Leveraging past FEA cases can improve accuracy and efficiency in analyzing new complex tasks. However, current engineering case retrieval methods struggle to measure semantic similarity between FEA cases. Therefore, this article proposed a method for measuring the similarity of FEA cases based on ontology semantic trees. FEA tasks are used as indexes for FEA cases, and an FEA case ontology is constructed. By using named entity recognition technology, pivotal entities are extracted from FEA tasks, enabling the instantiation of the FEA case ontology and the creation of a structured representation for FEA cases. Then, a multitree algorithm is used to calculate the semantic similarity of FEA cases. Finally, the correctness of this method was confirmed through an FEA case retrieval experiment on a pressure vessel. The experimental results clearly showed that the approach outlined in this article aligns more closely with expert ratings, providing strong validation for its effectiveness.

Type
Research Article
Copyright
© Zhejiang University of Technology, 2024. Published by Cambridge University Press

Introduction

As a widely used numerical calculation method in the engineering field, finite-element analysis (FEA) technology is of great importance for improving innovation ability, ensuring product quality, and reducing production costs (Hughes, Reference Hughes2012). The FEA process often involves the following stages: problem identification, preprocessing, computation, and post-processing (Wriggers et al., Reference Wriggers, Siplivaya, Joukova and Slivin2007; Xu et al., Reference Xu, Xiao, Lou, Lu, Yang and Cheng2019). At each stage of the analysis process, engineers need to determine many types of decision-making tasks, such as classifying problems, selecting computation model parameters (e.g., geometric simplification, and finite element type and size), determining the numeric algorithm type and parameters, and evaluating the numeric results (Wriggers et al., Reference Wriggers, Siplivaya, Joukova and Slivin2007). The accuracy and reliability of the FEA result are highly dependent on the quality of the decisions made at each stage of the analysis process.

One approach to improve the quality of decision-making is to obtain decision-making-related references from existing solved cases (Zhan et al., Reference Zhan, Jayaram, Kim and Zhu2010; Badin et al., Reference Badin, Chamoret, Chamoret, Gomes and Culley2011; Numthong and Butdee, Reference Numthong and Butdee2012; Kestel et al., Reference Kestel, Gler, Zirngibl, Schleich and Wartzack2019). The references include how to simplify the geometric model, how to determine the type and size of the finite element, how to determine the boundary conditions, and how to select the analysis algorithm. Therefore, the decision-making results in solved cases can help the decision-making process of the current analysis task.

The case-based reasoning (CBR) method that models the process of solving a problem by establishing the analogy relation between the current problem and previously solved problem(s) is proposed to facilitate the use of information from previously solved FEA cases (Wriggers et al., Reference Wriggers, Siplivaya, Joukova and Slivin2007; Zhao et al., Reference Zhao, Cui, Zhao, Qiu and Chen2009; Khan et al., Reference Khan, Chaudhry and Sarosh2014; Khan and Chaudhry, Reference Khan and Chaudhry2015). Wriggers et al. (Reference Wriggers, Siplivaya, Joukova and Slivin2007) proposed a knowledge-based system for the intelligent support of the preprocessing stage of engineering analysis in the contact mechanics domain. Khan and Chaudhry (Reference Khan and Chaudhry2015) proposed an adaptive FEA-integrated system based on the CBR method for mesh selection. Wang and Rong (Reference Wang and Rong2008) presented a CBR method for welding fixture design. According to the above research results, the key tasks of the CBR-based FEA process are representation in terms of FEA cases and retrieval of solved cases that are most similar to the current case.

One of the most promising approaches to represent the FEA case is through the use of ontologies. Yoshioka et al. (Reference Yoshioka, Umeda, Takeda, Shimomura, Nomaguchi and Tomiyama2004) demonstrated a physical ontology-based support system for knowledge-intensive engineering called the Knowledge-Intensive Engineering Framework to integrate multiple engineering models and allow more flexible use of them. Sun et al. (Reference Sun, Ma and Chen2009) proposed an ontology-based framework that included a hierarchy transfer approach and a three-stage automated FEA method for automated FEA to help users to define the appropriate finite element model more easily. Grosse et al. (Reference Grosse, Milton–Benoit and Wileden2005) proposed a formal set of ontologies for classifying analysis modeling knowledge to enable robust knowledge sharing. Xu et al. (Reference Xu, Xiao, Lou, Lu, Yang and Cheng2019) proposed that FEA modeling processes can be expressed as the entities and relations among entities in an ontology tree to obtain the FEA script grammar. These results suggest that the FEA process can be modeled as ontologies in terms of a set of concepts within a domain and the relationships between them.

Another key technology of the CBR-based FEA system is the similarity estimation between FEA cases. A common similarity retrieval method in the engineering field is to express the case using the word level-based method,, such as keyword mathching (Salton et al., Reference Salton, Wong and Yang1975), feature vectors (Korenius et al., Reference Korenius, Laurikkala and Juhola2007), topic extraction, (Lin, Reference Lin2020) and vector space model (VSM) (Figueiras et al., Reference Figueiras, Costa, Paiva, Celson, Ricardo, Joaquim and Jan2012), and then calculate the similarity between the two vectors using the Euclidean or cosine distance (Hu et al., Reference Hu, Wang, Yong, Yong and Paul2013; Ke et al., Reference Ke, Jiang, Zhang, Wang and Zhu2020). The elements in the vector are either keywords or entities within the ontologies. Similarity measures based on VSM have been applied to a variety of tasks, such as the retrieval of information in collaborative engineering projects (Figueiras et al., Reference Figueiras, Costa, Paiva, Celson, Ricardo, Joaquim and Jan2012), relaxed lightweight assembly (Hu et al., Reference Hu, Wang, Yong, Yong and Paul2013) customer demand data for remanufacturing processes (Ke et al., Reference Ke, Jiang, Zhang, Wang and Zhu2020), and the data in IoT (Internet of Things) (Sang et al., Reference Sang, Pang, Zha and Yang2019). However, this type of calculation method lacks the structure information between entities, making it difficult to measure the similarity between the two cases from a semantic point of view. Another option to define similarity measures for cases is through the use of embeddings (Zou et al., Reference Zou, Yang, Zhang, Rehman, Huang, Zhao, Shi, Piva and Kim2020). For example, Xu et al. (Reference Xu, Chen, Zhou, Chang, Ji, Wei and Hou2021) used the word2vec model to obtain semantic information from fault data for fault classification. Cordeiro et al., (Reference Cordeiro, Gomes, Gomes and Texeira2019) proposed that the doc2vec model can be used to measure the semantic similarity of text in the oil and gas domain. Cai et al. (Reference Cai, Palazoglu, Zhang and Hu2019) proposed a deep learning and word embedding method to represent industrial alarm data to predict alarm information. Reimers and Gurevych (Reference Reimers and Gurevych2019) used the Sentence-BERT (SBERT) model to improve the document retrieval system for the supply chain domain (Sant Albors, Reference Sant Albors2021). Furthermore, with the advancement of cross-modal embedding techniques (Rehman et al., Reference Rehman, Tu, Huang and Rehman2018, Reference Rehman, Huang, Tu, Ahmad, Leong and Hady2019), embedding the multimodal data within cases can also lead to further enhancements in case retrieval performance. However, although the embedding method can express context information, structure information between objects in cases is not considered in the similarity calculation.

For these reasons, a semantic similarity calculation method for FEA cases based on ontology is proposed that aims at resolving the difficulty of representing and retrieving solved FEA cases. Considering that FEA models are often encapsulated and difficult to represent directly, the FEA task is used as the index for the FEA case. The FEA tasks are expressed structurally using ontologies. Through instantiating the ontology based on the named entity recognition (NER) method, the structural FEA tasks are constructed automatically. These structural FEA tasks are organized into a semantic tree. By comparing the structural similarity of the semantic trees, a similarity comparison algorithm for FEA cases is proposed. Finally, the most relevant FEA cases can be obtained by comparing the structural similarity of the semantic trees.

Methods

With the accumulation of FEA cases, how to identify the most similar cases from these solved cases has become one of the key technologies in the CBR-based FEA system. An FEA case mainly includes the data of the FEA modeling process that consists of the FEA task, FEA solution, FEA model, and FEA result (Saarelainen et al., Reference Saarelainen, Buda and Juhanko2014; Joshi, Reference Joshi and Ashwin2004). The FEA task is the textual description of the analysis problem, which includes the mechanical device classification, aim of the analysis, contact pair identification, material properties definition, and contact problem properties definition. Therefore, the description of the FEA task can be considered as a requirement of a specific FEA modeling process, and this description is used as the index for the FEA case.

The FEA task description is typically text described in natural language, which needs to be transferred into a structured representation. Based on the structured representation of the FEA task description, an FEA case retrieval method is proposed. Figure 1 provides an insight into the overall process. In the first step, the ontology of the FEA modeling process is constructed and then the structured representation of the FEA task description is obtained by instantiating the ontology according to the task description. Finally, based on semantic tree technologies, a similarity comparison algorithm is proposed.

Figure 1 Framework of FEA case retrieval.

Representation of the description of FEA tasks

The description of FEA tasks is generally in the form of natural language, which is particularly applicable to large-scale enterprises where the process of design and analysis requires the collaboration of several departments (Saarelainen et al., Reference Saarelainen, Buda and Juhanko2014; Nosenzo et al., Reference Nosenzo, Tornincasa, Bonisoli and Brino2014), and to the engineering field where the analysis results need to be checked, such as pressure vessels (Gupta and Vora, Reference Gupta and Vora2014; Niranjana et al., Reference Niranjana, Patel and Dubey2018). To form a representation that can be processed by a computer, first, an ontology of the FEA modeling process is constructed, then text processing technology is used to obtain the entities of the task description, and finally, the ontology is instantiated to form a structured description of the FEA task based on the entities.

An ontology is a clear and accurate description of a conceptualization (Uschold and Gruninger, Reference Uschold and Gruninger1996). An ontology can be easily transformed into a storage form that can be understood by a computer. Additionally, an ontology can express domain knowledge through the semantic definition of terms and axioms. Many research results have demonstrated that an ontology can be used to formally define the FEA task (Wriggers et al., Reference Wriggers, Siplivaya, Joukova and Slivin2007; Sun et al., Reference Sun, Ma and Chen2009). Figure 2 shows the ontology of the FEA task. The ontology is divided into a set of classes, where each class represents information about the product, the aim of the analysis, the material, and the working conditions. The Product information class describes the analysis object, including its equipment name, part name, and design requirements. The Analysis aim class describes the purpose of FEA in the case, such as stress analysis or fatigue analysis. The class of Material and Physical data describes material information and its physical data about the analysis object, such as Material designation and Material characteristic. The Working condition class can be subdivided into the Design condition and Operating condition classes. The Design condition represents the condition used in the product design stage, such as design pressure or design temperature. The Operating condition represents the various loads and constraints that act on the analysis object during the operating state. The class hierarchy forms the ontology, which defines the general terms and state relations between the classes of the FEA modeling process.

Figure 2 Ontology of FEA task.

The semantic representation of the FEA task of the buffer tank of a reciprocating piston compressor is shown in Figure 3. In the figure, each entity is obtained by instantiating the ontology according to the textual FEA task of the buffer tank.

Figure 3 Semantic representation of the FEA task of a buffer tank.

Structural representation of an FEA case

Based on the ontology of the FEA task, the structural description of the FEA problem could be represented by instantiating the ontology. In the instantiation process of the FEA task, the specific entities that are contained in the textual description of the analysis problem are added to the classes of ontology. As described in the previous section, these classes are the leaf nodes of the ontology, which include Equipment name, Part name, Analysis requirement, Analysis aim, Material designation, Material characteristic, Design condition, and Operating condition. To accurately identify the entities in the textual description and their subordinate class, the NER method is adopted to establish the connection between the entity and the class of the ontology.

In this study, the Bert-BiLSTM-CRF model (Xie T et al., Reference Xie, Yang and Liu2020) is adopted to instantiate the FEA task ontology. Initially, the labeled corpus is transformed into the word vector through the BERT pretraining language model. Then the word vector is input into the BiLSTM module for further processing. The conditional random field (CRF) module is used to decode the output result of the BiLSTM module to obtain a predictive annotation sequence. Finally, each entity in the sequence is extracted and classified to complete the entire NER process.

The BIO mode is used to label entities, that is, B (Begin) represents the starting position of an entity, I (Inside) indicates that the word is inside the entity, and O (Outside) indicates that the word does not belong to any entity. For each entity, types are also designed to describe the entity. All entity types are listed in Table 1. These entity types are the classes corresponding to the leaf nodes in the FEA task ontology. Specifically, the entity types in the product information are “Equipment name,” “Part name,” and “Design requirement.” The entity types of the material and physical data are “Material designation” and “Material characteristic.” The entity types of the working conditions are “Design condition” and “Operating condition.”

Table 1. Labels of the FEA task description

For instance, B-PN indicates that the entity is the starting word of the Product name entity. For the sentence “Carrying out stress analysis on the exhaust buffer tank”, the entity labels are shown in Table 2.

Table 2. Example of entity labels

Finally, the identified and classified entities of the FEA problem description are added as individuals to the class of the FEA task ontology. Consider the representation of the description of the FEA problem of the buffer tank as an example, as shown in Figure 4. The top panel of Figure 3 shows the FEA task ontology. Additionally, the bottom panel of Figure 4. shows the textural description of the FEA problem of the buffer tank (part). After NER, the entities are identified, as shown in the green box, and the value of the entity is shown in the orange box.

Figure 4 Instantiation of the FEA task description.

Semantic comparison method

After the FEA task description is instantiated according to the FEA ontology, the tree-structured representation of the task description is obtained. To obtain the semantic similarity comparison between two task descriptions, a comparison method is proposed based on the multi-tree structure (Hajian B et al., Reference Hajian and White2011).

Suppose that two given trees T 1 and T 2, as shown in Figure 5a and Figure 5b, respectively, represent two FEA task descriptions. The main steps in comparing the similarity of the two trees are as follows:

  1. (1) Merge T 1 and T 2 into tree Tm, as shown in Figure 5c.

  2. (2) Obtain the similarity of T 1 and T 2 according to the relationship between each node of tree Tm.

Figure 5 Merge two trees of FEA task description T 1 and T 2 into tree Tm. T 1 represent the FEA case ontology semantic tree 1; T 2 represent the FEA case ontology semantic tree 2; Tm represent the Merged tree between T 1 and T 2.

The algorithm for merging two trees is shown in Algorithm 1. First, tree Tm is created with empty nodes. Then, each node Ni 1 in T 1 is selected from bottom to top, and Ni 1 is compared with each node in tree T 2. If Nj 2 is the same as Ni 1 in T 2, nodes Ni 1 and Nj 2 are combined into a new node. The new node and its child nodes are added to tree Tm. Finally, the merged tree Tm is returned, as shown in Figure 5c dotted (solid) white nodes only come from T 1 (T 2), and blue nodes belong to the two trees.

Algorithm 1. Merge two trees

Input: trees T 1 and T 2

Output: merged tree Tm

T1 _nodes ← T1.all_node()

T2_nodes ← T2.all_node()

T1_nids ← T1_nodes.identifier

T1_nids ← T2_nodes.identifier

Tm ← T1.tree()

add_node ← List()

FOR T1_nid in T1_nids:

IF T1_nid in T2_nids:

Tm _node ← Tm.get_node(T1_nid)

ELSE:

IF T1_nid not in add_node:

parent_nid ← T1.parent(T1_nid)

Tm _node ← Tm.get_node(parent_nid)

NT ← T1.subtree(T1_nid)

NT_nodes ← NT.all_node()

NT_nids ← NT_nodes.identifier

add_node ← add_node + NT_nids

Tm.paste(Tm _node, NT)

RETURN Tm

After the two trees are merged, the similarity between the FEA task descriptions represented by the two trees can be calculated according to the merging results of the two trees. The leaf nodes on the merged tree are derived from the entities of the FEA task document, and the internal nodes on the merged tree represent the conceptual description of the FEA problem. Therefore, each node of the merged tree can be traversed according to the bottom-up process, and the corresponding combination value of each node can be calculated in turn. The combination value of the root node of the merged tree is the similarity result.

The combination value of nodes in the merged tree is calculated, respectively, according to their node types. There are three types of nodes in the merged tree: leaf nodes, internal nodes, and the root node. Additionally, the value of the FEA task can be represented as follows: $ {v}_{\mathrm{ss}} $ stands for a single scalar value; $ {v}_{\mathrm{mm}} $ is a min–max scalar value that has a maximum scalar value ( $ {v}_{\mathrm{mm}}^{\mathrm{max}} $ ) and minimum scalar value ( $ {v}_{\mathrm{mm}}^{\mathrm{min}} $ ); $ {v}_{\mathrm{str}} $ is a string-type value. The single scalar value is the most widespread, such as pressure, temperature, and so forth The range value represents the minimum and maximum type value needed to describe a boundary such as 5–35 °C. The string-typed value is used for non-dimension-related values such as material designations.

When the type of node ni belongs to the leaf node of the merged tree, the combination value is calculated using Eq. (1). Node ni on the merged tree is from tree T 1 and tree T 2, which indicates that the two nodes of tree T 1 and tree T 2 are exactly the same; hence, the combination value of node ni is set to 1. When the nodes are only from tree T 1 and tree T 2, the combination value of the node ni is set to 0:

(1) $$ {V}^1\left({n}_i\right)=\left\{\begin{array}{c}1,\mathbf{if}\;{n}_i\in {T}_1\;\mathbf{and}\;{n}_i\in {T}_2\qquad \\ {}0,\mathbf{otherwise}.\hskip2.4em \qquad \end{array}\right. $$

When the type of node ni belongs to the internal node of the merged tree, the combination value V 2(ni) of node ni is calculated using Eqs. (2) and (3). The calculation of V 2(ni) includes two parts. The first part is calculating the average of the combination values of all the children’s nodes of ni. The second part is related to the type of node ni; the value V 1(ni) in Eq. (2) is given by Eq. (1). In Eq. (2), $ \alpha $ is the adjustment factor, in this article $ \alpha $ was set to e = 2.71. The weight between node ni and node ci is calculated using the function weight(ni, ci) in Eq. (3), which indicates the importance of the relationship between nodes to the similarity calculation. Intuitively, the node on the merged tree belongs to tree T 1 and tree T 2, and the greater the depth of this node, the greater the combination value of this node. Additionally, the greater the depth of the node on the merged tree, the smaller the average value passed to the node by each child node of the node:

(2) $$ {V}^2\left({n}_i\right)=\left(1-\frac{1}{\alpha^{height\left({n}_i\right)}}\right)A\left({n}_i\right)+\left(\frac{1}{\alpha^{height\left({n}_i\right)}}\right){V}^1\left({n}_i\right) $$
(3) $$ A\left({n}_i\right)=\frac{1}{\mid children\left({n}_i\right)\mid}\sum \limits_{\forall {c}_j\in children\left({n}_i\right)} weight\left({n}_i,{c}_j\right){V}^2\left({c}_j\right). $$

The height function of ni in Eq. (2) recursively calculates the height of node ni, and the calculation function is given by

(4) $$ height\left({n}_i\right)=\left\{\begin{array}{c}0,\hskip15em \boldsymbol{if}\;{n}_i\; is\ leaf\ node\;\qquad \\ {}\max \left[ height\left({c}_i\right)\right],{c}_i\in children\left({n}_i\right),\;\boldsymbol{otherwise}.\qquad \end{array}\right. $$

When the type of node ni belongs to the root node of the merged tree, the combination value can be obtained by calculating the average value of the combination value of each child node of the root node using Eq. (3). The combination value of the root node is the similarity between tree T 1 and tree T 2.

Additionally, the measuring equations (Mun D et al., Reference Mun and Ramani2011) are defined, as shown in Eq. (5), to calculate the similarity of the values.

(5) $$ Sim\left({v}_1,{v}_2\right)\left\{\begin{array}{c}1-\frac{\left|\left|{v}_{ss1}\right|-\left|{v}_{ss2}\right|\right|}{\left|{v}_{ss1}+{v}_{ss2}\right|}\hskip6.0em \boldsymbol{if}\;{v}_{ss1}\ne {v}_{ss2}\hskip2.04em \qquad \\ {}1,\hskip15.88em \boldsymbol{if}\;{v}_{mm}^{min}\le {v}_{ss}\le {v}_{mm}^{max}\hskip0.24em \qquad \\ {}1,\hskip15.879996em \boldsymbol{if}\;{v}_{str1}={v}_{str2}\hskip1.8em \qquad \\ {}0,\hskip15.879996em \boldsymbol{otherwise}\hskip2.639999em \qquad \end{array}\right. $$

Algorithm 2. Similarity of the two trees based on their merged tree

Input: trees T1 and T2, merged tree Tm

Output: similarity of T1 and T2

node ← ni

stack ← empty stack

lastNodeVisited ← null

while not stack.isEmpty() or node ≠ null

if node ≠ null

stack.push(node)

node ← node.left

else

peekNode ← stack.peek()

if peekNode.nextchild ≠ null and lastNodeVisited ≠ peekNode.nextchild

node ← peekNode.nextchild

else

if peekNode == LeafNode:

if LeafNode.value ≠ null:

V[peekNode.value] = Eq. (5)

V[peekNode] = V[peekNode.value] * Eq. (1)

else:

V[peekNode] = Eq. (1)

elif peekNode == RootNode:

V[peekNode] = Eq. (2)

else:

V[peekNode] = Eq. (3)

lastNodeVisited ← stack.pop()

The algorithm for the similarity of two trees based on their merged tree is shown in Algorithm 2. The algorithm traverses each node of the merged tree in turn according to the post-order traversal process and then calculates the combination value. The combination values are propagated from bottom to top in layers on the merged tree to integrate the relationship between the entity and the ontology into the similarity calculation; that is, if two trees representing FEA task descriptions have different entities, but these entities belong to the same class in the ontology, the similarity of the two trees can be improved through the above calculation process. The combination values and the calculation formulas of the nodes in the merged tree Tm are shown in Figure 5c.

Example

Case study

The proposed method was validated using the typical pressure vessel of the reciprocating piston compressor (Figure 6) as an example. The reciprocating piston compressor is one of the most widely used items of process equipment in the field of natural gas compression (Ribas et al., Reference Ribas, Deschamps, Fagotti, Morriesen and Dutra2008; Farzaneh-Gord et al., Reference Farzaneh-Gord, Niazmand, Deymi-Dashtebayaz and Rahbari2015). Among the equipment in the reciprocating piston compressor, the buffer tank is the most common structure used to reduce the pulsation of gas flow. During the design process for the buffer tank using the design-by-analysis methodology, the preliminary design (including the product model, material, and dimensions) needs to be continuously adjusted according to the result of the FEA, such as changing the material, adjusting the wall thickness, and optimizing the structure.

Figure 6 Geometry model of the buffer tank.

The basic FEA modeling process can be divided into three phases: preprocessing, solving, and post-processing. Specifically, the analyst is required to perform operations that include constructing the geometry model, simplifying the model, ensuring the analysis method, setting the boundary according to the working condition, setting the material, meshing, choosing the solver, outputting the analysis result data, evaluating the results according to standards, and writing the analysis report. Several analytical operations are involved in decision-making, such as simplifying the model, choosing the analysis method, meshing, and choosing the solver. These analytical operations require a great deal of accumulated expertise, which makes it possible for analysts to improve efficiency and reduce the difficulty of analysis modeling by reusing the decision knowledge of FEA modeling in similar FEA cases. To retrieve a similar FEA case, the method proposed in Section title “Methods” is adopted to structurally represent the textual description of the FEA task for the buffer tank. The textual description is shown in the bottom panel of Figure 7. After the NER method is applied to instantiate the FEA task ontology according to the textual description of the FEA task, a structural FEA task of the buffer tank is generated, as shown in Figure 8. Subsequently, the analyst can use the represented FEA task to retrieve a similar FEA case.

Figure 7 FEA task representation.

Figure 8 Existing solved case provides references for a new analysis task.

Based on the retrieval result, the FEA operations of the retrieved case can provide references for the new task (Figure 8). For instance, the solving method of the new task can consider using the sparse director solver, and the element type of Solid82 can also be considered for the new task. Therefore, retrieving the case similar to the current task from the existing cases” library can provide important references for the decision-making of the FEA process.(Detailed examples of reusing retrieved FEA cases are provided in Tables A2 to A4 of the Appendix).

Performance of the similarity between FEA cases

To validate the proposed method for retrieving FEA cases, 396 FEA cases in the pressure vessel field were built, where each case contained a problem description, FEA scheme, FEA model, and FEA result. Additionally, the problem description of each FEA case was structurally represented and used as the index of the case. The typical FEA cases used in this study are shown in the Appendix.

Considering the difficulty of directly evaluating the pros and cons of FEA case semantic retrieval methods, a dataset was constructed, which included several FEA case pairs. These FEA case pairs were obtained by matching the FEA cases in pairs randomly. Specifically, all the 396 FEA cases were cross-matching and duplication eliminated, and 78,210 FEA case pairs were obtained (all data set used in this article can be downloaded from https://github.com/song885280/FEASimData). Then, the similarity of the FEA case pairs was scored by four graduate students in the mechanical engineering department who had FEA capabilities. The similarity of each FEA case pair was scored in three degrees: similar, generally similar, and dissimilar. The degrees of similarity are scored according to the reusability of the retrieval FEA cases. The reusability of the FEA cases was considered to be the most relevant for the analysis purpose, followed by product information and working conditions. Table 3 shows the principle for scoring the similarity of FEA cases. We use a real number between 0 and 1 to measure the similarity of cases. That is, the number between 0.75 and 1 represents “Similar”, the number between 0.45 and 0.75 represent “Generally similar”, and the number less than 0.45 and greater than 0 represents “Dissimilar”.

Table 3. The principle for scoring the similarity of FEA cases

The average error (Manning et al., Reference Manning, Raghavan and Sch Tze2010) was used to evaluate the retrieval performance. The formula is as follows:

(6) $$ E=\frac{1}{N}\sum \limits_{j=1}^N\left|{S}_j-{S}_j^p\right|, $$

where E is the average error, N is the number of case pairs, j is the index of the case pairs, j = 1, 2, 3,…, N; S is the case similarity obtained by the algorithm; and $ {S}^p $ is the case similarity scored by the 4 graduate students in Mechanical Engineering. They all have obtained the qualification of FEA granted by the Zhejiang University of Technology.

In order to compare the retrieval performance, three types of retrieval methods were performed: (1) the VSM (Figueiras et al., Reference Figueiras, Costa, Paiva, Celson, Ricardo, Joaquim and Jan2012); (2) the method of sentence BERT (SBERT) Reference Kallmeyer and Osswald2013 and (3) the method of multitree without weight. The performance of these methods is shown in Table 4, and the relative performance of each retrieval model is compared graphically in Figure 9. The average errors demonstrate that the VSM outperforms the SBERT and the multi-tree outperforms the VSM. The proposed method (multi-tree with weight) outperforms the method of multi-tree without weight. The weights (Eq. 3) of four sub-nodes of the root node are set as follows: 0.25, 0.5, 0.075, and 0.175. The main reason for the high performance of the multi-tree is the method’s ability to measure the similarity between FEA cases at the semantic level. The semantic representation of FEA cases can be expressed by instantiating FEA ontology. On the other hand, this result points out the limitation of keyword-based searches of FEA cases. In addition, the Bert-based method cannot achieve a better method when three are no large-scale training corpus.

Table 4. Average error of each task and total average error

Figure 9 The average error of each method for the similarity measurement of FEA cases. VSM represent the method of vector space model; SBERT represent the method of sentence BERT; Multi_Tree represent the method of muilt-tree without weight; Multi_Tree_Weight represent the method of muilt-tree with weight.

Discussion

This article proposed to index the textual description of the analysis problem to retrieve the FEA cases. Compared with the existing retrieval methods, VSM and SBERT, our method can not only improve the efficiency of FEA case retrieval but also improve the accuracy of FEA case retrieval. VSM is a widely used method for text retrieval in the engineering domain (Figueiras et al., Reference Figueiras, Costa, Paiva, Celson, Ricardo, Joaquim and Jan2012; Hu et al., Reference Hu, Wang, Yong, Yong and Paul2013; Ke et al., Reference Ke, Jiang, Zhang, Wang and Zhu2020; Sang et al., Reference Sang, Pang, Zha and Yang2019). With the development of natural language processing technology, The STOA model of SBERT (Reimers and Gurevych, Reference Reimers and Gurevych2019) has achieved superior performance in many tasks. The SBERT can embed the text into the vector, which can be used to measure the similarity of text. However, the SBERT does not work well in this article. The reason is that the SBERT model needs a large-scale training corpus, whereas the domain covered in this article lacks the training corpus. That is why we implement an ontology-based method to measure the similarity of FEA cases. To avoid the complexity of the ontology for the complete FEA cases, which include analysis problem, geometric model, and analysis report, we build the ontology of the FEA problem description. The FEA problem description mainly includes the analysis objectives, the materials, and the working conditions (Wriggers et al., Reference Wriggers, Siplivaya, Joukova and Slivin2007; Khan and Chaudhry, Reference Khan and Chaudhry2015). The text description of analysis problems can be expressed semantically by instantiating FEA ontology. This representation method can accurately extract the key variables that describe the FEA analysis process contained in the text description, and form a tree structure index according to the FEA ontology. Some related research (Han et al., Reference Han and Hedberg2008; Morinaga et al., Reference Morinaga, Arimura, Ikeda, Yosuke, Susumu and Robert2005; Kallmeyer and Osswald, Reference Kallmeyer and Osswald2013; Plank and Moschitti, Reference Plank, Moschitti, Hinrich, Pascale and Massimo2013) has shown that the tree structure can fuse semantic information. The tree structure of the FEA case not only contains the keywords but also represents the relation between the keywords. Certainly, if the data of the CAD model can be integrated into the index description, we believe it can improve the accuracy of FEA case retrieval.

We also designed a simple algorithm to calculate the similarity between two FEA cases based on the multi-tree data structure. The multi-tree measures similarity based on hierarchical relations that exist between attributes of the entities in an ontology. This method provides a useful tool to embed the domain knowledge of domain engineers into the retrieval model. The main reason that our method can improve retrieval performance is that the key factors affecting the similarity of FEA cases are embedded in the weight of multi-tree nodes. It should be noted that these weights are currently given by domain experts. In the follow-up study, we will consider using the machine learning algorithm to automatically calculate these weights. In addition, although the source language is assumed as Chinese, the proposed method is language-independent. We have translated all the corpus of FEA cases into English and given the results of case similarity to verify the generalizable of our method. The experiment results (Figure 9b) show that our method has also obtained good performance in these English datasets.

The purpose of FEA case retrieval is to reuse the cases. The existing research on FEA case reuse was focused on automatic FEA modeling by the method of CBR (Wriggers et al., Reference Wriggers, Siplivaya, Joukova and Slivin2007; Numthong and Butdee, Reference Numthong and Butdee2012; Khan and Chaudhry, Reference Khan and Chaudhry2015). However, the application of these methods is highly dependent on the relevance of the case to the modeling task. For example, it is difficult to construct the FEA model by CBR if the geometric models of the case and the modeling task are highly dissimilar, which has prevented the application of the FEA case reuse methods. Therefore, this article proposes to improve the reusability of FEA cases by retrieving the FEA cases efficiently and accurately. In a retrieved FEA model, many components can be reused, such as model simplification methods, meshing methods, and post-processing methods, and so forth After obtaining the relevant case, the analyst can quickly complete the current FEA modeling task by consulting the reusable information in the cases. The proposed approach not only exploits the value of the knowledge contained in the FEA cases but also can offer support to less experienced simulation users.

Conclusion

This article proposed an approach for FEA case retrieval by taking the textual description of the analysis problem as the index. The analysis problem document generated in the early analysis stage contains abstract analysis descriptions. The structural analysis problem is expressed semantically by instantiating the FEA ontology. Based on the tree structure of the analysis problem, the similarity between two FEA cases is calculated by using the multi-tree-based similarity comparison algorithm. The experimental results clearly show that the proposed approach outperforms the compared existing retrieval methods. This demonstrated that the semantic representations of the text description of analysis problems can be captured accurately by instantiating FEA ontology.

Although the proposed method has significant advantages in FEA case retrieval, some limitations remain, such as (1) the textual description of the analysis problem does not include the geometric model, and (2) the ontologies involved in the experiment in the study were constructed manually. Therefore, in follow-up research, the retrieval of the geometric model will be integrated into case retrieval to further improve the retrieval accuracy. Furthermore, the automatic construction of structured FEA case ontology, as well as conducting CBR or parameterizing FEA modeling on the retrieved cases, are both areas worthy of in-depth research.

Acknowledgments

This work was supported by the National Natural Science Foundation of China [No. 61976193, 52205291] and the Science and Technology Key Research Planning Project of Zhejiang Province, China (No. 2021C03136, 2023C01215, 2023C01022). The authors would like to thank Maxine Garcia, Ph.D., from Liwen Bianji (Edanz) (www.liwenbianji.cn/) for editing the English text of a draft of this article.

Appendix

Table A1. Example of FEA case retrieval

Table A2. Example 1 of reusing retrieved FEA case

Table A3. Example 2 of reusing retrieved FEA case

Table A4. Example 3 of reusing retrieved FEA case

References

Badin, J, Chamoret, D, Chamoret, D and Gomes, S (2011) Knowledge configuration management for product design and numerical simulation. In Culley, S (ed.), Proceedings of the 18th International Conference on Engineering Design (ICED 11). Denmark, August 15–18.Google Scholar
Cai, S, Palazoglu, A, Zhang, L and Hu, J (2019) Process alarm prediction using deep learning and word embedding methods. ISA Transactions 85, 274283.CrossRefGoogle ScholarPubMed
Cordeiro, FC, Gomes, DDSM, Gomes, FAM and Texeira, RC (2019) Technology intelligence analysis based on document embedding techniques for oil and gas domain. In Proceedings of the Offshore Technology Conference. Brazil, October 29–31.Google Scholar
Farzaneh-Gord, M, Niazmand, A, Deymi-Dashtebayaz, M and Rahbari, H (2015) Thermodynamic analysis of natural gas reciprocating compressors based on real and ideal gas models. International Journal of Refrigeration 56, 186197.CrossRefGoogle Scholar
Figueiras, P, Costa, R, Paiva, L, Celson, L and Ricardo, J (2012) Information retrieval in collaborative engineering projects: A vector space model approach. In Joaquim, F and Jan, D (eds.), Proceedings of the International Conference on Knowledge Engineering and Ontology Development. Spain, October 4–7.Google Scholar
Grosse, IR, Milton–Benoit, JM and Wileden, JC (2005) Ontologies for supporting engineering analysis models. Artificial Intelligence for Engineering Design Analysis and Manufacturing 19(1), 118.CrossRefGoogle Scholar
Gupta, SR and Vora, CP (2014) A review paper on pressure vessel design and analysis. International Journal of Engineering Research & Technology 3(3), 17.Google Scholar
Hajian, B and White, T (2011) Measuring semantic similarity using a multi-tree model. In Proceedings of the ITWP@ IJCAI.Google Scholar
Han, C-H and Hedberg, N (2008) Syntax and semantics of IT-CLEFTS: A tree adjoining grammar analysis. Journal of Semantics 25(4), 345380.CrossRefGoogle Scholar
Hu, K-M, Wang, B, Yong, J-H, Yong, J and Paul, J (2013) Relaxed lightweight assembly retrieval using vector space model. Computer-Aided Design 45(3), 739750.CrossRefGoogle Scholar
Hughes, TJ (2012) The Finite Element Method: Linear Static and Dynamic Finite Element Analysis. Courier Corporation.Google Scholar
Joshi, AA (2004) CAE data management using traditional PDM systems. In Ashwin, A (ed.), Proceedings of the International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. USA, September 28–October 2.Google Scholar
Kallmeyer, L and Osswald, R (2013) Syntax-driven semantic frame composition in lexicalized tree adjoining grammars. Journal of Language Modelling 1(2), 267330.Google Scholar
Ke, C, Jiang, Z, Zhang, H, Wang, Y and Zhu, S (2020) An intelligent design for remanufacturing method based on vector space model and case-based reasoning. Journal of Cleaner Production 277, 123269.CrossRefGoogle Scholar
Kestel, PK, Gler, P, Zirngibl, C, Schleich, B and Wartzack, S (2019) Ontology-based approach for the provision of simulation knowledge acquired by data and text mining processes. Advanced Engineering Informatics 39(22), 292305.CrossRefGoogle Scholar
Khan, AA and Chaudhry, IA (2015) Object oriented case representation for CBR application in structural analysis. Applied Artificial Intelligence 29(4), 335352.Google Scholar
Khan, AA, Chaudhry, IA and Sarosh, A (2014) Case based reasoning support for adaptive finite element analysis: mesh selection for an integrated system. Applied Physics Research 6(3), 2139.CrossRefGoogle Scholar
Korenius, T, Laurikkala, J and Juhola, M (2007) On principal component analysis, cosine and Euclidean measures in information retrieval. Information Sciences 177(22), 48934905.CrossRefGoogle Scholar
Lin, K-S (2020) A case-based reasoning system for interior design using a new cosine similarity retrieval algorithm. Journal of Information and Telecommunication 4(1), 91104.CrossRefGoogle Scholar
Manning, C, Raghavan, P and Sch Tze, H (2010) Introduction to information retrieval. Natural Language Engineering 16(1), 100103.Google Scholar
Morinaga, S, Arimura, H, Ikeda, T, Yosuke, S and Susumu, A (2005) Key semantics extraction by dependency tree mining. In Robert, G (ed.), Proceedings of the Eleventh ACM SIGKDD International Conference on Knowledge Discovery in Data Mining. USA, August 21–24.CrossRefGoogle Scholar
Mun, D and Ramani, K (2011) Knowledge-based part similarity measurement utilizing ontology and multi-criteria decision making technique. Advanced Engineering Informatics 25(2), 119130.CrossRefGoogle Scholar
Niranjana, S, Patel, SV and Dubey, AK (2018) Design and analysis of vertical pressure vessel using ASME code and FEA technique. In Lokesha (ed.), Proceedings of the IOP Conference Series: Materials Science and Engineering. India, March 2–3.Google Scholar
Nosenzo, V, Tornincasa, S, Bonisoli, E and Brino, M (2014) Open questions on product lifecycle management (PLM) with CAD/CAE integration. International Journal on Interactive Design and Manufacturing (IJIDeM) 8(2), 91107.CrossRefGoogle Scholar
Numthong, C and Butdee, S (2012) The knowledge based system for forging process design based on case-based reasoning and finite element method. Applied Science and Engineering Progress 5(2), 4554.Google Scholar
Plank, B and Moschitti, A (2013) Embedding semantic similarity in tree kernels for domain adaptation of relation extraction. In Hinrich, S, Pascale, F and Massimo, P (eds.), Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics. Bulgaria, August 4–9.Google Scholar
Rehman, S, Tu, S, Huang, Y and Rehman, O (2018) A benchmark dataset and learning high-level semantic embeddings of multimedia for cross-media retrieval. IEEE Access 6, 6717667188.CrossRefGoogle Scholar
Rehman, S, Huang, Y, Tu, S and Ahmad, B (2019) Learning a semantic space for modeling images, tags and feelings in cross-media search. In Leong, H and Hady, W (eds.), Proceedings of the Trends and Applications in Knowledge Discovery and Data Mining: PAKDD 2019 Workshops. China, April 14–17.Google Scholar
Reimers, N and Gurevych, I (2019) Sentence-bert: Sentence embeddings using siamese bert-networks. Preprint, arXiv:1908.10084.CrossRefGoogle Scholar
Ribas, FA, Deschamps, CJ, Fagotti, F, Morriesen, A and Dutra, T (2008) Thermal analysis of reciprocating compressors: A critical review.Google Scholar
Saarelainen, T, Buda, A and Juhanko, J (2014) Open loops in CAE data management and simulation-based design. International Journal of Product Lifecycle Management 7(4), 318339.CrossRefGoogle Scholar
Salton, G, Wong, A and Yang, CS (1975) A vector-space model for information retrieval. Communications of the ACM 18(11), 613620.CrossRefGoogle Scholar
Sang, J, Pang, S, Zha, Y and Yang, F (2019) Design and analysis of a general vector space model for data classification in internet of things. EURASIP Journal on Wireless Communications and Networking 2019(1), 110.Google Scholar
Sant Albors, I (2021) Improvement of a document retrieval system for the supply chain domain. Universitat Politècnica de Catalunya.Google Scholar
Sun, W, Ma, Q and Chen, S (2009) A framework for automated finite element analysis with an ontology-based approach. Journal of Mechanical Science & Technology 23(12), 32093220.CrossRefGoogle Scholar
Uschold, M and Gruninger, M (1996) Ontologies: Principles, methods and applications. Knowledge Engineering Review 11(2), 93136.CrossRefGoogle Scholar
Wang, H and Rong, YM (2008) Case based reasoning method for computer aided welding fixture design. Computer-Aided Design 40(12), 11211132.CrossRefGoogle Scholar
Wriggers, P, Siplivaya, M, Joukova, I and Slivin, R (2007) Intelligent support of engineering analysis using ontology and case-based reasoning. Engineering Applications of Artificial Intelligence 20(5), 709720.CrossRefGoogle Scholar
Xie, T, Yang, J and Liu, H (2020) Chinese entity recognition based on BERT-BiLSTM-CRF model. Computer Systems & Applications 29(7), 4855 (in Chinese).Google Scholar
Xu, X, Xiao, G, Lou, G, Lu, J, Yang, J and Cheng, Z (2019) Flexible parametric FEA modeling for product family based on script fragment grammar. Computers in Industry 111, 1525.CrossRefGoogle Scholar
Xu, Z, Chen, B, Zhou, S, Chang, W, Ji, X, Wei, C and Hou, W (2021) A text-driven aircraft fault diagnosis model based on a Word2vec and priori-knowledge convolutional neural network. Aerospace 8(4), 112.CrossRefGoogle Scholar
Yoshioka, M, Umeda, Y, Takeda, H, Shimomura, Y, Nomaguchi, Y and Tomiyama, T (2004) Physical concept ontology for the knowledge intensive engineering framework. Advanced Engineering Informatics 18(2), 95113.CrossRefGoogle Scholar
Zhan, P, Jayaram, U, Kim, O and Zhu, J (2010) Knowledge representation and ontology mapping methods for product data in engineering applications. Journal of Computing and Information Science in Engineering 10(2), 114.CrossRefGoogle Scholar
Zhao, J, Cui, L, Zhao, L, Qiu, T and Chen, B (2009) Learning HAZOP expert system by case-based reasoning and ontology. Computers & Chemical Engineering 33(1), 371378.CrossRefGoogle Scholar
Zou, J, Yang, Z, Zhang, S, Rehman, S and Huang, Y (2020) High-performance linguistic steganalysis, capacity estimation and steganographic positioning. In Zhao, X, Shi, Y, Piva, A and Kim, H (eds.), Proceedings of the International Workshop on Digital Watermarking. Australia, November 25–27.Google Scholar
Figure 0

Figure 1 Framework of FEA case retrieval.

Figure 1

Figure 2 Ontology of FEA task.

Figure 2

Figure 3 Semantic representation of the FEA task of a buffer tank.

Figure 3

Table 1. Labels of the FEA task description

Figure 4

Table 2. Example of entity labels

Figure 5

Figure 4 Instantiation of the FEA task description.

Figure 6

Figure 5 Merge two trees of FEA task description T1 and T2 into tree Tm. T1 represent the FEA case ontology semantic tree 1; T2 represent the FEA case ontology semantic tree 2; Tm represent the Merged tree between T1 and T2.

Figure 7

Figure 6 Geometry model of the buffer tank.

Figure 8

Figure 7 FEA task representation.

Figure 9

Figure 8 Existing solved case provides references for a new analysis task.

Figure 10

Table 3. The principle for scoring the similarity of FEA cases

Figure 11

Table 4. Average error of each task and total average error

Figure 12

Figure 9 The average error of each method for the similarity measurement of FEA cases. VSM represent the method of vector space model; SBERT represent the method of sentence BERT; Multi_Tree represent the method of muilt-tree without weight; Multi_Tree_Weight represent the method of muilt-tree with weight.

Figure 13

Table A1. Example of FEA case retrieval

Figure 14

Table A2. Example 1 of reusing retrieved FEA case

Figure 15

Table A3. Example 2 of reusing retrieved FEA case

Figure 16

Table A4. Example 3 of reusing retrieved FEA case