Introduction

Product conceptual design plays an important role in the product lifecycle, which determines product’s primary cost with a small investment1. Defects caused by insufficient product conceptual design are difficult to be remedied in the manufacturing and maintenance stages. This stage starts from the customer requirements analysis, then gradually realizes the mapping from product functional to physical structure, and obtains the design scheme through evaluation and optimization in final2. Customer-centered product design philosophy is widely recognized by manufacturing enterprises nowadays. Therefore, narrowing the gap between product design and customer requirements is a pivotal goal from beginning to end. Previous published studies conduct customer investigations by questionnaire or interview to gather data for analyzing customer requirements. These offline methods are complicated and time-consuming. For the past few years, a large quantity of literature has researched the extraction of customer requirements from online comments3,4. However, traditional studies failed to consider customer requirements representation in the analogical reasoning environment so that it is not effortless to gain latent and innovative customer requirements. Some studies have indicated that accommodating customers into the analogical reasoning environment is essential5,6. In addition to explicit customer requirements, latent customer requirements are extremely crucial to product innovation and success. Understanding and acquisition of latent customer requirements have prominent effect on satisfying customers. While explicit customer requirements tend to be self-evident as long as we can obtain raw data like questionnaire and interview, latent customer requirements are usually hidden in the semantics of customer requirements information and customers may not be distinctly aware of them. Analogical reasoning in the product conceptual design is the process of solving current design problems based on the solutions of past design problems5. Design process can be supported using analogical stimuli by assisting participants to overcome fixation and generate abundant solutions with more positive characteristics during ideation. In the customer requirements analysis stage, customers have established a preliminary perceptual cognition when they interact with product function and structure. If customers are placed in the analogical reasoning environment, the product feedback information is processed ulteriorly in their brain according to the given analogical stimuli, invisible and creative product requirements existing in the customer’s brain can be discovered. Meanwhile, it is inevitable that analogical stimuli results are not always positive and can be incorrect in the far-domain stimuli environment especially. In order to ensure a maximal utility for analogical stimuli, near-domain stimuli are provided to guarantee the feasibility and usefulness of the customer requirements and far-domain stimuli are selected to assure the novelty of the customer requirements. Besides, the collected customer requirements should be carefully evaluated and filtered by the domain experts.

VPA is an effective research paradigm that can reflect the cognitive process by thinking aloud7,8. The subjects elaborate their detailed thinking content when performing a specific cognitive task truthfully. The obtained recordings are converted into texts so that the thinking process of subjects can be displayed explicitly. Besides, an analogy-inspired VPA experiment is a practical engineering method, which acquires inspiration for current cognition task by utilizing the solution space of other domains9,10. In this paper, the subjects express the functional, behavioral and structural customer requirements of elevator based on the existing design scheme and the visual stimuli including near and far domain. The specific definitions of function, behavior and structure are demonstrated here11. Firstly, the function represents the abstract description about product’s working capability. Then, the behavior indicates the scientific principle of realizing the function. Lastly, structure refers to the physical entity that produces the desired behavior and performs the predetermined function. Designers always follow the function-behavior-structure design process model, which helps designers to solve specific design tasks by describing the relationship among function, behavior and structure11. The reason why customer requirements are classified into the above three categories is that it is beneficial to express customer requirements in accordance with the cognition of designers. Hence, this can not only discover the potential product internal requirements involving function, behavior, structure as well as associated constraints, but also accelerate the subsequent solving process of product design schemes for designers. There are plentiful common features between near-domain stimuli and design problem, and they all come from the same or similar fields. It is easier for subjects to propose feasible customer requirements. On the contrary, there are few common features between far-domain stimuli and design problem, which contributes to the generation of innovative customer requirements9. Hence, the text data transformed from VPA data is segmented with natural sentences as the unit and then input into the established BERT deep transfer model. The functional, behavioral and structural customer requirements are classified by fine-tuning the BERT deep transfer model. In fact, a large amount of requirement texts is hard to provide direct assistance to designer after the customer requirements classification. Thus, acquiring the practicable customer requirements is actually a topic analysis task. Functional customer requirements representing customer intention maximally are chosen as the main data source of customer requirements mining. Topic modeling is carried out for the functional requirement semantics. Aiming at existing problems such as the frequency representation of document-word matrix and the empirical selection of topic quantity about traditional latent Dirichlet allocation, ILDA is proposed to enhance the application efficacy in the customer requirements mining.

To overcome the limitations of prior approaches, this paper proposed a semantic analysis-driven customer requirements mining method for product conceptual design based on deep transfer learning and ILDA. In comparison with the existing studies, our work makes three distinctive contributions: (1) An analogy-inspired VPA experiment providing near-domain and far-domain stimuli is presented for obtaining feasible and innovative customer requirement descriptions of elevator. Latent customer requirements are usually hidden in the semantics of customer requirements information and customers may not be distinctly aware of them. Traditional studies failed to consider customer requirements representation in the analogical reasoning environment so that it is not effortless to gain potential and innovative customer requirements. (2) A BERT deep transfer model is established to realize the customer requirements classification among functional domain, behavioral domain and structural domain by fine-tuning training. It can be seen from the final results that satisfying overall classification accuracy and well-pleasing efficacy for Precision, Recall and F1 can be obtained. (3) The ILDA is proposed to mine the functional customer requirements representing customer intention maximally. The optimal topic quantity is determined based on the proposed measurable indicator Perplexity-AverKL. The implicit customer requirements of elevator can be elaborated by topic-word distribution explicitly. In general, this work can provide a novel research perspective on customer requirements mining for product conceptual design by analogy-inspired VPA experiment and natural language processing algorithms.

The overall implementation framework is displayed in Fig. 1. The rest of this paper is organized as follows. “Literature review” section reviews the related work. “Method” section illustrates the customer requirements classification based on BERT and customer requirements mining based on ILDA. A case study is presented in “Case study” section. Finally, “Conclusions” section concludes our work.

Figure 1
Figure 1
Full size image

The overall implementation framework.

Literature review

This Section is closely divided into three parts. The one subsection describes the research situation of customer requirements classification, and another subsection introduces the deep transfer learning in the natural language processing, and a third subsection elaborates the customer requirements mining.

Customer requirements classification

Due to the diversity, dynamics and fuzziness of customer requirement semantics, it is inevitable to classify them systematically in order to understand and further analyze them. Scholars at home and abroad have studied it from different perspectives. Kano model as well as its derivatives is an available requirements analysis tool, which distinguishes the different nonlinear relationships between customer requirements fulfillment and customer satisfaction12. Xu et al.13 presented an analytical Kano model to classify functional requirements into logical groups. This leads to an optimal trade-off between customer classification and producer capability. Lou et al.14 proposed a data-driven approach for customer requirements discernment via Kano model, intuitionistic fuzzy sets theory and electroencephalogram technology. The vagueness of requirements is handled at the semantic expression and neurocognitive level. Shi et al.15 utilized big data of online customer reviews and improved Kano model to classify customer requirements accurately and efficiently. Polynomial modeling and least square methods are adopted to define customer satisfaction and function implementation of customer requirements. Customer requirements are classified based on the slope of the fitted function curves. In addition, customer requirements can generally be divided into dominant and implicit requirements. Obviously, whether enterprises can meet the implicit requirements becomes an important consideration for improving product quality and retaining customers. Xi et al.16 used the triangular fuzzy sets to realize fuzzy semantic quantization of customers and constructed the implicit requirements classification model based on self-organizing mapping neural network. Onyeka et al.17 developed a software tool called COTIR that integrates commonsense knowledge, ontology knowledge and text mining for implicit requirements identification. As a matter of fact, customer requirements can be divided into functional, behavioral and structural requirements. Function-Behavior-Structure design process model is a general design solution framework, which assists designers to solve the design task by describing the relationship among product function, behavior and structure18. The cognition of designers still follows the mapping process corresponding to the functional domain, behavioral domain and structural domain. Therefore, the customer requirements expression are satisfactory when they are consistent with the cognition of designers. Latent product functional, behavioral and structural requirements are obtained through an analogy-inspired VPA experiment. Those feasible and innovative customer requirements will provide support for designers.

Deep transfer learning in the natural language processing

With the development of deep learning and high-performance GPU, plentiful neural network models with more layers and more parameters are proposed. However, it is difficult to achieve satisfying result without a large number of data for model training. In terms of this issue, transfer learning method was proposed19. Namely, the neural network structure parameters are trained in advance through a large amount of data, and then the trained neural network is fine-tuned under the current specific task. The idea of transfer learning was widely applied in the field of natural language processing when word2vec was displayed20. Nevertheless, the word vectors obtained by word2vec are static, which is hard to solve polysemy problem. In response to the polysemy problem, ELMo based on bi-directional long short-term memory structure was presented21. Nonetheless, the difficulty of parallel training in ELMo prevents its network depth increasing. Since Transformer network was proposed, the high parallelism of multi-head attention mechanism can learn relevant information in different subspaces and it is designed into a deeper network structure to acquire stronger semantic representation ability22. The BERT pre-training language model based on Transformer unit has reached the leading level in many natural language processing tasks due to its excellent semantic representation and transfer generalization ability23,24. It is unnecessary for specific tasks to rebuild network structure and basic neural network can be directly designed in the last layer of BERT. Deep transfer learning in the natural language processing is widely utilized in the product design. Wang et al.25 explored a method for smart customization service based on configurators. The ELMo was adopted to encode the review text and the mapping between customer requirements and product specifications was built by a multi-task learning-based neural network. Qie et al.26 analyzed product textual requirements and created the related models with deep learning and natural language processing skills. In addition, it is worth noting that the dataset is usually imbalanced. On the one hand, granular computing27,28,29 and data resampling30,31 are utilized to change the imbalance rate of training dataset. On the other hand, ensemble learning methods can enhance the classification efficacy of imbalanced data by combining a series of weak classifiers32,33. Indicators including Precision, Recall and F1 are often applied to evaluate the classifier performance for imbalanced data. In this paper, the text data transformed from VPA data is segmented with natural sentences as the unit and then input into the established BERT deep transfer model. The functional, behavioral and structural customer requirements are classified by fine-tuning the BERT deep transfer model and classifier efficacy for imbalanced text data is evaluated.

Customer requirements mining

Previously, customer requirements were analyzed by offline ways like questionnaire or interview. In contrast, the voice of customer is contained in a large number of online reviews at present. Sun et al.4 proposed a dynamical mining method about ever-changing customer requirements. The changing behavior of product attributes was analyzed and an improvement strategy for next-generation product design was shown based on the changing behavior of attributes. Li et al.34 applied general rough set concepts to reveal the association between historical customer needs and design specifications. Jin et al.35 identified the product features and sentiment polarities from big consumer requirements data and employed kalman filter method to forecast the consumer requirement trends. As we know from the “Customer requirements classification” section, customer requirements actually involve multi-domain information and functional customer requirements represent customer intention maximally. However, there is a lack of detailed elaboration on the acquisition of functional customer requirements topic-word distribution. Hence, a series of topic models like latent semantic analysis (LSA), probabilistic latent semantic analysis (PLSA) and latent Dirichlet allocation (LDA)36,37,38 can be widely applied to make implicit and fuzzy customer intention explicitly. Topic-word distribution about functional requirements descriptions in the analogy-inspired VPA experiment can be confirmed. Nevertheless, the LSA is not a probabilistic language model so that ultimate results are hard to be explained intuitively. Although the PLSA endows the LSA with probabilistic interpretation, it is prone to overfit due to the solving complexity. Subsequent, the LDA is proposed by introducing Dirichlet distribution into the PLSA. However, traditional LDA is faced with some defects such as the empirical selection of topic quantity, which results in the algorithm performance degradation. To avoid the above problems, this paper presents an ILDA method.

Method

For this research involving human participants, we identify that this research has been approved by the ethical committee of the Zhejiang University and the state key laboratory of fluid power and mechatronic systems, confirm that this research was performed in accordance with the declaration of Helsinki. Meanwhile, informed consent was obtained from all participants and they agreed to the publication of identifying information/images in an online open-access publication.

Customer requirements classification based on BERT deep transfer model

With the development of deep learning and high-performance GPU, plentiful neural network models with more layers and more parameters are proposed. However, it is difficult to achieve satisfying result without a large number of data for model training. In fact, the collected customer requirements from the analogy-inspired VPA experiment is a small dataset. In terms of this issue, we adopt deep transfer learning method based on model pre-training. Research on the interpretability of deep neural networks shows that the shallow layers of the networks extract general features, while the deep layers of networks extract high-level semantic information related to specific tasks39. Hence, it is not difficult to obtain satisfying results for small dataset under specific tasks by transferring the powerful language representation ability of deep language model to the customer requirements classification. The BERT is an autoencoder language model, which gains semantic representation knowledge by adding noise to the input data and restoring masked information. The Transformer network is the core of the BERT and detailed introduction about the Transformer network can be acquired in the literature21. Compared with the previous mainstream approaches, the Transformer has the following advantages: (1) The limitation of parallel computation of recurrent neural network as well as its derivatives is broken. (2) The number of operations for calculating the association between two positions in the network is not effected by distance. (3) A more explanatory deep learning model is generated through the multi-head self-attention mechanism.

In this paper, there are 12 layers Transformer network in the base-model. The dimension of hidden layer is 768 and there are 12 attention heads in total. Large-model consists of 24 layers Transformer network. The dimension of hidden layer is 768 and there are 16 attention heads in total. The input Chinese sentences are converted into word vectors including token, position and segment, which respectively represent the word itself, word position and sentence dependency. The obtained vector representations are input into the BERT model, and the bi-directional Transformer structure can effectively extract semantic associations in the text data. The Gaussian error linear unit (GELU) is used as a nonlinear activation function inside BERT, which is presented as follows.

$$ GELU(x) \approx 0.5x\left( {1 + \tanh \left[ {\sqrt {2{/}\pi } \left( {x + 0.044715x^{3} } \right)} \right]} \right) $$
(1)

In the BERT pre-training process, all texts are divided into sentences, in which any two sentences constitute a training data and 15% words are masked randomly in the training data. The pre-training is implemented based on two unsupervised tasks, which are masked language model and next sentence prediction. Multi-source pre-training data makes BERT model more powerful in the semantic representation of Chinese text. In fact, the original Chinese BERT model proposed by Google only uses Chinese Wikipedia as the pre-training corpus. Considering the huge influence of Baidu baike in the Chinese knowledge community, choosing a parallel corpus is more conducive to the domain knowledge transfer. Hence, the BERT pre-training model is carried out on the Chinese Wikipedia and Baidu baike so that the Chinese semantic representation can be fully learned40. Moreover, the above two enormous and universal corpus contain abundant textual data related to the functional, behavioral and structural requirements of elevator. Namely, there are sufficient semantic connections between customer requirements and training corpus. In the fine-tuning stage, full connection layers and a softmax layer are added to the output-end of BERT for fine-tuning training. The cross entropy loss function is utilized for back-propagation training and the accuracy is employed to demonstrate the model classification ability. The accuracy and cross entropy loss function are calculated as follows40.

$$ {\text{Accuracy = }}\frac{1}{m}\sum\limits_{i = 1}^{m} {F\left( {f(x_{i} ) = y_{i} } \right)} $$
(2)
$$ F\left( {f(x_{i} ) = y_{i} } \right) = \left\{ {\begin{array}{*{20}l} {1,} \hfill & {f(x_{i} ) = y_{i} } \hfill \\ { - 1,} \hfill & {otherwise} \hfill \\ \end{array} } \right. $$
(3)
$$ CE{ = } - \frac{1}{m}\sum\limits_{i = 1}^{m} {y\log \left( {f(x_{i} )} \right)} $$
(4)

The voice of customers is obtained by accomplishing the analogy-inspired VPA experiment and converted into text data. The text data is segmented as many sentences and input into the BERT deep transfer model for fine-tuning, so as to classify customer requirements as functional domain, behavioral domain and structural domain. The implementation process of customer requirements classification based on BERT deep transfer model is shown in Fig. 2.

Figure 2
Figure 2
Full size image

The implementation process of customer requirements classification based on BERT.

Customer requirements mining based on ILDA

Customer requirements mining is actually a topic analysis task aiming at functional customer requirements representing customer intention maximally on the basis of the customer requirements classification. However, traditional LDA is faced with some defects like the empirical selection of topic quantity, which results in the algorithm performance degradation. This paper adopts ILDA method to acquire the topic-keyword distribution representing customer intention. Here are some basic definitions involving the ILDA method primarily39. Three sets used in the ILDA are listed as follows. (1) Word set W is {w1, …, wv, …, wV}, wv represents the vth word (v = 1, 2, …, V) and V is the total number of the words. (2) Document set D is {w1, …, wm, …, wM}, wm indicates the mth document (m = 1, 2, …, M) and M is the total number of the documents. Document wm is regarded as a sequence of words. Namely, wm is (wm1, …, wmn, …, wmNm), wmn describes the nth word in the document wm (n = 1, 2, …, Nm) and Nm is the total number of the words in the document wm. (3)Topic set Z is {z1, …, zk, …, zK}, zk means the kth topic (k = 1, 2, …, K) and K is the total number of the topics. Therefore, the generation process of the document set D is expressed as follows. (1) Generate word distributions of K topics randomly. Specifically, a vector φk (φk ~ Dir(β)) is randomly produced as the parameter of the word distribution p(w| zk) of topic zk. (2) Generate topic distributions of M documents randomly. Namely, a vector θm (θm ~ Dir(α)) is randomly produced as the parameter of the topic distribution p(z|wm) of document wm. (3) Generate a sequence of words of M documents according the multinomial distributions Mult(θm) and Mult(φk). In summary, the joint probability distribution of the model consists of observed variables and latent variables, which is presented in the Eq. (5)39.

$$ p({\mathbf{w}},{\mathbf{z}}{,}\theta {,}\varphi |\alpha ,\beta ) = \prod\limits_{k = 1}^{K} {p(\varphi_{k} |\beta )} \prod\limits_{m = 1}^{M} {p(\theta_{m} |\alpha )} \prod\limits_{n = 1}^{{N_{m} }} {p(z_{mn} |\theta_{m} )p(w_{mn} |z_{mn} ,\varphi )} $$
(5)

where observed variable w is a sequence of words of M documents, latent variable z is a sequence of topics of M documents, latent variable θ is the parameters of the topic distributions of M documents, latent variable φ is the parameters of the word distributions of K topics, α and β are the hyper-parameters. Appropriate values for hyper-parameters α and β depend on the number of topics and the number of words in the corpus. For most applications, good results can be obtained by setting α = 50/K + 1 and β = 200/V41.

Hence, the algorithm learning process is to estimate the latent variables z, θ and φ of the joint probability distribution according to the observed variable w. In fact, it is a complicated optimization problem and we can only obtain the approximation solutions. Approximate solutions are mainly Gibbs sampling42 and variational inference43. This paper applies the collapsed Gibbs sampling because of its simple and feasible implementation42. The implementation process of the collapsed Gibbs sampling can be briefly described as follows. (1) The marginal probability distribution p(w, z|α, β) is acquired by integrating the latent variables θ and φ. (2) The posterior probability distribution p(z| w, α, β) is sampled to gain the sample set of p(z| w, α, β). (3) The above sample set is utilized to estimate the latent variables z, θ and φ. The estimation equations for parameter θ and φ are shown as follows42.

$$ \theta_{mk} = \frac{{n_{mk} + \alpha_{k} }}{{\sum\nolimits_{k = 1}^{K} {(n_{mk} + \alpha_{k} )} }} $$
(6)
$$ \varphi_{kv} = \frac{{n_{kv} + \beta_{v} }}{{\sum\nolimits_{v = 1}^{V} {(n_{kv} + \beta_{v} )} }} $$
(7)

where θmk represents the probability that topic zk appears in the document wm, φkv expresses the probability that word wv appears in the topic zk, nmk is the count of the document-topic and nkv is the count of the topic-word. In addition to the hyper-parameters setting, an essential factor affecting the efficacy of the topic analysis is the optimal topic number K. At present, some measurable indicators like Perplexity and KL divergence are adopted to measure the optimal topic quantity44. However, Perplexity focuses on the prediction ability of the LDA model for new documents, which often leads to larger topic quantity. Meanwhile, KL divergence pays attention to the difference and stability among topics so that the optimal topic quantity is fewer. Thus, this paper proposes an improved measurable indicator Perplexity-AverKL for gaining the optimal topic quantity by combining the advantages of Perplexity and KL divergence. The detailed calculation equations are displayed as Eqs. (8)–(11).

$$ Perplexity = \exp \left( { - \frac{{\sum {\log p(w)} }}{{N_{d} }}} \right) $$
(8)
$$ AverKL = \frac{{\sum\nolimits_{i = 1}^{K} {\sum\nolimits_{j = 1}^{K} {KL\left( {\left. {z_{i} } \right\|z_{j} } \right)} } }}{{C_{K}^{2} }} $$
(9)
$$ KL\left( {\left. {z_{i} } \right\|z_{j} } \right) = \sum {p(z_{i} {|}{\mathbf{w}})} \log \frac{{p(z_{i} {|}{\mathbf{w}})}}{{p(z_{j} {|}{\mathbf{w}})}} $$
(10)
$$ Perplexity - AverKL = \frac{Perplexity}{{AverKL}} $$
(11)

where p(w) is the occurrence probability of each word in the corpus, Nd is the total number of words in the corpus, zi and zj is the extracted topics, KL(zizj) is the KL divergence between topic zi and zj, CK2 is the calculation times of KL divergence between two topics when the topic quantity is K. As can be seen from the above Eqs. (8)–(11), the generalization ability of the ILDA model is stronger when the Perplexity is smaller. Moreover, the difference among topics is greater when the AverKL is bigger. Small average similarity among topics means valid topics extraction. Namely, the optimal topic quantity K is determined when Perplexity-AverKL is the smallest.

From what has been discussed above, the specific ILDA implementation procedures are described as follows. (1) Preprocess all documents in the functional requirement corpus by carrying out Chinese words segmentation and deleting the Chinese stop-words. (2) Select the appropriate topic quantity K and initialize the hyper-parameters α and β. (3) Utilize Gibbs sampling to estimate the latent variables z, θ and φ. (4) Change the topic quantity K, then repeat step (2) and (3). Different topic analysis results under different topic quantity can be acquired. (5) The optimal topic quantity K is determined based on the proposed measurable indicator Perplexity-AverKL. Finally, the first-rank topic analysis result is output. Namely, multiple types of customer intentions and corresponding keywords can be obtained.

Case study

In order to conduct the proposed method, this paper implements an analogy-inspired VPA experiment. The experiment goal is to put forward as many functional, behavioral and structural requirements about elevator as possible based on the existing elevator design schemes and analogical inspiration. Ten Chinese graduate students majoring in mechanical engineering are selected as the experiment subjects and numbered from S1 to S10. They participated in the elevator design project previously and had a deep understanding on the function and structure of elevator. The average age of ten subjects is 24 year-olds and informed consent is obtained from all participants. The experiment is executed in a quiet room so that subjects can think deeply. Experiment instruction and tools like pens and paper are provided to the subjects. The experiment starts after the subjects are fully aware of the experimental specifications, problems and procedures. The spoken data of thinking aloud is collected by video recording during the experiment, and a retrospective discussion is conducted before the end of the experiment so that some errors in the spoken data preprocessing can be avoided. Particularly, the thinking aloud of subjects and experiment materials are all in Chinese in order to reduce the cognitive load of subjects. The analogy-inspired VPA experiment process is displayed in Fig. 3.

Figure 3
Figure 3
Full size image

The analogy-inspired VPA experiment process.

The detailed procedures of the experiment are described below: (1) Experiment preparation stage. The experimenter explains the verbal protocol analysis method to the subjects and assists them become familiar with the experiment process. Besides, the subjects are required to read a formal instruction, which introduces an example about how to express elevator customer requirements under an analogical stimulus normatively. The subjects will imitate the expression style of the example during the experiment process so that we can collect more normative VPA data relatively. (2) Open goal stage. The subjects are asked to describe the elevator requirements based on their existing knowledge freely. This stage lasts 3 min and ensures that the subjects have enough time to understand the experimental task and their knowledge memory in the relevant field is awakened. (3) Analogical stimuli stage. There are plentiful common features between near-domain stimuli and design problem, and they all come from the same or similar fields9. Product images such as stereo garage, spring shock absorber and vehicle display screen are randomly presented to the subjects as near-domain stimuli. There are few common features between far-domain stimuli and design problem, and they all come from the irrelevant fields9. Product images like magnetically levitated train, sneakers with air cushion and automobile airbag are randomly provided to the subjects as far-domain stimuli. Near-domain stimuli are provided to guarantee the feasibility and usefulness of the customer requirements and far-domain stimuli are selected to assure the novelty of the customer requirements. The subjects can associate stimuli source with existing knowledge in the aspects of function, behavior and structure. Meanwhile, the subjects elaborate elevator requirements by thinking aloud. This stage lasts four rounds and spends 3 min each round. (4) Evaluation and discussion stage. The subjects score the proposed elevator requirements in each round and the experimenter reviews the proposed elevator requirements and discusses them with the subjects. The specific experimental environment and process are shown in Fig. 4. (The recognizable face in Fig. 4 is a participant in the study).

Figure 4
Figure 4
Full size image

The experimental environment and process: (a) experiment preparation, (b) requirement statement, (c) evaluation and discussion.

The spoken data is converted into text data by using the Web API based on deep full sequence convolutional neural network provided by iFLYTEK open platform45,46,47. The text data is segmented as natural sentences, which are labeled into functional domain, behavioral domain and structural domain according to the specific sentence semantics. Then, Chinese text normalization tool provided by iFLYTEK open platform45,46,47 is utilized to optimize the colloquial text data in order to ensure the classifier performance. Non-fluent factors such as word repetition and semantic redundancy in the colloquial text data are removed, while the linguistic expression style of the colloquial text data is corrected to make it more similar to formal written language. Since the number of the colloquial text data is not large, three experts, who are familiar with elevator conceptual design and possess high-level Chinese writing skill, are invited to check the above text data according to previous instruction example as well as their own understanding ulteriorly. A part of text data is shown in the Table 1. The obtained text data is translated into English due to the Chinese experimental environment.

Table 1 A part of text data.

There are a total of 453 labeled sentences, which are randomly divided into training set and test set in a ratio of 8:2. The overall test classification accuracy can reach 88.7% by using the customer requirements classification method proposed in the “Customer requirements classification based on BERT deep transfer model” section. The results of customer requirements classification are presented in Fig. 5. It can be seen from the above results that effective overall classification accuracy has been obtained. Nevertheless, the identification results of functional and behavioral requirements are still not as satisfying as structural requirements, which is toilless to be confused with other types of requirements and lead to misjudgment. The reason for such problem lies in the small-scale semantic dataset currently and the subjective data annotation by some researchers in related fields. Furthermore, it is inevitable that there is potential imbalanced data problem for customer requirements classification. As a matter of fact, the number of functional, behavioral and structural labeled sentences is 103, 161 and 189 respectively. Thus, the evaluation indicators including Precision, Recall and F1 are computed and displayed in the Table 2. The analysis results indicate that the customer requirements classification based on BERT has gained well-pleasing efficacy for potential imbalanced data.

Figure 5
Figure 5
Full size image

The results of customer requirements classification: (a) training accuracy, (b) testing accuracy, (c) confusion matrix in the testing.

Table 2 Precision, Recall and F1 for customer requirements classification.

The ILDA method is applied to acquire the functional requirements topic-word distribution representing customer intention maximally. The stop-words method is utilized in order to filter out the words in the functional requirement texts that are not related to the product function. In order to ensure the excellent generalization ability of the ILDA model and the maximal difference among topics, the topic quantity is chosen as five by calculating the Perplexity-AverKL for models with different topic quantity. The relationship between the Perplexity-AverKL and the topic quantity is depicted in Fig. 6. The efficacy comparison among Perplexity-AverKL, Perplexity and KL divergence is presented in Fig. 7. Perplexity focuses on the prediction ability of the LDA model for new documents, which often leads to larger topic quantity. Meanwhile, KL divergence pays attention to the difference and stability among topics so that the optimal topic quantity is fewer. Perplexity-AverKL achieves appropriate topic quantity by combining the advantages of Perplexity and KL divergence. However, although the number of functional requirements texts is not large in this work, different phenomena regarding three different indicators may be presented since the topics from ILDA may mirror the distribution of words rather than the concept about “topic” in our daily conversation. Therefore, it is necessary to further evaluate the performance of the ILDA model with more topic quantity, which is shown in Fig. 8. The results denote that setting more topic quantity does not lead to better model performance due to worse measurable indicator values. We consider why large number of topics are not appropriate from two aspects. On the one hand, the number of types of main functional customer requirements for conceptual design of elevator is not too large. On the other hand, the amount of functional requirements texts is small. Thus, it is not easy to acquire lots of differentiated topics. Finally, five kinds of functional requirements and corresponding keywords as well as their weight coefficients in the topic-word distribution are shown in the Table 3. It can be seen from the table that customer requirements from topic1 to topic 5 are mainly aimed at the elevator operating state, elevator intelligence, elevator internal environment and elevator stability optimization and elevator sightseeing. The requirements analysis results will be beneficial for elevator conceptual design.

Figure 6
Figure 6
Full size image

The relationship between the Perplexity-AverKL and the topic quantity.

Figure 7
Figure 7
Full size image

The efficacy comparison among Perplexity-AverKL, Perplexity and KL divergence.

Figure 8
Figure 8
Full size image

The efficacy comparison among Perplexity-AverKL, Perplexity and KL divergence while setting more topic quantity.

Table 3 Five kinds of functional requirements.

Conclusions

This paper presents a semantic analysis-driven customer requirements mining method for product conceptual design based on deep transfer learning and ILDA. Firstly, an analogy-inspired VPA experiment providing cross-domain stimuli is conducted to obtain feasible and innovative customer requirement descriptions of elevator. Secondly, a BERT deep transfer model is constructed to realize the customer requirements classification among functional domain, behavioral domain and structural domain in terms of the customer requirement descriptions of elevator. Last but not least, the ILDA is proposed to mine the functional customer requirements representing customer intention maximally. Hence, this paper provides a novel research perspective on feasible and innovative customer requirements mining in the product conceptual design through natural language processing algorithm.

There are still some limitations and shortcomings in this work, which should be addressed in the future. On the one hand, the customer requirements acquired from the analogy-inspired VPA experiment are not abundant. More experiments are necessary to be implemented for providing massive and high-quality data. On the other hand, the topic-word distribution of customer requirements is extracted without considering the behavioral and structural customer requirements. The reason is that the semantic presentation in the behavioral and structural domain is usually manifested as “verb-noun” matching form, which is difficult to be extracted directly.