Product Form Feature Selection for Numerical Definition-Based Design Chih-Chieh Yang

advertisement
Product Form Feature Selection for Numerical Definition-Based Design
Chih-Chieh Yang
Department of Multimedia and Entertainment Science, Southern Taiwan University,
Tainan County, Taiwan 71005
Hung-Yuan Chen
Department of Visual Communication Design, Southern Taiwan University,
Tainan County, Taiwan 71005
Abstract
For developing appealing products, it is important for designers to pin point
critical product form features (PFFs) that influence consumers’ affective responses
(CARs) toward a product design. Manual inspection of critical form features based on
expert opinions has not proved to meet with the acceptance of consumers. In this
paper, a product form feature selection model based on CARs and a numerical
definition-based systematic approach (NSDA) is proposed. First, NSDA was used to
generate an explicit numerical definition of product form design. Next, CARs were
described using single adjectives as affective dimensions. The evaluation data of
consumers can be gathered by a semantic differential (SD) experiment. The prediction
models of CARs were constructed using support vector regression (SVR) and
multiple linear regression (MLR). Two feature selection methods, namely SVR with
support vector machine recursive feature elimination (SVM-RFE) and MLR with the
stepwise procedure, were used to select critical form features. A case study of knife
design is given to demonstrate the experimental results of the two feature selection
methods. The results of our experiment show that the feature ranking obtained from
SVM-RFE is very helpful to determine the importance of the form features. It is also
possible to select a subset of features with a given predictive performance of the SVR
model. On the other hand, the results of MLR with the stepwise procedure provide
several useful statistics to analyze the form features. The effects of the form features
for inducing specific CAR of knife design can also be interpreted by the standardized
regression coefficients. Since only a small data set of knife design was adopted in this
study, a further study using different kinds of product samples is worthy of ongoing
investigation.
Keywords: Feature selection; support vector regression; support vector machine
recursive feature elimination (SVM-RFE); Multiple linear regression; Kansei
engineering; Product form design.
1
1. Introduction
A product’s appearance is one of the most important factors affecting a
consumer’s purchasing decision. Many systematic product design studies have been
carried out to get a better insight into consumers’ subjective perceptions. The most
notable research is Kansei engineering (KE) (Jindo et al., 1995). The basic
assumption of KE studies is that there exists a cause-and-effect relationship between
product form features (PFFs) and consumers’ affective responses (CARs) (Han and
Hong, 2003). Therefore, a prediction model can be constructed from collected product
samples and the CAR data. Using these models, the relationship between PFFs and
CARs can be analyzed and a specially designed product form for specific target
consumer groups can be produced more objectively and efficiently. The construction
of the prediction model can be regarded as a regression problem taking PFFs of
product samples as input data and the CAR data as output values. Various methods
can be used to build the prediction model including multiple linear regressions (MLR)
(Chen and Chang, 2009), quantification theory type I (QT1) (Lai et al., 2006), partial
least squares regression (PLSR) (Han and Kim, 2003), neural networks (NN) (Hsiao
and Huang, 2002) and support vector regression (SVR) (Yang and Shieh, 2010). Of
these methods, SVR’s remarkable performance makes it the first choice in a number
of real-world applications but it is seldom adapted to the product design field.
In the study of Yang and Shieh (2010), SVR was introduced to develop a model
that predicts CARs for product form design with a very satisfactory predictive
performance. However, the PFFs of product samples (mobile phone design) were
analyzed by a morphological analysis (Jones, 1992), which results in a product form
representation with a mixed of continuous and discrete design attributes. This kind of
product form encoding is actually an abstract and simplified representation of the
form design and not directly linked to the form features of a product. A convincing
remedy, proposed by Chang and Chen (2008), to this problem is to employ a
numerical definition-based systematic approach (NSDA) to generate an explicit
numerical definition for product form design. Using NSDA, it is beneficial that not
only the importances of the form features obtained from feature selection methods can
be directly interpreted in the domain of the form features, but also the integration with
hybrid Kansei engineering system, such as the automatic product design systems
proposed by Hsiao and Tsai (2005) and Tsai and Chou (2007), to generate new
product form design according to the specified affective responses is quite
straightforward.
Although being an important research topic for KE, the problem of product form
2
feature selection according to CARs has not been intensively investigated. CARs are
often influenced by a wide variety of form features. The number of form features
could be many and might be highly correlated to each other. The relative importance
of each form feature is hard to identify so the selection of the critical form features
that please the consumer is a difficult task. When designing new products, critical
design features are often arrived at based on the opinions of expert (such as product
designers) or focus groups. However, the selection of features based on expert opinion
has its drawbacks such as a lack of objectivity and expert availability (Han and Kim,
2003). Moreover, the designers often need to combine or re-organize form features.
Therefore, if the relative importance and the ranking of the form features can be
analyzed, it can facilitate the decision making of designers. Since the feature selection
methods are connected with the prediction model of CAR, not only the critical form
feature can be selected but also their influence to produce specific CARs can be
extracted. Only a few attempts have been made to overcome these above-mentioned
shortcomings in the product form feature selection process. Han and Kim (2003) used
several statistical methods for screening critical design features including principal
component regression (PCR), cluster analysis, and PLSR. In Han and Yang (2004), a
genetic algorithm-based partial least squares method (GA-based PLS) was applied to
screen design variables.
In the present study, a product form feature selection model is constructed based
on CARs and NSDA in the context of KE. CARs were described using single
adjectives and NSDA was used to define the product form design. In order to gather
the evaluation data of consumers, a semantic differential (SD) experiment was
conducted using questionnaire investigation. The prediction models of CARs were
constructed using both SVR and MLR. Two feature selection methods, including SVR
with support vector machine recursive feature elimination (SVM-RFE) and MLR with
the stepwise procedure, were used to pin point critical form features. A case study of
knife design is given to demonstrate the analysis result. The reminder of the paper is
organized as follows: Section 2 reviews the problem of product form feature selection
and introduces the two feature selection methods, namely SVR with SVM-RFE and
MLR with the stepwise procedure. Section 3 presents the proposed model for product
form feature selection. Section 4 describes the experimental results and analyses of
the two feature selection methods. Section 5 offers some brief conclusions.
2. Background review
2.1. The problem of product form feature selection
3
Although there are only sparse literatures existing for product form feature
selection, the problem of feature selection can be found in many other fields besides
that of product design. The crux of the problem is how to find the subset of PFFs with
the least possible generalization errors and to select the smallest possible subset with a
high predictive capability. Different approaches have been proposed for solving the
feature selection problem including rough sets (Wakaki et al., 2004), rough-fuzzy
(Jensen, 2005), neuro-fuzzy (Pal et al., 1996), and support vector machine (SVM)
(Liu and Zheng, 2006). Another crucial issue in solving the PFFS problem is how to
deal with the nonlinear correlations between PFFs and CARs effectively (Park and
Han, 2004). The most widely adapted techniques such as MLR and QT1 do not
handle nonlinear relationships very well. In contrast, SVM and SVR are known for
their elegance in solving nonlinear problems by applying the kernel trick, which
automatically maps a feature space nonlinearly.
Based on whether or not feature selection is performed independently of the
learning algorithm that constructs the prediction model, feature selection methods can
be grouped into two categories: the filter method and the wrapper method (Kohavi
and John, 1997). The wrapper method is model-dependent. Based on the predictive
performance, the method evaluates directly the goodness of the selected feature’s
subset, which should intuitively yield a better performance. Many reported
experimental results also favored the wrapper method (Juang and Katagiri, 1992;
Wakaki et al., 2004). A well-known wrapper method, SVM-RFE, which was proposed
to select a relevant set of features for cancer classification problem (Guyon et al.,
2002). SVM-RFE, as a sequential backward feature elimination method, has been
successfully applied for product form feature selection based on a multiclass fuzzy
SVM classification model (Shieh and Yang, 2008b). In this study, SVM-RFE was
integrated into the SVR model to identify critical form features and obtain the feature
ranking. However, from a product designer’s point of view, the predictive
performance of the constructed model or which subset of the form features was
selected may not be main concern. On the contrary, it is more appealing for a design
practitioner to get better insight into the relationships of the form features and to
examine the effects of the features for inducing specific affective responses. Therefore,
MLR, as a linear model, is also adapted for product form feature selection in this
study. MLR equipped with a couple of well-developed analyzing statistics yields a
simple but intuitive manner to study the form features.
2.2. Feature selection using SVR with SVM-RFE
For a regression problem, a set D of l data samples, each represented are
4
given as ( xi , yi ) where xi  Rn is the input feature vector, yi  R is the desired
output value. In a standard SVR with a linear  -insensitive loss function the
objective is to find a function y  f ( x) , where  is a precision parameter
representing the radius of the tube located around f ( x) . Then, the SVR problem can
be regarded as the solution to
minimize
l
1 2
w  C (i  i* )
2
i 1

 yi  w   ( xi )  bi    i , i  0

subject to  w   ( xi )  bi  yi    i* , i*  0
i  1,..., l

(1)
where  ( x) is the high dimensional feature space that is nonlinearly mapped from
the input space R n . The parameter C  0 is a regularized constant determining the
trade-off between the empirical error and the regularization term.  i is the upper
training error ( i* is lower), subject to the  -insensitive tube y  (w   ( x)  b)   .
SVR fits f ( x) to the data by minimizing the empirical error (  i and  i* ) and the
regularization term
1 2
w . The regularization term is a measure for the model’s
2
complexity, and can be interpreted as a measure for the flatness of the regression
hyperplane in feature space. By introducing the Lagrange multipliers  , * , a kernel
function K and exploiting the optimality constraints, the constrained optimization
problem in Eq. (1) can be solved and the values of multipliers  and  * can be
obtained. For detailed derivations the reader is referred to Vapnik (1995). The weight
vector w can be expressed as follows:
l
l
w   (i  i* )( j   *j ) K ( xi , x j )
(2)
i 1 j 1
An efficient wrapper approach called SVM-RFE was used to conduct product
form feature selection in this study. The selection criterion of SVM-RFE was
developed according to Optimal Brain Damage (OBD) which has proved to be better
than earlier methods (Rakotomamonjy, 2003). OBD was first proposed by LeCun et al.
(1990) and uses the change of the cost function as the feature selection criterion,
which is defined as the second order term in the Taylor series of the cost function:
1 2 L
(3)
( D f )2
f 2
2 ( )
in which L is the cost function of any learning machine, and  is the weight of
cf 
features. OBD uses c f to approximate the change in the cost function caused by
5
removing a given feature f by expanding the cost function in the Taylor series.
Therefore, for SVR, the measure of OBD can be considered as the removed feature
that has the least influence on the weight vector norm w
2
in Eq. (1). The ranking
criterion can be written as
c f  w  w(  f )
2

1
2
l
l
 (
i 1 j 1
l
l
i
 (
i 1 j 1
2
  i* )( j   *j )K ( xi , x j ) 
( f )
i
(4)
  i*(  f ) )( (j  f )   *(j  f ) ) K (  f ) ( xi , x j )
where  and  * is the corresponding solution of Eq. (1), the notation  f means
that the feature f has been removed, and K is the kernel function calculated using
xi and x j . SVM-RFE starts with all the features. At each step, feature weights are
obtained by comparing the training samples with the existing features. Then, the
feature with the minimum c f is removed. This procedure continues until all features
are ranked according to the removed order.
2.3. Feature selection using MLR with the stepwise procedure
MLR is a traditional statistical method for modeling the linear relationship
between the input vector (independent variables) xi  R n and a desired output value
(dependent variable) yi  R . The regression equation of MLR constructed with l
data samples is
yi  b0  b1 xi ,1  ...  bn xi ,n  ei , i  1,..., l
(5)
where b0 is the regression constant, b1 ,..., bn are the partial regression coefficients
corresponding to the n input variables, and ei is the error term. The coefficients of
MLR are often estimated using ordinary least squares (OLS) procedure such that the
l
sum of the squared errors (  ei ) is minimized. The coefficient of determination, R2 ,
i 1
indicates the percentage of the variation in yi explained by the independent variables
xi . The value of R2 (ranged from 0 to 1) determines the goodness of fit for the
6
regression model. The coefficients b1 ,..., bn in Eq. (5), which are also denoted as
unstandardized regression coefficients, can be used to construct the regression
equation directly. However, the standardized regression coefficients 1 ,..., n , instead
of unstandardized regression coefficients, are more suitable for analyzing the
importance of variables. Each standardized regression coefficient represents the
change in response per standard deviation change in one output value y . Since the
input variables are often highly correlated with one another, the variance inflation
factor (VIF) can be used to assess the problem of multicollinearity.
For the task of product form feature selection in this study, MLR is combined
with the stepwise procedure to select significant independent variables based on the
statistic of F-test. In the stepwise procedure a variable that added into the model in the
earlier stage of selection may be deleted at the later stages. At each step, the
independent variable not in the model which has the smallest probability of F is added,
and the probability is sufficiently small (say 0.05). On the other hand, the variables
already included in the model are removed if the probability of F is sufficiently large
(say 0.1). In fact, the inclusion and exclusion of variables in the stepwise procedure is
the same as forward selection and backward elimination.
3. The proposed product form feature selection model
This study aims to construct a product form feature selection model based on
CARs and NSDA. First, NSDA is used to generate an explicit numerical definition of
product form design. Next, CARs are described using single adjectives as affective
dimensions. The representative affective dimensions are extracted using factor
analysis. In order to gather the evaluation data of consumers, an SD experiment is
conducted using questionnaire investigation. The SVR prediction models according to
the representative adjectives can be constructed using the form features of NSDA as
input data and the evaluated SD scores as output. Optimal training parameters of the
SVR model are determined using a grid search with leave-one-out cross validation
(LOOCV). In order to streamline the selection of critical product form features,
SVM-RFE is capable to identify the smallest feature subset with a given predictive
performance. Finally, MLR with the stepwise procedure is adopted to analyze the
importance of the form features and to interpret their effects to the designated CARs.
3.1. Preparing numerical definition-based product form samples
A numerical definition-based form representation for generic knife form, which
was previously proposed by Chang and Chen (2008), was adopted in this study. The
7
development of this form representation was based on an extensive investigation of
commercially-available knives. As shown in Fig. 1, the main structure of the knife
form was defined using three NURBS curves and 15 control points, namely the blade
profile (P1~P7), the lower edge of the handle (P8~P12), and the shape of the
tail-section (P13~P15). These NURBS curves were used to construct a extrude
surface with a fixed height for the knife blade. Note that control point p1 , located at
the intersection of the two blade curves, is specified as the origin (0, 0, 0) . As for the
modeling of the knife handle, it can be achieved by manipulating six adjustable design
elements, namely a centric curve to define the contour of the lower edge of the handle,
the width-to-height proportion, the proportional scale of the sectional area, the
rounded radius parameter, the sectional inclined angle, and a tail arc curve. In addition,
five sectional contours were uniformly distributed along the axis of the knife handle
and construct a loft surface. The resulting 39 form features for the numerical
definition of the knife form are listed in Table 1. A total of 43 knife model were
created using 3D software by three experienced product designers. The corresponding
form features for each 3D model were recorded. In order to reduce the perceived
differences caused by the materials and textures of the 3D model, the blade was
assigned with a gray material and the handle a black material. All models were
rendered under the same lighting and camera settings.
< Insert Fig. 1 about here >
< Insert Table 1 about here >
3.2. Extracting representative affective dimensions
In product design research, adjectives are often used to describe CARs. Prior to
conducting the adjective evaluation experiment, it is important to reduce the number
of adjectives and extract the most representative ones to allow consumers to describe
their subjective perceptions. By doing so, not only the phenomena of relevancy,
dependency, redundancy, cause/effect, and similarity existing among adjectives can be
reduced, but the experimental results of questionnaire investigation can also be
improved with better mental consistency for the subjects. 31 candidate adjectives,
which can be used to describe the affective responses toward knife design, were
collected in a previous study (Chen and Chang, 2009). These adjectives were used as
initial affective dimensions for consumers to evaluate a small amount of
representative knife samples in an SD experiment. The evaluated SD data of the initial
adjectives were then analyzed using factor analysis. The extracted three factors
8
account for 37.5%, 32.9% and 21.5% of explained-variance, respectively. For the task
to select representative adjectives, a simple and intuitive way is to pick up the
adjectives with the largest absolute factor loadings in each factor. Three extracted
adjectives including elegant, modern, and individualized were used in the following
adjective evaluation experiment. For detailed reports to extract these adjectives for
knife design the authors refer to the article of Chen and Chang (2009).
3.3. Questionnaire investigation for adjective evaluation
In order to collect the CAR data for knife design, 56 subjects (31 male and 25
female) were asked to evaluate each knife using a 9-point Likert scale using three
representative adjectives. The rendered images of the knife samples were presented to
the subjects in the format of questionnaires. In order to collect the evaluation data in a
more effective way, a user-friendly interface for the questionnaire was provided (see
Fig. 2). The images of the knife samples are displayed in the interface. The
presentation order of the knife samples was randomized to avoid any systematic
effects. After each section of the experiment, the evaluation data of each subject can
be recorded directly, thus simplifying the data post-processing time. The adjective
data was normalized to the range [0, 1]. All the subjects’ evaluation scores toward
each knife were averaged to reach a final utility score. These scores were applied as
the output values in the SVR and MLR models.
< Insert Fig. 2 about here >
3.4. Constructing the SVR prediction model
SVR was used to construct a prediction model based on the form features of the
knife samples and the average CAR ratings obtained from all the respondents. Since
SVR can only deal with one output value at a time, each prediction model need to be
constructed according to different adjectives. The training scheme of a single
prediction model based on SVR is depicted in Fig. 3. First, the knife samples consist
of a series of input vectors are fed into the training model. These input vectors are
mapped into feature space by a map function  . The mapping  is usually
nonlinear and unknown. Instead of calculating  , the kernel function K is used to
compute the inner product of two vectors xi and x j in the feature space  ( xi ) and
 ( x j ) , that is, K ( xi , x j )   ( xi )   ( x j ) . SVR is also known as its elegance in solving
9
the nonlinear problem with the kernel functions that automatically do a nonlinear
mapping to a feature space. Any function satisfying Mercer’s condition can be used as
the kernel function (Vapnik, 1995). Of the commonly used kernel functions, the
Gaussian kernel is favored for many applications due to its good features (Wang et al.,
2003); and thus was adapted as follows:
Gaussian kernel: K ( xi , x j )  exp( xi  x j
2
/ 2 2 ) ,
where  is the spread width parameter determining the influence of squared
distance between xi and x j to the kernel value. Using Gaussian kernel, dot
products are computed with the output values of the training samples under the map
 . The weights in the SVR represent the knowledge acquired from the knife samples.
Finally, the dot products are added up using the weights in terms of the Lagrange
multipliers,  i and  i* . This, plus a constant term b yields the final predictive
output value as follows:
l
y  f ( x)   ( i   i* ) K ( x, xi )  b .
i 1
The following procedure was used to train the knife samples:
(1) Transform data to the format of SVR;
(2) Normalize the form features linearly to the range [0, 1];
(3) Conduct a grid search with LOOCV to optimize the best parameter C ,  and
the spread width  of Gaussian kernel;
(4) Use the best parameter set to train the whole training set.
<Insert Fig. 3 about here>
3.5. Selecting critical form features using SVM-RFE
The SVM-RFE process was conducted based on the optimal SVR model using
the parameters obtained from a grid search, thus enabling critical product form
features according to each affective dimension to be selected. The relative importance
of the form features can be identified by analyzing the weight distribution in each
iterative step of the SVM-RFE process. The ranking criterion c f of SVR was
computing using Eq. (4). Since the SVM-RFE is a wrapper approach, the SVR model
was retrained and the weight values of features were updated during each iterative
step. Consequently, product designers can not only obtain the information about the
10
feature ranking but also that on the relative importance of the form features in each
iterative step of the SVM-RFE process. The complete SVM-RFE procedure for
product form feature selection, modified from Shieh and Yang (2008b), is described
as follows:
(1) Start with an empty ranked features list R  [] and the selected feature list
F  [1,..., n] ;
(2) Repeat until all features are ranked:
(a) Train SVR with all the training samples, with all features in F ;
(b) Compute the ranking criterion c f of SVR for features in F using Eq. (4);
(c) Find the feature with the smallest ranking criterion: a  arg min f c f ;
(d) Add the feature a into the ranked feature list R : R  [a, R] ;
(e) Remove the feature a from the selected feature list F : F  F  [a] ;
(3) Output: Ranked feature list R .
3.6. Analyzing the importance of form features using MLR
4. Experimental results
4.1. Results of form feature selection using SVR with SVM-RFE
4.1.1. Analysis of feature ranking
Table 2 shows the results of the feature ranking obtained by SVM-RFE. For the
sake of clarity, only the top 12 selected form features for the three adjectives are listed.
Also, if a feature from v1~ v30 is selected, the feature name is assigned with an
additional subscript x or y to indicate the direction of the NURBS control points
explicitly. As shown in Table 2(a), for the adjective “elegant”, there are 9 features
belong to the knife handle except for the features v8 y (6th), v14 y (11th) and
v12 y (12th). Another characteristic for the results of “elegant” is that the dominant
dimension of the NURBS control points is the y axis. According to the top 5
selected features a design suggestion for producing an “elegant” knife can be made. A
designer can first manipulate the features v18 y (3rd), v 20 y (1st) and v 26 y (4th)
to determine the main form of the knife handle; then adjust the features v31 (2nd)
and v32 (5th) to refine the knife form by tuning the inclined angle of the fore- and
tail- sections. The feature ranking of the adjective “modern” is shown in Table 2(b).
11
The top 6 selected features include v6 y (5th), v7 x (1st), v9 x (3rd), and v14 y
(4th) on the control points of the knife blade, plus v 26 y (6th) and the inclined angle
v32 (2nd) of the knife handle. Note that the two features v 26 y and v32 are
important for both the adjectives “elegant” and “modern”. The feature ranking of the
adjective “individualized” is shown in Table 2(c). The obtained feature ranking using
SVM-RFE is very helpful to product designers when they are determining the
importance of the form features while designing a knife.
< Insert Table 2 about here >
4.1.2. Predictive performance of selected feature subset
As described in the previous section, the feature ranking can be obtained from
the process of SVM-RFE. Since the number of form features is large, it is vital to
select the more important features according to the predictive performance of the SVR
model. During the elimination steps, each feature is screened out one at a time
according to the criterion c f . Therefore, the features which have less influence on the
prediction results will be eliminated in the earlier steps than those which have more
influence. Fig. 4 shows the predictive performance, measured in the root mean
squared error (RMSE), of the selected feature subset for the three adjectives. In the
beginning of the process before removing any feature, the SVR model performs best
with the lowest RMSE for the three adjectives. During the following steps, the
tendency of the RMSE increased gradually until a certain critical step, for example,
the 34th step of the adjective “elegant” (drawn in red line). Product designers can
select the smallest feature subset with a given predictive performance. If the allowable
RMSE for the adjective “elegant” is 0.08, the top 5 features can be selected without
deducting the performance significantly. It also seems to be a suitable threshold of the
RMSE using the value of 0.08 for the adjective “modern” (drawn in green line).
According to this threshold, the top 6 feature can be selected. As for the adjective
“individualized” (drawn in blue line), the threshold is difficult to determine since the
increase of the RMSE is not obvious.
< Insert Fig. 4 about here >
4.2. Results of form feature selection using MLR with the stepwise procedure
4.2.1. Selected features by the stepwise procedure
12
The results of the selected form features using MLR with the stepwise procedure
are shown in Table 3. The selection criterion of the form features is determined by the
probability of F in the stepwise procedure. For the adjective “elegant”, 11 features are
selected including v13x , v19 x , v 20 y , v23x , v25x , v27 x , v31 , v33 , v34 ,
v35 , and v38 . It can be observed that the features selected by MLR are very
different from the top 12 features selected by SVM-RFE (see Table 2[a]). More
specifically, there are only two selected features are identical ( v 20 y and v31 ). The
results for the other two adjective, modern and individualized, also exhibit the same
phenomenon. This difference is mainly due to the linear or nonlinear nature of these
two methods. MLR, as a linear model, does not have good predictive performance.
The coefficients of determination ( R2 ) for the three adjectives are 0.758, 0.685, and
0.759, respectively. Another reason for the unsatisfactory predictive performance of
MLR is that the number of knife samples (43) is not sufficient enough for a relative
larger number of features (39), that is, suffered from the problem of “curse of
dimensionality”. Nevertheless, the unstandardized regression coefficients still can be
used to construct the regression equation, as a simplified model neglecting the
underlying nonlinear relationships between the form features. The problem of
multicollinearity is not found in the models of the three adjectives since the VIF
values of all the selected features are less than 10, following to the suggestion of Hair
et al. (1995).
< Insert Table 3 about here >
4.2.2. Effects of form features for inducing affective responses
The effects of the form features for inducing specific CAR of knife design can be
interpreted by the standardized regression coefficients shown in Table 3. Recall that
each standardized coefficient represents the change of a form feature in response per
standard deviation change in the value of CAR. Moreover, a positive value indicates
the change of the feature and the CAR value are in the same direction, vice versa. For
the adjective “elegant”, for instance, the selected feature v13x , which represents the
x -axis coordinate of the NURBS control point P7 , has a positive value 0.379 for the
standardized coefficient. To explain the effects of the features in a graphical manner,
as shown in Fig. 5, if P7 on the curve of the blade edge was translated along the
x -axis for 1 standard deviation distance, it can produce a more “elegant” feeling in a
magnitude of 0.379. Similarly, a designer can consider the following suggestions:
translating the control point P10 ( v 20 y ) along the y -axis; rotating the fore-section
13
( v31 ) in a clockwise direction; reducing the proportional scale of the 2nd section
shape ( v34 ); reducing the width-to-height proportion ( v38 ). By interpreting the
standardized regression coefficients of MLR, not only the critical form features can be
identified but it also provides direct suggestions on how to produce a specific
affective response. However, since this analysis is based on a linear model, the joint
effects for adjusting these features at the same time need to be further examined.
< Insert Fig. 5 about here >
5. Conclusions
In the product design field, selecting critical form features and analyzing their
effects for inducing designated affective responses are very helpful to produce
successful products. However, there are only sparse literatures existing for this
problem. In this study, a product form feature selection model based on CARs and
NSDA is proposed. NSDA was employed to generate an explicit numerical definition
for product form design and CARs were described using single adjectives. The
adjective evaluation can be gathered by the SD experiment. Two feature selection
methods, namely SVR with SVM-RFE and MLR with the stepwise procedure, were
used to analyze the form features using knife design as an example. The SVM-RFE
process, as a state-of-the-art machine learning approach, is capable to obtain the form
feature ranking and select a subset of features with a given performance of the SVR
model. MLR combined with the stepwise procedure, as a traditional statistical method,
can be used to select significant form features based on the F-test. Further analysis
can be made using the statistics such as the regression coefficients and the VIF values.
The standardized regression coefficients also provide a useful means how to produce
a specific affective response. The proposed method for product form feature selection
provides product designers a potential tool for systematically studying the form
features and their effects on affective responses.
References
Chang, Y. M. and Chen, H. Y. (2008). Application of novel numerical definition-based
systematic approach (NDSA) to the design of knife forms. Journal of the
Chinese Institute of Industrial Engineers, 25(2), 148-161.
Chen, H. Y. and Chang, Y. M. (2009). Extraction of product form features critical to
determining consumers' perceptions of product image using a numerical
definition-based systematic approach. International Journal of Industrial
14
Ergonomics, 39(1), 133-145.
Guyon, I., Weston, J., Barnhill, S. and Bapnik, V. (2002). Gene selection for cancer
classification using support vector machines. Machine Learning, 46(1-3),
389-422.
Hair, J., Anderson, R., Tatham, R. and Black, W. (1995). Multivariate data with
readings. US America: Prentice Hall Inc.
Han, S. H. and Hong, S. W. (2003). A systematic approach for coupling user
satisfaction with product design. Ergonomics, 46(13/14), 1441-1461.
Han, S. H. and Kim, J. (2003). A comparison of screening methods: selecting
important design variables for modeling product usability. International
Journal of Industrial Ergonomics, 32, 189-198.
Han, S. H. and Yang, H. (2004). Screening important design variables for building a
usability model. International Journal of Industrial Ergonomics, 33, 159-171.
Hsiao, S. W. and Huang, H. C. (2002). A neural network based approach for product
form design. Design studies, 23, 67-84.
Hsiao, S. W. and Tsai, H. C. (2005). Applying a hybrid approach based on fuzzy
neural network and genetic algorithm to product form design. International
Journal of Industrial Ergonomics, 35, 411-428.
Jensen, R. (2005). Combining rough and fuzzy sets for feature selection. School of
Informatics, University of Edinburgh. Doctor of Philosophy.
Jindo, T., Hirasago, K. and Nagamachi, M. (1995). Development of a design support
system for office chairs using 3-D graphics. International Journal of Industrial
Ergonomics, 15, 49-62.
Jones, J. C. (1992). Design Methods. New York, Van Nostrand Reinhold.
Juang, B. H. and Katagiri, S. (1992). Discriminative learning for minimum error
classification. IEEE Transactions of Signal Process, 40(12), 3043-3054.
Kohavi, R. and John, G. (1997). Wrapper for feature subset selection. Artificial
Intelligence, 97(1-2), 273-324.
Lai, H. H., Lin, Y. C., Yeh, C. H. and Wei, C. H. (2006). User-oriented design for the
optimal combination on product design. International Journal of Production
Economics, 100, 253-267.
LeCun, Y., Denker, J., Solla, S., Howard, R. and Jackel, L. D. (1990). Optimal brain
damages. Advances in Neural Information Processing Systems. D. S.
Touretzky. CA: Mateo, Morgan Kaufmann. 2: 598-605.
Liu, Y. and Zheng, Y. F. (2006). FS_SFS: A novel feature selection method for support
vector machines. Pattern Recognition, 39, 1333-1345.
Pal, S. K., Basak, J. and De, R. K. (1996). Feature selection: a neuro-fuzzy approach.
IEEE International Conference on Neural Networks, Washington, DC.
15
Park, J. and Han, S. H. (2004). A fuzzy rule-based approach to modeling affective
user satisfaction towards office chair design. International Journal of Industrial
Ergonomics, 34, 31-47.
Rakotomamonjy, A. (2003). Variable selection using SVM-based criteria. Journal of
Machine Learning Research, 3, 1357-1370.
Shieh, M. D. and Yang, C. C. (2008b). Multiclass SVM-RFE for product form feature
selection. Expert Systems with Applications, 35(1-2), 531-541.
Tsai, H. C. and Chou, J. R. (2007). Automatic design support and image evaluation of
two-colored products using color association and color harmony scales and
genetic algorithm. Computer-Aided Design, 39(9), 818-828.
Vapnik, V. N. (1995). The nature of statistical learning theory. New York, Springer.
Wakaki, T., Itakura, H. and Tamura, M. (2004). Rough set-aided feature selection for
automatic web-page classification. Proceedings of the IEEE/WIC/ACM
International Conference on Web Intelligence.
Wang, W., Xu, Z., Lu, W. and Zhang, X. (2003). Determination of the spread
parameter in the Gaussian kernel for classification and regression.
Neurocomputing, 55, 643-663.
Yang, C. C. and Shieh, M. D. (2010). A support vector regression based prediction
model of affective responses for product form design. Computers & Industrial
Engineering, doi:10.1016/j.cie.2010.07.019.
16
Fig. 1. The numerical definition of generic knife form.
Origin (0,0)
P2 (v3x, v4y)
P3 (v5x, v6y)
P4 (v7x, v8y)
P9 (v17x, v18y)
P1 (v1x, v2y)
P11 (v21x, v22y)
p8 (v15x, v16y)
P13 (v25x, v26y)
P5 (v9x, v10y)
P10 (v19x, v20y)
P7 (v13x, v14y)
P6 (v11x, v12y)
P12 (v23x, v24y)
P14 (v27x, v28y)
P15 (v29x, v30y)
The fore-section shape (v33)
and its inclined angle (v31)
3rd section shape
(v35)
2nd section shape
(v34)
The tail-section shape (v37)
and its inclined angle (v32)
4th section shape
(v36)
Proportion of width and height (v38); Rounded size (v39)
17
Fig. 2. An interface of questionnaire for adjective evaluation.
18
Fig. 3. Training scheme of the SVR model.
Input
vectors
Mapped
vectors
x1
 ( x1 )
x2
 ( x2 )
Kernel
function
Weights
K ( x, x1 )
K ( x, x2 )
1  1*
 2   2*
y
...
...
...
xl
 ( xl )
K ( x, xl )
19
 l   l*
predictive
output
Fig. 4. Performance of feature subset (RMSE) of the three adjectives for SVM-RFE: y1 (elegant), y 2 (modern), and y3 (individualized).
20
Fig. 5. Effects of the form features in creating an “elegant” feeling.
Reducing the width-to-height proportion (v38)
creates a more “elegant” feeling.
Reducing the proportional scale of the 2nd section
(v34) creates a more “elegant” feeling.
Curve of blade back
Translating control point P10 (v20y)
in the upward direction creates a
more “elegant” feeling.
(0, 0)
Rotating the fore-section (v31) in a clockwise
direction produces a more “elegant” feeling.
Curve of blade edge
Translating control point P7 (v13x) in the rightward direction to reduce the radian of the arc
curve of the blade edge (as shown by dotted line) produces a more “elegant” feeling.
21
Table 1. The list of form features for the numerical definition of the knife form.
Knife form decomposed descriptions
Feature specifications
Features
Curve of blade back
( x, y ) coordinate of 4 NURBS control points (P1~P4)
v1 ~ v8
Curve of blade edge
( x, y ) coordinate of 3 NURBS control points (P5~P7)
v9 ~ v14
Centric curve of knife handle
( x, y ) coordinate of 5 NURBS control points (P8~P12)
v15 ~ v24
Tail arc curve of knife handle
( x, y ) coordinate of 3 NURBS control points (P13~P15)
v25 ~ v30
Knife blade
Knife handle
Fore-section inclination of knife handle
Inclined angle (60°/90°/120°)
v31
Tail-section inclination of knife handle
Inclined angle (60°/90°/120°)
v32
Fore-section shape
Proportional scale (1/0.75/0.5)
v33
2nd section shape
Sectional
variation 3rd section shape
of knife
handle
4th section shape
Proportional scale (1/0.75/0.5)
v34
Proportional scale (1/0.75/0.5)
v35
Proportional scale (1/0.75/0.5)
v36
Tail-section shape
Proportional scale (1/0.75/0.5)
v37
Proportion of all sections
Width to height proportion (1/0.67/0.5)
v38
Rounded radius of all sections
Rounded radius (1mm/3mm/5mm)
v39
22
Table 2. Top 12 of feature ranking of the adjectives obtained by SVM-RFE: (a) elegant (b) modern and (c) individualized.
Elegant
(a)
Rank
1
2
3
4
5
6
7
8
9
10
11
12
(b)
Modern
(c)
Individualized
Feature
v 20 y
*b/h
Rank
Feature
*b/h
Rank
h
1
b
1
v31
v18 y
v 26 y
h
h
h
h
b
h
h
2
3
4
5
6
7
8
v7 x
v32
v9 x
v14 y
v6 y
v 26 y
h
b
b
b
h
h
h
2
3
4
5
6
7
8
h
h
b
b
9
10
11
12
h
h
h
h
9
10
11
12
v32
v8 y
v 28 y
v 24 y
v30 y
v 22 y
v14 y
v12 y
v29x
v39
v25x
v30 y
v33
v36
*b/h denotes that the form feature belongs to a knife blade or handle (b: blade; h: handle).
23
Feature
v30
v14 y
v36
v38
v37
v23x
v2 y
v1x
v17 x
v34
v35
v33
*b/h
h
b
h
h
h
h
b
b
h
h
h
h
Table 3. Regression coefficients of selected features obtained from MLR with the
stepwise procedure.
Form
features
Elegant
(Constant)
v13x
v19 x
v 20 y
v23x
v25x
v27 x
v31
v33
v34
v35
v38
Unstandardized
coefficients
B
Std. Error
4.380
0.034
0.640
0.033
0.102
0.009
0.489
0.011
-7.840
-0.077
0.041
Standardized
coefficients
Beta
-
Sig.
Collinearity
statistics
VIF
-
0.379
0.131
0.275
0.000
0.001
0.020
0.007
1.361
1.285
1.168
3.796
0.040
0.021
-0.226
-0.195
0.188
0.047
0.062
0.056
1.529
1.298
1.154
-0.009
1.714
-2.157
1.043
-2.941
0.005
0.536
0.556
0.537
0.536
-0.171
0.298
-0.382
0.180
-0.501
0.085
0.003
0.001
0.016
0.000
1.192
1.108
1.241
1.104
1.068
4.235
0.042
0.039
0.749
-0.747
-10.043
0.175
-1.918
0.122
0.010
0.017
0.590
0.423
4.436
0.047
0.659
-
-
0.439
0.264
0.142
-0.193
-0.268
0.416
-0.315
0.000
0.000
0.024
0.013
0.087
0.030
0.001
0.006
1.251
1.307
1.316
1.249
1.469
1.291
1.226
1.998
-2.376
0.642
0.651
0.316
-0.375
0.004
0.001
1.081
1.108
5.732
-0.008
-0.044
0.052
0.004
0.020
-
0.000
0.028
0.035
-
R 2  0.758
Modern
(Constant)
v13x
v14 y
v19 x
v21x
v23x
v29x
v34
v37
v38
R 2  0.685
Individualized
(Constant)
v3x
v9 x
-0.211
-0.243
24
1.705
1.454
v17 x
v19 x
v 20 y
v21x
v 24 y
v25x
v 28 y
v33
v34
v36
v38
4.549
1.829
0.289
0.039
1.630
-0.620
-0.021
0.666
-0.089
0.218
0.087
-0.790
0.492
0.475
0.428
0.262
0.007
0.206
0.027
0.034
0.023
0.301
0.282
0.269
0.287
-0.259
-0.356
0.378
-1.497
1.122
1.803
-0.280
0.178
0.168
0.149
0.025
0.007
0.003
0.003
0.000
0.001
0.014
0.032
0.059
0.047
1.446
1.776
1.638
2.488
1.747
1.566
1.367
1.247
1.096
1.193
R 2  0.759
25
Download