Research Journal of Applied Sciences, Engineering and Technology 6(10): 1794-1798, 2013 ISSN: 2040-7459; e-ISSN: 2040-7467 © Maxwell Scientific Organization, 2013 Submitted: November 12, 2012 Accepted: January 01, 2013 Published: July 20, 2013 Improved Fast ICA Algorithm Using Eighth-Order Newton’s Method Tahir Ahmad, Norma Alias, Mahdi Ghanbari and Mohammad Askaripour IbnuSina Institute for Fundamental Sciences, Universiti Teknologi Malaysia 81310 Skudai, Johor, Malaysia Abstract: Independent Component Analysis (ICA) is a computational method to solve Blind Source Separation (BSS) problem. In this study, an improved Fast ICA based on eighth-order Newton’s method is proposed to solve BSS problems. Eight-order Newton’s method for finding the solution of nonlinear equations is much faster than ordinary Newton’s iterative method. The improved FastICA algorithm is applied to separate sound signals. The simulation results show the method has fewer iterations and faster convergence speed. Keywords: FastICA, Independent Component Analysis (ICA), Newton’s method 2001; Chio et al., 2005). The basic linear mixture model (without noise) for ICA is as follows: INTRODUCTION Independent Component Analysis (ICA) or Blind Source Separation (BSS) is a signal processing method that extracts statistically independent components from a set of measured signals (Hyvarinen, 1998; Lee et al., 1999; Ye et al., 2006). There are various ICA approaches including maximum likelihood estimation (Shia et al., 2004; Cardoso, 1998), mutual information minimization (Amari et al., 1996; Bell and Sejnowski, 1995; Pham, 2004) and negentropy maximization (Hevarinen et al., 2001). The FastICA algorithm was presented by Hyvarinen (1999). A kind of Newton-type algorithm was used in FastICA and the algorithm was based on fixed-point algorithm iteration to maximize nongaussianity as a measure of statistical independence. In FastICA algorithm, the Newton iteration method with two-order convergence was employed in a fixedpoint algorithm for estimating the separation matrix. In this study, we first briefly introduce ICA algorithms with Kurtosis and negentropy contrast function. Then, an improved FastICA based on an eighth-order convergence Newton’s Method (Li et al., 2011) is proposed. Using this new algorithm, the update equations inside the FastICA algorithm is developed for estimating separation matrix with less iterative. Then, the new algorithm is applied to separation sound signals. Finally, the performance of the newalgorithm is compared with FastICA algorithm in terms of evaluate performance rate and convergence rates. INDEPENDENT COMPONENT ANALYSIS AND FASTICA Independent Component Analysis (ICA) has become a powerful method for signal processing in last decade (Cichochi and Amari, 2004; Hevarinen et al., X = AS (1) where, S = (S1 , … , Sm )T denotes the original independent sources and X = (X1 , … , Xn )T denoted mixtures of original sources and A is the unknown m × n constant mixing matrix. ICA aims to recover or estimate the independent components (ICs) from their mixtures. An estimated IC is denoted by y and a separation vector is denoted byw, such that y = w T X = w T AS (2) πΎπΎπΎπΎπΎπΎπΎπΎ(π¦π¦) = πΈπΈ[π¦π¦ 4 ] − 3(πΈπΈ[π¦π¦ 2 ])2 (3) πΎπΎπΎπΎπΎπΎπΎπΎ(π¦π¦) = πΈπΈ[π¦π¦ 4 ] − 3 (4) The signal mixtures tend to have Gaussian probability density functions (pdfs) and the source signals have nongaussianpdfs (Stone, 1999). In terms of central limit theorem (CLT), Gaussian signals often consist of a mixture of nongaussian signals (Stone, 1999). Given a set of such Gaussian mixtures signal X = (ππ1 , … , ππππ )ππ . Therefore, the process of ICA is to find source signal by finding separation vector π€π€ that extract signal π¦π¦, which has to be the most nongaussian. One of the classic measures for estimation of nongaussianity of random variable is Kurtosis. We denote the Kurtosis of y by πΎπΎπΎπΎπΎπΎπΎπΎ(π¦π¦) and it is defined by: where, π¦π¦ is a random variable with zero mean. If π¦π¦ is normalized, variance of π¦π¦ is equal to one i.e.πΈπΈ[π¦π¦ 2 ] = 1. Then, (3) is simplified to: Corresponding Author: Tahir Ahmad, Department of Mathematics, Faculty of Science, Universiti Teknologi Malaysia 81310 Skudai, Johor, Malaysia 1794 Res. J. Appl. Sci. Eng. Technol., 6(10): 1794-1798, 2013 Negentropy for Nongaussianity: An important measure of nongaussianity is negentropy. The concept of negentropy strongly depended on entropy that it is another fundamental concept in information theory. Given ππ as a discrete value random variable, Entropy π»π» for random variable ππ is defined by: π»π»(ππ) = − ∑ππ ππ(ππ = π₯π₯ππ ) ln ππ(ππ = π₯π₯ππ ) (5) In Eq. (5), the π₯π₯ππ are the value of random variable ππ and ππ is density function of the random variable ππ (Cover and Thomas, 1991; Papoulis, 1991). The definition of entropy for continues random variable is obtained from generalizing (5). Given ππ as a continue random variable with density function ππππ (.). The differential entropy π»π» for continues random variable ππ is defined by: π»π»(ππ) = − ∫ ππππ (πΎπΎ) ln(πΎπΎ)ππππ (6) π½π½(ππ) = π»π»οΏ½ππππππππππππ οΏ½ − π»π»(ππ) (7) Let that ππ as a vector random variable. Negentropy of ππ is showed by π½π½(ππ) and is defined by: where, ππππππππππππ is gaussian random variable with same covariance of the ππ. Negentropy is another measure for nongaussianity. But, for the estimate of negentropy, we must first estimate the pdf of the vector random variable. Therefore, it is very hard to obtain it from the computation. One method to approximate negentropy by nonpolynomial function was proposed by Hyvarinen (1998). In related to this study, a method using expansion of pdf to approximate negentropy was given by Hevarinen et al. (2001). Let π¦π¦ be a whitened random variable. Then, approximate of π½π½(π¦π¦) was given: π½π½(π¦π¦) ∝ (πΈπΈ[πΊπΊ(π¦π¦)] − πΈπΈ[πΊπΊ(ππ)])2 (8) where, G is a nonquadratic function and variable ππ is a gaussian variable with unit variance and zero mean. Note that "proportion to" is the means of symbol ∝ . In order to achieve a very strongly estimator, choose G that grow slowly. Whereby: πΊπΊ1 (π¦π¦) = 1 ππ 1 log cosh ππ1 π¦π¦ πΊπΊ2 (π¦π¦) = − exp οΏ½− π¦π¦ 2 2 οΏ½ and 1 ≤ ππ1 ≤ 2 (Hevarinen et al., 2001). (9) π€π€ ← π€π€ − οΏ½πΈπΈοΏ½π§π§π§π§ οΏ½π€π€ ππ π§π§οΏ½οΏ½−π½π½π½π½ οΏ½ (11) οΏ½πΈπΈοΏ½ππ β«οΏ½οΏ½π§π§ ππ π€π€οΏ½ Χ³β¬−π½π½ οΏ½ where, π€π€ is separation vector, ππ is function as Eq. (9) and (10), π§π§ is the whitening observation vector, ππ ππ π§π§π§π§οΏ½π€π€ππππππ π§π§οΏ½οΏ½and π€π€ππππππ corresponds to π½π½ = πΈπΈοΏ½π€π€ππππππ optimum π€π€ . Equation (11) can be simplified by multiplying both sides byπ½π½ − πΈπΈοΏ½ππβ«οΏ½)π§π§ ππ π€π€( Χ³β¬and: π€π€ ← πΈπΈοΏ½ππβ« οΏ½)π§π§ ππ π€π€( Χ³β¬− πΈπΈ{π§π§π§π§(π€π€ ππ π§π§)}π€π€ (12) Let ππ is a vector random variable. Assume that π§π§ is given by the whitening of ππ. the fixed-point algorithm using negentropy is given as follow (Hevarinen et al., 2001): Step 1 : Step2 : Step 3 : Step 4 : Let ππ = 0 and π€π€(0) is an initial random vector with unit norm Let π€π€(ππ) = πΈπΈ{π§π§π§π§(π€π€(ππ − 1)ππ π§π§)} − β«ππ(π€π€( Χ³β¬ πΈπΈοΏ½ππ (13) − 1)ππ π§π§)οΏ½π€π€(ππ − 1) π€π€(ππ) Let π€π€(ππ) = βπ€π€(ππ)β (14) If the algorithm is not converged, then ππ = ππ + 1 and go back to step 2. From this algorithm, we obtain one component. Therefore, in order to estimate several independent components, the algorithm needs to be run for several times. EIGHT-ORDER NEWTON’S METHOD The classic Newton’s method for single nonlinear equation can be expressed as follows: π₯π₯ππ+1 = π₯π₯ππ − ππ(π₯π₯ ππ ) (15) ππ β«) ππ π₯π₯( Χ³β¬ This is an important and basic method, which converge quadratically.Pota and Petak (Potra and Ptak, 1984) proposed a modification Newton’s method with third-order convergence defined by: (10) π₯π₯ππ+1 = π₯π₯ππ − ππ(π₯π₯ ππ )+πποΏ½π₯π₯ ππ − ππ β«) ππ π₯π₯( Χ³β¬ ππ(π₯π₯ ππ ) ππ β«) ππ π₯π₯( Χ³β¬ οΏ½ (16) King (1973) developed a one- parameter family of forth-order methods, which is written as follows: A fixed-point algorithm using negentropy: A FastICA algorithm based on maximizing negentropy was introduced by Hyvarinen (1999). In this algorithm, fixed-point iteration and Newton’s method with twoorder convergence is used to find separation π€π€ vector as follows: 1795 π¦π¦ππ = π₯π₯ππ − ππ(π₯π₯ ππ ) ππ β«) ππ π₯π₯( Χ³β¬ π₯π₯ππ+1 = π¦π¦ππ − ππ(π₯π₯ ππ )+π½π½π½π½ (π¦π¦ππ ) (17) ππ(π¦π¦ππ ) ππ(π₯π₯ ππ )+(π½π½−2)ππ(π¦π¦ππ ) ππ β«) ππ π₯π₯( Χ³β¬ (18) Res. J. Appl. Sci. Eng. Technol., 6(10): 1794-1798, 2013 Such that ∈ R . Many methods have been proposed to improve the order of convergence such as (Kou, 2007; Chun, 2007; Parhi and Gupta, 2008; Bi et al., 2008). Li et al. (2011) proposed an eighth-order Newton’s method convergence defined by: π’π’ππ+1 = π₯π₯ππ − ππ(π₯π₯ ππ )+πποΏ½π₯π₯ ππ − π£π£ππ+1 = π’π’ππ+1 − π₯π₯ππ+1 = π£π£ππ+1 − ππ β«) ππ π₯π₯( Χ³β¬ ππ(π₯π₯ ππ ) ππ β«) ππ π₯π₯( Χ³β¬ οΏ½ ππ(π’π’ ππ +1 ) ππ(π₯π₯ ππ ) ππ β« ππ π₯π₯οΏ½ Χ³β¬− οΏ½ ππ β«) ππ π₯π₯( Χ³β¬ ππ(π£π£ππ +1 ) (19) and π½π½ ∈ π π in (28) will be updated in every iterations by β = πΈπΈ{π€π€ ππ π§π§π§π§(π€π€ ππ π§π§)} . Hence, we propose following fixed-point algorithm for estimate the separation matrix: Step1: Let ππ = 0 and given π€π€0 as an initial random vector with unit norm. Assume β = πΈπΈ{π€π€0ππ π§π§π§π§(π€π€0ππ π§π§)} Step 2: Let: IMPROVED FASTICA USING EIGHTH-ORDER NEWTON’S METHOD With regards to Eq. (11) can be derived new equations for estimating the separation matrix (unmixing matrix) by using Eq. (19) until (21) as follows: πΈπΈοΏ½π§π§π§π§ οΏ½π€π€ π£π£ππ π§π§οΏ½οΏ½−π½π½ π€π€ π£π£ (22) πΈπΈοΏ½ππ β«οΏ½οΏ½π§π§ πππ’π’ π€π€οΏ½ Χ³β¬−π½π½ Such that: π€π€π£π£ ← π€π€π’π’ − π€π€π’π’ ← π€π€ − πΈπΈοΏ½π§π§π§π§ οΏ½π€π€ π’π’ππ π§π§οΏ½οΏ½−π½π½ π€π€ π’π’ πΈπΈοΏ½ππ β«οΏ½οΏ½π§π§ πππ¦π¦ π€π€οΏ½ Χ³β¬−π½π½ πΈπΈοΏ½π§π§π§π§ οΏ½π€π€ ππ π§π§οΏ½οΏ½−π½π½π½π½ +πΈπΈοΏ½π§π§π§π§ οΏ½π€π€ π¦π¦ππ π§π§οΏ½οΏ½−π½π½ π€π€ π¦π¦ π€π€π¦π¦ ← π€π€ − πΈπΈοΏ½πποΏ½π€π€ ππ π§π§οΏ½οΏ½−π½π½ πΈπΈοΏ½π§π§π§π§ οΏ½π€π€ ππ π§π§οΏ½οΏ½−π½π½π½π½ (23) (24) where, − π€π€ π€π€π£π£ ← ππ +1 Step 4: If the algorithm is not converged then, β = ππ π§π§π§π§(π€π€ππππ+1 π§π§)}, ππ = ππ + 1 ππππππ go back πΈπΈ{π€π€ππ+1 to step 2. SIMULATION RESULTS We used three sound signals from MATLAB database as shown in Fig. 1. Each size of simple signals is 1000. These source signals were randomly mixed. The mixed signals are shown in Fig. 2. After whitening the mixed signals, we ran the proposed improved FastICA algorithm to separate the mixed signals .The results are shown in Fig. 3. 5 0 πΈπΈοΏ½ππβ«π£π£π€π€οΏ½)π§π§ πππ’π’π€π€( Χ³β¬ − (33) Step 3: π€π€ππ +1 = βπ€π€ ππ +1 β -5 β«Χ³β¬ πΈπΈοΏ½ππ οΏ½π€π€π¦π¦ππ π§π§οΏ½οΏ½π€π€π’π’ π€π€π’π’ ← πΈπΈ{π§π§π§π§(π€π€ ππ π§π§)} + πΈπΈοΏ½π§π§π§π§οΏ½π€π€π¦π¦ππ π§π§οΏ½οΏ½ − πΈπΈ{ππ(π€π€ ππ π§π§)}π€π€ + π½π½π€π€π¦π¦ π€π€π¦π¦ ← πΈπΈ{π§π§π§π§(π€π€ ππ π§π§)} − πΈπΈοΏ½ππβ«π€π€οΏ½)π§π§ ππ π€π€( Χ³β¬ 0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000 5 (26) 0 -5 πΈπΈ{π§π§π§π§(π€π€π’π’ππ π§π§)} (32) π€π€π¦π¦ = πΈπΈ{π§π§π§π§(π€π€ππ ππ π§π§)} − πΈπΈοΏ½ππβ«πππ€π€οΏ½)π§π§ ππ πππ€π€( Χ³β¬ Equations (22) to (25) can be simplified to: π€π€ ← (31) (25) πΈπΈοΏ½ππ β«οΏ½οΏ½π§π§ ππ π€π€οΏ½ Χ³β¬−π½π½ πΈπΈ{π§π§π§π§(π€π€π£π£ππ π§π§)} π€π€π£π£ = πΈπΈ{π§π§π§π§(π€π€π’π’ππ π§π§)} − πΈπΈοΏ½ππβ«π’π’π€π€οΏ½οΏ½π§π§ πππ¦π¦π€π€οΏ½ Χ³β¬ π€π€π’π’ = πΈπΈ{π§π§π§π§(π€π€ππ ππ π§π§)} + πΈπΈοΏ½π§π§π§π§οΏ½π€π€π¦π¦ππ π§π§οΏ½οΏ½ − πΈπΈ{ππ(π€π€ππ ππ π§π§)}π€π€ + π½π½π€π€π¦π¦ It is observed that this method required less number of iterations then traditional Newton’s method: π€π€ ← π€π€π£π£ − (30) where, (21) ππ β« πππ£π£( Χ³β¬+1 ) π€π€ππ+1 = πΈπΈ{π§π§π§π§(π€π€π£π£ππ π§π§)} − πΈπΈοΏ½ππβ«π£π£π€π€οΏ½)π§π§ πππ’π’π€π€( Χ³β¬ (20) (27) 5 0 (28) -5 (29) Fig. 1: Three source signals 1796 Res. J. Appl. Sci. Eng. Technol., 6(10): 1794-1798, 2013 1 οΏ½∑ππππ=1 ππππππππ + ∑ππππ=1 ππππππππ οΏ½ ππ 2 οΏ½ππ ππππ οΏ½ 1 οΏ½∑ππππ=1 οΏ½∑ππππ=1 − 1οΏ½ + max ππ |ππ ππππ | ππ 2 ππππ = 5 0 -5 100 0 200 300 400 500 600 700 800 900 ππ=1ππππ=1ππππππππmaxππππππππ−1 1000 5 where, οΏ½∑ππππ=1 0 -5 0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000 0 Fig. 2: Three mixed signals Fig. 3: Separation the signals Table 1: Comparison of the separation performance NO 1 2 3 4 5 FastICA 0.2156 19340 0.1873 0.1721 0.1964 Improved 0.2155 19350 0.1871 0.1724 0.1961 algorithm 6 0.2143 0.2146 P P P max ππ οΏ½ππ ππππ οΏ½ −1 and ππππππππ = − 1οΏ½ in which ππππππ is the ijth element of In this study, improved Fast ICA using eight-order Newton’s method was proposed for BSS. The separation performance of the proposed algorithm was almost the same as the Fast ICA algorithm. Besides that, the mixed sound signal was separated by new algorithm with fewer iteration and faster convergence than the Fast ICA. P P οΏ½ππ ππππ οΏ½ οΏ½ππ ππππ οΏ½ max ππ |ππ ππππ | CONCLUSION Table 2: Comparison of iteration numbers and CPU system Fast ICA Improved algorithm The 1th component 23.11 2.35 The 2th component 9.45 4.71 The 3th component 4 3.2N Total iteration 6824. 10.26 CPU time 0.2412 sec. 0.1381 sec. P ππππππππ = ∑ππππ=1 (34) ππ × ππ matrix ππ = ππππππ, where ππis a separation matrix, ππ is a whitening matrix and π΄π΄ isa mixing matrix. The large value ππππ is the poor statistical performance of the BSS algorithm (Amari et al., 1996; Shi and Zhang, 2006). These algorithms were run six times to evaluate their performances. The results are presented in Table 1. Whereby the separation performance of proposed algorithm is almost the same as the FastICA algorithm. Furthermore, we compared the convergence speed of these two algorithms. We randomly ran each of these two methods for 15 times executively. We then calculated the average iterative numbers and presented in Table 2. As a result, the improved Fast ICA algorithm performances with less iteration than Fast ICA, while maintaining comparable separation performance. This is due to the fact that convergence of eighth-order Newton’s method is much faster than the traditional Newton’s method with second-order convergence to find the solution for non-linear equation. To calculate the CPU time, two algorithms 15 times was run by using the MATLAB version R2010 on the Intel(R) Core(TM) 2 DUE CPU T6400 2.00 GHZ processor with 2.00 GB RAM. Table 2 describes iteration numbers and the CPU times required for sound signal separation from three mixed sound signal, whereby the average CPU times are 0.2412 s and 0.1381s, respectively. This demonstrated that the proposed method exhibited faster convergence with less iteration than Fast ICA algorithm, with comparable separation performance. 5 -5 = We can also get the same result when using the Fast ICA algorithm (Hevarinen et al., 2001). In order to measure the accuracy of separation, we calculated the performance index (Amari et al., 1996; Shi and Zhang, 2006). ACKNOWLEDGMENT The researchers would like to thanks from School of post graduate Studies (SPS) of Universiti Teknologi Malaysia (UTM) for getting Fellowship (IDF). 1797 Res. J. Appl. Sci. Eng. Technol., 6(10): 1794-1798, 2013 REFERENCES Amari, S., A. Cichocki and H. Yang, 1996. A new learning algorithm for blind separation of sources. Adv. Neural Inform. Process. Syst., 8: 757-763. Bell, A. and T. Sejnowski, 1995. An informationmaximization approach to blind separation and blind deconvolution. Neural Comp., 7: 1129-1159. Bi, W., H. Ren and Q. Wu, 2008. New family of seventh-order methods for nonlinear equations. Appl. Math. Comp., 203: 408-412. Cardoso, J.F., 1998. Blind signal separation: Statistical principles. Proceeding of the IEEE, 86: 2009-2025. Chun, C., 2007. Some improvements of Jarratt’s method with sixth-order convergence. Appl. Math. Comp., 190: 1432-1437. Cover, T.M. and J.A. Thomas, 1991. Elements of Information Theory. Publisher John Wiley & Sons, Hoboken, pp: 640, ISBN: 0471748811. Cichochi, A. and S. Amari, 2004. Adaptive Blind Signal and Image Processing: Learning Algorithms and Applications. Wiley, New York. Chio, S., A. Cichocki and H.M. Park, 2005. Blind source separation and independent component analysis: A review. Neural Inform. Process. Lett. Rev., 6(1): 1-57. Hyvarinen, A., 1998. New approximations of differential entropy for independent component analysis and projection pursuit. Adv. Neural Inform. Proc. Syst., 10: 273-279. Hyvarinen, A., 1999. Fast and robust fixed-point algorithms for independent component analysis. IEEE T. Neural Networ., 10(3): 626-634. Hevarinen, A., J. Karhunen and E. Oja, 2001. Independent Component Analysis. John Wiley & Sons, New York. Stone, J.V., 1999. Independent Component Analysis a Tutorial Introduction. The MIT Press, Cambridge, Mass, pp: 193, ISBN: 0262693151. King, R., 1973. A family of fourth order methods for nonlinear equations, SIAM J. Numer. Anal., 10: 876-879. Kou, J., 2007. The improvements of modified Newton’s method. Appl. Math. Comp., 189: 602-609. Lee, T.W., M. Girolami and T.J. Sejnowski, 1999. Independent component analysis using an extended infomax algorithm for mixed sub-Gaussian and super- Gaussian source. Neural Comp., 11: 417-441. Li, Z., C. Peng, T. Zhou and J. Gao, 2011. A new newton-type method for solving nonlinear equations with any integer order of convergence. J. Comp. Inform. Syst., 7: 2371-2378. Parhi, S.K. and D.K. Gupta, 2008. A sixth order method for nonlinear equations. Appl. Math. Comput., 203: 50-55. Pham, D.T., 2004. Fast algorithms for mutual information based independent component analysis. IEEE T. Signal Proc., 52(10). Potra, F.A. and V. Ptak, 1984. Nondiscrete Induction and Iterative Processes. Pitman, Boston, pp: 207, ISBN: 0273086278. Papoulis, A., 1991. Probability, Random Variables and Stochastic Processes. 3rd Edn., McGraw-Hill, New York. Shia, Z., H. Tanga and Y. Tang, 2004. A new fixedpoint algorithm for independent component analysis. Neurocomputing, 56: 467-473. Shi, Z. and C. Zhang, 2006. Energy predictability to blind source separation. Elec. Lett., 42(17): 1006-1008. Ye, Y., Z.L. Zhang, S. Wu and X. Zhou, 2006. Improved multiplicative orthogonal-group based ICA algorithm for separating mixed sub-Gaussian and super-Gaussian sources. Proceeding of International Conference on Communications, Circuits and Systems, Guilin, China (ICCCAS), pp: 340-343. 1798