Copyright © 2009 KSAE 1229−9138/2009/045−11 International Journal of Automotive Technology, Vol. 10, No. 2, pp. 219−228 (2009) DOI 10.1007/s12239−009−0026−0 SENSOR FUSION-BASED LANE DETECTION FOR LKS+ACC SYSTEM H. G. JUNG , Y. H. LEE , H. J. KANG and J. KIM 1,2)* 1) 1) 1) 2) MANDO Global R&D H.Q., 413-5 Gomae-dong, Giheung-gu, Yongin-si, Gyeonggi 446-901, Korea School of Electrical and Electronic Engineering, Yonsei University, Seoul 120-749, Korea 2) (Received 20 May 2008; Revised 23 September 2008) ABSTRACT−This paper discusses the market trends and advantages of a safety system integrating LKS (Lane Keeping System) and ACC (Adaptive Cruise Control), referred to as the LKS+ACC system, and proposes a method utilizing the range data from ACC for the sake of lane detection. The overall structure of lane detection is the same as the conventional method using monocular vision: EDF (Edge Distribution Function)-based initialization, sub-ROI (Region Of Interest) for left/right and distance-based layers, steerable filter-based feature extraction, and model fitting in each sub-ROI. The proposed method adds only the system for confining lane detection ROI to free space that is established by range data. Experimental results indicate that such a simple adaptive ROI can overcome occlusion of lane markings and disturbance of neighboring vehicles. KEY WORDS : Lane detection, Sensor fusion, Lane keeping system, Adaptive cruise system 1. INTRODUCTION CHAUFFEUR II is a European project that was completed in 2003 and aimed at developing truck platoon and integration of LKS and ACC. In particular, the project proposed a system integrating LKS and Smart Distance Keeping (SDK), corresponding to ACC, and named it CHAUFFEUR Assistance (Fritz et al., 2004). 1.1. Background: Popularization of the LKS+ACC System Adaptive cruise control (ACC) is a driver convenience system adding headway time control, which maintains distance to the preceding vehicle within a preset headway time, to conventional cruise control that maintains preset speed if there is no preceding vehicle. The lane keeping system (LKS) is a driver convenience system which assists a vehicle in maintaining its driving lane. These two systems have been developed as two separate systems (Bishop, 2005). However, as the adoption rate of ACC is rising and various marketable embedded vision systems are emerging, the LKS+ACC system integrating both functions have attracted more interest. Major Japanese automakers have already produced LKS+ACC systems. LKS of Toyota (or Lexus) maintains its driving lane only if ACC is operating. If ACC is not operating, it will warn the driver of lane departure by a torque pulse (Toyota, 2004; The Tundra Solutions, 2006). Application vehicles include Lexus LS460 (Lexus, 2008) and Crown Majesta (The Auto Channel, 2004). Nissan has also developed a system integrating LKS and ACC (University of Twente, 2003), which it has applied to the Cima (Nissan, 2001). Honda developed the Honda Intelligent Driver Support System (HiDS) integrating Intelligent Highway Cruise Control (IHCC), corresponding to ACC, and Lane Keeping Assist System (LKAS), corresponding to LKS (Honda, 2006a). Application vehicles are the Accord (Pintonhead, 2006), the Legend (Honda, 2006b), and the Inspire (Honda, 2006c). 1.2. Advantages of the LKS+ACC System 1.2.1. Reduction of driver’s workload A considerable portion of traffic accidents are caused by driver carelessness and improper driving maneuvers. In particular, the burden of long hours of driving causes drivers to be fatigued, resulting in traffic accidents. Although conventional ACC and LKS can relieve the driver’s workload, the LKS+ACC system is expected to provide greater workload relief. Analyzing the effect of CHAUFFEUR Assistance on the driver using a driving simulator certified that driving stability was enhanced and the driver’s weariness was reduced compared with separate systems (Hogema, 2003). The result of vehicle testing of the Honda HiDS showed that 88% of test subjects felt their workload was reduced. Eye gaze pattern analysis indicated that drivers with HiDS observed a wider field of view (FOV) (Bishop, 2008). 1.2.2. Increase of traffic system capacity The results of analysis of CHAUFFEUR Assistance on the driver showed that drivers tended to maintain smaller headway time and change lanes less (Hogema, 2003). It was found that the LKS+ACC system gave a greater increase of traffic capacity than either LKS or ACC alone. Experts have predicted that the LKS+ACC system would provide a remarkable increase of traffic capacity when the lane width *Corresponding author. e-mail: hgjung@mando.com 219 220 H. G. JUNG, Y. H. LEE, H. J. KANG and J. KIM is narrow (Arem and Schermers, 2003). 1.2.3. Enhancement of control performance In regards to lane keeping control, if ACC is not operational, it is hard to predict time to cross (TTC). Contrarily, if ACC controls vehicle speed, LKS can easily design and follow the driving trajectory. As a result, the control performance will be enhanced (Cho , 2006). In regards to ACC, if preceding roadway information acquired by LKS is provided, ACC can implement proper speed control taking into consideration the shape of the road. For example, speed control on curves realizes cruise control that suits the driver’s preference by controlling the speed according to the road’s curve shape. Speed control at exits contributes to a reduction of the driver’s operating load by controlling deceleration when a car enters an exit lane (Denso, 2004a). et al. 1.2.4. Enhancement of recognition performance Using lane information acquired by LKS, ACC can recognize the preceding vehicle in a curved road. Preceding vehicle detection by using only long range radar (LRR) is supposed to be complicated by the need to eliminate noise from vehicle movement and vibration. Radar-based obstacle recognition can be enhanced by using image portion corresponding to the obstacle’s position. One of the major disturbances of lane detection is occlusion by the preceding vehicle. Therefore, the position information of the preceding vehicle makes the lane detection algorithm simpler and more robust. Otherwise, the lane detection algorithm may be complicated by the need to handle various cases including those in which the preceding vehicle occludes lane markings. three layers according to distance and again divided into the left hand side (LHS) and the right hand side (RHS). In these six regions, lane features are searched locally. Lane feature pixels are detected by a steerable filter and are approximated into a line or a parabola. The orientation of the steerable filter is initialized by peak detection of the edge distribution function (EDF), and then established according to the lane feature state predicted by temporal tracking. Regions of the lowest layer are fixed but regions of the second and third layer are set dynamically. The conventional lane detection system works well when there is no obstacle in the vicinity. With the recent adoption of the high dynamic range CMOS (HDRC) camera, traditional problems such as driving against the sun and tunnels have been overcome (Hoefflinger, 2007). However, if the preceding vehicle occludes lane markings or a vehicle in the adjacent lane approaches, lane features become lost or too small. As a result, the edges of the obstacle start to disturb lane detection. To overcome such problems, ROI establishment based on precise trajectory prediction using vehicle motion sensors and lane feature verification-based outlier rejection are incorporated (McCall and Trivedi, 2006). Assuming that the disturbance of neighboring vehicles occurs because the system has no knowledge about free space, our previous paper proposed that simple confine- 1.2.5. Benefits of ECU integration In order to enhance recognition performance of LKS and ACC, low-level fusion between image information and range information is essential. Low-level fusion between separate LKS and ACC requires over-weighted traffic load on the communication channel. In order to enhance control performance, an extended vehicle model incorporating lateral and longitudinal motion is needed and the vehicle trajectory should be designed comprehensively. Therefore, it is expected that high performance integrated electronic control unit (ECU) implement one vehicle model. Denso supplied LKS+ACC ECU to Toyota, who in turn developed the fusion ECU, which processes all sensor information including the vision sensor, the radar sensor, and the Lidar sensor and sends control commands to the active steering system and active braking system (Denso, 2004a, 2004b). 1.3. Adaptive ROI-Based Lane Detection The lane detection system proposed by this paper is fundamentally based on the monocular vision-based lane detection system published by McCall and Trivedi, (2006) and Guo (2006). The forward scene is divided into et al. Figure 1. Lane shape depends on road shape. SENSOR FUSION-BASED LANE DETECTION FOR LKS+ACC SYSTEM ment of ROI to free space can efficiently prevent the disturbance of neighboring vehicles (Jung , 2008). In this paper, this idea is referred to as adaptive ROI-based lane detection. Furthermore, in the case of the LKS+ACC system, because a range sensor is already installed for the sake of ACC function, lane detection performance can be improved without sensor addition. Experimental results confirm that the proposed method can detect lanes successfully, even in the case when conventional methods fail because of neighboring vehicles. Compared with our previous paper, this paper adds an algorithm which can account for traffic signs on the ground and quantitative evaluation. In particular, our experimental results show not only daytime performance but also nighttime performance. et al. 2. CONVENTIONAL SYSTEM: MONOCULAR VISION-BASED LANE DETECTION and sub-ROIs of the second and third layers are established using the lane detection results of their lower layer (Guo ., 2006). In other words, the detected lane of sub-ROI I determines the location of sub-ROI II and the detected lane of sub-ROI II determines the location of sub-ROI III. et al 2.2. Steerable Filtering Lanes appear as slanted edge lines in the lane searching region. If the slope of the lane feature is known , the steerable filter can detect lane features more efficiently than general edge detection methods (McCall , 2006; Guo , 2006). The steerable filter is defined using the two dimensional (2D) Gaussian function of equation (1). If the lane marking is regarded as a line having width, then the second derivative is used (McCall , 2006). If the inner edge of the lane marking is regarded as a lane feature, then the first derivative is used (Guo , 2006; Mineta, 2003). In our research, first derivatives, defined as in equations (2) and (3), are used. Equation (2) is the derivative of (1) in the xaxis direction (θ =0°) and (3) is the derivative of (1) in the y-axis direction (θ =90°). It is noteworthy that the equations (1) to (3) define 2D masks. a priori el at. et al. et al. The basic lane detection algorithm is implemented based on state of the art methods, McCall (2006) and Guo (2006). et al. 221 et al. 2.1. Three Layered ROI Structure Lane markings have different shapes according to the shape of the road, as shown in Figure 1. If the road is straight, as in Figure 1(a), all lane markings, both near and far, can be approximated as a straight line. If the road is curved, as in Figure 1(b), lane markings at near and far distances should be approximated as a straight line and a curve, respectively. ROI should be established such that the searching area is minimized but still contains the lane features. Desirable ROI is expected to include lane features and exclude image portion belonging to other objects. Considering the fact that the lane becomes smaller as distance increases, the searching area is divided into three layers whose size decrease gradually, and then divided into a LHS and a RHS. Consequently, six sub-ROIs are established. The height of the available searching area changes according to camera configuration and the height of each layer is defined as the ratio to the height of available searching area. Sub-ROI I and IV near to the subject vehicle is established and fixed, et al. G ( x y ) e– , = ∂ – G01 ----e ∂x o = (x ∂ – G901 ----e ∂y o = (x 2 (x 2 +y 2 +y 2 +y ) 2 = –2 2 ) (1) ) xe– =–2 (x ye– 2 (x +y 2 2 +y (2) ) 2 (3) ) The first derivative of equation (1) in a specific direction θ is defined using equations (2) and (3), as in equation (4) (Guo , 2006; Mineta, 2003). The filter defined in equation (4) outputs a strong response to edges perpendicular to the specific direction θ and outputs a weaker response as the angular difference increases. Therefore, because the possibility that edges of the shadow and stain have the same orientation as a lane feature is very low, a steerable filter tuned using known lane feature direction can selectively detect lane features, i.e. lane feature pixels. Figure 3(a) is an input image and (b) and (c) are the outputs of the steerable filter tuned to −45o and 45o, respectively. (4) Gθ1 ( θ ) ⋅ G01 ( θ ) ⋅ G901 et al. a priori o =cos o +sin 2.3. Edge Distribution Function The EDF is used to initialize the orientation parameter of the steerable filter. The EDF is the histogram of edge pixel directions with respect to angle (Guo , 2006; Nishida , 2005). Equation (5) defines the gradient of pixel ( , ). denotes the intensity variation with respect to the xaxis and denotes the intensity variation with respect to the y-axis. The gradient is approximated by the Sobel operator. With and , edge direction at pixel ( , ) is defined as in equation (6). After edge direction of all pixels et al. et al. y x Dx Dy Figure 2. Three layered ROI structure. Dx Dy x y 222 H. G. JUNG, Y. H. LEE, H. J. KANG and J. KIM Figure 4. EDF construction and peak detection. After dividing the EDF into two regions with respect to 90o, the maximum peak of each region is detected as shown in Figure 4(b). The left portion corresponds to the sub-ROI I of Figure 2 and the right portion corresponds to sub-ROI IV. As mentioned above, lane features in the lowest layer can be approximated by a line and the angle of the detected peak represents the direction of the lane feature of each sub-ROI. Therefore, the angle corresponding to the detected peak is used for the initialization of the orientation parameter of the steerable filter. Figure 3. Lane feature pixels detected by tuned steerable filter. in ROI is calculated using equation (6), the EDF can be constructed by accumulating the pixel occurrence with respect to edge direction. Figure 4(a) shows the Sobel operator result of Figure 3(a) and Figure 4(b) is the constructed EDF. ∂ I ----∂ I ⎞ T ≈ ( Dx ∇ I ( x y ) ⎛⎝ ----∂ x ∂ y⎠ Dy ⎞ ⎛ -----θ(x y) ⎝ Dx-⎠ , , = =tan , –1 , Dy )T (5) (6) 2.4. Lane Feature Detection Lane feature detection consists of steerable filtering, Hough transformation, inner edge point detection, and model fitting. Figure 5 presents the procedure of initial lane feature detection. A steerable filter tuned to an known lane direction and binarization detects lane feature pixels. Using the lane feature pixels, a Hough transform finds the lane feature. The second column of Figure 5 shows the Hough transform results; the horizontal and vertical axes represent parameters of the linear lane model. In the case of the lowest layer, the orientation of the steerable filter is set using the EDF and in the case of the other layers, it is set by lane feature state tracking, which is explained below. The initial lane feature found using the Hough transform a priori SENSOR FUSION-BASED LANE DETECTION FOR LKS+ACC SYSTEM 223 Figure 5. Initial lane feature detection. Figure 6. Detected inner edge points. is a linear approximation of pixels showing strong response to the steerable filter tuned to a specific direction using the voting method. By searching the edge point from the lane feature to the image center, the inner edge points are detected as shown in Figure 6 (McCall , 2006). The inner edge points detected in the first layer are fitted to a line using least square (LS) linear regression. The line is represented by two parameters, as shown in equation (7). The horizontal image direction is the x-axis and the vertical image direction is the y-axis. The cross-point of the line and a border between the first and the second layer is used as the center x coordinates of the second layer sub-ROI (McCall , 2006). y=a x+b (7) In the second layer sub-ROI, lane feature pixels are detected by the steerable filter and then the inner edge points are detected. The detected inner edge points are fitted to a curve using LS quadratic regression. The curve is represented by the three parameters of a parabola as given by equation (8). The intercept of the curve defined by the quadratic fitting and a border between the second and the third layer is used as the center x coordinates of the third layer sub-ROI (McCall , 2006). Figure 7. Dynamically established second and third subROIs. et al. et al. ⋅ et al. Figure 8. The tracked lane feature state is the output of lane detection. y=a ⋅ x 2 +b ⋅ x+c (8) 2.5. Lane Feature Tracking The left and right lines are determined by fitting inner edge points detected in the three layers, respectively, for the LHS and RHS of an image. Then, the orientation and offset of the left lane and the orientation and offset of the right 224 H. G. JUNG, Y. H. LEE, H. J. KANG and J. KIM lane are used as the lane feature state. The lane feature state is tracked by a Kalman filter such that it is robust to external disturbance (McCall ., 2006). The confidence level of the detected lane feature is measured, and if it drops below a pre-defined threshold, the EDF-based initialization procedure is called. Therefore, when the subject vehicle changes lanes, the lane features can be detected in spite of the abrupt state variable change. The lane feature state is tracked in such a way to be used for the setting of the direction parameter of the steerable filter in the next frame. Furthermore, it is used as lane information for lane keeping control and preceding vehicle recognition. Using the lane feature state instead of the instantaneous lane feature detected in each frame prevents performance degradation in the case when the lane marking is disconnected or occluded by neighboring vehicles. Figure 8 presents an example of a tracked lane feature state. In our previous paper (Jung , 2008), one of the open problems was traffic signs on the ground, which were drawn on the road and could not be distinguished by range data. A horizontal edge count between the LHS and RHS lane features enables the system to recognize the existence of a traffic sign. As the time duration to pass the traffic sign is short, the system can eliminate the effects of the traffic et al et al. Figure 9. Narrowly established ROI to cope with traffic signs on the ground. sign by using the area about the previously recognized lane as the new ROI. Figure 9(a) shows a correctly recognized lane feature using the proposed algorithm and Figure 9(b) shows the result without it. 3. RANGE DATA-BASED ROI ESTABLISHMENT According to a recently published survey concerning visionbased lane detection, vision-based lane detection generally consists of five components: road marking extraction, postprocessing, road modeling, vehicle modeling, and position tracking (McCall , 2006). Reviewing the development direction of each component, one common objective can be realized. The main challenge of road marking extraction is overcoming external disturbances such as shadow and stain, and focusing only on the lane feature. The steerable filter used in this et al. Figure 10. Acquired range data in the world coordinate system and the image coordinate system. SENSOR FUSION-BASED LANE DETECTION FOR LKS+ACC SYSTEM 225 an adjacent vehicle are almost parallel to lane markings, they can be falsely recognized as lane features when the adjacent vehicle approaches near the subject vehicle or an incorrectly established ROI is used. The shadow of an adjacent vehicle causes many problems, even when the adjacent vehicle does not approach near the subject vehicle. In particular, a cutting-in vehicle is an external disturbance which is difficult to identify as it is related to the update speed of lane feature tracking (i.e. the response time). However, it has been found that once the road surface covered by vehicles is rejected using range data, lane detection can simply ignore all edges generated by the appearance of neighboring objects. Furthermore, it is noteworthy that such a procedure can be implemented by simple operation, which is finding the image position corresponding to range data and masking off the area from the ROI. Denoting the image pixel coordinates ( , ) and world coordinates of range sensor by ( , , ), these two coordinates are related by homography H as follows: xi yi Xw Yw Zw Figure 11. Recognized free space. paper is developed to improve lane detection performance by focusing on edges having the expected orientation. Postprocessing is aimed at eliminating falsely detected lane features caused by external disturbances using knowledge regarding the road and lane. Road modeling, vehicle modeling, and position tracking are aimed at efficiently narrowing the searching area by formularizing lane marking shape, vehicle motion, and lane marking motion. In other words, they are developed to establish the ROI only at a region where the lane feature is expected to appear in the next frame considering the current position of the lane marking, vehicle motion, and lane marking structure. Consequently, external disturbances can be ignored and lane detection performance can be improved. The common objective of component development is minimizing the effect of external disturbances. We pay attention to the fact that external disturbances are inevitable because they are caused from dimension reduction from a 3D world to a 2D image. This means that once the external disturbance can be identified in advance, complicated post-processing and modeling can be simplified. Assuming the most important external disturbance of lane detection is neighboring objects including the preceding vehicle, adjacent vehicles, and guide rail, it can be expected that simply by confining the lane feature searching area to a free space ensured by range data, lane detection performance will be improved. When the preceding vehicle approaches near to the subjective vehicle, it occludes lane markings, and the edges of its appearance can be falsely detected as a lane feature. Because the side surface edges of a priori Xb xi Yb = H ⋅ yi Zb 1 (9) Xw = Xb / Zb Zw Yb / Zb (10) In order to acquire coordinates of the road surface, is set to 0. The homography H of equation (9) is defined as in equation (11). denotes the camera height and θ and ϕ denote the yaw angle and tilt angle of camera, respectively (Jung ., 2004). Yw hc et al H= –hc ⋅ cos θ –hc ⋅ sin θ ⋅ sin ϕ f ⋅ cos ϕ ⋅ hc ⋅ sin θ hc ⋅ cos θ –hc ⋅ cosθ ⋅ sin ϕ f ⋅ cos ϕ ⋅ hc ⋅ cos θ 0 cos ϕ f ⋅ sin ϕ (11) Figure 10(a) shows range data acquired by a scanning laser radar. The range data are acquired in the polar coordinate system and then transformed into the Cartesian coordinate system. Figure 10(b) indicates range data projected onto the input image. It can be seen that the positions where the vehicles and the guide rail meet, the road surface is successfully detected. However, the range data are disconnected in several positions and contains noise. Clustering range data eliminates disconnected and added noise. Scanning consecutive range data points, if two range data points are far more than a threshold apart, e.g. 50 cm, then they are recognized as a border between two range data clusters. Among recognized clusters, clusters with too small points or too short length are eliminated and then the deleted region is interpolated by adjacent clusters. The area below the border line consisting of recognized range data clusters and the sky line is recognized as free space, to which the lane feature searching region is confined. Figure 11 provides an example of recognized free space. Figure 226 H. G. JUNG, Y. H. LEE, H. J. KANG and J. KIM Table 1. Detection performance of the proposed method and conventional method. Proposed Conventional method method Daytime 93.17% 92.13% (2,491 frames) Nighttime 99.04% 98.82% (5,100 frames) 11(a) shows horizontal line-based clusters and range databased clusters. By combining these two kinds of clusters, the free space is constructed, as shown in Figure 11(b). Each of the six sub-ROIs is defined by a rectangle whose four corners are established so as to be located in the free space. 4. EXPERIMENTAL RESULTS In order to verify the feasibility of the proposed range databased adaptive ROI establishment, we installed scanning laser radar and a camera on the test vehicle and compared lane detection performance of the proposed method with the conventional method. A brief summary of the specifications of the scanning laser radar (SICK LD-OEM) is as follows: the FOV is 360o, the angular resolution is 0.125o, the range resolution is 3.9 mm, the maximum range is 250 m, the data interface is a controller area network (CAN), and the laser class is 1 (eye-safe). The resolution of the image is 640×480. Each image and range data was recorded with a speed of 10 frames per second. In total, 2,491 data images were recorded for daytime and 5,100 data images for nighttime. Table 1 compares detection performance. It can be seen that the proposed method shows better detection performance than the conventional method. Although the difference in detection performance is small, it is significant because situations related to the difference are potentially dangerous, e.g. when there are closing vehicles or when a vehicle cuts in suddenly. Figure 12 shows that the proposed adaptive ROI can overcome a disturbance from an adjacent vehicle. Figure 12(a) displays the input image and (b) and (d) show lane feature pixels detected on the LHS of the input image by the conventional method and the proposed method, respectively. It can be seen that the bottom edge of the adjacent vehicle looks similar to the lane feature. Figure 12(c) and (e) indicate the lane feature state detected by the conventional method and the proposed method, respectively. They show that the problem that the left lane feature is incorrectly detected by the conventional method in Figure 12(c) can be solved by the proposed method, as shown in Figure 12(e). It is noteworthy that neighboring vehicles are excluded in the free space, which is depicted in Figure 12(e). Figure 13 and Figure 14 show examples when the Figure 12. Comparison when adjacent vehicle approaches. Figure 13. Comparison when the preceding vehicle occludes left lane markings wholly. SENSOR FUSION-BASED LANE DETECTION FOR LKS+ACC SYSTEM 227 Figure 15. Successful cases. Figure 14. Comparison when little lane markings are observable. proposed method overcomes problems caused by the preceding vehicle. Figure 13 is a situation when lane markings are disconnected at the current location and the preceding vehicle occludes the remaining lane markings so that there is no useful information available on the left lane feature. The proposed method realizes that there is no useful information and then maintains a tracked lane feature state to output proper lane information. Figure 14 shows a situation where there are few observable lane markings. As the proposed method eliminates the image portion occupied by the preceding vehicle, it can focus on observable lane markings. In contrast, the conventional method fails because of the vehicle’s edges. Figure 15 demonstrates that the proposed method can successfully detect lanes in various situations. Figure 15(a) shows a situation where there is wide free space in front of the subjective vehicle. Figure 15(b) and (c) show situations where there are a lot of shadows on the road surface. Figure 15(d) indicates a situation where the cutting-in vehicle occludes right lane markings. In this case, although lane markings in the near area are occluded, the tracked lane feature state helps in finding the lane markings in the far area. Figure 16 shows results during nighttime; Figure 16(a) and (c) shows the results of the conventional method and Figure 16(b) and (d) shows the results of the proposed Figure 16. The recognition results during nighttime. method. As shown in Table 1, the detection performance during nighttime is better than during daytime. This is because lane markings become distinctive by narrow headlamp light-beams while other objects, which may disturb the system during the daytime, can not be observed in the dark. In other words, as the headlamp established an effective ROI for lane detection, the disturbance caused by environmental objects was reduced without any additional operations. 5. CONCLUSION This paper proposes a method which prevents external disturbance caused by neighboring vehicles by confining lane detection ROI to free space confirmed by range data. Experimental results show that the detection performance of the proposed method is better than that of the conventional method. Although the difference between the detection performances is small, the proposed method is 228 H. G. JUNG, Y. H. LEE, H. J. KANG and J. KIM expected to significantly enhance the safety of the system, as it correctly recognizes lane features even with an adjacent vehicle and a cutting-in vehicle. The main contribution of this paper is showing that a range sensor can enhance lane detection performance and simplify the lane detection algorithm. In particular, the proposed approach confining the ROI based on range data can be implemented by CAN communication even if ACC and LKS are implemented in separated ECUs, as in conventional implementation. Therefore, this approach requires only a minimal change and may be easily adopted. Future studies will focus on 1) real time implementation of the proposed method on an embedded platform and 2) replacement of the high angular resolution scanning laser radar used in this paper with Lidar or Radar with low angular resolution. REFERENCES Arem, B. and Schemers, G. (2003). Exploration of the traffic flow impacts of combined lateral and longitudinal support. Available at http://www.aida.utwente.nl/en/ research/Seminar/slides_GS.ppt. Accessed on 17 May 2008. Bishop, R. (2005). Intelligent Vehicle Technology and Trends. Artech House Inc. Norwood. MA. Bishop, R. (2008). Societal benefits of in-car technology. Available at http://www.ivsource.net/public/ppt/ BishopDutchRttF.ppt. Accessed on 17 May 2008. Cho, J. H., Nam, H. K. and Lee, W. S. (2006). Driver Behavior with Adaptive Cruise Control. Int. J. Automotive Technology 7, 5, 603−608. Denso (2004 a). Sensing system. Available at http://www. denso.co.jp/en/its/2004/products/pdf/sensingsystem_e.pdf. Accessed on 17 May 2008. Denso (2004 b). 11th ITS world cong. – Exhibited product lineup. Available at http://www.denso.co.jp/en/its/2004/ products/safety2.html. Accessed on 17 May 2008. Fritz, H., Gern, A., Schiemenz, H. and Bonnet, C. (2004). CHAUFFEUR assistant: a driver assistance system for commercial vehicles based on fusion of advanced ACC and lane keeping. 2004 IEEE Intelligent Vehicle Symp. 495−500. Guo, L., Li, K., Wang, J., and Lian, X. (2006). A robust lane detection method using steerable filters. The 8th Int. Symp. Advanced Vehicle Control, 2006. Hoefflinger, B. (2007). High-Dynamic-Range (HDR) Vision. Springer Berlin Heidelberg. Hogema, J. (2003). Driving behavior effects of the chauffeur assistant. Available at http://www.aida.utwente.nl/ en/research/Seminar/slides_JH.ppt. Accessed on 17 May 2008. Honda (2006 a). Safety for everyone in our mobile society. Available at http://world.honda.com/CSR/pdf/CSR-067-Safety-Initiatives.pdf. Accessed on 17 May 2008. Honda (2006 b). Honda introduces the all-new legend. Available at http://world.honda.com/news/2004/4041007. html. Accessed on 17 May 2008. Honda (2006 c). Honda announces a full model change for the inspire. Available at http://world.honda.com/news/ 2003/4030618_2.html. Accessed on 17 May 2008. Jung, C. R. and Kelber, C. R. (2004). A robust linearparabolic model for lane follow. Proc. XVII Brazilian Symp. Computer Graphics and Image Processing, 10, 72−79. Jung, H. G., Lee, Y. H., Yoon, P. J. and Kim, J. H. (2008). Forward sensing system for LKS+ACC. SAE Paper No. 2008-01-0205. Lexus (2008). The new LEXUS LS460. Available at http:/ /www.garagecordi.com/ls460.pdf. Accessed on 17 May 2008. McCall, J. C. and Trivedi, M. M. (2006). Video-based lane estimation and tracking for driver assistance: survey, system, and evaluation. IEEE Trans. Intelligent Transportation Systems 7, 1, 20−37. Mineta, K. (2003). Development of a lane mark recognition system for a lane keeping assist system. SAE Paper No. 2003-01-0281. Nishida, M., Kawakami, S. and Watanabe, A. (2005). Development of lane recognition algorithm for steering assistance system. SAE Paper No. 2005-01-1606. Nissan (2001). Nissan releases all-new cima. Avaiable at http://www.nissan-global.com/EN/NEWS/2001/_STORY/ 010112-01.html. Accessed on 17 May 2008. PistonHeads (2006). HONDA ADAS. Available at http:// www.pistonheads.com/doc.asp?c=52&i=15032. Accessed on 17 May 2008. The Auto Channel (2004). Toyota crown majesta undergoes complete redesign. Available at http://www. theautochannel.com/news/2004/07/07/202727.html. Accessed on 17 May 2008. The Tundra Solutions (2006). 4th-generation lexus flagship luxury sedan features …. Available at http:// www.tundrasolutions.com/forums/toyota-scion-andlexus-news/58916-fourth-generation-lexus-flagshipluxury-sedan/. Accessed on 17 May 2008. Toyota (2004). Environmental & social report 2004. Avaiable at http://www.toyota.co.jp/en/environmental_rep/ 04/download/pdf/e_s_report_2004.pdf. Accessed on 17 May 2008. University of Twente (2003). Summary_seminar: combination and/or integration of longitudinal and lateral support. Available at http://www.aida.utwente.nl/en/research/ Summary_seminar/. Accessed on 17 May 2008.