This paper presents control system design with autonomous control elements focused on field of automotive industry. The main objective of this document is description of the control and monitoring system with integrated image processing from the camera. The images obtained from the camera are used for recognizing routing and traffic situation. The communication interface between driver and car integrated system is carried out by process visualization on the LCD touch panel.
This paper presents control system design with autonomous control elements focused on field of automotive industry. The main objective of this document is description of the control and monitoring system with integrated image processing from the camera. The images obtained from the camera are used for recognizing routing and traffic situation. The communication interface between driver and car integrated system is carried out by process visualization on the LCD touch panel.
This paper presents control system design with autonomous control elements focused on field of automotive industry. The main objective of this document is description of the control and monitoring system with integrated image processing from the camera. The images obtained from the camera are used for recognizing routing and traffic situation. The communication interface between driver and car integrated system is carried out by process visualization on the LCD touch panel.
123456748There is presented control system design with
autonomous control elements focused on field of automotive
industry in this paper. The main objective of this document is description of the control and monitoring system with integrated image processing from the camera. The images obtained from the camera are used for recognizing routing and traffic situation. During system proposal we focused our attention on integration of components for the car localization using GPS and navigation system as well. The implemented embedded system communicates with other car control units using CAN bus and industrial Ethernet. The communication interface between driver and car integrated system is carried out by process visualization on the LCD touch panel. I. INTRODUCTION esults Ior automotive industry development has been the main aim oI the presentation in this paper. Our re- search has Iocused on process control and monitoring oI a process states and values and the speciIic solution Ior intelli- gent car control with autonomous elements has been pre- sented in this contribution. The solution was implemented in a prototype oI an electric car, where video-cameras Ior im- age processing and objects recognition in a real time were used. Stream data oI the recognized objects allow identiIy route lane, people, other cars, traIIic signs and situations. In addition to objects recognition, the accurate position must be obtained Irom GPS module and then transIerred to control system. InIormation oI exact position is useIul Ior the car lo- calization process and system navigation. The position data can be helpIul during estimation oI a distance between car and surrounding objects where supersonics sensors cannot be used. R The actuators, sensors, control units and video-cameras communicate with each other using industrial interIaces. Implemented industrial interIaces are CAN bus, wire or wireless industrial Ethernet and Bluetooth Ior extended Iea- tures. The communication system is able to transIer secured data Irom one car to another by using open wireless industri- al network. Car view inIormation, captured on the camera, is transIerred Irom the car to intelligent highway system. There is also space and possibility to use wireless communication connection Ior remote diagnostic test. |6|
This work is supported by grant oI the Grant Agency oI the Czech
Republic GA102/08/1429 - SaIety and security oI networked embedded system applications. InIormation oI the system is displayed on the touch screen panel that is connected into central development board based on 32.bits processor i.MX35. This type oI processor is suit- able Ior communication, graphic and multimedia applica- tions. In virtue oI using a lot oI hardware interIaces and due to complexity oI realization, we made a decision to use two types oI operating system. Both oI them are based on Unix platIorm. The Iirst one is RTAI Linux which supports a lot oI drivers Ior devices like cameras or modules Ior wireless connection or Bluetooth. On the other hand there is a lack oI preIerable memory protection. The kernel runs in the same memory space like drivers and the bug in driver can conse- quently cause the Iailure oI the whole operating system. The new Linux reboot lasts rather long and it is annoying Ior a car driver or other passengers in the car. The second solution oIIers applying the system with great memory protection and memory adaptive partitioning management. The mi- cro-kernel and drivers or other processes run in separate memory space. The Iailure oI one driver or process does not cause operating system crash but only automatic process restart in milliseconds. Subject oI concern is the operating system QNX Neutrino RTOS. Disadvantage oI QNX Neutri- no RTOS is weak support Ior wireless network drivers.|4| II. CAR SYSTEM ARCHITECTURE DESIGN The system architecture design is composed oI embedded modules with special or universal Iunction Ior the car con- trol process. The central control embedded unit has many Iunctions such as the system Ior battery management, mo- tion control system, image recognizing system, global posi- tion system and others. In this paper we have Iocused our at- tention to the system Ior image processing and recognition oI objects. The system embedded design is proposed Ior mo- bile devices like robots, car or industrial devices Ior product testing or measurement. Actual development design is com- posed oI low cost products. I. Embedded Car Svstem We propose embedded car system with 8.bits, 16.bits and 32.bits processors according to its Iunctionality. The central unit is based on ARM architecture with build-in 32.bits pro- cessor i.MX35. This embedded central unit is reserved Ior graphic applications, image processing with objects recogni- tion and communication processes. We preIer low power Intelligent Car Control and Recognition Embedded System Vilem Srovnal Jr., Zdenek Machacek, Radim Hercik, Roman Slaby, Vilem Srovnal VSB Technical University oI Ostrava, Measurement and Control, FEECS Ostrava Poruba, Czech Republic Email: vilem.srovnal, zdenek.machacek, radim.hercik, roman.slaby, vilem srovnal }vsb.cz Proceedings of the International Multiconference on Computer Science and Information Technology pp. 831836 ISBN 978-83-60810-27-9 ISSN 1896-7094 978-83-60810-27-9/09/$25.00 c 2010 IEEE 831 consumption Ior all our embedded units and that is why we have chosen ARM architecture. |5|
Fig. 1 Embedded car system architecture Figure 1 presents some oI the basic embedded modules implemented in the car control and monitoring system. II. Software Implementation for Jision Car Svstem The vision system capability is to react to various types oI traIIic situations. Together with processing image data there has to be artiIicial intelligence implemented with ability to predict a critical traIIic situation. Our development has been divided into 3 phases. The Iirst one is testing our objects recognition algorithms design in Matlab with cheap webcam on platIorm x86. Standard USB interIace was used Ior web- cam connection. The next development phase was a test on mobile devices like PDA. The last phase was a transIer oI the recognition system in embedded device that is part oI car system. A new communication interIace was used. We de- cided to choose a new camera with GigaEthernet interIace on industrial Ethernet protocol. There is high data stream oI data Ior a new 4 Megapixel camera with resolution up to 2336 x 1752l thereIore we need very Iast communication in- terIace. III.Real Time Operating Svstem Implementation Central control unit carries many Iunctions which present a lot oI processes running at the same time. ThereIore we decided to use the real time operating system. There are 2 types oI operating system mentioned in this paper RTAI Linux and QNX Neutrino RTOS. |1| RTAI Linux was modiIied Ior our purpose and Ior our hardware architecture. For this operating system, communi- cation drivers have been adjusted Ior wire and wireless con- nection, based on industrial protocols - FlexCAN, EtherCAT and Bluetooth. QNX Neutrino RTOS was rather easier to implement due to BSP supplied by QNX Company, but there was a problem with wireless communication driver. We still continue work- ing on it with QNX community cooperation. Following Iig- ures 2 and 3 show the diIIerence between Embedded Linux and QNX. Fig. 2 Architecture oI embedded Linux operating system Fig. 3 Architecture oI QNX Neutrino RTOS IJ. QT cross-platform graphic framework The soItware Qt is the cross-platIorm application and UI Iramework used Ior embedded applications. Graphics inter- Iace was chosen with regards to both operating systems. Our selection was inIluenced by our experiences with graphical QT system Irom Nokia. QT embedded graphical interIace is applicable Ior Linux and QNX as well. Figure 4 shows the example oI car graphic user interIace. Fig. 4 LCD touch panel with QT embedded car system III. RECOGNITION DRIVER SYSTEM The recognition driver system was developed Ior traIIic signs and traIIic lanes detection. The system processes real- time image captured by a camera. Implemented algorithms are based on the idea oI standardized Iorm and appearance oI traIIic signs. 832 PROCEEDINGS OF THE IMCSIT. VOLUME 5, 2010 Nowadays there are two production technologies oI image scanners available, namely scanners based on CMOS and CCD technologies. Better quality cameras use scanners based primarily on CCD technology. The advantage oI this technology is a high luminous sensitivity ensuring better im- age quality with low brightness. In comparison, the CMOS technology is much cheaper, because it is based on standard technology, which is used in mass production oI memory chips. By virtue oI this technology, the scanning element can be placed on one chip together with other electronic cir- cuit elements. Moreover, the advantage oI CMOS technolo- gy is lower energy consumption in comparison with CCD technology. Generally, there is a reasonable argument that when the scanner is heated up, the undesirable noise increas- es and reduces the quality oI Iinal record. For the presented application determined Ior image recog- nition, the VGA camera has satisIactory resolution. Better resolution would extend the process time, as each point oI image matrix prolongs recognition process. In this case, a web camera with VGA resolution based on CMOS technolo- gy is used Ior image recording. These cameras communicate with in-built systems with the help oI multi-purpose stan- dard interIace. The advantage oI these cameras is a low price and very good availability, while undesirable distur- bance, which can be Iixed by developed soItware, is a disad- vantage. Methods Ior reducing the undesirable noise distur- bance are based on using Iilters, such as linear and median Iilters, which are very simple, reliable and quickly imple- mented. The developed system retrieves images in real time directly Irom a webcam and they are evaluated immediately. Webcams Logitech with resolution 352 * 288 pixels were used Ior the test. Image quality reIlects necessary quality oI traIIic signs identiIication. Low image resolution oI traIIic signs causes that recognition success will decrease to 90. It is about 5 less than iI the original high resolution image is evaluated. Application speed is convenient, as it is able to process more than 10 Irames per second with presented camera resolution. An application is created Ior Linux, Windows CE, Windows Mobile QNX. Testing was perIormed on a development kit i.MX35. Use oI 32-bit multimedia applications processor based on iMX357 ARM11 core is run on 532MHz Irequency and size oI RAM is 128 megabytes. Implemented algorithms are suitable Ior wide range oI applications in existing PDA devices and mobile phones. These devices are limited by low Irequency computing power and memory space compared to being commonly used in desktop computers. The application was tested in devices E-TEN GloIiish X650. This device contains a Samsung S3C2442 processor 500 oI MH, 128 MB Flash ROM, 64 MB RAM and a VGA TFT display 2.8-inch. The device has also a camera 2Mpix. Installed Operating System is Windows Mobile 6 ProIessional. Testing has shown that reduced perIormance oI the device has largely aIIected algorithms time consumption. Acceleration oI algorithms is likely by more eIIicient use oI resource devices and optimizing memory management. The developed application shows examples oI detection system Ior traIIic signs in Fig.5. Fig. 5 Application Ior devices with operating system Windows CE and Windows Mobile Execution time Ior implemented algorithms depends on several parameters. The Iirst parameter is size oI the input image, size and number oI objects that are inside image. Number oI inside objects is input criteria which aIIects the number oI cycles that must be done during patterns comparison. The execution time, which consumes the algorithm, depends on the perIormance oI computer on which it has been executed. Processing time code implemented in C on a desktop PC in test events, did not exceed the period oI 1 second. The veriIication oI the developed soItware was accom- plished approximately on 50 traIIic signs. The majority oI test results are correct, i. e. in majority oI cases traIIic signs were detected correctly, but rarely there was not detected any traIIic sign. This error was caused by damage to traIIic signs, the excessive pollution or poor light conditions. TraI- Iic signs were successIully recognized with patterns in range Irom 90 to 95. During the recognition process, the distance oI traIIic signs Irom the camera plays a big role. Due to a small number oI pixels oI traIIic sign that is located too Iar, it cannot be guaranteed that the system will Iind the required consensus. Threshold Ior identiIying traIIic signs is approximately equal to 50 meters, iI the camera zoom value is set to 1:1. IV. ALGORITHMS FOR TRAFFIC SIGNS RECOGNITION The algorithms are based on ideas oI standardized appearance and shape oI traIIic signs. Parameters are deIined by traIIic signs in the Czech Republic, which are stated in the regulation no.30/2001, Ministry oI Transport The algorithms consist oI two main parts. The Iirst part implements correction and segmentation oI the input image. The second part implements object searching, analysis and user inIormation system. J. Image segmentation and conversion The basic aim oI image segmentation is to search continuous parts in whole Iigure Irom a camera. From these VILEM SROVNAL, ZDENEK MACHACEK, ET AL.: INTELLIGENT CAR CONTROL 833 analysed parts the objects are created, which are explored by parameters and the similarity oI the patterns. The method chosen Ior continuous parts searching is based on conversion oI the input colour image into the binary structure, taking into account the colour matrix with limited number oI colours. Converted traIIic signs image is composed oI 5 basic colours, where 4 colours (red, blue, black and yellow) determine the motive oI the label and white colour is chosen as the background oI the label. The analysis oI algorithms also solves the segmentation problem oI traIIic signs, which are composed just oI red and blue colours, so it is converted to only one object, which has to be divided to separated objects. There are not green colour analyses, because traIIic signs do not contain this color. For converting the image to a binary algorithm deIined by next presented Iunction is used. The example oI image segmentation is presented in Fig.6. f 1 p 1 x , v 2 23 4 0 for R 1 x , v2 ,G 1 x , v2 , B 1 x , v 2 ~ h 1 for R 1 x , v2 ,G 1 x , v2 , B 1 x , v2 5h (1) where: p is pixel h is colour limit R,G,B are colour components 1 2 2 2 1 1 1 1 1 3 2 2 2 4 4 4 4 4 1 1 2 2 2 2 2 1 1 3 3 2 2 2 2 2 4 4 1 1 2 2 2 2 2 2 2 3 3 2 2 2 2 2 2 2 1 1 2 1 1 1 2 2 2 3 3 2 5 5 5 2 2 2 1 1 2 1 2 1 2 2 2 3 3 2 5 2 5 2 2 2 1 1 2 1 2 1 2 1 1 3 3 2 5 2 5 2 6 6 1 2 2 1 2 1 2 1 1 3 2 2 5 2 5 2 6 6 Fig. 6 Continuous parts searching (leIt) and Individual part segmentation with diIIerent number oI each segment (right) JI. Image rotation and angle correction The camera images, which are obtained Irom real camera output, can be rotated to incorrect angles. Rotation is related to one point by an angle u. The simplest cases are: o390 6 , o3180 6 , o3790 6 . In these cases it is practical and easy to change the representation oI x and y axis. In reality, the angle u does not take exactly predictable values, thus we need to use trigonometric Iunctions conversion to new pixel location. Special situation occurs when the traIIic signs not only need to correct their orientation, but also to determine the angle oI which are deIlected Irom the vertical position. TraIIic signs are not always installed completely vertically. This Iact leads to lower conIormity oI recognition. The solution image has to be rotated to known angle. This angle can be calculated by algorithms based on symmetry oI traIIic signs as shown in Fig.7. The Iigure shows the geometric layout oI the problem where the presented trigonometric Iunctions is used Ior determination the angle u. The presented algorithm Ior rotation calculation cannot serve Ior traIIic signs oI circular shape. Angle u is calculated by next equation. |3| tg o3 dY X (2) The algorithm Ior rotation chooses reIerence point, which is located in the lower leIt corner oI the editing image. The calculation oI new coordinates is based on the next presented equations, where x,y are new coordinates and x, y are current coordinates. x 3cos 1 o8tg 71 v x 2 9 A x 2 8v 2 (3) v 3sin 1 o8tg 71 v x 2 9 A x 2 8v 2 (4) The rotation method gets images oI traIIic signs in a verti- cal position. New calculated pixel coordinates appear in the original matrix, but not always. ThereIore there is necessary to consider boundary data Iields so as not to exceed their limits. JII. Image pattern analvsis AIter image segmentation and image rotation, there is possibility to analyze patterns with the actual edited image Irom camera. TraIIic signs patterns are stored in a binary matrix. Each oI the patterns is size-standardized at 100 x 100 pixels. The resolution is chosen on the basis oI compromise between size oI the matrix patterns and quality oI the patterns. Fig. 8 Example oI binary matrix with patterns oI traIIic signs Each object has location in the original matrix and is given its size. In order to apply the correlation Iunction, size uniIication securing oI objects and patterns is needed. Function re-calculates the object's size on deIined dimension. The resize algorithm can be used aIter object continues parts recognition. The captured image can contain more traIIic labels with other recognized noise, which is not desirable. ThereIore, the method Ior object segmentation is supplemented with object centres identiIication. In case the centre is outside the other object area, it is solved separately. This recognition method eliminates problems with noise parts oI objects, which disturb object comparison with Fig. 7 Geometric expression oI rotation traIIic signs 834 PROCEEDINGS OF THE IMCSIT. VOLUME 5, 2010 patterns by correlation. This is processed by analyzing the geometric centre oI coordinates and objects. The example oI described method object separation is shown in Fig.9. Fig. 9 Example oI recognition basic points oI image objects Obtained object is then compared with correlation Iunction used to all patterns. In this process the biggest matching oI the object with patterns in percentage is expressed. The correct result is evaluated Irom the chosen pattern, which is the most identical to traIIic sign in actual image. Practical tests have Iound that matching in more than 87 is suIIicient to search traIIic signs. Learning the patterns number and rate oI compliance is an adequate criterion Ior the correct Iormulation oI the result recognition. Each number corresponds with speciIic patterns traIIic signs. Correlation Iunction is used to determine the relationship between two signals (in this case signal represents the analysed image), the similarity oI their histories, depending on their mutual displacement. Correlation can be expressed Ior linear and discrete signals. These signals cannot just be one-dimensional (vector) signal, but can be expressed by multidimensional signals. Correlation Iunction oI two signals is called a peer or cross-correlation Iunctions. The result oI correlation is a new signal, which has been displaced amplitude proportions have similar signals. This Iact can be expressed as correlation coeIIicient R, which takes values between zero and one and reIlects the similarity oI two signals. For discreet binary two-dimensional signals can be expressed by the correlation coeIIicient R values range Irom zero to one. R3 B x31 x B v31 v f 1 x , v29g 1 x , v 2 x9v (5) Function I(x,y) represented examined image and g(x,y) represented image pattern. Multiplying the correlation coeIIicient value by 100 represents the unity signal in percentage. Correlation is oIten used Ior the detection oI known signals. |2| JIII. Number recognition and analvsis from image Optical character recognition OCR is a method that enables the digitization oI texts Irom retrieved images. The developed program converts the image either automatically or must learn to recognize characters. The converted text is almost always dependent on the quality oI the draIt should undergo thorough prooIreading, because OCR program does not recognize all the letters correctly. For the detection oI traIIic signs that inIorm about the maximum speed limit a simpliIied version oI OCR algorithms are applied. OCR algorithm is applied to each traIIic sign that is probably Iound with the help oI the correlation Iunction in comparison to the patterns. Segments oI the traIIic signs are systematically compared with known patterns oI numbers. |3| Fig. 10 Patterns Ior implemented OCR algorithm Neural networks are useIul Ior solving problems in image and signal recognition or diagnosis. The neural network is generally designed Ior a structure oI spread parallel processing inIormation, which consists oI certain (usually very high) number oI simple computing elements. Each element is named as the neuron. Neuron receives a Iinite number oI inputs and their input inIormation and passes its output to a Iinite number oI outputs inIormation. Formalized algorithm model oI neuron is shown in Fig. 11. Fig. 11 Formalized algorithm model Activity oI this neuron can be expressed by: v3 f 1 2 , (6) where v is input oI neuron and is so-called potential, which may be expressed by equation 3 B f31 R w f x f 7w 0 (7) In this relationship x f symbol indicates the value oI j-th entry oI the neuron, w f denotes the value oI j-th entry, and w 0 denotes the threshold oI neurons and the Iunction f 1 2 shows generally nonlinear signal transIer Iunction oI neurons. Neurons are arranged in networks, this arrangement is known as network topology. Neural network does not contain any inIormation about Iinding the object. At the moment, neural methods and algorithms Ior OCR analyses are prepared, but implemented and successIully tested OCR recognition by correlation Iunction is available. The problem oI correlation Iunction, compared to other method, is insuIIicient speed oI algorithm execution, which increases with number oI patterns. VILEM SROVNAL, ZDENEK MACHACEK, ET AL.: INTELLIGENT CAR CONTROL 835 V. ALGORITHMS FOR TRAFFIC LANES RECOGNITION For the detection oI lanes on a road, Hough transIorm is implemented. At the moment the implementation is solved in MATLAB environment and it is prepared Ior implementation to embedded system. This transIormation is analytical method used to Iind a parametric description oI the objects in the picture. This method is used Ior the detection oI simple objects in the picture, which are example lines, circles and ellipses. The main advantage oI this method Ior lane detection is robust against irregularities and inIringements looking curve, which is particularly suitable Ior the detection oI dashed centre line. To Iind a mathematical model oI the object in a picture we used Hough transIorm as an input pixel input image. For example, detection oI lines in the image edited by the Hough transIorm was used in the equation: x9cos 08v9sin 03r (8) Where: r Length oI normal Irom the origin oI coordinates to the line. 0 - Angle between the normal and the x. 11111111 Fig. 12 Parametric description oI detected traIIic lane Hough transIorm is applied here, by an adjustment oI the input image. The adjustment consists oI the image transIer to binary image, and Iinding local maxima using edge detector. One oI the most important edge detection is Canny edge detector, which is implemented by algorithm Ior edge detection in two-dimensional discrete image. VI. CONCLUSION The main goal oI this paper is to show development oI car system Ior object recognition. The basic car system architec- ture and algorithms Ior traIIic signs and lanes recognition are presented here. Neural networks were used in order to solve problems in image and signal recognition or diagnosis. The problem oI image processing and object recognition was discussed in this contribution. We presented develop- ment stages oI recognition system. ACKNOWLEDGMENT This work was supported by projects GA102/08/1429 - SaIety and security oI networked embedded system applica- tions. REFERENCES |1| K. Arnold, Embedded Controller Hardware Design, San Diego, USA, 2001. 320 p. ISBN 978-1878707529 |2| J. D. Gibson, A. Bovik, Image and video processing, Academic Press, Orlando, USA, 2000. 891 p. ISBN 0121197905 |3| V. Hlavac, M. Sedlacek, Zpracovani signalu a obrazu, BEN, Praha, CZ, 2007. 255 p. ISBN 978-80-01-03110-0. |4| J. Kotzian, V. Srovnal Jr,: Distributed embedded system Ior ultralight airplane monitoring, ICINCO 2007, Intelligent Control Systems and Optimization, Anger, France, 2007 p.448-451 ISBN 978-972-8865- 82-5 |5| T. Sridhar, Design Embedded Communications SoItware. CMP Books, San Francisco, USA, 2003, ISBN 1-57820-125-X. |6| V. Srovnal Jr., Z. Machacek, V. Srovnal, Wireless Communication Ior Mobile Robotics and Industrial Embedded Devices, In proceedings ICN 2009 8th International ConIerence on Networks, Cancun, Mexico, 2009 p.253-258 ISBN 978-0-7695-3552-4 x y r 0 Fig. 13 Examples oI road lanes recognition system implementation 836 PROCEEDINGS OF THE IMCSIT. VOLUME 5, 2010