3488 四履帶搜救機(jī)器人的結(jié)構(gòu)設(shè)計(jì)—移動(dòng)平臺(tái)設(shè)計(jì)
3488 四履帶搜救機(jī)器人的結(jié)構(gòu)設(shè)計(jì)—移動(dòng)平臺(tái)設(shè)計(jì),履帶,搜救,機(jī)器人,結(jié)構(gòu)設(shè)計(jì),移動(dòng),挪動(dòng),平臺(tái),設(shè)計(jì)
河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文1附錄:外文資料與中文翻譯外文資料:Scene recognition for mine rescue robot localization based on visionCUI Yi-an(崔益安)1, 2, CAI Zi-xing(蔡自興)1, WANG Lu( 王 璐)11. School of Information Science and Engineering, Central South University, Changsha 410083, China;2. School of Info Physics Engineering, Central South University, Changsha 410083, ChinaReceived 18 April 2007; accepted 13 September 2007Abstract: A new scene recognition system was presented based on fuzzy logic and hidden Markov model(HMM) that can be applied in mine rescue robot localization during emergencies. The system uses monocular camera to acquire omni-directional images of the mine environment where the robot locates. By adopting center-surround difference method, the salient local image regions are extracted from the images as natural landmarks. These landmarks are organized by using HMM to represent the scene where the robot is, and fuzzy logic strategy is used to match the scene and landmark. By this way, the localization problem, which is the scene recognition problem in the system, can be converted into the evaluation problem of HMM. The contributions of these skills make the system have the ability to deal 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文2with changes in scale, 2D rotation and viewpoint. The results of experiments also prove that the system has higher ratio of recognition and localization in both static and dynamic mine environments.Key words: robot location; scene recognition; salient image; matching strategy; fuzzy logic; hidden Markov model1 IntroductionSearch and rescue in disaster area in the domain of robot is a burgeoning and challenging subject[1]. Mine rescue robot was developed to enter mines during emergencies to locate possible escape routes for those trapped inside and determine whether it is safe for human to enter or not. Localization is a fundamental problem in this field. Localization methods based on camera can be mainly classified into geometric, topological or hybrid ones[2]. With its feasibility and effectiveness, scene recognition becomes one of the important technologiesof topological localization.Currently most scene recognition methods are based on global image features and have two distinct stages: training offline and matching online.During the training stage, robot collects the images of the environment where it works and processes the images to extract global features that represent the scene. Some approaches were used to analyze the data-set of image directly and some primary features were found, such as the PCA method[3]. However, the PCA method is not effective in distinguishing the classes of features. Another type of 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文3approach uses appearance features including color, texture and edge density to represent the image. For example, ZHOU et al[4] used multidimensional histograms to describe global appearance features. This method is simple but sensitive to scale and illumination changes. In fact, all kinds of global image features are suffered from the change of environment.LOWE[5] presented a SIFT method that uses similarity invariant descriptors formed by characteristic scale and orientation at interest points to obtain the features. The features are invariant to image scaling,translation, rotation and partially invariant to illumination changes. But SIFT may generate 1 000 or more interest points, which may slow down the processor dramatically.During the matching stage, nearest neighbor strategy(NN) is widely adopted for its facility and intelligibility[6]. But it cannot capture the contribution of individual feature for scene recognition. In experiments,the NN is not good enough to express the similarity between two patterns. Furthermore, the selected features can not represent the scene thoroughly according to the state-of-art pattern recognition, which makes recognition not reliable[7].So in this work a new recognition system is presented, which is more reliable and effective if it is used in a complex mine environment. In this system, we improve the invariance by extracting salient local image regions as landmarks to replace the whole image to deal with large changes in scale, 2D rotation and viewpoint.And the number of interest points is reduced effectively,which makes the processing 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文4easier. Fuzzy recognition strategy is designed to recognize the landmarks in place of NN, which can strengthen the contribution of individual feature for scene recognition. Because of its partial information resuming ability, hidden Markov model is adopted to organize those landmarks, which can capture the structure or relationship among them. So scene recognition can be transformed to the evaluation problem of HMM, which makes recognition robust.2 Salient local image regions detectionResearches on biological vision system indicate that organism (like drosophila) often pays attention to certain special regions in the scene for their behavioral relevance or local image cues while observing surroundings[8]. These regions can be taken as natural landmarks to effectively represent and distinguish different environments. Inspired by those, we use center-surround difference method to detect salient regions in multi-scale image spaces. The opponencies of color and texture are computed to create the saliency map.Input is provided in the form of static color image named as G0. Multi-scale image spaces G1?G4 (1?1 to 1?64) are created by Eqns.(1) and (2).Gn0=w*Gn?1 (1)Gn=Subsampled Gn0 n∈[1,4] (2)where w is a Gaussion low-pass filter, and “*”denotes convolution operation. Let centers are {G1, G2}and surroundings are {G3, G4}, the definition of opponency among scales is the feature difference between centers and surroundings denoted by “ Θ ”,which 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文5means that the surroundings are interpolated andthen subtract the centers pixel by pixel.To compute the desired color opponencies, it is necessary to convert the RGB space into RGBY space for emphasizing the opponencies of red/green and blue/yellow[9]. The space is calculated byR=r?(g+b)/2G=g?(r+b)/2B=b?(r+g)/2Y=(r+g)/2?|r?g|/2?bSo, the color opponencies are computed by RG(c, s)=|(R(c)?G(c)) Θ (G(s)?R(s))|BY(c, s)=|(B(c)?Y(c)) Θ (Y(s)?B(s))|where c∈Centers, s∈Surroundings. RG(c, s) denotes the opponency between red and green; BY(c, s) denotes the opponency between blue and yellow.To compute the texture opponencies, Gabor filter is selected because of its ability of acquiring local optimum either in the time domain or in the frequency domain.Researches on human psychophysics and vision physiology show that it resembles human attention mechanism[10].Gabor filter is defined by h(x, y)=g(x, y)e2πj(ux+vy). Because Gabor filter is polar symmetric in the frequency domain, the orientation 0?π can cover the whole frequency domain. Generally only 4 orientations 0°, 45°,90° and 135° are considered.The texture of 4 orientations are computed by 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文6Tθ(x,y)=|Gn(x, y)*hθ(x, y)|Now we can compute the texture opponencies by T(c, s, θ)=|Tcθ(x, y) Θ Tsθ(x, y)|where c∈Centers, s ∈Surroundings, θ∈{0°, 45°, 90°, 135°}.Then all opponencies are combined according to Eqns.(3)?(5) to create the saliency map S. The definition of normalizing operator N( ) can be found in Ref.[9].where w1 and w2 are weights that denote the significance of color and texture. We also design an algorithm using LMS to learn w1 and w2 offline. Fig.1(b) shows the saliency obtained landmark regions where the position with more lightness has more salience.Follow-up, sub-image centered at the salient position in S is taken as the landmark region. The size of the landmark region can be decided adaptively according to the changes of gradient orientation of the local image[11].Mobile robot navigation requires that natural landmarks should be detected stably when environments change to some extent. To validate the repeatability on landmark detection of our approach, we have done some experiments on the cases of scale, 2D rotation and viewpoint 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文7changes etc. Fig.2 shows that the door is detected for its saliency when viewpoint changes. More detailed analysis and results about scale and rotation can be found in our previous works[12].Fig.1 Saliency detection on real mine images: (a) Original image, (b) Obtained landmark regionsFig.2 Experiment on viewpoint changes3 Scene recognition and localizationDifferent from other scene recognition systems, our system doesn’t need training offline. In other words, our scenes are not 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文8classified in advance. When robot wanders,scenes captured at intervals of fixed time are used to build the vertex of a topological map, which represents the place where robot locates. Although the map’s geometric layout is ignored by the localization system, it is useful for visualization and debugging[13] and beneficial to path planning. So localization means searching the best match of current scene on the map. In this paper hidden Markov model is used to organize the extracted landmarks from current scene and create the vertex of topological map for its partial informationresuming ability.Resembled by panoramic vision system, robot looks around to get omni-images. From each image, salient local regions are detected and formed to be a sequence,named as landmark sequence whose order is the same as the image sequence. Then a hidden Markov model is created based on the landmark sequence involving k salient local image regions, which is taken as the description of the place where the robot locates. In our system EVI-D70 camera has a view field of ±170°.Considering the overlap effect, we sample environment every 45°to get 8 images.Let the 8 images as hidden state Si (1≤i≤8), the created HMM can be illustrated by Fig.3. The parameters of HMM, aij and bjk, are achieved by learning, using Baulm-Welch algorithm[14]. The threshold of convergence is set as 0.001.As for the edge of topological map, we assign it河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文9Fig.3 HMM of environmentwith distance information between two vertices. The distances can be computed according to odometry readings.To locate itself on the topological map, robot must run its ‘eye’ on environment and extract a landmark sequence L1′ ? Lk′ , then search the map for the best matched vertex (scene). Different from traditional probabilistic localization[15], in our system localization problem can be converted to the evaluation problem of HMM. The vertex with the greatest evaluation value,which must also be greater than a threshold, is taken asthe best matched vertex, which indicates the mostpossible place where the robot is.For evaluation, firstly we must prepare an observation sequence V T from L1′ ? Lk′ as below.V T={Li|Li∈ vertex to be recognized and matches one of L1? Lk′ }To compute the similarity between Li and Li′ , we design a new match strategy based on fuzzy logic, which can be found in the following section. Once the observation sequence is generated, evaluation process is to compute a posterior probabilistic value P(VT) according to Eqn.(6), whereωr T is a set of hidden state sequence and r is the index of a special hidden state sequence[3]. The detailed analysis and algorithm can be found in Ref.[4].河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文104 Match strategy based on fuzzy logicOne of the key issues in image match problem is to choose the most effective features or descriptors to represent the original image. Due to robot movement,those extracted landmark regions will change at pixel level. So, the descriptors or features chosen should be invariant to some extent according to the changes ofscale, rotation and viewpoint etc. In this paper, we use 4 features commonly adopted in the community that are briefly described as follows.GO: Gradient orientation. It has been proved that illumination and rotation changes are likely to have less influence on it[5].ASM and ENT: Angular second moment and entropy, which are two texture descriptors.H: Hue, which is used to describe the fundamental information of the image.Another key issue in match problem is to choose a good match strategy or algorithm. Usually nearest neighbor strategy (NN) is used to measure the similarity between two patterns. But we have found in the experiments that NN can’t adequately exhibit the individual descriptor or feature’s contribution to similarity measurement. As indicated in Fig.4, the input image Fig.4(a) comes from different view of Fig.4(b). But the distance between Figs.4(a) and (b) computed by Jefferey divergence is larger than Fig.4(c).河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文11To solve the problem, we design a new match algorithm based on fuzzy logic for exhibiting the subtle changes of each features. The algorithm is described as below.Fig.4 Similarity computed using Jefferey divergence: (a) Input image; 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文12(b) Landmark in database whose index is 6 and Jefferey divergence d=7.305 2; (c) Landmark in database whose index 14 and its Jefferey divergence d=4.662 61) First all features are fuzzifyid as belowIn these equations N k represents the number of pixels with gray level k, N pixels the total number of pixels of the image, N m_GO the number of pixels with angle degree m in {GOij}, Nm_H in {Hij}. Aij represents the 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文13gray value of the pixel, and the averaged degree attributed through the fuzzy classification to the gray level k,to the angular degree m. k is equal to 256 and m is equal to 360.2) The similarity between two landmarks is computed using individual feature, respectively. The similarity degree about the lth feature among the fuzzy set {ASM, ENT, GO, H} is defined asThen we compare the local image with every one in the database. Rmax and rmean are recorded.3) All similarity degrees of each feature are fused to obtain a judgement, which can be formalized by Eqn.(7).The weights wl are decided according to rmax?rmean of each feature. The deviations are sorted, then wl is assigned to be 0.4, 0.3, 0.2, 0.1, respectively according to the order.And the landmark in the database whose fused imilarity degree is higher than any others is taken as the best match. The match results of Figs.4(b) and (c) are demonstrated by Fig.5.As indicated, this method canmeasure the similarity effectively between two patterns.河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文14Fig.5 Similarity computed using fuzzy strategy5 Experiments and analysisThe localization system has been implemented on a mobile robot, which is built by our laboratory. The vision system is composed of a CCD camera and a frame-grabber IVC-4200. The resolution of image is set to be 400×320 and the sample frequency is set to be 10 frames/s. The computer system is composed of 1 GHz processor and 512 M memory, which is carried by the robot. Presently the robot works in indoor environments. Because HMM is adopted to represent and recognize the scene, our system has the ability to capture the discrimination about distribution of salient loca image regions and distinguish similar scenes effectively. Table 1 shows the recognition result of static environments including 5 laneways and a silo. 10 scenes are selected 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文15from each environment and HMMs are created for each scene. Then 20 scenes are collected when the robot enters each environment subsequently to match the 60 HMMs above.In the table, “truth” means that the scene to be localized matches with the right scene (the evaluation value of HMM is 30% greater than the second high evaluation). “Uncertainty” means that the evaluation value of HMM is greater than the second high evaluation under 10%. “Error match” means that the scene to be localized matches with the wrong scene. In the table, the ratio of error match is 0. But it is possible that the scene to be localized can’t match any scenes and new vertexes are created. Furthermore, the “ratio of truth” about silo islower because salient cues are fewer in this kind of environment.In the period of automatic exploring, similar scenes can be combined. The process can be summarized as: when localization succeeds, the current landmark sequence is added to the accompanying observation sequence of the matched vertex un-repeatedly according to their orientation (including the angle of the image from which the salient local region and the heading of the robot come). The parameters of HMM are learned again. 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文16Compared with the approaches using appearance features of the whole image (Method 2, M2), our system (M1) uses local salient regions to localize and map, which makes it have more tolerance of scale, viewpoint changes caused by robot’s movement and higher ratio of recognition and fewer amount of vertices on the topological map. So, our system has better performance in dynamic environment. These can be seen in Table 2. Laneways 1, 2, 4, 5 are in operation where some miners are working, which puzzle the robot.6 Conclusions1) Salient local image features are extracted toreplace the whole image to participate in recognition,which improve the tolerance of changes in scale, 2D rotation and viewpoint of environment image.2) Fuzzy logic is used to recognize the local image, and emphasize the individual feature’s contribution to recognition, which improves the reliability of landmarks.3) HMM is used to capture the structure or relationship of those 河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文17local images, which converts the scene recognition problem into the evaluation problem of HMM.4) The results from the above experiments demonstrate that the mine rescue robot scene recognition system has higher ratio of recognition and localization.Future work will be focused on using HMM to deal with the uncertainty of localization.References[1] QIAN Shan-hua, GE Shi-rong, WANG Yong-sheng, LIUChang-qing.Research status of the disaster rescue robot and its applications to the ine rescue [J]. Robot, 2006, 28(3): 350?354.[2] ULRICH I, NOURBAKHSH I. Appearance based place recognitionfor topological localization [C]// KHATIB Q. Proc of the IEEEInternational Conference on Robotics and Automation. San Francisco,2000: 1023?1029.[3] ARTAC M, JOGAN M, LEONARDIS A. Mobile robot localizationusing an incremental eigenspace model [C]// DUBEY R V. Proc ofthe IEEE International Conference on Robotics and Automation.Washington, 2002: 1025?1030.[4] ZHOU C, WEI Y C, TAN T N. Mobile robot self-localization basedon global visual appearance features [C]// LUO R C. Proc of the河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文18IEEE International Conference on Robotics and Automation. Taipei,2003: 1271?1276.[5] LOWE D. Object recognition from local scale invariant features [C]//BOYER K L. Proc of the International Conference on ComputerVision. Greece, 1999: 1150?1157.[6] CHEN Yong-sheng, HUNG Yi-ping, YEN Ting-fang, FUH Chioushan.Fast and versatile algorithm for nearest neighbor search basedon a lower bound tree [J]. Pattern Recognition, 2007, 40(2):360?375.[7] RAPTIS S N, TZAFESTAS S G. Fuzzy-probabilistic objectrecognition based on edge and texture descriptors fusion [C]// JENGMu-de. IEEE International Conference on Systems, Man, andCybernetics. Nashville, USA. 2000: 1621?1626.[8] HEISENBERG M, WOLF R. Studies of brain function [M]. Berlin, Germany: Springer-Verlag Press, 1984.[9] EDUARDO T, TORRAS C. Detection of nature landmarks throughmultiscale opponent features [C]// LOOG M. Proc of InternationalConference on Pattern Recognition. Barcelona, 2000: 3988?3991.[10] SHENG W, XIA B. Texture segmentation method based on Gaborring filtering [J]. Infrared and Laser Engineering, 2003, 32(5):484?488.河南理工大學(xué)萬方科技學(xué)院本科畢業(yè)論文19[11] MIKOLAJCZYK K, SCHMID C. Scale and affine invariant interest point detectors [J]. International Journal of Computer Vision, 2004, 60(1): 63?86.[12] WANG L, CAI Z X. Saliency based natural landmarks detectionunder unknown environments [J]. Pattern Recognition and Artificial Intelligence, 2006, 1(1): 52?56.[13] GUILHERME N D, AVINASH C K. Vision for mobile ro
收藏