图像处理中值滤波器中英文对照外文翻译文献

合集下载

文献翻译(中文)

文献翻译(中文)

对多脉冲噪声的自适应阈值中值滤波通信与信息工程学院,电子科技大学成都中国中国61005与技术学院抽象衰减噪声在图像处理中起重要作用。

几乎所有的传统中值滤波器涉及去除具有单个层,其噪声灰度值是恒定的脉冲噪音。

在本文中,一种新的自适应中值滤波,提出了处理这些不仅是单层噪声的图像。

自适应阈值滤波器(ATMF)已开发通过组合自适应中值过滤器(AMF)和两个动态阈值。

动态门限的,因为正在使用,ATMF是能够平衡在除去多脉冲噪声和图像的质量。

提供该方法与传统的中值滤波的比较。

一些视觉实施例用来表明所提出的滤波器的性能。

关键词:中值滤波;自适应中值过滤器(AMF);自适应阈值中值滤波器(ATMF);多脉冲噪声;影像处理图像往往是由脉冲噪声是由于来自传感器或交际渠道产生的错误损坏。

它的边缘检测,图像分割和目标识别过程之前,以消除图像中的噪点是非常重要的。

众所周知的中值滤波器(MF)和它的衍生物已被确认为去除脉冲噪声的有效手段。

中值滤波器的成功是基于两个主要性能:边缘保持高效的噪声衰减,随着对冲动型噪声的鲁棒性。

边缘保持在图像处理必不可少由于视觉感知[7]的性质。

尽管它在平滑噪声效能,MF倾向于当应用于图像均匀地除去细的细节。

为了消除这个缺点,一个著名的改性的中值滤波,自适应中值过滤器(AMF),已经提出了。

它具有可变的窗口大小去除脉冲同时保留锐度同时。

以这种方式,边缘信息和详细信息的完整性变得更好。

上面提到的过滤器不善于去除多脉冲噪声。

然而,实际情况是,图像是由多脉冲噪声,包括单层噪声经常被破坏。

在本文中,一个基于决策的和信号自适应中值滤波算法。

它不仅实现脉冲噪声均强检测和视觉质量恢复的结果,但也确实很好地抗多的噪音。

对于噪声的识别,新的标准已在AMF加入,以使效果处理多个噪声。

此后,新的过滤器,命名为自适应阈值中值滤波器(ATMF),增加了当地的内核区域的两个动态阈值来帮助检测噪音。

仿真结果表明,该过滤器是一样好AMF的一层脉冲噪声,但比其他许多中值滤波器更好的为多脉冲噪声。

数字图像处理英文文献翻译参考

数字图像处理英文文献翻译参考

…………………………………………………装………………订………………线…………………………………………………………………Hybrid Genetic Algorithm Based Image EnhancementTechnologyMu Dongzhou Department of the Information Engineering XuZhou College of Industrial TechnologyXuZhou, China ****************.cnXu Chao and Ge Hongmei Department of the Information Engineering XuZhou College of Industrial TechnologyXuZhou, China ***************.cn,***************.cnAbstract—in image enhancement, Tubbs proposed a normalized incomplete Beta function to represent several kinds of commonly used non-linear transform functions to do the research on image enhancement. But how to define the coefficients of the Beta function is still a problem. We proposed a Hybrid Genetic Algorithm which combines the Differential Evolution to the Genetic Algorithm in the image enhancement process and utilize the quickly searching ability of the algorithm to carry out the adaptive mutation and searches. Finally we use the Simulation experiment to prove the effectiveness of the method.Keywords- Image enhancement; Hybrid Genetic Algorithm; adaptive enhancementI. INTRODUCTIONIn the image formation, transfer or conversion process, due to other objective factors such as system noise, inadequate or excessive exposure, relative motion and so the impact will get the image often a difference between the original image (referred to as degraded or degraded) Degraded image is usually blurred or after the extraction of information through the machine to reduce or even wrong, it must take some measures for its improvement.Image enhancement technology is proposed in this sense, and the purpose is to improve the image quality. Fuzzy Image Enhancement situation according to the image using a variety of special technical highlights some of the information in the image, reduce or eliminate the irrelevant information, to emphasize the image of the whole or the purpose of local features. Image enhancement method is still no unified theory, image enhancement techniques can be divided into three categories: point operations, and spatial frequency enhancement methods Enhancement Act. This paper presents an automatic adjustment according to the image characteristics of adaptive image enhancement method that called hybrid genetic algorithm. It combines the differential evolution algorithm of adaptive search capabilities, automatically determines the transformation function of the parameter values in order to achieve adaptive image enhancement.…………………………………………………装………………订………………线…………………………………………………………………II. IMAGE ENHANCEMENT TECHNOLOGYImage enhancement refers to some features of the image, such as contour, contrast, emphasis or highlight edges, etc., in order to facilitate detection or further analysis and processing. Enhancements will not increase the information in the image data, but will choose the appropriate features of the expansion of dynamic range, making these features more easily detected or identified, for the detection and treatment follow-up analysis and lay a good foundation.Image enhancement method consists of point operations, spatial filtering, and frequency domain filtering categories. Point operations, including contrast stretching, histogram modeling, and limiting noise and image subtraction techniques. Spatial filter including low-pass filtering, median filtering, high pass filter (image sharpening). Frequency filter including homomorphism filtering, multi-scale multi-resolution image enhancement applied [1].III. DIFFERENTIAL EVOLUTION ALGORITHMDifferential Evolution (DE) was first proposed by Price and Storn, and with other evolutionary algorithms are compared, DE algorithm has a strong spatial search capability, and easy to implement, easy to understand. DE algorithm is a novel search algorithm, it is first in the search space randomly generates the initial population and then calculate the difference between any two members of the vector, and the difference is added to the third member of the vector, by which Method to form a new individual. If you find that the fitness of new individual members better than the original, then replace the original with the formation of individual self.The operation of DE is the same as genetic algorithm, and it conclude mutation, crossover and selection, but the methods are different. We suppose that the group size is P, the vector dimension is D, and we can express the object vector as (1):xi=[xi1,xi2,…,xiD] (i =1,…,P)(1) And the mutation vector can be expressed as (2):()321rrriXXFXV-⨯+=i=1,...,P (2) 1rX,2rX,3rX are three randomly selected individuals from group, and r1≠r2≠r3≠i.F is a range of [0, 2] between the actual type constant factor difference vector is used to control the influence, commonly referred to as scaling factor. Clearly the difference between the vector and the smaller the disturbance also smaller, which means that if groups close to the optimum value, the disturbance will be automatically reduced.DE algorithm selection operation is a "greedy " selection mode, if and only if the new vector ui the fitness of the individual than the target vector is better when the individual xi, ui will be retained to the next group. Otherwise, the target vector xi individuals remain in the original group, once again as the next generation of the parent vector.…………………………………………………装………………订………………线…………………………………………………………………IV. HYBRID GA FOR IMAGE ENHANCEMENT IMAGEenhancement is the foundation to get the fast object detection, so it is necessary to find real-time and good performance algorithm. For the practical requirements of different systems, many algorithms need to determine the parameters and artificial thresholds. Can use a non-complete Beta function, it can completely cover the typical image enhancement transform type, but to determine the Beta function parameters are still many problems to be solved. This section presents a Beta function, since according to the applicable method for image enhancement, adaptive Hybrid genetic algorithm search capabilities, automatically determines the transformation function of the parameter values in order to achieve adaptive image enhancement.The purpose of image enhancement is to improve image quality, which are more prominent features of the specified restore the degraded image details and so on. In the degraded image in a common feature is the contrast lower side usually presents bright, dim or gray concentrated. Low-contrast degraded image can be stretched to achieve a dynamic histogram enhancement, such as gray level change. We use Ixy to illustrate the gray level of point (x, y) which can be expressed by (3).Ixy=f(x, y) (3) where: “f” is a linear or nonline ar function. In general, gray image have four nonlinear translations [6] [7] that can be shown as Figure 1. We use a normalized incomplete Beta function to automatically fit the 4 categories of image enhancement transformation curve. It defines in (4):()()()()10,01,111<<-=---⎰βαβαβαdtttBufu(4) where:()()⎰---=1111,dtttBβαβα(5) For different value of α and β, we can get response curve from (4) and (5).The hybrid GA can make use of the previous section adaptive differential evolution algorithm to search for the best function to determine a value of Beta, and then each pixel grayscale values into the Beta function, the corresponding transformation of Figure 1, resulting in ideal image enhancement. The detail description is follows:Assuming the original image pixel (x, y) of the pixel gray level by the formula (4),denoted byxyi,()Ω∈yx,, here Ω is the image domain. Enhanced image is denoted by Ixy. Firstly, the image gray value normalized into [0, 1] by (6).minmaxminiiiig xyxy--=(6)where:maxi andm ini express the maximum and minimum of image gray relatively.Define the nonlinear transformation function f(u) (0≤u≤1) to transform source image…………………………………………………装………………订………………线…………………………………………………………………Finally, we use the hybrid genetic algorithm to determine the appropriate Beta function f (u) the optimal parameters αand β. Will enhance the image Gxy transformed antinormalized.V. EXPERIMENT AND ANALYSISIn the simulation, we used two different types of gray-scale images degraded; the program performed 50 times, population sizes of 30, evolved 600 times. The results show that the proposed method can very effectively enhance the different types of degraded image.Figure 2, the size of the original image a 320 × 320, it's the contrast to low, and some details of the more obscure, in particular, scarves and other details of the texture is not obvious, visual effects, poor, using the method proposed in this section, to overcome the above some of the issues and get satisfactory image results, as shown in Figure 5 (b) shows, the visual effects have been well improved. From the histogram view, the scope of the distribution of image intensity is more uniform, and the distribution of light and dark gray area is more reasonable. Hybrid genetic algorithm to automatically identify the nonlinear…………………………………………………装………………订………………线…………………………………………………………………transformation of the function curve, and the values obtained before 9.837,5.7912, from the curve can be drawn, it is consistent with Figure 3, c-class, that stretch across the middle region compression transform the region, which were consistent with the histogram, the overall original image low contrast, compression at both ends of the middle region stretching region is consistent with human visual sense, enhanced the effect of significantly improved.Figure 3, the size of the original image a 320 × 256, the overall intensity is low, the use of the method proposed in this section are the images b, we can see the ground, chairs and clothes and other details of the resolution and contrast than the original image has Improved significantly, the original image gray distribution concentrated in the lower region, and the enhanced image of the gray uniform, gray before and after transformation and nonlinear transformation of basic graph 3 (a) the same class, namely, the image Dim region stretching, and the values were 5.9409,9.5704, nonlinear transformation of images degraded type inference is correct, the enhanced visual effect and good robustness enhancement.Difficult to assess the quality of image enhancement, image is still no common evaluation criteria, common peak signal to noise ratio (PSNR) evaluation in terms of line, but the peak signal to noise ratio does not reflect the human visual system error. Therefore, we use marginal protection index and contrast increase index to evaluate the experimental results.Edgel Protection Index (EPI) is defined as follows:…………………………………………………装………………订………………线…………………………………………………………………(7)Contrast Increase Index (CII) is defined as follows:minmaxminmax,GGGGCCCEOD+-==(8)In figure 4, we compared with the Wavelet Transform based algorithm and get the evaluate number in TABLE I.Figure 4 (a, c) show the original image and the differential evolution algorithm for enhanced results can be seen from the enhanced contrast markedly improved, clearer image details, edge feature more prominent. b, c shows the wavelet-based hybrid genetic algorithm-based Comparison of Image Enhancement: wavelet-based enhancement method to enhance image detail out some of the image visual effect is an improvement over the original image, but the enhancement is not obvious; and Hybrid genetic algorithm based on adaptive transform image enhancement effect is very good, image details, texture, clarity is enhanced compared with the results based on wavelet transform has greatly improved the image of the post-analytical processing helpful. Experimental enhancement experiment using wavelet transform "sym4" wavelet, enhanced differential evolution algorithm experiment, the parameters and the values were 5.9409,9.5704. For a 256 × 256 size image transform based on adaptive hybrid genetic algorithm in Matlab 7.0 image enhancement software, the computing time is about 2 seconds, operation is very fast. From TABLE I, objective evaluation criteria can be seen, both the edge of the protection index, or to enhance the contrast index, based on adaptive hybrid genetic algorithm compared to traditional methods based on wavelet transform has a larger increase, which is from This section describes the objective advantages of the method. From above analysis, we can see…………………………………………………装………………订………………线…………………………………………………………………that this method.From above analysis, we can see that this method can be useful and effective.VI. CONCLUSIONIn this paper, to maintain the integrity of the perspective image information, the use of Hybrid genetic algorithm for image enhancement, can be seen from the experimental results, based on the Hybrid genetic algorithm for image enhancement method has obvious effect. Compared with other evolutionary algorithms, hybrid genetic algorithm outstanding performance of the algorithm, it is simple, robust and rapid convergence is almost optimal solution can be found in each run, while the hybrid genetic algorithm is only a few parameters need to be set and the same set of parameters can be used in many different problems. Using the Hybrid genetic algorithm quick search capability for a given test image adaptive mutation, search, to finalize the transformation function from the best parameter values. And the exhaustive method compared to a significant reduction in the time to ask and solve the computing complexity. Therefore, the proposed image enhancement method has some practical value.REFERENCES[1] HE Bin et al., Visual C++ Digital Image Processing [M], Posts & Telecom Press,2001,4:473~477[2] Storn R, Price K. Differential Evolution—a Simple and Efficient Adaptive Scheme forGlobal Optimization over Continuous Space[R]. International Computer Science Institute, Berlaey, 1995.[3] Tubbs J D. A note on parametric image enhancement [J].Pattern Recognition.1997,30(6):617-621.[4] TANG Ming, MA Song De, XIAO Jing. Enhancing Far Infrared Image Sequences withModel Based Adaptive Filtering [J] . CHINESE JOURNAL OF COMPUTERS, 2000, 23(8):893-896.[5] ZHOU Ji Liu, LV Hang, Image Enhancement Based on A New Genetic Algorithm [J].Chinese Journal of Computers, 2001, 24(9):959-964.[6] LI Yun, LIU Xuecheng. On Algorithm of Image Constract Enhancement Based onWavelet Transformation [J]. Computer Applications and Software, 2008,8.[7] XIE Mei-hua, WANG Zheng-ming, The Partial Differential Equation Method for ImageResolution Enhancement [J]. Journal of Remote Sensing, 2005,9(6):673-679.…………………………………………………装………………订………………线…………………………………………………………………基于混合遗传算法的图像增强技术Mu Dongzhou 徐州工业职业技术学院信息工程系 XuZhou, China****************.cnXu Chao and Ge Hongmei 徐州工业职业技术学院信息工程系 XuZhou,********************.cn,***************.cn摘要—在图像增强之中,塔布斯提出了归一化不完全β函数表示常用的几种使用的非线性变换函数对图像进行研究增强。

数字图像检测中英文对照外文翻译文献

数字图像检测中英文对照外文翻译文献

中英文对照外文翻译(文档含英文原文和中文翻译)Edge detection in noisy images by neuro-fuzzyprocessing通过神经模糊处理的噪声图像边缘检测AbstractA novel neuro-fuzzy (NF) operator for edge detection in digital images corrupted by impulse noise is presented. The proposed operator is constructed by combining a desired number of NF subdetectors with a postprocessor. Each NF subdetector in the structure evaluates a different pixel neighborhood relation. Hence, the number of NF subdetectors in the structure may be varied to obtain the desired edge detection performance. Internal parameters of the NF subdetectors are adaptively optimized by training by using simple artificial training images. The performance of the proposed edge detector is evaluated on different test images and compared with popular edge detectors from the literature. Simulation results indicate that the proposed NF operator outperforms competing edge detectors and offers superior performance in edge detection in digital images corrupted by impulse noise.Keywords: Neuro-fuzzy systems; Image processing; Edge detection摘要针对被脉冲信号干扰的数字图像进行边缘检测,提出了一种新型的NF边缘检测器,它是由一定数量的NF子探测器与一个后处理器组成。

中值滤波器脉冲噪声中英文对照外文翻译文献

中值滤波器脉冲噪声中英文对照外文翻译文献

中英文资料外文翻译文献Improved 2-D Median Filter for On-Line Impulse Noise Suppressiom Abstract-An inproved 2-D median filter employing multishell concept to suppress impulse noise ,is presented.The performance of proposed filter is evaluated over image ‘LENA’,The impulsive noise is added using MATLAB utility.The modified strategy reduces the mnuber of replacement and results in better performance and simple hardware realization that is suitable for on-line implementation.Index terms-Median Filter , Multi-shell Median Filter, Impulse NoiseI.INTRODUCTIONIn TV and other imaging systems,impulse noise is a common impairment . The standard T.V.Broadcast signal is often contaminated with impulsive noise arising from various sources such as household electrical appliance and atmospheric disturbances.Broad banding of the signal further increases the level of impulsive noise. V arious filters are proposed to suppress such impairments[1].The median filter(MF)[1-2] is widely usedfor impulse noise suppression and the multishell median filter(MMF)[3] introduces the concept of missing line recovery. Although these filters have satisfactory performance, MMF failsto filter two impulse noises in the same prossing window. Moveover,these filters tend to blur the images due to too many replacements. C.J.Juan proposed a modified multishell median filter (MMMF)[4], which removes most of the shortcomings associated with the MF and the MMF. However, it is observed that under certain condions, to be discussed in the follow sections, MMMF fails to perform the desired filtering operation .Moreover,the number of calculations/replacements invoved on the basis of MIN/MAX conditions is still too large and makes the filter difficult to realize,particulariy for real time applications.In this paper, the threshold strtegy of MMMF is modified so that:(a)effective noise filtering operations are performed under allconditions,and(b)number of calculations/replacements is reduced and simplified. This results in a simple hardware realization of the filter.II.PROPOSED MODIFICATIONConsider a 3x3-processing window, with P5 as the central pixel,as shown in Figure 1.P1 P2 P3P4 P5 P6P7 P8 P9Fig.1. A 3x3 processing windowThe output of MMMF as proposed in [4] isOutput (X,Y)= Max(P2,P8)if P5﹥Max[S]P5 if Min [s]﹤Max[S]Min(P2,P8) if P5﹤Max[S] (1)Where S is the set of samples surrounding central pixels except(P4.P6)i.e.S={P1,P2,P3,P7,P8,P9} (2) The principle invoved in the replacement strategy of Equation(1) is that if P5 is corrupted by noise ,it is better to replaceits gray level by P2 or P8 than by using Min[S] orMax[S] .also,due to missing lines error,since P4 and P6 may belost, they are not considered in Equation(2).The limitation of Equation(1) is that when Min[S] or Max[S] arealso corrupted by impulse noise,i.e.either Min[S] or Max[S] isequal to P5,Equation(1)fails to perform the desired filtering operation.To overcome this limitation following modificationsin the replacement strategy of Equation(1),are proposed.Output (X,Y)= Max(P2,P8)if P5≥Max[S]P5 if Min [s]<P5<Max[S] Min(P2,P8) if P5≤Max[S] (3)It has been observed that more than 70-80% points in an image,the gray level diatances of P5 from(P2 or P8) and from Max[S] are below 16.This is shown in Fig.2 for the image ‘LENA’.This fact is used to further reduce unnessary replacements,thereby reducing the bluring of the images.Thus taking into considertion of Figure(3) can be further modified asOutput (X,Y)= Max(P2,P8)if P5-Max[S]≥16Max(P2,P8) if Min [s]-P5≥16P5 otherwise (4)Equation 4 indicates that replacing action takes place only when the distance between P5 and Min[S] or Max[S] is no smaller than 16. This strtegy thus avoids the necessary replacements and reduces blurring of the images.Moreover,it can be implemented using simple comparators and subtractors.Gray level distancesFig.2. Gray level distances between central point and its neighboring points for the image ‘LENNA’Ⅲ .RESULTSFigure 3 shows the original image ‘LENNA’and Figure 4 shows the same image when corrupted with impulse noise. Results of median filter and the proposed filter are given in Figures 5 and 6, paring Figures 5 and 6, it is observed that the result of the proposed filter is much better than those obtained using the median filter. Aithough,the median filter remove the impulsive moise effectively, however,the image gets blurred.The proposed filter removes the impulsive noise and also preserves the details of the image.A multishell filter employing the modified replacement strategy is presentde in this paper.The modified filter effectively suppresses the inpulse moise.It uses threshold conditions that require fewer comparisons and replacements and is faster as compared to the other multishell median filters.moreover,it can be realized using simple comparators and subtractors and subtractors and hence can be effectively used in real time applications改进二维中值滤波器在线脉冲噪声的抑制摘要:一种改进二维中值滤波器,采用多壳的概念,以抑制脉冲噪声,拟定的过滤器的性能进行评估超过图像“LENNA”的中值滤波,脉冲噪声被添加使用到MATLAB的实用工具中。

数字图像处理 外文翻译 外文文献 英文文献 数字图像处理

数字图像处理 外文翻译 外文文献 英文文献 数字图像处理

Digital Image Processing1 IntroductionMany operators have been proposed for presenting a connected component n a digital image by a reduced amount of data or simplied shape. In general we have to state that the development, choice and modi_cation of such algorithms in practical applications are domain and task dependent, and there is no \best method". However, it is interesting to note that there are several equivalences between published methods and notions, and characterizing such equivalences or di_erences should be useful to categorize the broad diversity of published methods for skeletonization. Discussing equivalences is a main intention of this report.1.1 Categories of MethodsOne class of shape reduction operators is based on distance transforms. A distance skeleton is a subset of points of a given component such that every point of this subset represents the center of a maximal disc (labeled with the radius of this disc) contained in the given component. As an example in this _rst class of operators, this report discusses one method for calculating a distance skeleton using the d4 distance function which is appropriate to digitized pictures. A second class of operators produces median or center lines of the digital object in a non-iterative way. Normally such operators locate critical points _rst, and calculate a speci_ed path through the object by connecting these points.The third class of operators is characterized by iterative thinning. Historically, Listing [10] used already in 1862 the term linear skeleton for the result of a continuous deformation of the frontier of a connected subset of a Euclidean space without changing the connectivity of the original set, until only a set of lines and points remains. Many algorithms in image analysis are based on this general concept of thinning. The goal is a calculation of characteristic properties of digital objects which are not related to size or quantity. Methods should be independent from the position of a set in the plane or space, grid resolution (for digitizing this set) or the shape complexity of the given set. In the literature the term \thinning" is not usedin a unique interpretation besides that it always denotes a connectivity preserving reduction operation applied to digital images, involving iterations of transformations of speci_ed contour points into background points. A subset Q _ I of object points is reduced by a de_ned set D in one iteration, and the result Q0 = Q n D becomes Q for the next iteration. Topology-preserving skeletonization is a special case of thinning resulting in a connected set of digital arcs or curves. A digital curve is a path p =p0; p1; p2; :::; pn = q such that pi is a neighbor of pi 1, 1 _ i _ n, and p = q. A digital curve is called simple if each point pi has exactly two neighbors in this curve. A digital arc is a subset of a digital curve such that p 6= q. A point of a digital arc which has exactly one neighbor is called an end point of this arc. Within this third class of operators (thinning algorithms) we may classify with respect to algorithmic strategies: individual pixels are either removed in a sequential order or in parallel. For example, the often cited algorithm by Hilditch [5] is an iterative process of testing and deleting contour pixels sequentially in standard raster scan order. Another sequential algorithm by Pavlidis [12] uses the de_nition of multiple points and proceeds by contour following. Examples of parallel algorithms in this third class are reduction operators which transform contour points into background points. Di_erences between these parallel algorithms are typically de_ned by tests implemented to ensure connectedness in a local neighborhood. The notion of a simple point is of basic importance for thinning and it will be shown in this report that di_erent de_nitions of simple points are actually equivalent. Several publications characterize properties of a set D of points (to be turned from object points to background points) to ensure that connectivity of object and background remain unchanged. The report discusses some of these properties in order to justify parallel thinning algorithms.1.2 BasicsThe used notation follows [17]. A digital image I is a function de_ned on a discrete set C , which is called the carrier of the image. The elements of C are grid points or grid cells, and the elements (p; I(p)) of an image are pixels (2D case) or voxels (3D case). The range of a (scalar) image is f0; :::Gmaxg with Gmax _ 1. The range of a binary image is f0; 1g. We only use binary images I in this report. Let hIi be the set of all pixel locations with value 1, i.e. hIi = I 1(1). The image carrier is de_ned on an orthogonal grid in 2D or 3Dspace. There are two options: using the grid cell model a 2D pixel location p is a closed square (2-cell) in the Euclidean plane and a 3D pixel location is a closed cube (3-cell) in the Euclidean space, where edges are of length 1 and parallel to the coordinate axes, and centers have integer coordinates. As a second option, using the grid point model a 2D or 3D pixel location is a grid point.Two pixel locations p and q in the grid cell model are called 0-adjacent i_ p 6= q and they share at least one vertex (which is a 0-cell). Note that this speci_es 8-adjacency in 2D or 26-adjacency in 3D if the grid point model is used. Two pixel locations p and q in the grid cell model are called 1- adjacent i_ p 6= q and they share at least one edge (which is a 1-cell). Note that this speci_es 4-adjacency in 2D or 18-adjacency in 3D if the grid point model is used. Finally, two 3D pixel locations p and q in the grid cell model are called 2-adjacent i_ p 6= q and they share at least one face (which is a 2-cell). Note that this speci_es 6-adjacency if the grid point model is used. Any of these adjacency relations A_, _ 2 f0; 1; 2; 4; 6; 18; 26g, is irreexive and symmetric on an image carrier C. The _-neighborhood N_(p) of a pixel location p includes p and its _-adjacent pixel locations. Coordinates of 2D grid points are denoted by (i; j), with 1 _ i _ n and 1 _ j _ m; i; j are integers and n;m are the numbers of rows and columns of C. In 3Dwe use integer coordinates (i; j; k). Based on neighborhood relations we de_ne connectedness as usual: two points p; q 2 C are _-connected with respect to M _ C and neighborhood relation N_ i_ there is a sequence of points p = p0; p1; p2; :::; pn = q such that pi is an _-neighbor of pi 1, for 1 _ i _ n, and all points on this sequence are either in M or all in the complement of M. A subset M _ C of an image carrier is called _-connected i_ M is not empty and all points in M are pairwise _-connected with respect to set M. An _-component of a subset S of C is a maximal _-connected subset of S. The study of connectivity in digital images has been introduced in [15]. It follows that any set hIi consists of a number of _-components. In case of the grid cell model, a component is the union of closed squares (2D case) or closed cubes (3D case). The boundary of a 2-cell is the union of its four edges and the boundary of a 3-cell is the union of its six faces. For practical purposes it is easy to use neighborhood operations (called local operations) on a digital image I which de_ne a value at p 2 C in the transformed image based on pixelvalues in I at p 2 C and its immediate neighbors in N_(p).2 Non-iterative AlgorithmsNon-iterative algorithms deliver subsets of components in specied scan orders without testing connectivity preservation in a number of iterations. In this section we only use the grid point model.2.1 \Distance Skeleton" AlgorithmsBlum [3] suggested a skeleton representation by a set of symmetric points.In a closed subset of the Euclidean plane a point p is called symmetric i_ at least 2 points exist on the boundary with equal distances to p. For every symmetric point, the associated maximal disc is the largest disc in this set. The set of symmetric points, each labeled with the radius of the associated maximal disc, constitutes the skeleton of the set. This idea of presenting a component of a digital image as a \distance skeleton" is based on the calculation of a speci_ed distance from each point in a connected subset M _ C to the complement of the subset. The local maxima of the subset represent a \distance skeleton". In [15] the d4-distance is specied as follows. De_nition 1 The distance d4(p; q) from point p to point q, p 6= q, is the smallest positive integer n such that there exists a sequence of distinct grid points p = p0,p1; p2; :::; pn = q with pi is a 4-neighbor of pi 1, 1 _ i _ n. If p = q the distance between them is de_ned to be zero. The distance d4(p; q) has all properties of a metric. Given a binary digital image. We transform this image into a new one which represents at each point p 2 hIi the d4-distance to pixels having value zero. The transformation includes two steps. We apply functions f1 to the image I in standard scan order, producing I_(i; j) = f1(i; j; I(i; j)), and f2 in reverse standard scan order, producing T(i; j) = f2(i; j; I_(i; j)), as follows:f1(i; j; I(i; j)) =8><>>:0 if I(i; j) = 0minfI_(i 1; j)+ 1; I_(i; j 1) + 1gif I(i; j) = 1 and i 6= 1 or j 6= 1m+ n otherwisef2(i; j; I_(i; j)) = minfI_(i; j); T(i+ 1; j)+ 1; T(i; j + 1) + 1gThe resulting image T is the distance transform image of I. Note that T is a set f[(i; j); T(i; j)] : 1 _ i _ n ^ 1 _ j _ mg, and let T_ _ T such that [(i; j); T(i; j)] 2 T_ i_ none of the four points in A4((i; j)) has a value in T equal to T(i; j)+1. For all remaining points (i; j) let T_(i; j) = 0. This image T_ is called distance skeleton. Now we apply functions g1 to the distance skeleton T_ in standard scan order, producing T__(i; j) = g1(i; j; T_(i; j)), and g2 to the result of g1 in reverse standard scan order, producing T___(i; j) = g2(i; j; T__(i; j)), as follows:g1(i; j; T_(i; j)) = maxfT_(i; j); T__(i 1; j) 1; T__(i; j 1) 1gg2(i; j; T__(i; j)) = maxfT__(i; j); T___(i + 1; j) 1; T___(i; j + 1) 1gThe result T___ is equal to the distance transform image T. Both functions g1 and g2 de_ne an operator G, with G(T_) = g2(g1(T_)) = T___, and we have [15]: Theorem 1 G(T_) = T, and if T0 is any subset of image T (extended to an image by having value 0 in all remaining positions) such that G(T0) = T, then T0(i; j) = T_(i; j) at all positions of T_ with non-zero values. Informally, the theorem says that the distance transform image is reconstructible from the distance skeleton, and it is the smallest data set needed for such a reconstruction. The used distance d4 di_ers from the Euclidean metric. For instance, this d4-distance skeleton is not invariant under rotation. For an approximation of the Euclidean distance, some authors suggested the use of di_erent weights for grid point neighborhoods [4]. Montanari [11] introduced a quasi-Euclidean distance. In general, the d4-distance skeleton is a subset of pixels (p; T(p)) of the transformed image, and it is not necessarily connected.2.2 \Critical Points" AlgorithmsThe simplest category of these algorithms determines the midpoints of subsets of connected components in standard scan order for each row. Let l be an index for the number of connected components in one row of the original image. We de_ne the following functions for 1 _ i _ n: ei(l) = _ j if this is the lth case I(i; j) = 1 ^ I(i; j 1) = 0 in row i, counting from the left, with I(i; 1) = 0 ,oi(l) = _ j if this is the lth case I(i; j) = 1^ I(i; j+ 1) = 0 ,in row i, counting from the left, with I(i;m+ 1) = 0 ,mi(l) = int((oi(l) ei(l)=2)+ oi(l) ,The result of scanning row i is a set of coordinates (i;mi(l)) of midpoints ,of the connected components in row i. The set of midpoints of all rows constitutes a critical point skeleton of an image I. This method is computationally eÆcient.The results are subsets of pixels of the original objects, and these subsets are not necessarily connected. They can form \noisy branches" when object components are nearly parallel to image rows. They may be useful for special applications where the scanning direction is approximately perpendicular to main orientations of object components.References[1] C. Arcelli, L. Cordella, S. Levialdi: Parallel thinning of binary pictures. Electron. Lett. 11:148{149, 1975}.[2] C. Arcelli, G. Sanniti di Baja: Skeletons of planar patterns. in: Topolog- ical Algorithms for Digital Image Processing (T. Y. Kong, A. Rosenfeld, eds.), North-Holland, 99{143, 1996.}[3] H. Blum: A transformation for extracting new descriptors of shape. in: Models for the Perception of Speech and Visual Form (W. Wathen- Dunn, ed.), MIT Press, Cambridge, Mass., 362{380, 1967.19}数字图像处理1引言许多研究者已提议提出了在数字图像里的连接组件是由一个减少的数据量或简化的形状。

外文翻译之滤波器外文原文及翻译

外文翻译之滤波器外文原文及翻译

This invention relates to mechanical wave filters and more particularly to those which comprise one or more transverse members adapted for flexural vibration.An object of the invention is to reduce the minimum width of transmission band obtainable in a mechanical wave filter which employs a transverse flexural vibratory member.Other objects of the invention are to simplify the mechanical structure and reduce the cost of filters of this type.One form of mechanical wave filter comprises a central rod of acoustic material and one or more centrally located transverse members adapted to be set into flexural vibration when longitudinal vibrations are impressed upon an end of the rod. For a given rod, the width of the transmission band decreases as the mass of the transverse members is increased. Peaks of attenuation occur at the antiresonant frequencies of the transverse members, and these peaks are usually located close to the band limits to obtain sharp cut-offs. Heretofore, the transverse members have been made in the form of crossbars. The flexural antiresonance of abar is directly proportional to its width and inversely proportional to the square of its length. Since the antiresonant frequency is fixed, to increase the mass, and thereby narrow the band, the width of the bar may be increased by a factor K and its length increased at the same time by the square root of K. However, a point is reached at which the ratio of the width to the length is so large that the bar will no longer vibrate satisfactorily in the flexural mode. There is, therefore, a fairly definite limit on the minimum band width obtainable with filters using crossbars.In accordance with the present invention this limitation on minimum width of band is overcome by using a flexurally vibrating disc, mounted at its center, as the transverse member. By using a disc the mass is greatly increased and therefore a much narrower band may be obtained. A single disc will provide a peak of attenuation either above or below the band. Two discs located near the mid-point of the central rod and close together will provide a peak above and a peak below the band. The central rod has a small cross-sectional dimensions compared to its length, which isapproximately equal to a half wave-length at a frequency within the band.The nature of the invention will be more fully understood from the following detailed description and by reference to the accompanying drawing, in which like reference characters refer to similar parts and in which: Fig. 1 is a perspective view of a mechanical wave filter in accordance with the invention employing a single disc, andFig. 2 is a perspective view of a two-disc filter. Taking up the figures in more detail, Fig. 1 shows a mechanical wave filter in accordance with the invention comprising a central rod 1 of circular cross section and a transverse disc 2,both made of suitable acoustic material. The diameter A of the rod 1 is small compared to its length B, which is approximately a half wave-length at a cut-off frequency. The disc 2 has a diameter C and thickness D and is centrally mounted near the center of the rod 1. Longitudinal vibrations impressed upon an end of the bar 1 by a suitable driving device, represented diagrammatically by the box 3 shown in broken outline, will cause the disc 2 to vibrate in the flexural mode. Thelongitudinal vibrations of the rod 1 may be picked up at the other end of the filter by some suitable device, represented by the box 4 shown in broken outline. A peak of attenuation will occur at each frequency at which the disc 2 is antiresonant. The dimensions C and D of the disc 2 are, therefore, so proportioned that the first flexural antiresonance f l , given approximately by the following formula, occurs at a frequency, usually near aband limit, at which a peak is desired:121.872D f G = (1) where C and D are in centimeters, M is the density of the material, P is Poisson's ratio and Y is Young's modulus. If f l is on the lower side of the transmission band, the length B of the rod 1 is made approximately equal to a half wavelength at the upper cut-off frequency. If f l is above the band, B is made approximately equal to a half wave-length at the lower cut-off.If two peaks of attenuation are desired, one on either side of the band and close to the band limits, the structure shown in Fig. 2 may be used. The filtercomprises a central rod 1 and two transverse, centrally mounted discs 5 and 6, located close together one on either side of the center of the rod 1. One of the discs has its first flexural antiresonance at a frequency close to one limit of the band and the other disc has its first flexural antiresonance close to the other limit of the band. In this case the length B of the rod 1 is approximately equal to a half wavelength at the-band frequency.In mechanical filters of the type shown in Figs. 1 and 2, employing transverse impedance members, for a given rod 1 the width of the band decreases with an increase in the mass of the transverse members. The use of discs, such as 2, 5 and 6, for these members allows their mass to be greatly increased as compared, for example, with crossbars, and therefore a much narrower band may be obtained. The former limit on the minimum band width obtainable with filters using transverse members is thus greatly lowered in the disc-type filters of the present invention. Another distinct advantage of the disc-type filter is its mechanical simplicity. The filter may, for example, be made from a single piece of metal andcheaply turned out on a lathe.The image impedance Z of the filter, which should match the image impedance of the driving means at themid-band frequency, is given by the formula:Z Z = (2)where w is the angular frequency, j is the quadrantal operator, V is the velocity of propagation, equal to the square root of the ratio of Y to M, Z 0 is the characteristic impedance of the rod 1 and Z D is the impedance of the disc 2, for the filter of Fig.1, or the sum of the impedances of the two discs 5 and 6, for the filter of Fig.2. This image impedance is of a type which can be readily matched by a driver comprising a piezoelectric crystal attached at its end to an end of the rod 1.The filter will have a transmission band below, and one or more bands above, the principal band which has been here considered. These extraneous bands, if objectionable, may be eliminated by attenuation provided by the driving means, which should be designed to have a transmission band coinciding ,with the principal band of the mechanical filter. Thediscrimination of the filter may, of course, be increased dy connecting in tandem two or more sections of the type shown in Fig. 1 or Fig. 2.What is claimed is:1. A mechanical wave filter for transmitting a band of frequencies comprising a rod and a transverse disc both made of acoustic material, said rod having a length approximately equal to a half wave-length at a frequency within said band and said disc being centrally mounted near the center of said rod and adapted to be set into flexural vibration by longitudinal vibrations impressed upon an end of said rod and having a flexural antiresonance at a frequency close to one limit of said band.2. A filter in accordance with claim 1 in which said rod has a circular cross section.3. A filter in accordance with claim 1 in which the cross-sectional dimensions of said rod are small compared to its length.4. A filter in accordance with claim 1 in which said rod has a circular , cross section the diameter of which is small compared to its length.5. A filter in accordance with claim 1 in which said flexural antiresonance of said disc is its first.6. A mechanical wave filter for transmitting a band of frequencies comprising a rod and a transverse disc both made of acoustic material, said rod having a length approximately equal to a half wave-length at a frequency within said band and said disc being centrally mounted near the center of said rod and adapted to be set into flexural vibration by longitudinal vibrations impressed upon an end of said rod and having its first flexural antiresonance at a frequency on one side of said band and said rod having a length approximately equal to a half wave-length at the band limit on the other side of said band.7. A filter in accordance with claim 1 in which said flexural antiresonance of said disc is its first and said rod has a length approximately equal to a half wave-length at the other limit of said band.8. A mechanical wave filter for transmitting a band of frequencies comprising a rod and a transverse disc both made of acoustic material, said rod having a length approximately equal to a half wave-length at afrequency within said band and said disc being centrally mounted near the center of said rod and adapted to be set into flexural vibration by longitudinal vibrations impressed upon an end of said rod and having its first flexural antiresonance at a frequency on the upper side of said band and said rod having a length approximately equal to a half wave-length at the lower limit of said band.9. A filter in accordance with claim 1 in which said flexural antiresonance of said disc is its first and is located above said band and said rod has a length approximately equal to a half wave-length at the lower limit of said band.10. A mechanical wave filter for transmitting a band of frequencies comprising a rod and a transverse disc both made of acoustic material, said rod having a length approximately equal to a half wave-length at a frequency within said band and said disc being centrally mounted near the center of said rod and adapted to be set into flexural vibration by longitudinal vibrations impressed upon an end of said rod and having its first flexural antiresonance at a frequency on the lower sideof said band and said rod having a length approximately equal to a half wave-length at the upper limit of said band.11. A filter in accordance with claim 1 in which said flexural antiresonance of said disc is its first and is located below said band and said rod has a length approximately equal to a half wave-length at the upper limit of said band.12. A mechanical wave filter for transmitting a band of frequencies comprising a rod and two transverse discs, said rod having a length approximately equal to a half wave-length at a frequency within said band, said discs being centrally mounted, located close together near the center of said rod and adapted to be set into flexural vibration by longitudinal vibrations impressed upon an end of said rod, one of said discs having a flexural antiresonance at a frequency below said band and the other of said discs having a flexural antiresonance at a frequency above said band.13. A filter in accordance with claim 12 in which said discs are located one on either side of the center of said rod.rod has a length approximately equal to a half wave-length at the mid-band frequency.15. A filter in accordance with claim 12 in which said flexural antiresonances are located,respectively, close to the limits of said band.16. A filter in accordance with claim 12 in which said flexural antiresonance of said one disc is the only one occurring below said band and said flexural antiresonance of said other disc is its first.17. A mechanical wave filter for transmitting a band of frequencies comprising a rod and two transverse discs, said rod having a length approximately equal to a half wave-length at a frequency within said band, said discs being centrally mounted, located close together near the center of said rod and adapted to be set into flexural vibration by longitudinal vibrations impressed upon an end of said rod, one of said discs having its first flexural antiresonance at a frequency close to one limit of said band and the other of said discs having its first flexural antiresonance at a frequency close to the other limit of said band.rod has a length approximately equal to a half wave-length at the mid-band frequency.WARREN P. MASON.翻译本发明涉及机械滤波器,特别是那些包含一个或多个适应弯曲振动的横向构件。

数字图像处理与边缘检测中英文对照外文翻译文献

数字图像处理与边缘检测中英文对照外文翻译文献

中英文资料对照外文翻译Digital Image Processing and Edge DetectionDigital Image ProcessingInterest in digital image processing methods stems from two principal applica- tion areas: improvement of pictorial information for human interpretation; and processing of image data for storage, transmission, and representation for au- tonomous machine perception.An image may be defined as a two-dimensional function, f(x, y), where x and y are spatial (plane) coordinates, and the amplitude of f at any pair of coordinates (x, y) is called the intensity or gray level of the image at that point. When x, y, and the amplitude values of f are all finite, discrete quantities, we call the image a digital image. The field of digital image processing refers to processing digital images by means of a digital computer. Note that a digital image is composed of a finite number of elements, each of which has a particular location and value. These elements are referred to as picture elements, image elements, pels, and pixels. Pixel is the term most widely used to denote the elements of a digital image.Vision is the most advanced of our senses, so it is not surprising that images play the single most important role in human perception. However, unlike humans, who are limited to the visual band of the electromagnetic (EM) spec- trum, imaging machines cover almost the entire EM spectrum, ranging from gamma to radio waves. They can operate on images generated by sources that humans are not accustomed to associating with images. These include ultra- sound, electron microscopy, and computer-generated images. Thus, digital image processing encompasses a wide and varied field of applications.There is no general agreement among authors regarding where image processing stops and other related areas, such as image analysis and computer vi- sion, start. Sometimes a distinction is made by defining image processing as a discipline in which both the input and output of a process are images. We believe this to be a limiting and somewhat artificial boundary. For example, under this definition, even the trivial task of computing the average intensity of an image (which yields a single number) would not be considered an image processing operation. On the other hand, there are fields such as computer vision whose ultimate goal is to use computers to emulate human vision, including learning and being able to make inferences and take actions based on visual inputs. This area itself is a branch of artificial intelligence (AI) whose objective is to emulate human intelligence. The field of AI is in its earliest stages of infancy in terms of development, with progress having been much slower thanoriginally anticipated. The area of image analysis (also called image understanding) is in be- tween image processing and computer vision.There are no clearcut boundaries in the continuum from image processing at one end to computer vision at the other. However, one useful paradigm is to consider three types of computerized processes in this continuum: low-, mid-, and highlevel processes. Low-level processes involve primitive opera- tions such as image preprocessing to reduce noise, contrast enhancement, and image sharpening. A low-level process is characterized by the fact that both its inputs and outputs are images. Mid-level processing on images involves tasks such as segmentation (partitioning an image into regions or objects), description of those objects to reduce them to a form suitable for computer processing, and classification (recognition) of individual objects. A midlevel process is characterized by the fact that its inputs generally are images, but its outputs are attributes extracted from those images (e.g., edges, contours, and the identity of individual objects). Finally, higherlevel processing involves “makin g sense”of an ensemble of recognized objects, as in image analysis, and, at the far end of the continuum, performing the cognitive functions normally associated with vision.Based on the preceding comments, we see that a logical place of overlap between image processing and image analysis is the area of recognition of individual regions or objects in an image. Thus, what we call in this book digital image processing encompasses processes whose inputs and outputs are images and, in addition, encompasses processes that extract attributes from images, up to and including the recognition of individual objects. As a simple illustration to clarify these concepts, consider the area of automated analysis of text. The processes of acquiring an image of the area containing the text, preprocessing that image, extracting (segmenting) the individual characters, describing the characters in a form suitable for computer processing, and recognizing those individual characters are in the scope of what we call digital image processing in this book. Making sense of the content of the page may be viewed as being in the domain of image analysis and even computer vision, depending on the level of complexity implied by the statement “making sense.”As will become evident shortly, digital image processing, as we have defined it, is used successfully in a broad range of areas of exceptional social and economic value.The areas of application of digital image processing are so varied that some form of organization is desirable in attempting to capture the breadth of this field. One of the simplest ways to develop a basic understanding of the extent of image processing applications is to categorize images according to their source (e.g., visual, X-ray, and so on). The principal energy source for images in use today is the electromagnetic energy spectrum. Other important sources of energy include acoustic, ultrasonic, and electronic (in the form of electron beams used in electron microscopy). Synthetic images, used for modeling and visualization, are generated by computer. In this section we discuss briefly how images are generated in these various categories and the areas in which they are applied.Images based on radiation from the EM spectrum are the most familiar, es- pecially images in the X-ray and visual bands of the spectrum. Electromagnet- ic waves can be conceptualized as propagating sinusoidal waves of varyingwavelengths, or they can be thought of as a stream of massless particles, each traveling in a wavelike pattern and moving at the speed of light. Each massless particle contains a certain amount (or bundle) of energy. Each bundle of energy is called a photon. If spectral bands are grouped according to energy per photon, we obtain the spectrum shown in fig. below, ranging from gamma rays (highest energy) at one end to radio waves (lowest energy) at the other. The bands are shown shaded to convey the fact that bands of the EM spectrum are not distinct but rather transition smoothly from one to the other.Image acquisition is the first process. Note that acquisition could be as simple as being given an image that is already in digital form. Generally, the image acquisition stage involves preprocessing, such as scaling.Image enhancement is among the simplest and most appealing areas of digital image processing. Basically, the idea behind enhancement techniques is to bring out detail that is obscured, or simply to highlight certain features of interest in an image. A familiar example of enhancement is when we increase the contrast of an image because “it looks better.” It is important to keep in mind that enhancement is a very subjective area of image processing. Image restoration is an area that also deals with improving the appearance of an image. However, unlike enhancement, which is subjective, image restoration is objective, in the sense that restoration techniques tend to be based on mathematical or probabilistic models of image degradation. Enhancement, on the other hand, is based on human subjective preferences regarding what constitutes a “good”enhancement result.Color image processing is an area that has been gaining in importance because of the significant increase in the use of digital images over the Internet. It covers a number of fundamental concepts in color models and basic color processing in a digital domain. Color is used also in later chapters as the basis for extracting features of interest in an image.Wavelets are the foundation for representing images in various degrees of resolution. In particular, this material is used in this book for image data compression and for pyramidal representation, in which images are subdivided successively into smaller regions.Compression, as the name implies, deals with techniques for reducing the storage required to save an image, or the bandwidth required to transmi it.Although storage technology has improved significantly over the past decade, the same cannot be said for transmission capacity. This is true particularly in uses of the Internet, which are characterized by significant pictorial content. Image compression is familiar (perhaps inadvertently) to most users of computers in the form of image file extensions, such as the jpg file extension used in the JPEG (Joint Photographic Experts Group) image compression standard.Morphological processing deals with tools for extracting image components that are useful in the representation and description of shape. The material in this chapter begins a transition from processes that output images to processes that output image attributes.Segmentation procedures partition an image into its constituent parts or objects. In general, autonomous segmentation is one of the most difficult tasks in digital image processing. A rugged segmentation procedure brings the process a long way toward successful solution of imaging problems that require objects to be identified individually. On the other hand, weak or erratic segmentation algorithms almost always guarantee eventual failure. In general, the more accurate the segmentation, the more likely recognition is to succeed.Representation and description almost always follow the output of a segmentation stage, which usually is raw pixel data, constituting either the bound- ary of a region (i.e., the set of pixels separating one image region from another) or all the points in the region itself. In either case, converting the data to a form suitable for computer processing is necessary. The first decision that must be made is whether the data should be represented as a boundary or as a complete region. Boundary representation is appropriate when the focus is on external shape characteristics, such as corners and inflections. Regional representation is appropriate when the focus is on internal properties, such as texture or skeletal shape. In some applications, these representations complement each other. Choosing a representation isonly part of the solution for trans- forming raw data into a form suitable for subsequent computer processing. A method must also be specified for describing the data so that features of interest are highlighted. Description, also called feature selection, deals with extracting attributes that result in some quantitative information of interest or are basic for differentiating one class of objects from another.Recognition is the process that assigns a label (e.g., “vehicle”) to an object based on its descriptors. As detailed before, we conclude our coverage of digital image processing with the development of methods for recognition of individual objects.So far we have said nothing about the need for prior knowledge or about the interaction between the knowledge base and the processing modules in Fig2 above. Knowledge about a problem domain is coded into an image processing system in the form of a knowledge database. This knowledge may be as sim- ple as detailing regions of an image where the information of interest is known to be located, thus limiting the search that has to be conducted in seeking that information. The knowledge base also can be quite complex, such as an interrelated list of all major possible defects in a materials inspection problem or an image database containing high-resolution satellite images of a region in con- nection with change-detection applications. In addition to guiding the operation of each processing module, the knowledge base also controls the interaction between modules. This distinction is made in Fig2 above by the use of double-headed arrows between the processing modules and the knowledge base, as op- posed to single-headed arrows linking the processing modules.Edge detectionEdge detection is a terminology in image processing and computer vision, particularly in the areas of feature detection and feature extraction, to refer to algorithms which aim at identifying points in a digital image at which the image brightness changes sharply or more formally has discontinuities.Although point and line detection certainly are important in any discussion on segmentation,edge dectection is by far the most common approach for detecting meaningful discounties in gray level.Although certain literature has considered the detection of ideal step edges, the edges obtained from natural images are usually not at all ideal step edges. Instead they are normally affected by one or several of the following effects:1.focal blur caused by a finite depth-of-field and finite point spread function; 2.penumbral blur caused by shadows created by light sources of non-zero radius; 3.shading at a smooth object edge; 4.local specularities or interreflections in the vicinity of object edges.A typical edge might for instance be the border between a block of red color and a block of yellow. In contrast a line (as can be extracted by a ridge detector) can be a small number of pixels of a different color on an otherwise unchanging background. For a line, there may therefore usually be one edge on each side of the line.To illustrate why edge detection is not a trivial task, let us consider the problem of detecting edges in the following one-dimensional signal. Here, we may intuitively say that there should be an edge between the 4th and 5th pixels.5 76 4 152 148 149If the intensity difference were smaller between the 4th and the 5th pixels and if the intensity differences between the adjacent neighbouring pixels were higher, it would not be as easy to say that there should be an edge in the corresponding region. Moreover, one could argue that this case is one in which there are several edges.Hence, to firmly state a specific threshold on how large the intensity change between two neighbouring pixels must be for us to say that there should be an edge between these pixels is not always a simple problem. Indeed, this is one of the reasons why edge detection may be a non-trivial problem unless the objects in the scene are particularly simple and the illumination conditions can be well controlled.There are many methods for edge detection, but most of them can be grouped into two categories,search-based and zero-crossing based. The search-based methods detect edges by first computing a measure of edge strength, usually a first-order derivative expression such as the gradient magnitude, and then searching for local directional maxima of the gradient magnitude using a computed estimate of the local orientation of the edge, usually the gradient direction. The zero-crossing based methods search for zero crossings in a second-order derivative expression computed from the image in order to find edges, usually the zero-crossings of the Laplacian or the zero-crossings of a non-linear differential expression, as will be described in the section on differential edge detection following below. As a pre-processing step to edge detection, a smoothing stage, typically Gaussian smoothing, is almost always applied (see also noise reduction).The edge detection methods that have been published mainly differ in the types of smoothing filters that are applied and the way the measures of edge strength are computed. As many edge detection methods rely on the computation of image gradients, they also differ in the types of filters used for computing gradient estimates in the x- and y-directions.Once we have computed a measure of edge strength (typically the gradient magnitude), the next stage is to apply a threshold, to decide whether edges are present or not at an image point. The lower the threshold, the more edges will be detected, and the result will be increasingly susceptible to noise, and also to picking out irrelevant features from the image. Conversely a high threshold may miss subtle edges, or result in fragmented edges.If the edge thresholding is applied to just the gradient magnitude image, the resulting edges will in general be thick and some type of edge thinning post-processing is necessary. For edges detected with non-maximum suppression however, the edge curves are thin by definition and the edge pixels can be linked into edge polygon by an edge linking (edge tracking) procedure. On a discrete grid, the non-maximum suppression stage can be implemented by estimating the gradient direction using first-order derivatives, then rounding off the gradient direction to multiples of 45 degrees, and finally comparing the values of the gradient magnitude in the estimated gradient direction.A commonly used approach to handle the problem of appropriate thresholds forthresholding is by using thresholding with hysteresis. This method uses multiple thresholds to find edges. We begin by using the upper threshold to find the start of an edge. Once we have a start point, we then trace the path of the edge through the image pixel by pixel, marking an edge whenever we are above the lower threshold. We stop marking our edge only when the value falls below our lower threshold. This approach makes the assumption that edges are likely to be in continuous curves, and allows us to follow a faint section of an edge we have previously seen, without meaning that every noisy pixel in the image is marked down as an edge. Still, however, we have the problem of choosing appropriate thresholding parameters, and suitable thresholding values may vary over the image.Some edge-detection operators are instead based upon second-order derivatives of the intensity. This essentially captures the rate of change in the intensity gradient. Thus, in the ideal continuous case, detection of zero-crossings in the second derivative captures local maxima in the gradient.We can come to a conclusion that,to be classified as a meaningful edge point,the transition in gray level associated with that point has to be significantly stronger than the background at that point.Since we are dealing with local computations,the method of choice to determine whether a value is “significant” or not id to use a threshold.Thus we define a point in an image as being as being an edge point if its two-dimensional first-order derivative is greater than a specified criterion of connectedness is by definition an edge.The term edge segment generally is used if the edge is short in relation to the dimensions of the image.A key problem in segmentation is to assemble edge segments into longer edges.An alternate definition if we elect to use the second-derivative is simply to define the edge ponits in an image as the zero crossings of its second derivative.The definition of an edge in this case is the same as above.It is important to note that these definitions do not guarantee success in finding edge in an image.They simply give us a formalism to look for them.First-order derivatives in an image are computed using the gradient.Second-order derivatives are obtained using the Laplacian.中文对照数字图像处理与边缘检测数字图像处理数字图像处理方法的研究源于两个主要应用领域:其一是为了便于人们分析而对图像信息进行改进:其二是为使机器自动理解而对图像数据进行存储、传输及显示。

matlab图像处理外文翻译外文文献

matlab图像处理外文翻译外文文献

matlab图像处理外文翻译外文文献附录A 英文原文Scene recognition for mine rescue robotlocalization based on visionCUI Yi-an(崔益安), CAI Zi-xing(蔡自兴), WANG Lu(王璐)Abstract:A new scene recognition system was presented based on fuzzy logic and hidden Markov model(HMM) that can be applied in mine rescue robot localization during emergencies. The system uses monocular camera to acquire omni-directional images of the mine environment where the robot locates. By adopting center-surround difference method, the salient local image regions are extracted from the images as natural landmarks. These landmarks are organized by using HMM to represent the scene where the robot is, and fuzzy logic strategy is used to match the scene and landmark. By this way, the localization problem, which is the scene recognition problem in the system, can be converted into the evaluation problem of HMM. The contributions of these skills make the system have the ability to deal with changes in scale, 2D rotation and viewpoint. The results of experiments also prove that the system has higher ratio of recognition and localization in both static and dynamic mine environments.Key words: robot location; scene recognition; salient image; matching strategy; fuzzy logic; hidden Markov model1 IntroductionSearch and rescue in disaster area in the domain of robot is a burgeoning and challenging subject[1]. Mine rescue robot was developed to enter mines during emergencies to locate possible escape routes for those trapped inside and determine whether it is safe for human to enter or not. Localization is a fundamental problem in this field. Localization methods based on camera can be mainly classified into geometric, topological or hybrid ones[2]. With its feasibility and effectiveness, scene recognition becomes one of the important technologies of topological localization.Currently most scene recognition methods are based on global image features and have two distinct stages: training offline and matching online.。

图像处理中英文对照外文翻译文献

图像处理中英文对照外文翻译文献

中英文对照外文翻译文献(文档含英文原文和中文翻译)译文:基于局部二值模式多分辨率的灰度和旋转不变性的纹理分类摘要:本文描述了理论上非常简单但非常有效的,基于局部二值模式的、样图的非参数识别和原型分类的,多分辨率的灰度和旋转不变性的纹理分类方法。

此方法是基于结合某种均衡局部二值模式,是局部图像纹理的基本特性,并且已经证明生成的直方图是非常有效的纹理特征。

我们获得一个一般灰度和旋转不变的算子,可表达检测有角空间和空间结构的任意量子化的均衡模式,并提出了结合多种算子的多分辨率分析方法。

根据定义,该算子在图像灰度发生单一变化时具有不变性,所以所提出的方法在灰度发生变化时是非常强健的。

另一个优点是计算简单,算子在小邻域内或同一查找表内只要几个操作就可实现。

在旋转不变性的实际问题中得到了良好的实验结果,与来自其他的旋转角度的样品一起以一个特别的旋转角度试验而且测试得到分类, 证明了基于简单旋转的发生统计学的不变性二值模式的分辨是可以达成。

这些算子表示局部图像纹理的空间结构的又一特色是,由结合所表示的局部图像纹理的差别的旋转不变量不一致方法,其性能可得到进一步的改良。

这些直角的措施共同证明了这是旋转不变性纹理分析的非常有力的工具。

关键词:非参数的,纹理分析,Outex ,Brodatz ,分类,直方图,对比度2 灰度和旋转不变性的局部二值模式我们通过定义单色纹理图像的一个局部邻域的纹理T ,如 P (P>1)个象素点的灰度级联合分布,来描述灰度和旋转不变性算子:01(,,)c P T t g g g -= (1)其中,g c 为局部邻域中心像素点的灰度值,g p (p=0,1…P-1)为半径R(R>0)的圆形邻域内对称的空间象素点集的灰度值。

图1如果g c 的坐标是(0,0),那么g p 的坐标为(cos sin(2/),(2/))R R p P p P ππ-。

图1举例说明了圆形对称邻域集内各种不同的(P,R )。

LMS中英文对照表

LMS中英文对照表

LMS 中英文对照表位置原文简体中文繁体中文菜单File 文件New 新建Open 打开Reopen 重新打开Save 保存SaveAs 另存为Revert 返回Load QuickSet File 载入快速设置文件Save QuickSet File夹保存快速设置文件Load GraphSetup File 载入图表设置文件Save GraphSetup File 保存图表设置文件Print 打印Editor 编辑器Preferences 参数选择Exit 退出Graph 图表Parameters 参数Curve Library 曲线库Notes & Comments 附注&注释Analyzer 分析仪Sweep Start/Stop 扫频开始/停止Osc On/Off Osc 开/关RLC meter RLC 表Microphone Setup 麦克风设置PAC Interface PAC接口Micro Run 宏运行Calibration 剪贴板Processing 处理Unary Math Operations 一元数学操作Binary Math Operations 二元数学操作Minimum Phase Transform 最小相位转换Delay Phase Transform 延迟相位转换Group Delay Transform 组延迟转换Inv Fast Fourier Transform 快速反傅立叶转换Fast Fourier Transform 快速傅立叶转换Speaker Parameters 喇叭参数Tail Correction 尾部修正Data Transfer 数据合并Data Splice 数据接合Data Realign 数据重组Curve Averaging 数据提升Curve Compare 数据对照Curve Integration 数据综合Utilities 实用程序Import Curve Data File 导入曲线数据文件Export Curve Data File 导出曲线数据文件Export Graphics to File 导出图表到文件Export Graphics to Clipboard 导出图表到剪贴板Curve Capture 曲线捕捉Curve Editor 曲线编辑器Macro Editor 宏编辑器MDF Editor MDF 编辑器Polar Convertor 极性转换器View Clipboard 查看剪贴板Transducer Model Derivation 传感器模式导出Scale 刻度Parameters 参数Auto 自动Up 向上Down 向下View 查看Zoom In 放大Zoom Out 缩小Zoom 缩放Redraw 刷新ToolBars 工具栏Show All 全部显示Hide All 全部隐藏ToolBox 工具框Help 帮助Contents 内容Index 索引Glossary 术语表About Modules 关于模块About Program 关于软件对话框(Graph Parameters)Frame 边框Background 背景Note Underline 注释下划线Large Frame Line 大边框线Small Frame Line 小边框线Grid 格子Border Line 边界线Major Div 主格Minor Div 次格Font 字体Title Block 工程明细表Map Legend 映射图例Note List 注释列表Graph Title 图表标题Scale Vertical 垂直刻度Scale Horizontal 水平刻度Typeface 字体Style 字形Size 大小Color 颜色(Curve Library)Data Curve 数据曲线Same Line Type 统一线型Left(Magnitude) 左坐标(量级) Right Lighter 右坐标浅色Right(Phase) 右坐标(相位)Info 信息Horz Data Range 频宽范围Left Vert 左坐标Right Vert 右坐标Points 点Style 类型Width 宽度Color 颜色(Note&Comments)Left Page 左页面Right Page 右页面Title Block Data 工程图明细表Person 个人Company 公司Project 工程Automatic Curve Info Notes 自动曲线信息标注(Analyzer Parameters)Oscillator 振荡器Output Level 输出水平Frequency 频率Mode 扫频模式Hi Speed Data 高速数据Precision Data 精密数据Gating 门控Off 关On 开Meter 仪表Data 数据Value 值Source 来源Freq 频率Sweep 扫频Lo Freq 低频Hi Freq 高频Direction 方向Control 控制Pulse 脉冲Gate Time Calculator 门控计时器Meter Filter 滤波器Filter Function 滤波函数Track Ratio 音轨率Gate Timing 门控时间(RLC Meter)Measurement 测量Resistance 电阻Inductance 电感Capacitance 电容Impedance 阻抗Limit Testing 测试限制Enable 启用MinValue 最小值Max Value 最大值(Microphone Setup)Mic Input 麦克风输入Line Input 信号输入Model 型号Acoustic Ref 声学参数Electric Ref 电学参数Serial 序列号Author 制造商Date 日期Load 载入(PAC Interface)Serial Port 串行端口Linking 正在链接Start Link 开始链接Automatic Link 自动链接Baud Rate 波特率System Power 系统供电Power Source 供电电源Battery Status 电池状态Charge Amps 功放充电System Status 系统状态Link Status 链接状态Port 端口Voltage 电压Battery V oltage 电池电压External V oltage 外部电压(Analyzer Calibration)Internal 内部External 外部Parameter Under Test 测试参数Results 结果(Unary Math Operations)Library Curve 曲线库Operation 操作Magnitude Offset 幅度补偿Phase Offset 相位偏移Delay Offset 延迟偏移Exponentiation 求幂Smooth Curve 平滑曲线Frequency Translation 频率转换Multiply by 乘Divide 除Real 正弦Imag(sin) 余弦Execute 执行(Binary Math Operations)Mul 乘Div 除Add 加Sub 减Operand 操作数(Minimum Phase Transform)Asymptotic Slope at Hi Freq Limit 频率上限斜率Asymptotic Slope at Lo Freq Limit 频率下限斜率Automatic Tail Correction/Mirroring for Impedance Curves 自动尾部修正/阻抗曲线镜像(Delay Phase Transform)Source Curve with Delay Data 来源延迟数据曲线Result Curve for Phase Data 结果相位数据曲线(Group Delay Transform)Source Curve with Phase Data 来源相位曲线Result Curve for Group Delay 结果组延迟曲线(Inverse Fast Fourier Transform)Linear Frequency Points 线性频率点Frequency Domain Data 频率范围数据Result Impulse Curve 结果推动曲线Time Domain Data 时间范围数据Result Step Curve 结果步骤曲线(Speaker Parameters)Method 方法Single Curve 单曲线Double Curve 双曲线Reference Curve 参考曲线Standard 标准Estimate 估算Optimize 优化Simulate 模拟Model Simulation 模拟模型Copy Binary to Clipboard 复制二元数据到剪贴板Copy Text to Clipboard 复制文本到剪贴板(Tail Correction)Slope Hi 高频Slope Lo 低频(Data Splice)Source Curve for Higher Data 来源高数据曲线Source Curve for Lower Data 来源低数据曲线Horz Splice Transition 水平接合转换(Data Realign)Linear 线性Log 对数Horz Lo Limit 水平下限Horz Hi Limit 水平上限Interpolation 插补Quadratic 平方Cubic 立方(Curve Averaging)Scalar 标量Vector 矢量(Curve Compare)Test Parameters 测试参数Absolute 绝对Relative 相对Tolerance 公差Relative Flatness 相对平面(Curve Integration)Average 平均Integrate 叠加(Import Curve Data)Left Vert Data 左坐标数据Right Ver Data 右坐标数据Polar Freq 极性频率Units 单位Prefix 前缀Curve Entry 曲线条目Special Processing 特殊处理Skip First Data Column 跳过首个数据列(Export Graphics)Artwork 作品Graph Name 图表名称Format 格式Raster 光栅Image Pixel Width 图像宽度(像素) Image Pixel Height 图像高度(像素) Image Bits per Pixels 图像位每像素Image Bytes 图像字节数(Clipboard Graphics Transfer)Enhanced Metafile 增强元文件Bitmap 位图(Curve Capture)Polar Freq 极性频率Graph Image 图表图像Reference Point Data 参考点数据Upper Right 右上Scan Direction 扫描方向Top to Bottom 顶到底Bottom to Top 底到顶Color Match 颜色匹配(Curve Editor)Control 控制Node 节点Insert 插入Snap 快照Guidelines 参考线Ruler Grid 标尺格Smooth 平滑(Transducer Model Derivation)Measurements 度量Profile 配置文件(Scale Parameters)Axis 轴Range 范围Divisions 分格Current 当前Volume 音量Acceleration 加速度Velocity 速率Excursion 偏移。

数字图像处理论文中英文对照资料外文翻译文献

数字图像处理论文中英文对照资料外文翻译文献

第 1 页中英文对照资料外文翻译文献原 文To image edge examination algorithm researchAbstract :Digital image processing took a relative quite young discipline,is following the computer technology rapid development, day by day obtains th widespread application.The edge took the image one kind of basic characteristic,in the pattern recognition, the image division, the image intensification as well as the image compression and so on in the domain has a more widesp application.Image edge detection method many and varied, in which based on brightness algorithm, is studies the time to be most long, the theory develo the maturest method, it mainly is through some difference operator, calculates its gradient based on image brightness the change, thus examines the edge, mainlyhas Robert, Laplacian, Sobel, Canny, operators and so on LOG 。

滤波器 外文翻译 外文文献 英文文献 用改进的窗函数设计FIR数字滤波器

滤波器 外文翻译 外文文献 英文文献 用改进的窗函数设计FIR数字滤波器

Research on FIR Digital Filter Design Using an Improved Window FunctionTAN Jiajie , LUO Changyou, HUANG Sanwei ,DENG Xiaohui( Department of Physics and Electronic Information Science, Hengyang Normal University, Hengyang Hunan 421008, China)Abstract : Window function has been used to design a linear phase digital filer for long times, but the use o f optimization techniques for designing digital filter has become widespread in recent year A new met ho d has been proposed to improve FIR window function in this paper ,T he window function that combines with co sine sequences in linear is different from previous Hann ,Hamming and Blackman window function The paper also proposes linear programming to optimize characterization of FIR digital filter according to its magnitude condition, and gives out t he algorithm to design dig ital filter using the improved window function ,Finally , we have designed FIR filter using new window for simulation and compared w it h the filter designed by Hamming window, Blackman window function T he simulation results show that filters designed using this method to meet t he design specificationsKey words: improved window function; FIR digital filter ; window function; linear programming0 lead speechThe design method of FIR digital filters are mainly: window function method, frequency sampling method and the chebyshev etc corrugated approximation method [1-4]. Window function method is the most commonly used designing FIR digital filters, the simplest method of [4-5]. The essence of window function method is the truncated ideal impulse response to approximate the method petitions filter index. Frequency sampling method is a design optimization method for its shortcoming is when the design that use the variable is limited to a few samples values of transitional, cut-off frequency not easy control [3]. Chebyshev etc corrugated approximation method is a kind of optimization design, but existing computational complexity, big disadvantage computation [1-2].Window function method is simple in design, have closed form of formula, thus very practical. Defect is the stopband bandpass, cut-off frequency not easy control [2-3]. Digital filter, window function of auto-heating window function method of selecting, the key is: design to choose the appropriate window function, choose the right order number of digital filter, improve amplitude frequency characteristics, reduce Gibbs phenomenon, solve convergence problem [1-2]. [3] choose window function, through to Guass Guass window function improved, design a low-pass filter has better superiority; [4] the error information, using the known in the iteration process through the window function method continuously revised design result in filter order number, under the condition of invariable frequency response approximation, filter ideal frequency response. [5] use integer sequence, such as where Fibonacci sequence, Golomb sequence, ConwayHofstadter Recursive sequence, Triangular series produce window function to design the filter, its effect is better than that of classical design method. [6] choose dual window window function sequence of structure was system characteristics approximation error is the minimum; [7] will be well Saramaki Dolph - and Chebysheve window with the well designed, its effect FIR digital filters than Kaiser well; [8] put good effect in Hamming ReImann well well well and Kaiser window. [9] put forward a kind of exponential window function, this window function has the width can be adjusted with the window design characteristics, the digital filters have more centralized, Lord disc energy side-lobe less features. [10] using linear programming design linear phase fir filter ascending cosine to 100 % super bandwidth. [11] linear programming method is adopted to design digital filter. This paper USES the existing window function, and carry on the weighted combination, reference [10 or 11], and linear programming with long Hamming, Blackman window are compared. The advantages of this method is strong logicality, goal clear, easy to achieve, and to explore the best solutions.1 common window functionWindow function select principle: window function as focus on energy, Lord disc transitional steep; Reduce the window function spectrum side-lobe level, increase stopband attenuation, and reduce the stopband bandpass and ripple effect. Common window function have [1-4] : Rectangle window, Hanning window, Hamming window, Blackman window, Kaiser window. Window function method design idea of FIR filters is [1-2] : make sure the frequency response of ideal filter )(ωj d e H .The frequency response of practical design filter ∑-=-=10)()(N n j j e n h e H ωωTo approximate )(ωj d e H .For again ∑-=-=10)()(N n j j e n h e H ωωReverse transform get Finally use window function )(n w To truncate )(n h d ,mean h( n) =)(n h d )(n w .To truncate )(n h d ,Will produce gibbs phenomenon, all the window function choices to reduce this phenomenon for the purpose. Judge ideal window function mainly according to the following three criteria! The Lord is high double amplitude and its width should try to narrowThe amplitude side-lobe fast speed, the biggest drop side-lobe relative to the main valve should be as low as possible. #transitional requirements will try to narrow. Facts prove the two standard cannot simultaneously satisfy window function should be, so the twocompromise [1-3]. In order to reduce caused due to add window truncation ripple and transitional grows wider impact in engineering design common Hamming window and Kaiser window.2 improved window function [1-2] enumerated window function, Hanning window, Hamming window, Blackman window is cosine sequence andrectangular sequence of linear combination. In order to restrain the amplitude,side-lobe Hanning window, Hamming window on the basis of the second, add cosine, when the harmonic component design and ideal window function and related to the frequency response of different from Blackman window, window function improved form below )(14cos 12cos )(n N R N n c N n b a n ⎥⎦⎤⎢⎣⎡-+-+=ππω (1)Formula (1) of a, b, c undetermined, their size and given filter technology indexes related. For convenience, this window function length choice for odd. The next several special case discussion this type. Case 1, take a = 1, b = c = 0, for rectangular window. Condition 2, take a = 0.5, b = - 0.5, c = 0, for Hann window. Case 3, take a = 0.53, b = - 0.46, c = 0, for Hamming window.Situation, a = 4, b = 0 0.42 j c = 0.08, 5, Blackman window for. By aboveknowable, the improved window function with these four window function the nature, belong to the general form of the window function.3 improved window function algorithmAccording to the given filter technology index )(ωj d e H ,Determine the backlog filter unit, but by sampling response formula below ask out:ωπωππωd e e H n h j j d d ⎰-=)(21)( (2)Calculating the actual filter unit sampling response:h( n) = )(n h d )(n w (3)Filter the frequency response is:∑-=-=10)()(N n j j e n h e H ωω(4) Will formula (1) generation into the formula (4) :ξωππjn N n n d j e N n c N n b a h e H --=⎥⎦⎤⎢⎣⎡-+-+=∑14cos 12cos )(10)( (5) Reference [1-2] [10 or 11], consider FIR filters satisfy the first kind of linearphase conditions, For 21-N Accidentally symmetry, And N an odd number,ordering h ( n) = )(n h d ⨯⎥⎦⎤⎢⎣⎡-+-+14cos 12cos N n c N n b a ππ。

数字滤波器外文翻译

数字滤波器外文翻译

中文5590字毕业设计(外文翻译材料)2009年6月学 院: 专 业: 学生姓名: 指导教师: 电气与电子工程学院 电子信息工程0503DIGITAL FILTERSDigital filtering is one of the most powerful tools of DSP. Apart from the obvious advantages of virtually eliminating errors in the filter associated with passive component fluctuations over time and temperature, op amp drift (active filters), etc., digital filters are capable of performance specifications that would, at best, be extremely difficult, if not impossible, to achieve with an analog implementation. In addition, the characteristics of a digital filter can be easily changed under software control. Therefore, they are widely used in adaptive filtering applications in communications such as echo cancellation in modems, noise cancellation, and speech recognition.The actual procedure for designing digital filters has the same fundamental elements as that for analog filters. First, the desired filter responses are characterized, and the filter parameters are then calculated. Characteristics such as amplitude and phase response are derived in the same way. The key difference between analog and digital filters is that instead of calculating resistor, capacitor, and inductor values for an analog filter, coefficient values are calculated for a digital filter. So for the digital filter, numbers replace the physical resistor and capacitor components of the analog filter. These numbers reside in a memory as filter coefficients and are used with the sampled data values from the ADC to perform the filter calculations.The real-time digital filter, because it is a discrete time function, works with digitized data as opposed to a continuous waveform, and a new data point is acquired each sampling period. Because of this discrete nature, data samples are referenced as numbers such as sample 1, sample 2, sample 3, etc. Figure 1 shows a low frequency signal containing higher frequency noise which must be filtered out. This waveform must be digitized with an ADC to produce samples x(n). These data values are fed to the digital filter, which in this case is a lowpass filter. The output data samples, y(n), are used to reconstruct an analog waveform using a low glitch DAC.Digital filters, however, are not the answer to all signal processing filtering requirements. In order to maintain real-time operation, the DSP processor must be able to execute all the steps in the filter routine within one sampling clock period1/f s.A fast general purpose fixed-point DSP such as the ADSP-2189M at 75MIPS can 。

介绍数字图像处理外文翻译

介绍数字图像处理外文翻译

附录1 外文原文Source: "the 21st century literature the applied undergraduate electronic communication series of practical teaching planThe information and communication engineering specialty in English ch02_1. PDF 120-124Ed: HanDing ZhaoJuMin, etcText A: An Introduction to Digital Image Processing1. IntroductionDigital image processing remains a challenging domain of programming for several reasons. First the issue of digital image processing appeared relatively late in computer history. It had to wait for the arrival of the first graphical operating systems to become a true matter. Secondly, digital image processing requires the most careful optimizations especially for real time applications. Comparing image processing and audio processing is a good way to fix ideas. Let us consider the necessary memory bandwidth for examining the pixels of a 320x240, 32 bits bitmap, 30 times a second: 10 Mo/sec. Now with the same quality standard, an audio stereo wave real time processing needs 44100 (samples per second) x 2 (bytes per sample per channel) x 2(channels) = 176Ko/sec, which is 50 times less.Obviously we will not be able to use the same techniques for both audio and image signal processing. Finally, digital image processing is by definition a two dimensions domain; this somehow complicates things when elaborating digital filters.We will explore some of the existing methods used to deal with digital images starting by a very basic approach of color interpretation. As a moreadvanced level of interpretation comes the matrix convolution and digital filters. Finally, we will have an overview of some applications of image processing.The aim of this document is to give the reader a little overview of the existing techniques in digital image processing. We will neither penetrate deep into theory, nor will we in the coding itself; we will more concentrate on the algorithms themselves, the methods. Anyway, this document should be used as a source of ideas only, and not as a source of code. 2. A simple approach to image processing(1) The color data: Vector representation①BitmapsThe original and basic way of representing a digital colored image in a computer's memory is obviously a bitmap. A bitmap is constituted of rows of pixels, contraction of the word s “Picture Element”. Each pixel has a particular value which determines its appearing color. This value is qualified by three numbers giving the decomposition of the color in the three primary colors Red, Green and Blue. Any color visible to human eye can be represented this way. The decomposition of a color in the three primary colors is quantified by a number between 0 and 255. For example, white will be coded as R = 255, G = 255, B = 255; black will be known as (R,G,B)= (0,0,0); and say, bright pink will be : (255,0,255). In other words, an image is an enormous two-dimensional array of color values, pixels, each of them coded on 3 bytes, representing the three primary colors. This allows the image to contain a total of 256×256×256 = 16.8 million different colors. This technique is also known as RGB encoding, and is specifically adapted to human vision. With cameras or other measure instruments we are capable of “seeing”thousands of other “colors”, in which cases the RG B encoding is inappropriate.The range of 0-255 was agreed for two good reasons: The first is that the human eye is not sensible enough to make the difference between more than 256 levels of intensity (1/256 = 0.39%) for a color. That is to say, an image presented to a human observer will not be improved by using more than 256 levels of gray (256shades of gray between black and white). Therefore 256 seems enough quality. The second reason for the value of 255 is obviously that it is convenient for computer storage. Indeed on a byte, which is the computer's memory unit, can be coded up to 256 values.As opposed to the audio signal which is coded in the time domain, the image signal is coded in a two dimensional spatial domain. The raw image data is much more straightforward and easy to analyze than the temporal domain data of the audio signal. This is why we will be able to do lots of stuff and filters for images without transforming the source data, while this would have been totally impossible for audio signal. This first part deals with the simple effects and filters you can compute without transforming the source data, just by analyzing the raw image signal as it is.The standard dimensions, also called resolution, for a bitmap are about 500 rows by 500 columns. This is the resolution encountered in standard analogical television and standard computer applications. You can easily calculate the memory space a bitmap of this size will require. We have 500×500 pixels, each coded on three bytes, this makes 750 Ko. It might not seem enormous compared to the size of hard drives, but if you must deal with an image in real time then processing things get tougher. Indeed rendering images fluidly demands a minimum of 30 images per second, the required bandwidth of 10 Mo/sec is enormous. We will see later that the limitation of data access and transfer in RAM has a crucial importance in image processing, and sometimes it happens to be much more important than limitation of CPU computing, which may seem quite different from what one can be used to in optimization issues. Notice that, with modern compression techniques such as JPEG 2000, the total size of the image can be easily reduced by 50 times without losing a lot of quality, but this is another topic.②Vector representation of colorsAs we have seen, in a bitmap, colors are coded on three bytes representing their decomposition on the three primary colors. It sounds obvious to a mathematician to immediately interpret colors as vectors in athree-dimension space where each axis stands for one of the primary colors. Therefore we will benefit of most of the geometric mathematical concepts to deal with our colors, such as norms, scalar product, projection, rotation or distance. This will be really interesting for some kind of filters we will see soon. Figure 1 illustrates this new interpretation:Figure 1(2) Immediate application to filters① Edge DetectionFrom what we have said before we can quantify the 'difference' between two colors by computing the geometric distance between the vectors representing those two colors. Lets consider two colors C1 = (R1,G1,B1) and C2 = (R2,B2,G2), the distance between the two colors is given by the formula :D(C1, C2) =(R1+This leads us to our first filter: edge detection. The aim of edge detection is to determine the edge of shapes in a picture and to be able to draw a resultbitmap where edges are in white on black background (for example). The idea is very simple; we go through the image pixel by pixel and compare the color of each pixel to its right neighbor, and to its bottom neighbor. If one of these comparison results in a too big difference the pixel studied is part of an edge and should be turned to white, otherwise it is kept in black. The fact that we compare each pixel with its bottom and right neighbor comes from the fact that images are in two dimensions. Indeed if you imagine an image with only alternative horizontal stripes of red and blue, the algorithms wouldn't see the edges of those stripes if it only compared a pixel to its right neighbor. Thus the two comparisons for each pixel are necessary.This algorithm was tested on several source images of different types and it gives fairly good results. It is mainly limited in speed because of frequent memory access. The two square roots can be removed easily by squaring the comparison; however, the color extractions cannot be improved very easily. If we consider that the longest operations are the get pixel function and put pixel functions, we obtain a polynomial complexity of 4*N*M, where N is the number of rows and M the number of columns. This is not reasonably fast enough to be computed in realtime. For a 300×300×32 image I get about 26 transforms per second on an Athlon XP 1600+. Quite slow indeed.Here are the results of the algorithm on an example image:A few words about the results of this algorithm: Notice that the quality of the results depends on the sharpness of the source image. Ifthe source image is very sharp edged, the result will reach perfection. However if you have a very blurry source you might want to make it pass through a sharpness filter first, which we will study later. Another remark, you can also compare each pixel with its second or third nearest neighbors on the right and on the bottom instead of the nearest neighbors. The edges will be thicker but also more exact depending on the source image's sharpness. Finally we will see later on that there is another way to make edge detection with matrix convolution.②Color extractionThe other immediate application of pixel comparison is color extraction.Instead of comparing each pixel with its neighbors, we are going to compare it with a given color C1. This algorithm will try to detect all the objects in the image that are colored with C1. This was quite useful for robotics for example. It enables you to search on streaming images for a particular color. You can then make you robot go get a red ball for example. We will call the reference color, the one we are looking for in the image C0 = (R0,G0,B0).Once again, even if the square root can be easily removed it doesn't really improve the speed of the algorithm. What really slows down the whole loop is the NxM get pixel accesses to memory and put pixel. This determines the complexity of this algorithm: 2xNxM, where N and M are respectively the numbers of rows and columns in the bitmap. The effective speed measured on my computer is about 40 transforms per second on a 300x300x32 source bitmap.3.JPEG image compression theory(一)JPEG compression is divided into four steps to achieve:(1) Color mode conversion and samplingRGB color system is the most common ways that color. JPEG uses a YCbCr colorsystem. Want to use JPEG compression method dealing with the basic full-color images, RGB color mode to first image data is converted to YCbCr color model data. Y representative of brightness, Cb and Cr represents the hue, saturation. By the following calculation to be completed by data conversion. Y = 0.2990R +0.5870 G+0.1140 B Cb =- 0.1687R-0.3313G +0.5000 B +128 Cr = 0.5000R-0.4187G-0.0813B+128 of human eyes on the low-frequency data than high-frequency data with higher The sensitivity, in fact, the human eye to changes in brightness than to color changes should be much more sensitive, ie Y component of the data is more important. Since the Cb and Cr components is relatively unimportant component of the data comparison, you can just take part of the data to deal with. To increase the compression ratio. JPEG usually have two kinds of sampling methods: YUV411 and YUV422, they represent is the meaning of Y, Cb and Cr data sampling ratio of three components.(2)DCT transformationThe full name is the DCT-discrete cosine transform (Discrete Cosine Transform), refers to a group of light intensity data into frequency data, in order that intensity changes of circumstances. If the modification of high-frequency data do, and then back to the original form of data, it is clear there are some differences with the original data, but the human eye is not easy to recognize. Compression, the original image data is divided into 8 * 8 matrix of data units. JPEG entire luminance and chrominance Cb matrix matrix, saturation Cr matrix as a basic unit called the MCU. Each MCU contains a matrix of no more than 10. For example, the ratio of rows and columns Jie Wei 4:2:2 sampling, each MCU will contain four luminance matrix, a matrix and a color saturation matrix. When the image data is divided into an 8 * 8 matrix, you must also be subtracted for each value of 128, and then a generation of formula into the DCT transform can be achieved by DCT transform purposes. The image data value must be reduced by 128, because the formula accepted by the DCT-figure range is between -128 to +127.(3)QuantizationImage data is converted to the frequency factor, you still need to accept a quantitative procedure to enter the coding phase. Quantitative phase requires two 8 * 8 matrix of data, one is to deal specifically with the brightness of the frequency factor, the other is the frequency factor for the color will be the frequency coefficient divided by the value of quantization matrix to obtain the nearest whole number with the quotient, that is completed to quantify. When the frequency coefficients after quantization, will be transformed into the frequency coefficients from the floating-point integer This facilitate the implementation of the final encoding. However, after quantitative phase, all the data to retain only the integer approximation, also once again lost some data content.(4)CodingHuffman encoding without patent issues, to become the most commonly used JPEG encoding, Huffman coding is usually carried out in a complete MCU. Coding, each of the DC value matrix data 63 AC value, will use a different Huffman code tables, while the brightness and chroma also require a different Huffman code tables, it needs a total of four code tables, in order to successfully complete the JPEG coding. DC Code DC is a color difference pulse code modulation using the difference coding method, which is in the same component to obtain an image of each DC value and the difference between the previous DC value to encode. DC pulse code using the main reason for the difference is due to a continuous tone image, the difference mostly smaller than the original value of the number of bits needed to encode the difference will be more than the original value of the number of bits needed to encode the less. For example, a margin of 5, and its binary representation of a value of 101, if the difference is -5, then the first changed to a positive integer 5, and then converted into its 1's complement binary number can be. The so-called one's complement number, that is, if the value is 0 for each Bit, then changed to 1; Bit is 1, it becomes 0. Difference between the five should retain the median 3, the following table that lists the difference between the Bit to be retained and the difference between the number of content controls.In the margin of the margin front-end add some additional value Hoffman code, such as the brightness difference of 5 (101) of the median of three, then the Huffman code value should be 100, the two connected together shall be 100101. The following two tables are the brightness and chroma DC difference encoding table. According to these two forms content, you can add the difference for the DC value Huffman code to complete the DC coding.4. ConclusionsDigital image processing is far from being a simple transpose of audiosignal principles to a two dimensions space. Image signal has its particular properties, and therefore we have to deal with it in a specificway. The Fast Fourier Transform, for example, which was such a practical tool in audio processing, becomes useless in image processing. Oppositely, digital filters are easier to create directly, without any signal transforms, in image processing.Digital image processing has become a vast domain of modern signal technologies. Its applications pass far beyond simple aesthetical considerations, and they include medical imagery, television and multimedia signals, security, portable digital devices, video compression,and even digital movies. We have been flying over some elementarynotions in image processing but there is yet a lot more to explore. Ifyou are beginning in this topic, I hope this paper will have given you thetaste and the motivation to carry on.附录2 外文翻译文献出处:《21 世纪全国应用型本科电子通信系列实用规划教材》之《信息与通信工程专业英语》ch02_1.pdf 120-124页主编:韩定定、赵菊敏等正文:介绍数字图像处理1.导言有几个原因使数字图像处理仍然是一个具有挑战性的领域。

  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

中英文资料对照外文翻译一、英文原文A NEW CONTENT BASED MEDIAN FILTERABSTRACTIn this paper the hardware implementation of a contentbased median filter suitabl e for real-time impulse noise suppression is presented. The function of the proposed ci rcuitry is adaptive; it detects the existence of impulse noise in an image neighborhood and applies the median filter operator only when necessary. In this way, the blurring o f the imagein process is avoided and the integrity of edge and detail information is pre served. The proposed digital hardware structure is capable of processing gray-scale im ages of 8-bit resolution and is fully pipelined, whereas parallel processing is used to m inimize computational time. The architecturepresented was implemented in FPGA an d it can be used in industrial imaging applications, where fast processing is of the utm ost importance. The typical system clock frequency is 55 MHz.1. INTRODUCTIONTwo applications of great importance in the area of image processing are noise filtering and image enhancement [1].These tasks are an essential part of any image pro cessor,whether the final image is utilized for visual interpretation or for automatic an alysis. The aim of noise filtering is to eliminate noise and its effects on the original im age, while corrupting the image as little as possible. To this end, nonlinear techniques (like the median and, in general, order statistics filters) have been found to provide mo re satisfactory results in comparison to linear methods. Impulse noise exists in many p ractical applications and can be generated by various sources, including a number of man made phenomena, such as unprotected switches, industrial machines and car ign ition systems. Images are often corrupted by impulse noise due to a noisy sensor or ch annel transmission errors. The most common method used for impulse noise suppressi on n forgray-scale and color images is the median filter (MF) [2].The basic drawback o f the application of the MF is the blurringof the image in process. In the general case,t he filter is applied uniformly across an image, modifying pixels that arenot contamina ted by noise. In this way, the effective elimination of impulse noise is often at the exp ense of an overalldegradation of the image and blurred or distorted features[3].In this paper an intelligent hardware structure of a content based median filter (CBMF) suita ble for impulse noise suppression is presented. The function of the proposed circuit is to detect the existence of noise in the image window and apply the corresponding MFonly when necessary. The noise detection procedure is based on the content of the im age and computes the differences between the central pixel and thesurrounding pixels of a neighborhood. The main advantage of this adaptive approach is that image blurrin g is avoided and the integrity of edge and detail information are preserved[4,5]. The pro posed digital hardware structure is capable of processing gray-scale images of 8-bitres olution and performs both positive and negative impulse noise removal. The architectt ure chosen is based on a sequence of four basic functional pipelined stages, and parall el processing is used within each stage. A moving window of a 3×3 and 5×5-pixel im age neighborhood can be selected. However, the system can be easily expanded to acc ommodate windows of larger sizes. The proposed structure was implemented using fi eld programmable gate arrays (FPGA). The digital circuit was designed, compiled and successfully simulated using the MAX+PLUS II Programmable Logic Development S ystem by Altera Corporation. The EPF10K200SFC484-1 FPGA device of the FLEX1 0KE device family was utilized for the realization of the system. The typical clock fre quency is 55 MHz and the system can be used for real-time imaging applications whe re fast processing is required [6]. As an example,the time required to perform filtering of a gray-scale image of 260×244 pixels is approximately 10.6 msec.2. ADAPTIVE FILTERING PROCEDUREThe output of a median filter at a point x of an image f depends on the values of t he image points in the neighborhood of x. This neighborhood is determined by a wind ow W that is located at point x of f including n points x1, x2, …, xn of f, with n=2k+1. The proposed adaptive content based median filter can be utilized for impulse noisesu p pression in gray-scale images. A block diagram of the adaptive filtering procedure is depicted in Fig. 1. The noise detection procedure for both positive and negative noise is as follows:(i) We consider a neighborhood window W that is located at point x of the image f. Th e differences between the central pixel at point x and the pixel values of the n-1surr ounding points of the neighborhood (excluding thevalue of the central pixel) are co mputed.(ii) The sum of the absolute values of these differences is computed, denoted as fabs(x ). This value provides ameasure of closeness between the central pixel and its su rrounding pixels.(iii) The value fabs(x) is compared to fthreshold(x), which is anappropriately selected positive integer threshold value and can be modified. The central pixel is conside red to be noise when the value fabs(x) is greater than thethreshold value fthresho d(x).(iv) When the central pixel is considered to be noise it is substituted by the median val ue of the image neighborhood,denoted as fk+1, which is the normal operationof the median filter. In the opposite case, the value of the central pixel is not altered and the procedure is repeated for the next neighborhood window.From the noised etection scheme described, it should be mentioned that the noise detection level procedure can be controlled and a range of pixel values (and not only the fixedvalues of 0 and 255, salt and pepper noise) is considered asimpulse noise.In Fig. 2 the results of the application of the median filter and the CBMF in the gray-sca le image “Peppers” are depicted.More specifically, in Fig. 2(a) the original,uncor rupted image“Peppers” is depicted. In Fig. 2(b) the original imagedegraded by 5% both positive and negative impulse noise isillustrated. In Figs 2(c) and 2(d) the resultant images of the application of median filter and CBMF for a 3×3-pixel win dow are shown, respectively. Finally, the resultant images of the application of m edian filter and CBMF for a 5×5-pixelwindow are presented in Figs 2(e) and 2(f). It can be noticed that the application of the CBMF preserves much better edges a nddetails of the images, in comparison to the median filter.A number of different objective measures can be utilized forthe evaluation of these results. The most wi dely used measures are the Mean Square Error (MSE) and the Normalized Mean Square Error (NMSE) [1]. The results of the estimation of these measures for the two filters are depicted in Table I.For the estimation of these measures, the result ant images of the filters are compared to the original, uncorrupted image.From T able I it can be noticed that the MSE and NMSE estimatedfor the application of t he CBMF are considerably smaller than those estimated for the median filter, in all the cases.Table I. Similarity measures.3. HARDWARE ARCHITECTUREThe structure of the adaptive filter comprises four basic functional units, the mo ving window unit , the median computation unit , the arithmetic operations unit , and th e output selection unit . The input data of the system are the gray-scale values of the pi xels of the image neighborhood and the noise threshold value. For the computation of the filter output a3×3 or 5×5-pixel image neighborhood can be selected. Image input d ata is serially imported into the first stage. In this way,the total number of the inputpin s are 24 (21 inputs for the input data and 3 inputs for the clock and the control signalsr equired). The output data of the system are the resultant gray-scale values computed f or the operation selected (8pins).The moving window unit is the internal memory of the system,used for storing th e input values of the pixels and for realizing the moving window operation. The pixel values of the input image, denoted as “IMAGE_INPUT[7..0]”, areimported into this u nit in serial. For the representation of thethreshold value used for the detection of a no Filter Impulse noise 5% mse Nmse(×10-2) 3×3 5×5 3×3 5×5Median CBMF 57.554 35.287 130.496 84.788 0.317 0.194 0.718 0.467ise pixel 13 bits are required. For the moving window operation a 3×3 (5×5)-pixel sep entine type memory is used, consisting of 9 (25)registers. In this way,when the windoP1 P2 P3w is moved into the next image neighborhood only 3 or 5 pixel values stored in the memory are altered. The “en5×5” control signal is used for the selection of the size of th e image window, when“en5×5” is equal to “0” (“1”) a 3×3 (5×5)-pixel neighborhood is selected. It should be mentioned that the modules of the circuit used for the 3×3-pix el window are utilized for the 5×5-pixel window as well. For these modules, 2-to-1mu ltiplexers are utilized to select the appropriate pixel values,where necessary. The mod ules that are utilized only in the case of the 5×5-pixel neighborhood are enabled by th e“en5×5” control signal. The outputs of this unit are rows ofpixel values (3 or 5, respe ctively), which are the inputs to the median computation unit.The task of the median c omputation unit is to compute themedian value of the image neighborhood in order to substitutethe central pixel value, if necessary. For this purpose a25-input sorter is utili zeed. The structure of the sorter has been proposed by Batcher and is based on the use of CS blocks. ACS block is a max/min module; its first output is the maximumof the i nputs and its second output the minimum. The implementation of a CS block includes a comparator and two 2-to-1 multiplexers. The outputs values of the sorter, denoted a s “OUT_0[7..0]”…. “OUT_24[7..0]”, produce a “sorted list” of the 25 initial pixel val ues. A 2-to-1 multiplexer isused for the selection of the median value for a 3×3 or 5×5-pixel neighborhood.The function of the arithmetic operations unit is to computethe value fabs(x), whi ch is compared to the noise threshold value in the final stage of the adaptive filter.The in puts of this unit are the surrounding pixel values and the central pixelof the neighb orhood. For the implementation of the mathematical expression of fabs(x), the circuit of this unit contains a number of adder modules. Note that registers have been used to achieve a pipelined operation. An additional 2-to-1 multiplexer is utilized for the selec tion of the appropriate output value, depending on the “en5×5” control signal. From th e implementation point of view, the use of arithmetic blocks makes this stage hardwar e demanding.The output selection unit is used for the selection of the appropriateoutput value of the performed noise suppression operation. For this selection, the corresponding no ise threshold value calculated for the image neighborhood,“NOISE_THRES HOLD[1 2..0]”,is employed. This value is compared to fabs(x) and the result of the comparison Classifies the central pixel either as impulse noise or not. If thevalue fabs(x) is greater than the threshold value fthreshold(x) the central pixel is positive or negative impulse noise and has to be eliminated. For this reason, the output of the comparison is used as the selection signal of a 2-to-1 multiplexer whose inputs are the central pixel and the c orresponding median value for the image neighborhood. The output of the multiplexer is the output of this stage and the final output of the circuit of the adaptive filter.The st ructure of the CBMF, the computation procedure and the design of the four aforeme n tioned units are illustrated in Fig. 3.ImagewindoeFigure 1: Block diagram of the filtering methodFigure 2: Results of the application of the CBMF: (a) Original image, (b) noise corrupted image (c) Restored image by a 3x3 MF, (d) Restored image by a 3x3 CBMF, (e) Restored image by a 5x5 MF and (f) Restored image by a 5x5 CBMF.4. IMPLEMENTATION ISSUESThe proposed structure was implemented in FPGA,which offer an attractive com bination of low cost, high performance and apparent flexibility, using the software pa ckage+PLUS II of Altera Corporation. The FPGA used is the EPF10K200SFC484-1 d evice of the FLEX10KE device family,a device family suitable for designs that requir e high densities and high I/O count. The 99% of the logic cells(9965/9984 logic cells) of the device was utilized to implement the circuit . The typical operating clock frequ ency of the system is 55 MHz. As a comparison, the time required to perform filtering of a gray-scale image of 260×244 pixelsusing Matlab® software on a Pentium 4/2.4 G Hz computer system is approximately 7.2 sec, whereas the corresponding time using h ardware is approximately 10.6 msec.The modification of the system to accommodate windows oflarger sizes can be done in a straightforward way, requiring onlya small nu mber of changes. More specifically, in the first unit the size of the serpentine memory P4P5P6P7P8P9SubtractorarryMedianfilteradder comparatormuitiplexerf abc(x)valueand the corresponding number of multiplexers increase following a square law. In the second unit, the sorter module should be modified,and in the third unit the number of the adder devicesincreases following a square law. In the last unit no changes are requ ired.5. CONCLUSIONSThis paper presents a new hardware structure of a content based median filter, ca pable of performing adaptive impulse noise removal for gray-scale images. The noise detection procedure takes into account the differences between the central pixel and th e surrounding pixels of a neighborhood.The proposed digital circuit is capable ofproce ssing grayscale images of 8-bit resolution, with 3×3 or 5×5-pixel neighborhoods as op tions for the computation of the filter output. However, the design of the circuit is dire ctly expandableto accommodate larger size image windows. The adaptive filter was d eigned and implemented in FPGA. The typical clock frequency is 55 MHz and the sys tem is suitable forreal-time imaging applications.REFERENCES[1] W. K. Pratt, Digital Image Processing. New York: Wiley,1991.[2] G. R. Arce, N. C. Gallagher and T. Nodes, “Median filters:Theory and applicat ions,” in Advances in ComputerVision and Image Processing, Greenwich, CT: JAI, 1986.[3] T. A. Nodes and N. C. Gallagher, Jr., “The output distributionof median type filte rs,” IEEE Transactions onCommunications, vol. COM-32, pp. 532-541, May1984.[4] T. Sun and Y. Neuvo, “Detail-preserving median basedfilters in imageprocessing,” Pattern Recognition Letters,vol. 15, pp. 341-347, Apr. 1994.[5] E. Abreau, M. Lightstone, S. K. Mitra, and K. Arakawa,“A new efficient approachfor the removal of impulsenoise from highly corrupted images,” IEEE Transa ctionson Image Processing, vol. 5, pp. 1012-1025, June 1996.[6] E. R. Dougherty and P. Laplante, Introduction to Real-Time Imaging, Bellingham:SPIE/IEEE Press, 1995.二、英文翻译基于中值滤波的新的内容摘要在本设计中的提出了基于中值滤波的硬件实现用来抑制脉冲噪声的干扰。

相关文档
最新文档