APPROXIMATION OF IMAGES BY BASIS FUNCTIONS FOR MULTIPLE REGION SEGMENTATION WITH LEVEL SETS

合集下载

图像分析英语作文模板

图像分析英语作文模板

图像分析英语作文模板Title: A Comprehensive Guide to Writing an English Essay on Image Analysis。

Introduction:Image analysis is a crucial aspect of visual communication, requiring a keen eye for detail and an understanding of various techniques. In this essay, we will delve into the fundamentals of image analysis, discussing key components such as interpretation, context, and the significance of visual elements.1. Introduction to Image Analysis:Image analysis involves examining visual content to derive meaning, often in fields such as art, media, and communication. It encompasses a wide range of techniques, including semiotics, composition analysis, and visual rhetoric.2. Key Components of Image Analysis:a. Semiotics: This refers to the study of signs and symbols and their interpretation. In image analysis, semiotics help identify the connotations and meanings associated with visual elements.b. Composition Analysis: Analyzing the arrangement of visual elements within an image is crucial for understanding its message. Composition techniques such as framing, balance, and focal points influence how viewers perceive an image.c. Visual Rhetoric: Visual rhetoric examines how images persuade and influence their audience. Understanding visual rhetoric involves analyzing techniques such as color choice, lighting, and visual metaphors.3. Steps in Image Analysis:a. Observation: Begin by carefully observing theimage, noting its visual elements, composition, and context.b. Interpretation: Interpret the meaning behind the visual elements, considering cultural, historical, and contextual factors.c. Analysis: Analyze the image's composition, visual rhetoric, and the relationship between its elements to understand its message.d. Conclusion: Summarize your findings and reflecton the image's significance within its context.4. Example Analysis:Let's analyze an image of a protest rally. The image features a crowd of people carrying signs and chanting slogans. The composition is chaotic, with overlappingfigures and bold colors dominating the scene. The use of red, a color often associated with passion and activism, conveys the intensity of the protest. Additionally, thefocal point of the image is a banner held high by theprotesters, featuring a powerful slogan demanding social change. Through semiotic analysis, we interpret the image as a representation of collective activism and a call to action for societal reform.5. Conclusion:Image analysis is a multifaceted process that requires careful observation, interpretation, and analysis. By understanding the key components of image analysis and following a structured approach, one can uncover the deeper meanings embedded within visual content. Whether analyzing art, advertising, or media, the skills gained from image analysis are invaluable for navigating the visual landscape of the modern world.。

A New Approach of Image Denoising Based on Discrete Wavelet Transfer

A New Approach of Image Denoising Based on Discrete Wavelet Transfer

222)(2 21Gaussian V V S m x eA New Approach of Image Denoising Based onDiscrete Wavelet TransformSami Hussien IsmaelDr. Firas Mahmood MustafaDr.İbrahim Taner OKÜMÜŞBioengineering and Sciences Department Technical College of Engineering (TCE)Bioengineering and Sciences Department Kahramanmaraş Sütçü İmam UniversityDuhok Polytechnic University (DPU)Kahramanmaraş Sütçü İmam UniversityKahramanmaraş, Turkey Duhok, IraqKahramanmaraş, Turkey Samihu54@Firas.mah.m@iokumus@.trAbstract —Image denoising is a process that used to enhance the image quality after degraded by the noise. There are several methods have been proposed for image denoising. In this paper, a new method of image denoising has been proposed. The proposed method is based on using wavelet transform.W avelet transform is best method for analysis the image due to the ability to split the image into sub-bands and work on each sub-band frequencies separately. Also, the robust median estimator has been used to estimate the noise ratio at the noisy image. According to experimental results, the proposed method presents best values for MSE and PSNR for denoised images. Also, by using different types of wavelet transform filters is make the proposed approach can obtained best results for image denoising process.Keywords: AGN; DWT; MSE; Noisy image; PSNR; Threshold;I.I NTRODUCTIONThe most images have non-stationary properties because their contents can be graduated from smooth areas to sharpen areas, where the smooth areas represent low frequencies regions while the sharpen area defines as a high frequencies regions [1]. These regions can be obtained by using two dimensional discrete wavelet transform DWT,where the approximation sub-band describes the low frequencies regions, and the three details sub-bands describes the high frequencies regions. For this reason the wavelet transform is appropriate to be used for study the properties of digital images [2].In [2], there are several approaches of image denoising have been proposed. DWT filters are the most popular one among them.Noise can be defined as any unwanted interference image data and noise often can be small values with high frequencies. Noise appears in images from multiple sources due to processed or procedure, transmissions reception. In addition, noise takes various forms such as uniform noise, noise impulse and also called the (Salt-and-Pepper) and Additive Gaussian Noise (AGN) that will be used during this research. The reason of choosing AGN at this work, because of this type of noise is include other types of noise implicitly. In addition, the possibility of appearing this type of noise in images is larger than other types of noise andnoisy image with AGN becomes more difficult during image denoising process. Fig. 1 shows the distribution of Gaussian noise and the AGN equation is:(1) Where:V :Standard-Deviation. m :Mean.Fig. 1 the distribution of Gaussian noiseII.P ROPOSED D ENOISING S CHEMEThe noisy image can be represented by using natural type of AGN and its equation:),(),(),(j i j i I j i In H (2) Where:I (i,j): Original image without noise.),(j i H :AGN.In (i,j): noisy image.The goal is to obtain the original image I (i, j) that contains Gaussian noise from the In (i, j) noisy image by applying a suitable method of image denoising [1],[3],[4], This means that: H (i,j) = 0.Wavelet transform is appropriate for the digital image denoising methods. Because it can be transforming the images into sub-bands, each one contains a certain frequencies that analyzed and processed easily [1],[5].Image denoising using DWT approach can be summarized in the following steps:2016 World Symposium on Computer Applications & Research),(j i xc 1.Apply one level 2-dim DWT on noisy image (the noisy image divided into four sub-bands (Low-low LL, Low-high LH, High-low HL and High-high HH sub-bands). 2.Find the threshold value for each sub-band (LH, HL and HH).pare the threshold value of each sub-band with all pixel values of selected sub-band. If the pixel value less than threshold value, then set the pixel value to be zero. Otherwise, leave it and select the next pixel value. This step repeated until the last pixel value of selected sub-band.4.Apply one level inverse 2-dim DWT on the sub-bands to obtain a new image without noise.5.Find the values of mean square error (M SE) and peak signal-to-noise ratio (PSNR) for new image and original image to evaluate the proposed image denoising method. To compute the values of MSE and PSNR, the following equation is used:>@2101),(),(.1¦¦ M i N j j i I j i F N M MSE (3)MSEPSNR 2)255(log.10 (4)Where:I (i, j): Original image. F (i, j): New image M and N: size of image255: Max pixel value of grayscale image.The selection of the threshold value is important factor in decision making to delete or keep the pixel values of selected image sub-band [6]. Therefore, the threshold value must be high for denoising purpose and small value to maintain properties of the image. The threshold value can be found by applying the following equation [1].x Thx V V 2(5)Where:Thx : Threshold value.2V :Noise value of noisy image.x V :Standard deviation for one of these sub-bands (LH,HL and HH), and can be calculated by equation:NN i m i x ¦102)(V(6)The standard deviation value for each sub-band gives a uniform threshold value for its properties. TheVThthreshold value has an inverse relationship with thestandard deviation of the sub-band and it has positiverelationship with the standard deviation value of the noisyimage. In case (1 xV V) that means the pixel values of sub-band greater than the noise value. Therefore, the VThvalue must be select with minimum value to maintain the several properties of the pixel values of sub-band and denoise some pixel values of sub-band. While, in case (1!!xV V ) that means the pixel values of sub-band has a high value of noise. Therefore, the VTh must be select withmaximum value to denoise a several pixel values of sub-band and this can be lead to loss many properties of the pixel values of sub-band [1] .The noise power (2V ) can be computed by the estimation method that called (Robust Median Estimator) and applying the following formula [1].6745.0/),(j i x MedianV (7)Where x (i, j): represent analysis of diagonal sub-band for the first stage (HH or HL).In this paper, a new algorithm of image denoising has been proposed based on using DWT with two-stage of analysis. During the algorithm implementation, two grayscale images with sizes (256 × 256) that shown in Fig. 2 and noise type is additive Gaussian noise with zero rate (m = 0) and the standard deviation values (V =10, V =15, V =25). The four (LL, LH, HL and HH) sub-bands of the selected noisy image can be obtained by applying the 2-dim DWT.(a) Camera man (b) LenaFig. 2 the original imagesThe block diagram that shown in Fig. 3 is describe the image denoising process of the proposed method.The image denoising method not has any information about the amount of additive noise in contents of the noisy image. For this reason, the estimation method is used to estimate the amount of noise by using Robust M edian Estimator [1].After computing the noise value, 2-dim DWT with two stages is applied on the noisy image to obtain the (LH, HL, HH, LH1, HL1 and HH1) sub-bands. Then, the standard deviation for each sub-band is computed to use it in the equation (5) to calculate the threshold value for each sub-band and equation (8) is used for comparison process.°¿°¾½°¯°®­t c otherwise Thx x j i x j i x ,0),,(),( (8)Where,),(j i x : The pixel values of noisy image sub-band. :The pixel value of denoised image.Finally, 2-dim inverse DWT with two stages is applied on these sub-bands to obtain denoised image.According to analysis of these sub-bands, the image denoising algorithm works on the following states:State 1: Apply one level 2-dim DWT on noisy image and compute Thx for these (LH, HL and HH) sub-bands. Then, the equation (8) is applied on all pixelvalues of these sub-bands. An inverse 2-dim DWTis applied on all sub-bands to construct a denoisedimage. Calculate M SE and PSNR between thedenoised image and original image.State 2: Obtain a (LL) sub-band by applying one level 2-dim DWT on the noisy image. And then apply 2-dim DWT on (LL) sub-band to obtain (LH1, HL1and HH1) sub-bands and calculate Thx for thesesub-bands. Then, equation (8) is applied on allpixel values of these sub-bands. To obtain thedenoised image an inverse 2-dim DWT is appliedon all pute M SE and PSNRbetween the denoised image and the original image. State 3: Apply one level 2-dim DWT on noisy image to obtain these (LH and HL) sub-bands, find their thxvalue by applying threshold method and set thepixel values of HH sub-band to be all zero values.Equation (8) is applied on (LH and HL) sub-bands.Then, apply an inverse 2-dim DWT for all sub-bands to construct a denoised image. CalculateM SE and PSNR between the denoised image andoriginal image.State 4: Obtain (LH)sub-band by applying one level 2-dim DWT on the noisy image, find its thx value byapplying threshold method and set the pixel valuesof (HL and HH) sub-bands to be all zero values.Equation (8) is applied on (LH) sub-band. Then, adenoised image can be obtained by applying aninverse 2-dim DWT on all sub-bands. Find M SEand PSNR values for the denoised image andoriginal image.State 5: Divide the noisy image into (LL, HL, LH and HH) sub-bands by applying one level 2-dim DWT.Then, compute thx value for (HL)sub-bands andset the pixel values of (LH and HH) sub-bands tobe all zero values. Equation (8) is applied on (HL)sub-band. Then, apply an inverse 2-dim DWT onall sub-bands to construct a denoised image.Compare between the denoised image and originalimage according to their MSE and PSNR values. State 6: One level 2-dim DWT is applied on noisy image to obtain (LL, HL, LH and HH) sub-bands. And then,set (HL, LH and HH) sub-bands to be all zerovalues.An inverse 2-dim DWT is applied on allsub-bands to construct a denoised image. CalculateM SE and PSNR between the denoised image andoriginal image.Fig. 3 the proposed method block diagramIII.E XPERIMENTAL R ESULTSM atlab technical programming language has been used to implement the proposed method. AGN is added to testoriginal image and the quantity of this noise must be notmore than a half amount of the standard deviation of theseoriginal images. Otherwise, if the noise amount is great thatmeans the denoising process will be remove more pixel values of the original image.The implementation results of proposed algorithm onCamera man image shows in Fig. 4 with the standard deviation is the (62.341). After AGN with values (V=10, V=15 and V=25) is added to it, the value of standard deviation has been became (62.928,63.459 and 65.211).Then, 2-dim DWT with one stage is applied on noisy imageto obtain high frequency sub-bands (HH, HL and LH). Toestimate the value of noise, equation (7) has been applied on (HH) sub-band and the estimation values of additive Gaussian noise for Camera man and Lena images tabulated in Table (1). By applying the equation (6) to obtain the threshold value Thx for each sub-band.Table 1: Standard deviation Std values for Camera man and Lenaimages using HH sub-bandImageStdoriginalimageStdNoise=10StdNoise=15StdNoise=25Stdestimation noise=10Stdestimation noise=15Stdestimation noise=25Cameraman62.34162.92863.45965.21111.02515.56624.723Lena52.29153.12454.19957.07010.71715.59024.416 According to experimental results for Camera man image that shown in Table (2), the proposed method at state (one sub-band deletion (HH) sub-band) has been achieved best results at noise ratio (V=10) and type of DWT filter is Rbior. While, in the state (all sub-band deletion), the proposed method has been achieved best results at the noise ratio (V=15 and V=25) and type of DWT filter is Sym.Fig. 4 Camera man image (a) Image with noise 10 (b)Denoised image without deleting level one (c)Denoised image without deleting level two (d)Denoised Image with HH sub-band deletion (e) Denoised image with HH and HL sub-bands deletion (f) Denoised image with HH, HL and LH sub-band deletion.Also, the proposed method is presented best values of PSNR for Camera man image with V =10 in state HH sub-band deletion among other states as shown in Fig. 5. While, in state all sub-bands deletion with V =15 and V =25, the proposed method presents best PSNR values for Camera man image as shown in Fig. 6 and Fig. 7.Fig .5 PSNR values for camera man image at V =10Table 3. shows the comparison of PSNR and M SEvalues for camera man image at two levels of DWT with V=10 between the proposed method and methods thatproposed in [7]. According to results, the proposed method is presented best PSNR and MSE values than methods thatproposed in [7].Table 2:MSE and PSNR values for Camera man image at different waveletfiltersImage Denoising MethodsOne analysis level Two analysis levelDelete HHDelete HH and HLDelete HH, HL and LH Filter NameMSE 88.24287.86281.71299.308104.782CoifPSNR 28.67428.69329.00828.16127.928MSE74.32374.32275.98594.347101.248BiorPSNR 29.42029.42029.32428.38428.077MSE 74.93174.76068.23294.37498.715RbioPSNR 29.38429.39429.79128.38228.187MSE 88.08287.88780.37389.94098.189dbPSNR 28.68228.69229.08028.59128.210MSE 87.49387.14480.42289.94096.324SymPSNR 28.71128.72829.07728.59128.293MSE 135.265140.929105.967118.99389.694HaarPSNR 26.81926.64127.87927.37628.603MSE167.161173.368104.085151.99788.546DmeyPSNR25.89925.74127.95726.31228.659Table 3. Comparison of PSNR and MSE values for Camera man image attwo levels of DWT with V =10.Also, Fig. 8 shows Lena image after the proposedmethod has been applied on it. Depending on the experimental results that shown in table 4,the proposed method at state (two sub-bands deletion (HH and LH) sub-bands has been achieved best results at noise ratio (V =10) and type of DWT filter is Bior 5.5 , While, in the state (all sub-band deletion), the proposed method has been achieved best results at the noise ratio (V =15) and type of DWT filter is Sym4, While, in the state (all sub-band deletion), the proposed method has been achieved best results at the noise ratio (V =25) and type of DWT filter is Sym5.Fig. 8 Lena image (a) Image with noise 10 (b) Denoised image without deleting level one (c) Denoised image without deleting level two (d)DWT LevelUniversal Threshold Visu Shrink Proposed Method MSEPSNRMSEPSNRMSEPSNR1st Level 564.92720.610548.29720.74074.32329.4202nd Level813.40319.027742.40319.42474.32229.420(a) (b) (c)(d) (e) (f)(a) (b) (c)(d) (e) (f)Denoised Image with HH sub-band deletion (e) Denoised image with HH and LH sub-bands deletion (f) Denoised image with HH, HL and LH sub-band deletion.In addition, the proposed algorithm is achieved best values of PSNR for Lena image with V =10 in state HH and LH sub-band deletion among other states as shown in Fig. 9. While, in state all sub-bands deletion with V =15 and V =25, the proposed method presents best PSNR values for Camera man image as shown in Fig. 10 and Fig. 11.Fig .9 PSNR values for Lena image at V =10Fig .10 PSNR values for Lena image at V =15Fig .11 PSNR values for Lena image at V =25IV.C ONCLUSIONIn this research, a new method of image denoising hasbeen proposed. The proposed method based on DWT that represents a best method for analysis the image due to the ability to split the image into sub-bands and work on each sub-band frequency separately. Also, the robust median estimator has been used to estimate the noise ratio in the noisy image. According to experimental results, the proposed method presents best values of MSE and PSNR for denoised images.Also, by using different types of wavelet transform filters is make the proposed approach can obtained best results for image denoising process.In future, the proposed method can be modified by more accurate estimation function to enhance its ability at image denoising process.Table 4: MSE and PSNR values for Lena image at different wavelet filtersImage Denoising MethodsOne analysis levelTwo analysis levelDelete HHDelete HH and LHDelete HH, HL and LHFilter NameMSE 94.26093.95482.27070.90971.284CoifPSNR 28.38828.40228.97829.62429.601MSE 82.60782.60472.92366.05069.927BiorPSNR 28.96128.96129.50229.93229.684MSE 82.09982.39767.25266.35571.962RbioPSNR28.98728.97229.85429.91229.560MSE 93.94693.77480.89468.56669.473dbPSNR 28.40228.41029.05229.77029.713MSE 93.89693.72280.89468.56669.712SymPSNR28.40428.41229.05229.77029.698MSE 94.450106.30977.64784.97096.829HaarPSNR28.37927.86529.23028.83828.271MSE 81.00692.72174.01582.61894.359DmeyPSNR29.04628.45929.43828.96028.383R EFERENCES[1]S. G. Chang, B. Yu, and M. Vetterli, “Adaptive Wavelet Thresholding for Image Denoising and Compression”, IEEE Transactions on Image Processing, vol. 9, no. 9, September 2000.[2]R. K. Rai, J. Asnani and T. R. Sontakke, “Review of Shrinkage Techniques for Image Denoising”, International Journal of Computer Applications (0975 – 8887), vol. 42, no.19, March 2012, pp. 13-16[3]P. Moulin, “Multiscale Image Decompositions and Wavelets”,Handbook of Image and Video Processing, 2nd edition, Academic Press, 2005[4] C.S. Burrus, R.A. Gopinath and H. GUO , “Introduction to Wavelets and Wavelet Transforms: A Primer ”. Prentice Hall, 1998.[5]A. Hamza and H. Krim, “Image Denoising: A Nonlinear Rob ust Statistical Approach”, IEEE Transactions on Signal Proce ssing, vol. 49, no. 12, pp. 3045-3054, December 2001.[6]A. Al Jumah, “Denoising of an Image Using Discrete Stationary Wavelet Transform and Various Thresholding Techniques”, Journal of Signal and Information Processing, vol. 4, pp.33-41, February 2013.[7]Anuta m and Rajni, “Performance Analysis of Image Denoising with Wavelet Thresholding M ethods For Different Levels of Decomposition”, The International Journal of Multimedia & Its Applications (IJMA) vol.6, no.3, pp.35-46, June 2014.P S NR。

数字图像处理-冈萨雷斯-课件(英文)Chapter11-表示与描述可编辑全文

数字图像处理-冈萨雷斯-课件(英文)Chapter11-表示与描述可编辑全文
an image in other forms that are more suitable than the image itself.
Benefits: - Easier to understand - Require fewer memory, faster to be processed - More “ready to be used”
3 from
Lupper
Turn Right OK!
Turn Right OK!
Algorithm (cont.)
For the lower side of a convex hull
7. 8.
Put For
the i=
np-o2indtoswpnn
The First Difference of a Chain Codes
Problem of a chain code: a chain code sequence depends on a starting point.
Solution: treat a chain code as a circular sequence and redefine the starting point so that the resulting sequence of numbers forms an integer of minimum magnitude.
Remove the first and the last points from AReptpuernndLLlower to Lupper resulting in the list
LLlower
3 points from Llower
Turn Left NOK!

reader-response-theory

reader-response-theory

• 5 the ideal reader • Proposed by J. Culler • Someone who has possessed , or rather
internalized ,the literary conventions , the mastery of which would enable him to perform literary readings acceptable to other readers, for such conventions constitute the very institution of literature itself.
“horizon of expectations,” “the set of cultural ,ethical, and literary expectations of a work’s readers in the historical moment of its appearance.
7/15/2024
The messenger of the gods
A theory that sees history as a living dialogue between past , present and future ,and seeks patiently to remove obstacles to this endless mutual communication.
7/15ቤተ መጻሕፍቲ ባይዱ2024
The implied reader
• Against these “limitations“,Iser’s concept of the reader stands as a good contrast.

EEC机动车辆及其挂车的法定铭牌安装的位置和方法

EEC机动车辆及其挂车的法定铭牌安装的位置和方法
(2) OJ N° C 47, 27.2.1975, p. 4.
(3) OJ N° L 42, 23.2.1970, p. 1.
Article 1
第一条 For the purposes of this Directive, ’vehicle’ means any motor vehicle intended for use on the road, with or without bodywork, having at least four wheels and a maximum design speed exceeding 25 km/h, and its trailers, with the exception of vehicles which run on rails, agricultural or forestry tractors and machinery, and public works vehicles. 在本规定中,“车辆” 意谓任何打算在道路上使用的、 有或者没有车身、但是至少有四个轮 子同时最大设计速度超过 25 公里/小时的机动车辆及其拖车(除了在铁轨上运行的车辆)、 农用或林业拖拉机和机械 、公共建设工程车辆。
76/114/EEC Statutory plates - Location, method of attachment
021128
法定铭牌--安装的位置和方法
Deutscher Text:
76/114/EWG - Schilder - Angaben, Lage und Anbringungsart
对各成员国关于机动车辆及其挂车的法定铭牌及其内容,以及铭牌安装的位置和方法的法律的协调统一
76/114/EEC of 18.12.1975 OJ No. L24 of 30.01.1976 78/507/EEC of 19.05.1978 OJ No. L155 of 13.06.1978

专业外语数字图像基础Digital Image Fundamentals

专业外语数字图像基础Digital Image Fundamentals

Page
14
6. Dilation and Erosion
Dilation adds a layer of pixels to all objects in an image. Erosion removes one layer of pixels from all objects. When dilation is followed by erosion, gaps in broken boundaries identified through edge detection can be filled in. Conversely, when erosion is followed by dilation, spots of noise in an image are removed.
Successfully detecting the edges in an image is the first step towards confident identification of object boundaries and then objects recognition. From boundary information, shape characteristics like perimeter and area can be calculated, which can be used to classify an object.
由于屏幕上的点、线和面都是由像 素组成的,显示器可显示的像素越 多,画面就越精细,同样的屏幕区 域内能显示的信息也越多,所以分 辨率是个非常重要的性能指标之一 。

800×480 1280×720 1920×1080
Page
8

数字图像处理课件(冈萨雷斯第三版)复习材料

数字图像处理课件(冈萨雷斯第三版)复习材料

(1) 名词解释RGB Red Green Blue,红绿蓝三原色CMYK Cyan Magenta yellow blacK , 用于印刷的四分色HIS Horizontal Situation Indicator 水平位置指示器FFT Fast Fourier Transform Algorithm (method) 快速傅氏变换算法CWT continuous wavelet transform 连续小波变换DCT Discrete Cosine Transform 离散余弦变换DWT DiscreteWaveletTransform 离散小波变换CCD Charge Coupled Device 电荷耦合装置Pixel: a digital image is composed of a finite number of elements,each of which has a particular lication and value,these elements are called pixel 像素DC component in frequency domain 频域直流分量GLH Gray Level Histogram 灰度直方图Mather(basic)wavelet:a function (wave) used to generate a set of wavelets, 母小波,用于产生小波变换所需的一序列子小波Basis functions basis image 基函数基图像Multi-scale analysis 多尺度分析Gaussian function 高斯函数sharpening filter 锐化滤波器Smoothing filter/convolution 平滑滤波器/卷积Image enhancement /image restoration 图像增强和图像恢复(2)问答题1. Cite one example of digital image processingAnswer: In the domain of medical image processing we may need to inspect a certain class of images generated by an electron microscope to eliminate bright, isolated dots that are no interest.2.Cite one example of frequency domain operation from the following processing result, make a general comment about ideal highpass filter (figure B) and Gaussian highpass filter(figure D)A. Original imageB. ideal highpass filterIn contrast to the ideal low pass filter, it is to let all the signals above the cutoff frequency fc without loss, and to make all the signals below the cutoff frequency of FC without loss of.C. the result of ideal highpass filterD. Gaussian highpass filterHigh pass filter, also known as "low resistance filter", it is an inhibitory spectrum of the low frequency signal and retain high frequency signal model (or device). High pass filter can make the high frequency components, while the high-frequency part of the frequency in the image of the sharp change in the gray area, which is often the edge of the object. So high pass filter can make the image get sharpening processingE. The result of Gaussian filter3.The original image, the ideal lowpass filter and Gaussian lowpass filter are shown below B nd C .D and E are the result of the eitherfilter B or CA. Draw lines to connect the filter with their resultB. Explain the difference of the two filtersDue to excessive characteristics of the ideal low-pass filter too fast Jun, it will produce a ringing phenomenon.Over characteristics of Gauss filter is very flat, so it is not ringing4.What is the result when applying an averaging mask with the size 1X1?5.State the concept of the Nyquist sampling theorem from the figure belovyThe law of sampling process should be followed, also called the sampling theorem and the sampling theorem. The sampling theorem showsthe relationship between the sampling frequency and the signal spectrum, and it is the basic basis of the continuous signal discretization. In analog / digital signal conversion process, when the sampling frequency fs.max greater than 2 times the highest frequency present in the signal Fmax fs.max>2fmax, sampling digital signal completely retained the information in the original signal, the general practical application assurance sampling frequency is 5 ~ 10 times higher than that of the signal of the high frequency; sampling theorem, also known as the Nyquist theorem6.A mean filter is a linear filter but a median filter is not, why?Mean filter is a typical linear filtering algorithm, it is to point to in the target pixels in the image to a template, this template including its surrounding adjacent pixels and the pixels in itself.To use in the template to replace all the pixels of average pixelvalues.Linear filter, median filter, also known as the main method used in the bounded domain average method.Median filter is a kind of commonly used nonlinear smoothing filter and its basic principle is to put the little value in a digital image or sequence to use value at various points in the field of a point at which the value to replace, its main function is to let the surrounding pixel gray value differences between larger pixel change with the surrounding pixels value close to the values, which can eliminate the noise of the isolated points, so median filter to filter out the salt and pepper noise image is very effective.(3)算法题1.The following matrix A is a 3*3 image and B is 3*3 Laplacian mask, what will be the resulting image? (Note that the elements beyond the border remain unchanged)2.Develop an algorithm to obtain the processing result B from original image A3.Develop an algorithm which computes the pseudocolor image processing by means of fourier tramsformAnswer:The steps of the process are as follow:(1) Multiply the input image f(x,y) by (-1)x+y tocenter the transform;(2) Compute the DFT of the image from (1) to get power spectrumF(u,v) of Fourier transform.(3) Multiply by a filter function h(u,v) .(4) Compute the inverse DFT of the result in (3).(5) Obtain the real part of the result in (4).(6) Multiply the result in (5) by(-1)x+y4.Develop an algorithm to generate approximation image series shown in the following figure b** means of down sampling.(4)编程题There are two satellite photos of night as blew.Write a programwith MATLAB to tell which is brighterAn 8*8 image f(i,i) has gray levels given by the following equation:f(i,i)=|i-j|, i,j=0,1 (7)Write a program to find the output image obtained by applying a 3*3 median filter on the image f(i,j) ;note that the border pixels remain unchanged.Answer:1.Design an adaptive local noise reduction filter and apply it to an image with Gaussian noise. Compare the performance of the adaptive local noise reduction filter with arithmetic mean and geometric mean filter.Answer:clearclose all;rt=imread('E:\数字图像处理\yy.bmp');gray=rgb2gray(rt);subplot(2,3,1);imshow(rt);title('原图像') ;subplot(2,3,2);imshow(gray);title('原灰度图像') ;rtg=im2double(gray);rtg=imnoise(rtg,'gaussian',0,0.005)%加入均值为0,方差为0.005的高斯噪声subplot(2,3,3);imshow(rtg);title('高噪点处理后的图像');[a,b]=size(rtg);n=3;smax=7;nrt=zeros(a+(smax-1),b+(smax-1));for i=((smax-1)/2+1):(a+(smax-1)/2)for j=((smax-1)/2+1):(b+(smax-1)/2)nrt(i,j)=rtg(i-(smax-1)/2,j-(smax-1)/2);endendfigure;imshow(nrt);title('扩充后的图像');nrt2=zeros(a,b);for i=n+1:a+nfor j=n+1:b+nfor m1=3:2m2=(m1-1)/2;c=nrt2(i-m2:i+m2,j-m2:j+m2);%使用7*7的滤波器Zmed=median(median(c));Zmin=min(min(c));Zmax=max(max(c));A1=Zmed-Zmin;A2=Zmed-Zmax;if(A1>0&&A2<0)B1=nrt2(i,j)-Zmin;B2=nrt2(i,j)-Zmax;if(B1>0&&B2<0)nrt2(i,j)= nrt2(i,j);elsenrt2(i,j)=Zmed;endcontinue;endendendendnrt3=im2uint8(nrt2);figure;imshow(nrt3);title('自适应中值滤波图');2. Implement Wiener filter with “wiener2” function of MatLab to an image with Gaussian noise and compare the performance with adaptive local noise reduction filter.代码如下:>> I=imread('E:\数字图像处理\yy.bmp');>>J=rgb2gray(I);>>K = imnoise(J,'gaussian',0,0.005);>>L=wiener2(K,[5 5]);>>subplot(1,2,1);imshow(K);title('高噪点处理后的图像');>>subplot(1,2,2);imshow(L);title('维纳滤波器处理后的图像');3. Image smoothing with arithmetic averaging filter (spatial convolution).图像平滑与算术平均滤波(空间卷积)。

Value Function Approximation on Non-Linear Manifolds for Robot Motor Control

Value Function Approximation on Non-Linear Manifolds for Robot Motor Control

Value Function Approximation on Non-Linear Manifoldsfor Robot Motor ControlMasashi Sugiyama Hirotaka Hachiya Christopher Towell and Sethu VijayakumarAbstract—The least squares approach works efficiently in value function approximation,given appropriate basis func-tions.Because of its smoothness,the Gaussian kernel is a popular and useful choice as a basis function.However,it does not allow for discontinuity which typically arises in real-world reinforcement learning tasks.In this paper,we propose a new basis function based on geodesic Gaussian kernels, which exploits the non-linear manifold structure induced by the Markov decision processes.The usefulness of the proposed method is successfully demonstrated in a simulated robot arm control and Khepera robot navigation.I.I NTRODUCTIONV alue function approximation is an essential ingredient of reinforcement learning(RL),especially in the context of solving Markov Decision Processes(MDPs)using policy iteration methods[1].In problems with large discrete state space or continuous state spaces,it becomes necessary to use function approximation methods to represent the value functions.A least squares approach using a linear com-bination of predetermined under-complete basis functions has shown to be promising in this task[2].Fourier func-tions(trigonometric polynomials),Gaussian kernels[3],and wavelets[4]are popular basis function choices for general function approximation problems.Both Fourier bases(global functions)and Gaussian kernels(localized functions)have certain smoothness properties that make them particularly useful for modeling inherently smooth,continuous functions. Wavelets provide basis functions at various different scales and may also be employed for approximating smooth func-tions with local discontinuity.Typical value functions in RL tasks are predominantly smooth with some discontinuous parts[5].To illustrate this, let us consider a toy RL task of guiding an agent to a goal in a grid world(see Fig.1(a)).In this task,a state corresponds to a two-dimensional Cartesian position of the agent.The agent can not move over the wall,so the value function of this task is highly discontinuous across the wall.On the other hand,the value function is smooth along the maze since neighboring reachable states in the maze have similar values (see Fig.1(b)).Due to the discontinuity,simply employing Fourier functions or Gaussian kernels as basis functions The authors acknowledgefinancial support from MEXT(Grant-in-Aid for Y oung Scientists17700142and Grant-in-Aid for Scientific Research(B) 18300057),the Okawa Foundation,and EU Erasmus Mundus Scholarship.Department of Computer Science,Tokyo Institute of Technology,2-12-1,O-okayama,Meguro-ku,Tokyo,152-8552,Japan sugi@cs.titech.ac.jpSchool of Informatics,University of Edinburgh,The King’s Buildings, Mayfield Road,Edinburgh EH93JZ,UK.H.Hachiya@, C.C.Towell@,sethu.vijayakumar@ tend to produce undesired,non-optimal results around the discontinuity,affecting the overall performance significantly. Wavelets could be a viable alternative,but are over-complete bases—one has to appropriately choose a subset of basis functions,which is not a straightforward task in practice. Recently,the article[5]proposed considering value func-tions defined not on the Euclidean space,but on graphs induced by the MDPs(see Fig.1(c)).V alue functions which usually contain discontinuity in the Euclidean domain(e.g., across the wall)are typically smooth on graphs(e.g.,along the maze).Hence,approximating value functions on graphs can be expected to work better than approximating them in the Euclidean domain.The spectral graph theory[6]showed that Fourier-like smooth bases on graphs are given as minor eigenvectors of the graph-Laplacian matrix.However,their global nature implies that the overall accuracy of this method tends to be degraded by local noise.The article[7]defined diffusion wavelets,which posses natural multi-resolution structure on graphs.The paper[8]showed that diffusion wavelets could be employed in value function approximation,although the issue of choosing a suitable subset of basis functions from the over-complete set is not discussed—this is not straight-forward in practice due to the lack of a natural ordering of basis functions.In the machine learning community,Gaussian kernels seem to be more popular than Fourier functions or wavelets because of their locality and smoothness[3],[9],[10].Fur-thermore,Gaussian kernels have‘centers’,which alleviates the difficulty of basis subset choice,e.g.,uniform allocation [2]or sample-dependent allocation[11].In this paper,we therefore define Gaussian kernels on graphs(which we call geodesic Gaussian kernel),and propose using them for value function approximation.Our definition of Gaussian kernels on graphs employs the shortest paths between states rather than the Euclidean distance,which can be computed efficiently using the Dijkstra algorithm[12],[13].Moreover, an effective use of Gaussian kernels opens up the possibility to exploit the recent advances in using Gaussian processes for temporal difference learning[11].When basis functions defined on the state space are used for approximating the state-action value function,they should be extended over the action space.This is typically done by simply copying the basis functions over the action space [2],[5].In this paper,we propose a new strategy for this extension,which takes into account the transition after taking actions.This new strategy is demonstrated to work very well when the transition is predominantly deterministic.→→→↓→→→→→→→→→→↑→↑→→→→↓→↓→→→→→→→→→→↑→→→↓↓→↓↓→→→→→↑→↑→↑→→→→↓→→→→→→→↑↑→↑→→→→→↓↓→↓→→→→→→→→↑→→→→→→↓→→→→→→→→↑→→→→→→→→↓↓↓↓↓→→→↑↑→→→↑→↓↓→↓↓↓↓↓→→→↑→→↑→↑→↓↓↓↓↓↓↓↓→→↑↑↑↑↑↑↑↑→↑→→→→→↑↑→→→→↑↑↑↑↑→↑→→↑→→→→→↑↑↑→→→→→→↑→→↑→↑↑↑↑↑→→→↑↑→↑↑↑→→→→→↑→↑→↑↑↑↑↑↑↑↑↑→→→↑↑↑↑↑↑↑→↑↑↑↑↑↑↑→→↑↑↑↑↑↑↑↑↑↑→↑↑↑→↑→↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑→↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑123456789101112131415161718192021 1234567891011121314151617181920(a)(b)(c)Fig.1.An illustrative example of an RL task of guiding an agent to a goal in the grid world.(a)Black areas are walls over which the agent cannot move while the goal is represented in gray.Arrows on the grids represent one of the optimal policies.(b)Optimal state value function(in log-scale).(c)Graph induced by the MDP and a random policy.II.F ORMULATION OF THE R EINFORCE MENT L EARNINGP ROBLEMIn this section,we briefly introduce the notation and reinforcement learning(RL)formulation that we will use across the manuscript.A.Markov Decision ProcessesLet us consider a Markov decision process(MDP),where is afinite1 set of states,is afinite set of actions,is the joint probability of making a transition to state if action is taken in state,is an immediate reward for making a transition from to by action, and is the discount factor for future rewards.The expected reward for a state-action pair is given as(1)Let be a deterministic policy which the agent follows.In this paper,we focus on deterministic policies since there always exists an optimal deterministic policy[2]. Let be a state-action value function for policy,which indicates the expected long-term discounted sum of rewards the agent receives when the agent takes action in state and follows policy thereafter.satisfies the following Bellman equation:(2)The goal of RL is to obtain a policy which results in maximum amount of long-term rewards.The optimal policy is defined as,whereis the optimal state-action value function defined by.1For the moment,we focus on discrete state spaces.In Sec.III-D,we extend the proposed method to the continuous state space.B.Least Squares Policy IterationIn practice,the optimal policy can not be directly obtained since and are usually unknown; even when they are known,direct computation of is often computationally intractable.To cope with this problem,the paper[2]proposed ap-proximating the state-action value function using a linear model:(3)where is the number of basis functions which is usu-ally much smaller than the number of states,are the parameters to be learned, denotes the transpose,and are pre-determined basis functions.Note that and can depend on policy,but we do not show the explicit dependence for the sake of simplicity.Assume we have roll-out samples from a sequence of actions:,where each tuple denotes the agent experiencing a transition fromto on taking action with immediate reward.Under the Least Squares Policy Iteration(LSPI)formulation[2], the parameter is learned so that the Bellman equation (2)is optimally approximated in the least squares sense2. Consequently,based on the approximated state-action value function with learned parameter,the policy is updated as(4)Approximating the state-action value function and updating the policy is iteratively carried out until some convergence criterion is met.III.G AUSSIA N K ERNELS ON G RAPHSIn the LSPI algorithm,the choice of basis functionsis an open design issue.Gaussian kernels have traditionally been a popular choice[2],[11],but they 2There are two alternative approaches:Bellman residual minimization andfixed point approximation.We take the latter approach following the suggestion in the reference[2].can not approximate discontinuous functions well.Recently, more sophisticated methods of constructing suitable basis functions have been proposed,which effectively make use of the graph structure induced by MDPs[5].In this section, we introduce a novel way of constructing basis functions by incorporating the graph structure;while relation to the existing graph-based methods is discussed in the separate report[14].A.MDP-Induced GraphLet be a graph induced by an MDP,where states are nodes of the graph and the transitions with non-zero transition probabilities from one node to another are edges. The edges may have weights determined, e.g.,based on the transition probabilities or the distance between nodes. The graph structure corresponding to an example grid world shown in Fig.1(a)is illustrated in Fig.1(c).In practice, such graph structure(including the connection weights)are estimated from samples of afinite length.We assume that the graph is connected.Typically,the graph is sparse in RL tasks,i.e.,,where is the number of edges and is the number of nodes.B.Ordinary Gaussian KernelsOrdinary Gaussian kernels(OGKs)on the Euclidean space are defined as(5) where are the Euclidean distance between states and;for example,when the Cartesian positions of and in the state space are given by and,respectively.is the variance parameter of the Gaussian kernel.The above Gaussian function is defined on the state space ,where is treated as a center of the kernel.In order to employ the Gaussian kernel in the LSPI algorithm,it needs to be extended over the state-action space. This is usually carried out by simply‘copying’the Gaussian function over the action space[2],[5].More precisely:let the total number of basis functions be,where is the number of possible actions and is the number of Gaussian centers.For the-th action and for the-th Gaussian center,the-th basis function is defined as(6)where is the indicator function,i.e.,if otherwise.Gaussian kernels are shift-invariant,i.e.,they do not directly depend on the absolute positions and,but depend only on the difference between two positions;more specifically,Gaussian kernels depend only on the distance between two positions.C.Geodesic Gaussian KernelsOn graphs,a natural definition of the distance would be the shortest path.So we define Gaussian kernels on graphs based on the shortest path:(7) where denotes the shortest path from state to state.The shortest path on a graph can be interpreted as a discrete approximation to the geodesic distance on a non-linear manifold[6].For this reason,we call Eq.(7)a geodesic Gaussian kernel(GGK).Shortest paths on graphs can be efficiently computed using the Dijkstra algorithm[12].With its naive implementation, computational complexity for computing the shortest paths from a single node to all other nodes is,where is the number of nodes.If the Fibonacci heap is employed,the computational complexity can be reduced to[13],where is the number of edges.Since the graph in value function approximation problems is typically sparse (i.e.,),using the Fibonacci heap provides signifi-cant computational gains.Furthermore,there exist various approximation algorithms which are computationally very efficient(see[15]and and references therein). Analogous to OGKs,we need to extend GGKs to the state-action space for using them in the LSPI method.A naive way is to just employ Eq.(6),but this can cause a‘shift’in the Gaussian centers since the state usually changes when some action is taken.To incorporate this transition,we propose defining the basis functions as the expectation of Gaussian functions after transition,i.e.,(8) This shifting scheme is expected to work well when the transition is predominantly deterministic(see Sec.IV and Sec.V-A for experimental evaluation).D.Extension to Continuous State SpacesSo far,we focused on discrete state spaces.However,the concept of GGKs can be naturally extended to continuous state spaces,which is explained here.First,the continuous state space is discretized,which gives a graph as a discrete approximation to the non-linear manifold structure of the continuous state space.Based on the graph,we construct GGKs in the same way as the discrete case.Finally,the discrete GGKs are interpolated,e.g.,using a linear method to give continuous GGKs.Although this procedure discretizes the continuous state space,it must be noted that the discretization is only for the purpose of obtaining the graph as a discrete approximation of the continuous non-linear manifold;the resulting basis func-tions themselves are continuously interpolated and hence,the state space is still treated as continuous as opposed to other conventional discretization procedures.→→→→↓→→→→↑→→→→→→→↓→↓→↑→→↓→→↑↓↓↓↓→↑→→↓→↑↑↓↓↓↓→↑→→↓→→→→→→→→→→→→↓→→→↓→→↑→→→→→→→→→→→↑↑↑→→→→→→↓→→→→↓→→→→↓→↑→→↓↓↓↓→↓→↑↑→→→→→↓↓→↓→→→→→→↓↓↓↓↓↓→→→→↑→↓↓↓↓↓↓→→→↑→→→→→↑↑↑→→↑→↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑↑12345678910111213141516171819201234567891011121314(a)Sutton’s maze→→↓→→↓→↓→→→→↑→→→→↑→→→→→↓→→→→→→→→→→→↑→→→→→↓→→→→→↑→→→↑↑↑↓→→→→↓→↓→→→↑→↑↑↑↑↑↓↓↓↓↓↓↓↓→→↑↑↑↑↑↑↑↑↓→→↓↓↓↓↓↓↓↓→→→→→↑↑↑↑↑↓↓→→→↓↓↓→→↑→→→→↑↑↑↓↓↓↓↓↓↓↓↓→→→→→↑↑↑↑↓↓↓↓↓→→→→→→→→↑↑↑↑↑↓↓↓↓↓↓↓→→→↑↑↑↑↑↑↑↑↓↓→→→↓↓→→→→↓↓→→→→→→↓→↓→→→→→→↓→↓↓→→↓→→↓→↓↓↓→→→↓↓↓↓↓↓↓↓→→→→↓↓↓↓↓↓→→→→↓↓↓↓↓↓↓→↓↓↓↓↓↓↓↓↓↓↓↓→→↓→↓→↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓12345678910111213141516171819201234567891011121314151617181920(b)Three-room maze(a)Sutton’s maze(b)Three-room mazeFig.3.Mean squared error of approximated value functions averaged overtrials for the Sutton and three room mazes.In the legend,the standard deviation of GGKs and OGKs is denoted in the bracket.(b)Three-room mazeFig.4.Fraction of optimal states averaged over trials for the Sutton and three room mazes.IV.E XPERIMENTAL C OMPARISONIn this section,we report the results of extensive and systematic experiments for illustrating the difference between GGKs and other basis functions.We employ two standard grid world problems illustrated in Fig.2,and evaluate the goodness of approximated value functions by computing the mean squared error (MSE)with respect to the optimal value function and the goodness of obtained policies by calculating the fraction of states from which the agent can get to the goal optimally (i.e.,in the shortest number of steps).series of random walk of length are gathered as training samples,which are used for estimating the graph as well as the transition probability and expected reward.We set the edge weights in the graph to (which is equivalent to the Euclidean distance between two nodes).We test GGKs,OGKs,graph-Laplacian eigenfunctions (GLEs)[5],and diffusion wavelets (DWs)[8].This simulation is repeated times for each maze and each method,randomly changing training samples in each run.The mean of the above scores as a function of the number of bases is plotted in Fig.4.Note that the actual number of bases is four times more becauseof the extension of basis functions over the action space (see Eq.(6)and Eq.(8)).GGKs and OGKs are tested with small/medium/large Gaussian widths.Fig.3depicts MSEs of the approximated value functions for each method.They show that MSEs of GGKs with width ,OGKs with width ,GLEs,and DWs are very small and decrease as the number of kernels increases.On the other hand,MSEs of GGKs and OGKs with medium/large width are large and increase as the number of kernels increases.Therefore,from the viewpoint of approximation quality of the value functions,the width of GGKs and OGKs should be small.Fig.4depicts the fraction of optimal states in the obtained policy.They show that overall GGKs with medium/large width give much better policies than OGKs,GLEs,and DWs.An interesting finding from the graphs is that GGKs tend to work better if the Gaussian width is large,while OGKs show the opposite trend;this may be explained as follows.Tails of OGKs extend across the wall.Therefore,OGKs with large width tend to produce undesired value function and erroneous policies around the partitions.This tail effect can be alleviated if the Gaussian width is made small.However,this in turn makes the approximated value function fluctuating;so the resulting policies are still erroneous.The fluctuation problem with a small Gaussian width seems to be improved if the number of bases is increased,while the tail effect with a large Gaussian width still remains even when the number of bases is increased.On the other hand,GGKs do not suffer from the tail problem thanks to the geodesic construction.Therefore,GGKs allows us to make the width large without being affected by the discontinuity across the wall.Consequently,smooth value functions along the maze are produced and hence better policies can be obtained by GGKs with large widths.This result highlights a helpful property since it alleviates the practical issue of determining the values of the Gaussian width parameter.V.A PPL ICATIONSAs discussed in the previous section,the proposed GGKs bring a number of preferable properties for making value function approximation effective.In this section,we in-vestigate the application of the GGK-based method to the challenging problems of a(simulated)robot arm control and mobile robot navigation and demonstrate its usefulness. A.Robot Arm ControlWe use a simulator of a two-joint robot arm(moving in a plane)illustrated in Fig.5(a).The task is to lead the end effector(‘hand’)of the arm to an object while avoiding the obstacles.Possible actions are to increase or decrease the angle of each joint(‘shoulder’and‘elbow’)by degrees in the plane,simulating coarse stepper motor joints.Thus the state space is the-dimensional discrete space consisting of two joint angles as illustrated in Fig.5(b).The black area in the middle corresponds to the obstacle in the joint angle state space.The action space involves actions:increase or decrease one of the joint angles.We give a positive immediate reward when the robot’s end effector touches the object;otherwise the robot receives no immediate reward. Note that actions which make the arm collide with obstacles are not allowed.The discount factor is set to. In this environment,we can change the joint angle exactly by degrees,so the environment is deterministic.However, because of the obstacles,it is difficult to explicitly compute an inverse kinematic model;furthermore,the obstacles intro-duce discontinuity in value functions.Therefore,this robot arm control task is an interesting test bed for investigating the behaviour of GGKs.We collected training samples from series of random arm movements,where the start state is chosen randomly in each trial.The graph induced by the above MDP consists of nodes and we assigned uniform weights to the edges.There are totally goal states in this environment (see Fig.5(b)),so we put thefirst Gaussian centers at the goals and the remaining centers are chosen randomly in the state space.For GGKs,kernel functions are extended over the action space using the shifting scheme(see Eq.(8))since the transition is deterministic in this experiment.Fig.6illustrates the value functions approximated using GGKs and OGKs3.The graphs show that GGKs give a nice smooth surface with obstacle-induced discontinuity sharply preserved,while OGKs tend to smooth out the discontinuity. This makes a significant difference in avoiding the obstacle: from‘A’to‘B’in Fig.5(b),the GGK-based value function results in a trajectory that avoids the obstacle(see Fig.6(a)). On the other hand,the OGK-based value function yields a trajectory that tries to move the arm through the obstacle by following the gradient upward(see Fig.6(b)).The latter causes the arm to get stuck behind the obstacle.Fig.7summarizes the performance of GGKs and OGKs measured by the percentage of successful movements(i.e., the end effector reaches the target)averaged over indepen-dent runs.More precisely,in each run,totally training samples are collected using a different random seed,a policy is then computed by the GGK-or OGK-based method using LSPI,and the obtained policy is tested.This graph shows that GGKs remarkably outperform OGKs since the arm can successfully avoid the obstacle.The performance of OGK does not go beyond even when the number of kernels is increased.This is caused by the‘tail effect’of ordinary Gaussian functions;the OGK-based policy can not lead the end effector to the object if it starts from the bottom-left half of the state spaceWhen the number of kernels is increased,the performance of both GGKs and OGKs once gets worse at around.This would be caused by our kernel center allocation strategy:thefirst kernels are put at the goal states and the remaining kernel centers are chosen randomly.When is less than or equal to,the approximated value function tends to have a unimodal profile since all kernels are put at the goal states.However,when is larger than,this unimodality is broken and the surface of the approximated value function gets slightlyfluctuated.This smallfluctuation can cause an error in policies and therefore the performance is degraded at around.This performance degradation tends to be improved as the number of kernels is further increased.Overall,the above result shows that when GGKs are combined with our kernel center allocation strategy,almost perfect policies can be obtained with a very small number of kernels.Therefore,the proposed method is computationally very advantageous.B.Robot Agent NavigationThe above simple robot arm control simulation shows that the GGK method is promising.Here we apply GGKs to a more challenging task of a mobile robot navigation,which involves a high-dimensional and continuous state space. We employ a Khepera robot illustrated in Fig.8(a)on a navigation task.A Khepera is equipped with infra-red 3For illustration purposes,let us display the state value function ,which is the expected long-term discounted sum of rewards the agent receives when the agent takes actions following policy from state. From the definition,it can be confirmed that is expressed.(a)Aschematic(b)State space(a)Geodesic Gaussian kernels(b)Ordinary Gaussian kernelsFig.6.Approximated value functions.Fig.7.Number of successful trials.sensors (‘s1’to ‘s8’in the figure)which measure the strength of the reflected light returned from surrounding obstacles.Each sensor produces a scalar value between and(which may be regarded as continuous):the sensor obtains the maximum value if an obstacle is just in front of the sensor and the value decreases as the obstacle gets farther till it reaches the minimum value .Therefore,the state space is -dimensional and continuous.The Khepera has two wheels and takes the following defined actions:forward,left-rotation,right-rotation and backward (i.e.,the action space contains actions).The speed of the left and right wheels for each action is described in Fig.8(a)in the bracket (the unit is pulse per 10milliseconds).Note that the sensor values and the wheel speed are highly stochastic due to the change of the ambient light,noise,the skid etc.Furthermore,perceptual aliasing occurs due to the limited range and resolution of sensors.Therefore,the state transition is highly stochastic.We set the discount factor to .The goal of the navigation task is to make the Khepera explore the environment as much as possible.To this end,we give a positive reward when the Khepera moves forward and a negative reward when the Khepera collides with an obstacle.We do not give any reward to the left/right rotation and backward actions.This reward design encourages the Khepera to go forward without hitting obstacles,through which extensive exploration in the environment could be achieved.We collected training samples fromseries of random movements in a fixed environment with several ob-stacles (see Fig.9(a)).Then we constructed a graph from the gathered samples by discretizing the continuous state space using the Self-Organizing Map (SOM)[16].The number of nodes (states)in the graph is set to (equivalent with the SOM map size of );this value is computed by the standard rule-of-thumb formula [17],where is the number of samples.The connectivity of the graph is determined by the state transition probability computed from the samples,i.e.,if there is a state transition from one node to another in the samples,an edge is established between these two nodes and the edge weight is set according to the Euclidean distance between them.Fig.8(b)illustrates an example of the obtained graph structure—for visualization purposes,we projected the -dimensional state space onto a -dimensional subspace spanned by(9)The -th element in the above bases corresponds to the output of the -th sensor (see Fig.8(a)).Therefore,the projection onto this subspace roughly means that the horizontal axis corresponds to the distance to the left/right obstacle,while the vertical axis corresponds to the distance to the front/back obstacle.For clear visibility,we only displayed the edges whose weight is less than .This graph has a notable feature:the nodes around the region ‘B’in the figure are(a)A schematic(b)State space projected onto a -dimensional subspace for visualization.Fig.8.Khepera robot.(a)Training(b)TestFig.9.Simulation environment(a)Geodesic Gaussian kernels(b)Ordinary Gaussian kernels Fig.10.Examples of obtained policy .Fig.11.Average amount of exploration.putation time.directly connected to the nodes at ‘A ’,but are not directly connected to the nodes at ‘C’,‘D’,and ‘E’.This implies that the geodesic distance from ‘B’to ‘C’,‘D’,or ‘E’is large,although the Euclidean distance is small.Since the transition from one state to another is highly stochastic in the current experiment,we decided to simply duplicate the GGK function over the action space (see Eq.(6)).For obtaining continuous GGKs,GGK functions need to be interpolated (see Sec.III-D).We may employ a simple linear interpolation method in general.However,the current experiment has unique characteristics—at least one of the sensor values is always zero since the Khepera is never completely surrounded by obstacles.Therefore,samples are always on the surface of the -dimensional hypercube-shaped state space.On the other hand,the node centers determined by the SOM are not generally on the surface.This means thatany sample is not included in the convex hull of its nearest nodes and we need to extrapolate the function value.Here,we simply add the Euclidean distance between the sample and its nearest node when computing kernel values;more precisely,for a state that is not generally located on a node center,the GGK-based basis function is defined as(10)where is the node closest to in the Euclidean distance.Fig.10illustrates an example of actions selected at each node by the GGK-based and OGK-based policies.We usedkernels and set the width to .The symbols ‘’,’’,‘’,and ‘’in the figure indicates forward,backward,left rotation,and right rotation actions.This shows that there is。

SUBJECTIVE IMAGE QUALITY TRADEOFFS BETWEEN SPATIAL RESOLUTION AND QUANTIZATION NOISE

SUBJECTIVE IMAGE QUALITY TRADEOFFS BETWEEN SPATIAL RESOLUTION AND QUANTIZATION NOISE

SUBJECTIVE IMAGE QUALITY TRADEOFFS BETWEEN SPATIAL RESOLUTION ANDQUANTIZATION NOISESoo Hyun Bae,Thrasyvoulos N.Pappas†,Biing-Hwang Juang Center for Signal and Image Processing,Georgia Institute of Technology,Atlanta,GA30332†EECS Department,Northwestern University,Evanston,IL60208{soohyun,juang}@,pappas@ABSTRACTThe importance of tradeoffs between spatial resolution and quan-tization noise has been examined in our previous work.Subjec-tive experiments indicate that as the bitrate decreases,human ob-servers generally prefer to reduce image resolution in order to maintain image quality,but the amount of distortion they are will-ing to accept increases with decreasing resolution.In this paper, we conducted further experiments with several images,different encoders,and afiner set of bitrates to determine the preferred res-olution at each bitrate,and also the resolution at which there are no visible coding artifacts.Analysis of the subjective results using a wavelet-based perceptual quality metric verifies our earlier con-clusion that human observers tend to reduce resolution in order to maintain image quality,but are willing to accept more artifacts as image size decreases.1.INTRODUCTIONThe proliferation of display and capture devices with varying char-acteristics and spatiotemporal resolution necessitates a scalable ap-proach to image/video communication.The spatiotemporal reso-lution of the signal should depend on the transmission bandwidth and display device of each user,and should be determined with the help of an objective measure of image quality that takes into account the visibility of both the compression artifacts and the im-age/video signal.To gain an understanding of the tradeoffs be-tween spatial resolution and quantization noise,we conducted sub-jective experiments[1].We found that as the bitrate decreases, human observers generally prefer to reduce image resolution in order to maintain image quality,but the amount of distortion they are willing to accept increases with decreasing resolution.In this paper,we conducted further experiments with several images,dif-ferent encoders,and afiner set of bitrates to obtain more precise results across image resolutions and bitrates.Four test images are employed in these experiments,two of which contain complicated image details.For the JPEG encoding,we introduced the per-ceptually tuned visibility threshold for the discrete cosine trans-form(DCT)at six image heights,which was proposed by Wat-son[2].Also,we carefully determined a wider andfiner set of bitrates for image coders so as to avoid remarkable difference of perceived quality between each bitrate.In our previous work,the absolute perceived quality assessment was designed to obtain both numeric expression of the subjective image quality and the percep-tually transparent noise level,then the highest level of subjective quality was compared to the most preferable spatial resolution.In this work,we designed a new experiment named the critical noise perception assessment.Throughout the experiments,the subjects were asked to distinguish the original image from the coded one at various resolutions and bitrates for the four images.We pre-cisely drew the critical bitrates at which the human eye cannot or can recognize compression artifacts,then compared it to result of the relative perceived quality assessment.For obvious under-standing of the tendency of the most preferable resolutions,we selected an image quality metric different from the previous one, the wavelet-based metric by Watson et al.[3],which gives a nu-meric expression of image quality computed at the same spatial resolution.Several viewing conditions were also refined in the ex-perimental environment.Analysis of the subjective results using perceptual quality met-rics verifies our earlier conclusion that human observers tend to re-duce resolution in order to maintain image quality,but are willing to accept more artifacts as image size decreases.We are in progress of development of image quality metric incorporating both signal visibility and noise visibility.2.IMAGE QUALITY METRICSMedia signal processing inevitably involves distortion on the sig-nal.A measure that provides an evaluation of the incurred distor-tionfinds many applications in compression,transmission,and en-hancement.The measure is conventionally termed a quality met-ric or a quality measure and can be formulated within two extreme perspectives.On the one end of the spectrum is the subjective measure,in which the evaluation is accomplished through a pro-cess that reflects the human assessment.On the other end of the spectrum is the objective measure(s),which are customarily de-fined on the mean squared error between corresponding signals. In between,there are a number of hybrids that attempt to estab-lish a measure,which can be computed from the signal directly and yet draw a very close approximation of the subjective result without any cumbersome procedure in administering the human assessment process.It is important to note that these conventional measures are de-signed to quantify the error sensitivity between the original signal and the distorted one,while keeping most of the signal characteris-tics intact.For example,in image processing,conventional quality metrics are mostly defined over a squared difference between cor-responding pixel values;the sampling rate of the image remains the same.Since the human visual system(HVS)involves per-ception along several dimensions(i.e.visual area,viewing angle, viewing distance,etc.),a new class of quality measures should in-volve the various perceptions as well in order to incorporate these(a)Critical noise perception assessment(b)Relative perceived quality assessmentFig.1.Test images presented to observer at each test.Bank coded by JPEG at0.2bits/pixel(a),Lena by JPEG2000at0.1 bits/pixel(b)additional factors.To obtain a better understanding of the displaying and viewing parameters with an ultimate goal of designing image quality met-rics for scalable image coding applications,we conducted various subjective experiments upon tradeoffs between compression arti-facts and spatial resolution.First,a series of compressed images at different bitrates,which are carefully chosen for covering wider perceptual quality then the previous results[1],are generated and then downsampled by optimal sinc-function.The subject tests are designed along two aspects:critical noise perception assessment and relative perceived quality assessment.Thefirst aspect is mea-surement of the critical compression noise level at which human cannot or barely recognize compression artifact.The second one is measurement of the most preferable resolution.3.SUBJECTIVE TEST SETUPOne of psychophysical experiments for analyzing the effect of spa-tial resolution in image quality assessment is[4],which formed a basis of modern image quality analysis.However,the specific tradeoffs we examine in this work were not addressed in the paper. Exploring tradeoffs between spatial resolution and image com-pression artifacts to obtain the perceptually optimal compression conditions at a given coding algorithm and a bitrate is our goal. As we discussed above,two subjective tests were designed as fol-lows:The critical noise perception assessment aims at pointing the noise transparent bitrate at every image and its spatial resolu-tion.We showed an observer two images,the original image and the decoded one,then asked to differentiate the original one.A combination of two images are randomly ordered when displayed as shown in Figure1(a).Basically if the answer is correct,it indicates that noise at a combination of given image,coding bi-trate,and spatial resolution is visible to humaneyes.Otherwise,it(a)Bank(b)Lena(c)Bike(d)WomanFig.2.Images for subjective tests.Image Bitrates(unit:bits/pixel)1.0,0.8,0.7,0.6,0.5,0.45,0.4,0.3,0.25,0.2JPEG20001.0,0.8,0.6,0.5,0.4,0.3,0.27,0.23JPEG20001.0,0.5,0.4,0.35,0.3,0.27,0.25,0.23,0.2JPEG20001.0,0.6,0.5,0.4,0.35,0.3,0.25,0.2,0.15,0.14JPEG2000B i t r a t eFig.3.Analysis of the result of the critical noise perception assessment.The solid lines are corresponding to the medians of all votes.allowed enough time to make their decisions and to view the orig-inal test images before and during the test.The ordering of images and coders are randomized to avoid any biases,but the bitrates in the relative perceived quality assessment decreases at every test.Again,the bitrates in the critical noise perception assessment be-gin at the mid-level bitrates and remaining bitrates are actively de-termined according to the observers’answers in order to abridge a number of unnecessary measurements.Quantitative expressions of image quality are computed be-tween the originals and the decoded images at the same resolution by the Wavelet-based metric developed by Watson et al.[3].The linear-phase 9/7biorthogonal filters are used for signal decompo-sition,then the baseline sensitivity thresholds,τi,k ,for the wavelet decomposition were measured.Here k denotes the subband index and i the coefficient location in the subband image.The overall image distortion “Perceptual Masked Error (PME)”is then com-puted with D p =1τi,kQ 1D 2p.(2)A detailed description of this metric can be found in [5].4.EXPERIMENTAL RESULTSFor the analysis of the result of the critical noise perception as-sessment,the median of all votes were employed.The results of images,bank and woman,are given in Figure 3.At a given reso-lution,images coded over the bitrates above each median value are perceptually noise-transparent.Obviously,the images gen-erated by JPEG2000have lower critical bitrates than images by JPEG.Note that the bank image contains complicated details and the woman image has comparatively less details.The bank image thus has higher critical bitrates than woman image does.There-fore,in the woman coded by JPEG 2000,the images with resolu-tions lower than 128×128are nearly noise transparent even at 0.1bpp.For comparison of the result to an objective image quality,the critical bitrates are drawn (with shaded cells)in Table 2.Similar to the results in the subjective test,the objective quality decreasesas bitrate decreases or spatial resolution increases;conversely,itincreases as bitrate increases or spatial resolution decreases.In particular,at low resolution,the bitrate,which is closely related to quantization level in the encoder,does not seriously affect the subjective and objective quality.It explains two important find-ings that people are able to accept more distortion for low resolu-tion images and bitrate needs to be determined not only by a target quality but by other parameters,e.g.spatial resolution and viewing distance.To analyze the results of the relative perceived quality assess-ment,the median values of the most preferable resolution are ob-tained.Figure 4shows all the votes of observers and the median values for the image bike.On top of the human eye’s basic prefer-ence to a higher resolution and less distorted image,Figure 4also depicts their tradeoffs substantially.Comparison of the two sets of tests leads us to another result that people tends to maintain perceptual quality at every spatial resolution as presented in bold numbers in Table 2.A tendency to maintain quality around 59.00dB over various bitrates can be found in both (a)and (b)of Table parison of the noise-transparent level and the most preferable resolution at a fixed reso-lution implies that people are willing to accept more distortion.For example,at 128×128of Bank JPEG2000,the bitrates can be low-ered from 0.5bpp down to 0.18bpp.Therefore,the level differ-ence at a fixed resolution,i.e.the resolution at a noise transparent bitrate and most preferable resolutions,is equivalent to perceptual tolerance over noise-transparent condition.In this example,3.69dB of noise can be more added over the noise-transparent condi-tion without sacrifice of perceptual quality.5.CONCLUSIONNoise visibility of the compressed image over various spatial res-olutions and bitrates in various types of images is studied for a framework of image quality metric.Since most of the image qual-ity metric incorporate just the visibility of noise,not the visibility of signal itself,analysis of tradeoff between the spatial resolution and the quantization noise is highly necessary in the scalable im-age compression application.We designed subjective tests along two aspects,the critical noise perception and the relative perceived quality to explore such tradeoffs.A series of compressed images at different bitrates,which are carefully determined for covering wide perceptual qual-ity,is generated.Lower resolution images were then produced by optimal sinc-function upsampling and downsampling in integer ratios.The critical noise perception assessment is to obtain noise transparent bitrate at every image and spatial resolution.The rel-(a)Bank JPEG2000Resolution/bpp0.750.50.20.150.160.9759.8858.2255.5354.9054.56384x38459.1156.8955.3154.5753.1363.7960.7557.8556.4955.67192x19263.4659.3157.5556.3854.2665.7264.0863.0060.2258.8057.5296x9666.1063.6761.5060.2058.2055.7467.9467.3462.6962.0759.8310.40.30.250.2512x51257.9456.7456.0155.0463.3559.8558.2356.9156.15256x25662.5661.2259.5458.4757.0267.1161.8961.0959.2758.20128x12864.9261.5760.2858.5868.8463.9861.0659.8364x6466.9564.9763.8462.2960.32Table2.MPSNR values over different coding rates and spatial resolutions.The critical bitrates are in shaded cells,and the most preferable resolutions are specified in bold numbers.(unit:decibels)ative perceived quality assessment is to gain the most preferable resolution of each image with consideration of distortion artifacts and image size.Conducting the subjective tests with observers,we have found how image characteristic affects the critical bitrates, and the most preferable resolution at each noise level of every im-age.By the comparison of two different levels,a tendency that hu-man eyes try to maintain the subjective quality as image size de-creases is observed.We have also gained that when various view-ing parameters are considered,there exists perceptual noise toler-ance so that observers are willing to accept more artifacts as image size decreases.Estimation of the two results obtained by the sub-jective tests will provide a more important framework for image quality metric yielding objective perceptual quality over spatial resolution and quantization noise,which will be essential to the scalable image compression.We are in progress of developing the image quality metric incorporating signal visibility as well as noise visibility.6.REFERENCES[1]S.H.Bae,T.N.Pappas,and B.-H.Juang,“Spatial resolutionand quantization noise tradeoffs for scalable image compres-sion,”in Proc.IEEE Int.Conf.Acoustics,Speech,and Signal Processing,May2006,vol.1,pp.945–948.[2] A.B.Watson,“DCT quantization matrices visually optimizedfor individual images,”in Proc.SPIE,Human Vision,Visual Proc.,and Digital Display IV,Jan P.Allebach and Bernice E.Rogowitz,Eds.,1993,vol.1913.[3] A.B.Watson,G.Y.Yang,J.A.Solomon,and J.Villasenor,“Visibility of wavelet quantization noise,”IEEE Trans.on Image Processing,vol.6,no.8,pp.1164–1175,Aug.1997.[4]J.H.D.M.Westerink and J.A.J.Roufs,“Subjective imagequality as a function of viewing distance,resolution,and pic-ture size,”SMPTE Journal,vol.98,pp.113–119,Feb.1989.[5]T.N.Pappas,R.J.Safranek,and J.Chen,“Perceptual crite-ria for image quality evaluation,”in Handbook of Image and Video Processing,Alan C.Bovik,Ed.,pp.939–959.Academic Press,second edition,2005.Fig.4.Tradeoffs between spatial resolution and compression artifacts. Image is bike.。

模拟ai英文面试题目及答案

模拟ai英文面试题目及答案

模拟ai英文面试题目及答案模拟AI英文面试题目及答案1. 题目: What is the difference between a neural network anda deep learning model?答案: A neural network is a set of algorithms modeled loosely after the human brain that are designed to recognize patterns. A deep learning model is a neural network with multiple layers, allowing it to learn more complex patterns and features from data.2. 题目: Explain the concept of 'overfitting' in machine learning.答案: Overfitting occurs when a machine learning model learns the training data too well, including its noise and outliers, resulting in poor generalization to new, unseen data.3. 题目: What is the role of a 'bias' in an AI model?答案: Bias in an AI model refers to the systematic errors introduced by the model during the learning process. It can be due to the choice of model, the training data, or the algorithm's assumptions, and it can lead to unfair or inaccurate predictions.4. 题目: Describe the importance of data preprocessing in AI.答案: Data preprocessing is crucial in AI as it involves cleaning, transforming, and reducing the data to a suitableformat for the model to learn effectively. Proper preprocessing can significantly improve the performance of AI models by ensuring that the input data is relevant, accurate, and free from noise.5. 题目: How does reinforcement learning differ from supervised learning?答案: Reinforcement learning is a type of machine learning where an agent learns to make decisions by performing actions in an environment to maximize a reward signal. It differs from supervised learning, where the model learns from labeled data to predict outcomes based on input features.6. 题目: What is the purpose of a 'convolutional neural network' (CNN)?答案: A convolutional neural network (CNN) is a type of deep learning model that is particularly effective for processing data with a grid-like topology, such as images. CNNs use convolutional layers to automatically and adaptively learn spatial hierarchies of features from input images.7. 题目: Explain the concept of 'feature extraction' in AI.答案: Feature extraction in AI is the process of identifying and extracting relevant pieces of information from the raw data. It is a crucial step in many machine learning algorithms, as it helps to reduce the dimensionality of the data and to focus on the most informative aspects that can be used to make predictions or classifications.8. 题目: What is the significance of 'gradient descent' in training AI models?答案: Gradient descent is an optimization algorithm used to minimize a function by iteratively moving in the direction of steepest descent as defined by the negative of the gradient. In the context of AI, it is used to minimize the loss function of a model, thus refining the model's parameters to improve its accuracy.9. 题目: How does 'transfer learning' work in AI?答案: Transfer learning is a technique where a pre-trained model is used as the starting point for learning a new task. It leverages the knowledge gained from one problem to improve performance on a different but related problem, reducing the need for large amounts of labeled data and computational resources.10. 题目: What is the role of 'regularization' in preventing overfitting?答案: Regularization is a technique used to prevent overfitting by adding a penalty term to the loss function, which discourages overly complex models. It helps to control the model's capacity, forcing it to generalize better to new data by not fitting too closely to the training data.。

一般算子系统的张量积

一般算子系统的张量积

中图分类号:UDC:学校代码:10055密级:公开高恐大法博士学位论文一般算子系统的张量积TensorProductsofNon-unitalOperatorSystems评阅人生尚全:塾垫型!直国送整壹涸:篚丛笪南开大学研究生院二。

一三年五月Chapter2OperatorspaceandoperatorsystemChapter2Operatorspaceandoperator2.1Preliminariessystem_LetVbeacomplexvectorspace.WecallVa木-vectorspace.ifthereisamap爿c:V-÷Vsuchthat(z。

)+=zand(Ax+可)+=Ax‘+Y+forallz,可∈VandA∈C.Thesetoftheelementsz∈Vsatisfyingz+=zisarealvectorspace.calledthesetofhermitianelements,denotedbyK口.Thenv0∈Koiscalledaconeifz+可∈v0and7’z∈v0foranyr>0,z,Y∈v0.Supposethatn,m∈N.Wedenoteby^厶,m(y)t11evectorspaceofn×mmatriceswithelementsinV.Forsimplicity,wedenote螈,m:=‰m(C),螈(y):=%,几(y)and螈:=%∽Then螈(y)isalsoa爿c-vectorspacewiththecanonical爿c-operationdefinedby(ai,J)b:=(n釉t,J.Let乱∈^霸(y)andu∈Mm(y).Theirdirectsumistheelementin^磊+m(y)definedbyYeW:=降Z,:]∈%删一Moreover,ifQ∈Mn,mandp∈‰,n,wedefinetheirmatrixproductc℃vfi∈A靠(y)byQ"p:=[∑¨Q啪Vk,lp幻kLetV,WandXbevectorspaces.WedenotebyVOWtheftalgebraictensorproduct,whichislinearlyspannedbytheelementszOYwhereX∈VandY∈W.If圣:V×W--+Xisbilinear,thereisauniquelinearmappingePL:VQW斗Xsuchthatffz(zoY)=圣(z,Y)foranyz∈VandY∈W.4Thecompletelyboundednormofabilinearmappingisdefinedtobe1圣1c6:=sup{l[圣p;rp,7.∈卧Thenthisbilinearmappingiscalledcompletelyboundedif11垂11cb<∞.WedenotebyCB(V×彬x)thesetofsuchcompletelyboundedmappings.Indeed.CB(VX彬X)isanoperatorspaceviathecanonicalidentiffcationMn(CB(V×彬x))=CB(V×彬螈(x))Proposition2.13lf33.Proposition7.1.21)LetV.WandXbeoperatorspaces.ThenwehavethecompletelyisometricidentificationsCB(V圆W,X)=CB(V×彬x)=CB(V,CB(彬x))’rheorem2.14f[33,Proposition7.1.71)LetⅥ,%,胍and%beoperatorspaces.-Supposethat妒:Ⅵj%and妒:肌_÷w2arecompletequotientmappings.Thenthecorrespondingmapping妒@砂:Ⅵo阢一%ow2extendstoacompletequotientmapping妒园砂:M圆M_÷%4w2.2.4OperatorsystemWecalla木。

尺度上推 像元聚合方法

尺度上推 像元聚合方法

尺度上推像元聚合方法英文回答:Scaling up and pixel aggregation methods are commonly used techniques in image processing and computer vision. These methods aim to enhance the resolution and quality of images by combining multiple low-resolution images into a single high-resolution image.Scaling up refers to the process of increasing the size of an image while maintaining its aspect ratio. This is often done by interpolating the pixels of the original image to fill in the gaps in the enlarged image. The most commonly used scaling up method is bilinear interpolation, which calculates the values of the new pixels based on the average of the surrounding pixels. This method can produce smooth and visually pleasing results, but it may also introduce blurring and loss of details.On the other hand, pixel aggregation methods involvecombining multiple low-resolution images to create a single high-resolution image. This can be done by aligning the images and averaging the pixel values at each corresponding position. This technique takes advantage of the fact that each low-resolution image captures a slightly different perspective of the scene, which can be used to enhance the overall resolution and reduce noise. Examples of pixel aggregation methods include super-resolution and image stacking.中文回答:尺度上推和像元聚合方法是图像处理和计算机视觉中常用的技术。

On rational approximation of algebraic functions

On rational approximation of algebraic functions
ON RATIONAL APPROXIMATION OF ALGEBRAIC FUNCTIONS
JULIUS BORCEA∗ , RIKARD BØGVAD, AND BORIS SHAPIRO
arXiv:math/0409353v2 [math.CA] 17 Jun 2005
Abstract. We construct a new scheme of approximation of any multivalued algebraic function f (z ) by a sequence {rn (z )}n∈N of rational functions. The latter sequence is generated by a recurrence relation which is completely determined by the algebraic equation satisfied by f (z ). Compared to the usual Pad´ e approximation our scheme has a number of advantages, such as simple computational procedures that allow us to prove natural analogs of the Pad´ e Conjecture and Nuttall’s Conjecture for the sequence {rn (z )}n∈N in the complement CP1 \ Df , where Df is the union of a finite number of segments of real algebraic curves and finitely many isolated points. In particular, our construction makes it possible to control the behavior of spurious poles and to describe the asymptotic ratio distribution of the family {rn (z )}n∈N . As an application we settle the so-called 3-conjecture of Egecioglu et al dealing with a 4-term recursion related to a polynomial Riemann Hypothesis.

Deformable Medical Image Registration

Deformable Medical Image Registration

Deformable Medical Image Registration:A Survey Aristeidis Sotiras*,Member,IEEE,Christos Davatzikos,Senior Member,IEEE,and Nikos Paragios,Fellow,IEEE(Invited Paper)Abstract—Deformable image registration is a fundamental task in medical image processing.Among its most important applica-tions,one may cite:1)multi-modality fusion,where information acquired by different imaging devices or protocols is fused to fa-cilitate diagnosis and treatment planning;2)longitudinal studies, where temporal structural or anatomical changes are investigated; and3)population modeling and statistical atlases used to study normal anatomical variability.In this paper,we attempt to give an overview of deformable registration methods,putting emphasis on the most recent advances in the domain.Additional emphasis has been given to techniques applied to medical images.In order to study image registration methods in depth,their main compo-nents are identified and studied independently.The most recent techniques are presented in a systematic fashion.The contribution of this paper is to provide an extensive account of registration tech-niques in a systematic manner.Index Terms—Bibliographical review,deformable registration, medical image analysis.I.I NTRODUCTIOND EFORMABLE registration[1]–[10]has been,alongwith organ segmentation,one of the main challenges in modern medical image analysis.The process consists of establishing spatial correspondences between different image acquisitions.The term deformable(as opposed to linear or global)is used to denote the fact that the observed signals are associated through a nonlinear dense transformation,or a spatially varying deformation model.In general,registration can be performed on two or more im-ages.In this paper,we focus on registration methods that involve two images.One is usually referred to as the source or moving image,while the other is referred to as the target orfixed image. In this paper,the source image is denoted by,while the targetManuscript received March02,2013;revised May17,2013;accepted May 21,2013.Date of publication May31,2013;date of current version June26, 2013.Asterisk indicates corresponding author.*A.Sotiras is with the Section of Biomedical Image Analysis,Center for Biomedical Image Computing and Analytics,Department of Radi-ology,University of Pennsylvania,Philadelphia,PA19104USA(e-mail: aristieidis.sotiras@).C.Davatzikos is with the Section of Biomedical Image Analysis,Center for Biomedical Image Computing and Analytics,Department of Radi-ology,University of Pennsylvania,Philadelphia,PA19104USA(e-mail: christos.davatzikos@).N.Paragios is with the Center for Visual Computing,Department of Applied Mathematics,Ecole Centrale de Paris,92295Chatenay-Malabry,France,and with the Equipe Galen,INRIA Saclay-Ile-de-France,91893Orsay,France,and also with the Universite Paris-Est,LIGM(UMR CNRS),Center for Visual Com-puting,Ecole des Ponts ParisTech,77455Champs-sur-Marne,France. Digital Object Identifier10.1109/TMI.2013.2265603image is denoted by.The two images are defined in the image domain and are related by a transformation.The goal of registration is to estimate the optimal transforma-tion that optimizes an energy of the form(1) The previous objective function(1)comprises two terms.The first term,,quantifies the level of alignment between a target image and a source image.Throughout this paper,we in-terchangeably refer to this term as matching criterion,(dis)sim-ilarity criterion or distance measure.The optimization problem consists of either maximizing or minimizing the objective func-tion depending on how the matching term is chosen.The images get aligned under the influence of transformation .The transformation is a mapping function of the domain to itself,that maps point locations to other locations.In gen-eral,the transformation is assumed to map homologous loca-tions from the target physiology to the source physiology.The transformation at every position is given as the addition of an identity transformation with the displacementfield,or.The second term,,regularizes the trans-formation aiming to favor any specific properties in the solution that the user requires,and seeks to tackle the difficulty associ-ated with the ill-posedness of the problem.Regularization and deformation models are closely related. Two main aspects of this relation may be distinguished.First, in the case that the transformation is parametrized by a small number of variables and is inherently smooth,regularization may serve to introduce prior knowledge regarding the solution that we seek by imposing task-specific constraints on the trans-formation.Second,in the case that we seek the displacement of every image element(i.e.,nonparametric deformation model), regularization dictates the nature of the transformation. Thus,an image registration algorithm involves three main components:1)a deformation model,2)an objective function, and3)an optimization method.The result of the registration algorithm naturally depends on the deformation model and the objective function.The dependency of the registration result on the optimization strategy follows from the fact that image regis-tration is inherently ill-posed.Devising each component so that the requirements of the registration algorithm are met is a de-manding process.Depending on the deformation model and the input data,the problem may be ill-posed according to Hadamard’s definition of well-posed problems[11].In probably all realistic scenarios, registration is ill-posed.To further elaborate,let us consider some specific cases.In a deformable registration scenario,one0278-0062/$31.00©2013IEEEseeks to estimate a vector for every position given,in general, scalar information conveyed by image intensity.In this case,the number of unknowns is greater than the number of constraints. In a rigid setting,let us consider a consider a scenario where two images of a disk(white background,gray foreground)are registered.Despite the fact that the number of parameters is only 6,the problem is ill-posed.The problem has no unique solution since a translation that aligns the centers of the disks followed by any rotation results in a meaningful solution.Given nonlinear and nonconvex objective functions,in gen-eral,no closed-form solutions exist to estimate the registration parameters.In this setting,the search methods reach only a local minimum in the parameter space.Moreover,the problem itself has an enormous number of different facets.The approach that one should take depends on the anatomical properties of the organ(for example,the heart and liver do not adhere to the same degree of deformation),the nature of observations to be regis-tered(same modality versus multi-modal fusion),the clinical setting in which registration is to be used(e.g.,offline interpre-tation versus computer assisted surgery).An enormous amount of research has been dedicated to de-formable registration towards tackling these challenges due to its potential clinical impact.During the past few decades,many innovative ideas regarding the three main algorithmic registra-tion aspects have been proposed.General reviews of thefield may be found in[1]–[7],[9].However due to the rapid progress of thefield such reviews are to a certain extent outdated.The aim of this paper is to provide a thorough overview of the advances of the past decade in deformable registration.Never-theless,some classic papers that have greatly advanced the ideas in thefield are mentioned.Even though our primary interest is deformable registration,for the completeness of the presenta-tion,references to linear methods are included as many prob-lems have been treated in this low-degree-of-freedom setting before being extended to the deformable case.The main scope of this paper is focused on applications that seek to establish spatial correspondences between medical im-ages.Nonetheless,we have extended the scope to cover appli-cations where the interest is to recover the apparent motion of objects between sequences of successive images(opticalflow estimation)[12],[13].Deformable registration and opticalflow estimation are closely related problems.Both problems aim to establish correspondences between images.In the deformable registration case,spatial correspondences are sought,while in the opticalflow case,spatial correspondences,that are associ-ated with different time points,are looked for.Given data with a good temporal resolution,one may assume that the magnitude of the motion is limited and that image intensity is preserved in time,opticalflow estimation can be regarded as a small defor-mation mono-modal deformable registration problem.The remainder of the paper is organized by loosely following the structural separation of registration algorithms to three com-ponents:1)deformation model,2)matching criteria,and3)op-timization method.In Section II,different approaches regarding the deformation model are presented.Moreover,we also chose to cover in this section the second term of the objective function, the regularization term.This choice was motivated by the close relation between the two parts.In Section III,thefirst term of the objective function,the matching term,is discussed.The opti-mization methods are presented in Section IV.In every section, particular emphasis was put on further deepening the taxonomy of registration method by grouping the presented methods in a systematic manner.Section V concludes the paper.II.D EFORMATION M ODELSThe choice of deformation model is of great importance for the registration process as it entails an important compromise between computational efficiency and richness of description. It also reflects the class of transformations that are desirable or acceptable,and therefore limits the solution to a large ex-tent.The parameters that registration estimates through the op-timization strategy correspond to the degrees of freedom of the deformation model1.Their number varies greatly,from six in the case of global rigid transformations,to millions when non-parametric dense transformations are considered.Increasing the dimensionality of the state space results in enriching the de-scriptive power of the model.This model enrichment may be accompanied by an increase in the model’s complexity which, in turns,results in a more challenging and computationally de-manding inference.Furthermore,the choice of the deformation model implies an assumption regarding the nature of the defor-mation to be recovered.Before continuing,let us clarify an important,from imple-mentation point of view,aspect related to the transformation mapping and the deformation of the source image.In the in-troduction,we stated that the transformation is assumed to map homologous locations from the target physiology to the source physiology(backward mapping).While from a theoretical point of view,the mapping from the source physiology to the target physiology is possible(forward mapping),from an implemen-tation point of view,this mapping is less advantageous.In order to better understand the previous statement,let us consider how the direction of the mapping influences the esti-mation of the deformed image.In both cases,the source image is warped to the target domain through interpolation resulting to a deformed image.When the forward mapping is estimated, every voxel of the source image is pushed forward to its esti-mated position in the deformed image.On the other hand,when the backward mapping is estimated,the pixel value of a voxel in the deformed image is pulled from the source image.The difference between the two schemes is in the difficulty of the interpolation problem that has to be solved.In thefirst case,a scattered data interpolation problem needs to be solved because the voxel locations of the source image are usually mapped to nonvoxel locations,and the intensity values of the voxels of the deformed image have to be calculated.In the second case,when voxel locations of the deformed image are mapped to nonvoxel locations in the source image,their intensities can be easily cal-culated by interpolating the intensity values of the neighboring voxels.The rest of the section is organized by following coarsely and extending the classification of deformation models given 1Variational approaches in general attempt to determine a function,not just a set of parameters.SOTIRAS et al.:DEFORMABLE MEDICAL IMAGE REGISTRATION:A SURVEY1155Fig.1.Classi fication of deformation models.Models that satisfy task-speci fic constraints are not shown as a branch of the tree because they are,in general,used in conjunction with physics-based and interpolation-based models.by Holden [14].More emphasis is put on aspects that were not covered by that review.Geometric transformations can be classi fied into three main categories (see Fig.1):1)those that are inspired by physical models,2)those inspired by interpolation and ap-proximation theory,3)knowledge-based deformation models that opt to introduce speci fic prior information regarding the sought deformation,and 4)models that satisfy a task-speci fic constraint.Of great importance for biomedical applications are the con-straints that may be applied to the transformation such that it exhibits special properties.Such properties include,but are not limited to,inverse consistency,symmetry,topology preserva-tion,diffeomorphism.The value of these properties was made apparent to the research community and were gradually intro-duced as extra constraints.Despite common intuition,the majority of the existing regis-tration algorithms are asymmetric.As a consequence,when in-terchanging the order of input images,the registration algorithm does not estimate the inverse transformation.As a consequence,the statistical analysis that follows registration is biased on the choice of the target domain.Inverse Consistency:Inverse consistent methods aim to tackle this shortcoming by simultaneously estimating both the forward and the backward transformation.The data matching term quanti fies how well the images are aligned when one image is deformed by the forward transformation,and the other image by the backward transformation.Additionally,inverse consistent algorithms constrain the forward and backward transformations to be inverse mappings of one another.This is achieved by introducing terms that penalize the difference between the forward and backward transformations from the respective inverse mappings.Inverse consistent methods can preserve topology but are only asymptotically symmetric.Inverse-consistency can be violated if another term of the objective function is weighted more importantly.Symmetry:Symmetric algorithms also aim to cope with asymmetry.These methods do not explicitly penalize asym-metry,but instead employ one of the following two strategies.In the first case,they employ objective functions that are by construction symmetric to estimate the transformation from one image to another.In the second case,two transformation functions are estimated by optimizing a standard objective function.Each transformation function map an image to a common domain.The final mapping from one image to another is calculated by inverting one transformation function and composing it with the other.Topology Preservation:The transformation that is estimated by registration algorithms is not always one-to-one and cross-ings may appear in the deformation field.Topology preserving/homeomorphic algorithms produce a mapping that is contin-uous,onto,and locally one-to-one and has a continuous inverse.The Jacobian determinant contains information regarding the injectivity of the mapping and is greater than zero for topology preserving mappings.The differentiability of the transformation needs to be ensured in order to calculate the Jacobian determi-nant.Let us note that Jacobian determinant and Jacobian are in-terchangeably used in this paper and should not be confounded with the Jacobian matrix.Diffeomorphism:Diffeomoprhic transformations also pre-serve topology.A transformation function is a diffeomorphism,if it is invertible and both the function and its inverse are differ-entiable.A diffeomorphism maps a differentiable manifold to another.1156IEEE TRANSACTIONS ON MEDICAL IMAGING,VOL.32,NO.7,JULY2013In the following four subsections,the most important methods of the four classes are presented with emphasis on the approaches that endow the model under consideration with the above desirable properties.A.Geometric Transformations Derived From Physical Models Following[5],currently employed physical models can be further separated infive categories(see Fig.1):1)elastic body models,2)viscousfluidflow models,3)diffusion models,4) curvature registration,and5)flows of diffeomorphisms.1)Elastic Body Models:a)Linear Models:In this case,the image under deforma-tion is modeled as an elastic body.The Navier-Cauchy Partial Differential Equation(PDE)describes the deformation,or(2) where is the forcefield that drives the registration based on an image matching criterion,refers to the rigidity that quanti-fies the stiffness of the material and is Lamésfirst coefficient. Broit[15]first proposed to model an image grid as an elastic membrane that is deformed under the influence of two forces that compete until equilibrium is reached.An external force tries to deform the image such that matching is achieved while an internal one enforces the elastic properties of the material. Bajcsy and Kovacic[16]extended this approach in a hierar-chical fashion where the solution of the coarsest scale is up-sam-pled and used to initialize thefiner one.Linear registration was used at the lowest resolution.Gee and Bajscy[17]formulated the elastostatic problem in a variational setting.The problem was solved under the Bayesian paradigm allowing for the computation of the uncertainty of the solution as well as for confidence intervals.Thefinite element method(FEM)was used to infer the displacements for the ele-ment nodes,while an interpolation strategy was employed to es-timate displacements elsewhere.The order of the interpolating or shape functions,determines the smoothness of the obtained result.Linear elastic models have also been used when registering brain images based on sparse correspondences.Davatzikos[18]first used geometric characteristics to establish a mapping be-tween the cortical surfaces.Then,a global transformation was estimated by modeling the images as inhomogeneous elastic ob-jects.Spatially-varying elasticity parameters were used to com-pensate for the fact that certain structures tend to deform more than others.In addition,a nonzero initial strain was considered so that some structures expand or contract naturally.In general,an important drawback of registration is that when source and target volumes are interchanged,the obtained trans-formation is not the inverse of the previous solution.In order to tackle this shortcoming,Christensen and Johnson[19]pro-posed to simultaneously estimate both forward and backward transformations,while penalizing inconsistent transformations by adding a constraint to the objective function.Linear elasticity was used as regularization constraint and Fourier series were used to parametrize the transformation.Leow et al.[20]took a different approach to tackle the incon-sistency problem.Instead of adding a constraint that penalizes the inconsistency error,they proposed a unidirectional approach that couples the forward and backward transformation and pro-vides inverse consistent transformations by construction.The coupling was performed by modeling the backward transforma-tion as the inverse of the forward.This fact was also exploited during the optimization of the symmetric energy by only fol-lowing the gradient direction of the forward mapping.He and Christensen[21]proposed to tackle large deforma-tions in an inverse consistent framework by considering a se-quence of small deformation transformations,each modeled by a linear elastic model.The problem was symmetrized by consid-ering a periodic sequence of images where thefirst(or last)and middle image are the source and target respectively.The sym-metric objective function thus comprised terms that quantify the difference between any two successive pairs of images.The in-ferred incremental transformation maps were concatenated to map one input image to another.b)Nonlinear Models:An important limitation of linear elastic models lies in their inability to cope with large defor-mations.In order to account for large deformations,nonlinear elastic models have been proposed.These models also guar-antee the preservation of topology.Rabbitt et al.[22]modeled the deformable image based on hyperelastic material properties.The solution of the nonlinear equations was achieved by local linearization and the use of the Finite Element method.Pennec et al.[23]dropped the linearity assumption by mod-eling the deformation process through the St Venant-Kirchoff elasticity energy that extends the linear elastic model to the non-linear regime.Moreover,the use of log-Euclidean metrics in-stead of Euclidean ones resulted in a Riemannian elasticity en-ergy which is inverse consistent.Yanovsky et al.[24]proposed a symmetric registration framework based on the St Venant-Kir-choff elasticity.An auxiliary variable was added to decouple the regularization and the matching term.Symmetry was im-posed by assuming that the Jacobian determinants of the defor-mation follow a zero mean,after log-transformation,log-normal distribution[25].Droske and Rumpf[26]used an hyperelastic,polyconvex regularization term that takes into account the length,area and volume deformations.Le Guyader and Vese[27]presented an approach that combines segmentation and registration that is based on nonlinear elasticity.The authors used a polyconvex regularization energy based on the modeling of the images under deformation as Ciarlet-Geymonat materials[28].Burger et al.[29]also used a polyconvex regularization term.The au-thors focused on the numerical implementation of the registra-tion framework.They employed a discretize-then-optimize ap-proach[9]that involved the partitioning voxels to24tetrahedra.2)Viscous Fluid Flow Models:In this case,the image under deformation is modeled as a viscousfluid.The transformation is governed by the Navier-Stokes equation that is simplified by assuming a very low Reynold’s numberflow(3) These models do not assume small deformations,and thus are able to recover large deformations[30].Thefirst term of theSOTIRAS et al.:DEFORMABLE MEDICAL IMAGE REGISTRATION:A SURVEY1157Navier-Stokes equation(3),constrains neighboring points to de-form similarly by spatially smoothing the velocityfield.The velocityfield is related to the displacementfield as.The velocityfield is integrated in order to estimate the displacementfield.The second term al-lows structures to change in mass while and are the vis-cosity coefficients.Christensen et al.[30]modeled the image under deformation as a viscousfluid allowing for large magnitude nonlinear defor-mations.The PDE was solved for small time intervals and the complete solution was given by an integration over time.For each time interval a successive over-relaxation(SOR)scheme was used.To guarantee the preservation of topology,the Jaco-bian was monitored and each time its value fell under0.5,the deformed image was regridded and a new one was generated to estimate a transformation.Thefinal solution was the con-catenation of all successive transformations occurring for each regridding step.In a subsequent work,Christensen et al.[31] presented a hierarchical way to recover the transformations for brain anatomy.Initially,global affine transformation was per-formed followed by a landmark transformation model.The re-sult was refined byfluid transformation preceded by an elastic registration step.An important drawback of the earliest implementations of the viscousfluid models,that employed SOR to solve the equa-tions,was computational inefficiency.To circumvent this short-coming,Christensen et al.employed a massive parallel com-puter implementation in[30].Bro-Nielsen and Gramkow[32] proposed a technique based on a convolutionfilter in scale-space.Thefilter was designed as the impulse response of the linear operator defined in its eigen-function basis.Crun et al.[33]proposed a multi-grid approach towards handling anisotropic data along with a multi-resolution scheme opting forfirst recovering coarse velocity es-timations and refining them in a subsequent step.Cahill et al.[34]showed how to use Fourier methods to efficiently solve the linear PDE system that arises from(3)for any boundary condi-tion.Furthermore,Cahill et al.extended their analysis to show how these methods can be applied in the case of other regu-larizers(diffusion,curvature and elastic)under Dirichlet,Neu-mann,or periodic boundary conditions.Wang and Staib[35]usedfluid deformation models in an atlas-enhanced registration setting while D’Agostino et al. tackled multi-modal registration with the use of such models in[36].More recently,Chiang et al.[37]proposed an inverse consistent variant offluid registration to register Diffusion Tensor images.Symmetrized Kullback-Leibler(KL)diver-gence was used as the matching criterion.Inverse consistency was achieved by evaluating the matching and regularization criteria towards both directions.3)Diffusion Models:In this case,the deformation is mod-eled by the diffusion equation(4) Let us note that most of the algorithms,based on this transforma-tion model and described in this section,do not explicitly state the(4)in their objective function.Nonetheless,they exploit the fact that the Gaussian kernel is the Green’s function of the diffu-sion equation(4)(under appropriate initial and boundary condi-tions)to provide an efficient regularization step.Regularization is efficiently performed through convolutions with a Gaussian kernel.Thirion,inspired by Maxwell’s Demons,proposed to perform image matching as a diffusion process[38].The proposed algo-rithm iterated between two steps:1)estimation of the demon forces for every demon(more precisely,the result of the appli-cation of a force during one iteration step,that is a displace-ment),and2)update of the transformation based on the cal-culated forces.Depending on the way the demon positions are selected,the way the space of deformations is defined,the in-terpolation method that is used,and the way the demon forces are calculated,different variants can be obtained.The most suit-able version for medical image analysis involved1)selecting all image elements as demons,2)calculating demon forces by considering the opticalflow constraint,3)assuming a nonpara-metric deformation model that was regularized by applying a Gaussianfilter after each iteration,and4)a trilinear interpo-lation scheme.The Gaussianfilter can be applied either to the displacementfield estimated at an iteration or the updated total displacementfield.The bijectivity of the transformation was en-sured by calculating for every point the difference between its initial position and the one that is reached after composing the forward with the backward deformationfield,and redistributing the difference to eachfield.The bijectivity of the transformation can also be enforced by limiting the maximum length of the up-date displacement to half the voxel size and using composition to update the transformation.Variants for the contour-based reg-istration and the registration between segmented images were also described in[38].Most of the algorithms described in this section were inspired by the work of Thirion[38]and thus could alternatively be clas-sified as“Demons approaches.”These methods share the iter-ative approach that was presented in[38]that is,iterating be-tween estimating the displacements and regularizing to obtain the transformation.This iterative approach results in increased computational efficiency.As it will be discussed later in this section,this feature led researchers to explore such strategies for different PDEs.The use of Demons,as initially introduced,was an efficient algorithm able to provide dense correspondences but lacked a sound theoretical justification.Due to the success of the algo-rithm,a number of papers tried to give theoretical insight into its workings.Fischer and Modersitzki[39]provided a fast algo-rithm for image registration.The result was given as the solution of linear system that results from the linearization of the diffu-sion PDE.An efficient scheme for its solution was proposed while a connection to the Thirion’s Demons algorithm[38]was drawn.Pennec et al.[40]studied image registration as an energy minimization problem and drew the connection of the Demons algorithm with gradient descent schemes.Thirion’s image force based on opticalflow was shown to be equivalent with a second order gradient descent on the Sum of Square Differences(SSD) matching criterion.As for the regularization,it was shown that the convolution of the global transformation with a Gaussian。

计算机视觉常用术语中英文对照

计算机视觉常用术语中英文对照

---------------------------------------------------------------最新资料推荐------------------------------------------------------ 计算机视觉常用术语中英文对照计算机视觉常用术语中英文对照(1)人工智能 Artificial Intelligence 认知科学与神经科学Cognitive Science and Neuroscience 图像处理Image Processing 计算机图形学Computer graphics 模式识别 Pattern Recognized 图像表示 Image Representation 立体视觉与三维重建Stereo Vision and 3D Reconstruction 物体(目标)识别 Object Recognition 运动检测与跟踪Motion Detection and Tracking 边缘edge 边缘检测detection 区域region 图像分割segmentation 轮廓与剪影contour and silhouette1/ 10纹理 texture 纹理特征提取 feature extraction 颜色 color 局部特征 local features or blob 尺度 scale 摄像机标定 Camera Calibration 立体匹配stereo matching 图像配准Image Registration 特征匹配features matching 物体识别Object Recognition 人工标注Ground-truth 自动标注Automatic Annotation 运动检测与跟踪 Motion Detection and Tracking 背景剪除Background Subtraction 背景模型与更新background modeling and update---------------------------------------------------------------最新资料推荐------------------------------------------------------ 运动跟踪 Motion Tracking 多目标跟踪 multi-target tracking 颜色空间 color space 色调 Hue 色饱和度 Saturation 明度 Value 颜色不变性 Color Constancy(人类视觉具有颜色不变性)照明illumination 反射模型Reflectance Model 明暗分析Shading Analysis 成像几何学与成像物理学 Imaging Geometry and Physics 全像摄像机 Omnidirectional Camera 激光扫描仪 Laser Scanner 透视投影Perspective projection 正交投影Orthopedic projection3/ 10表面方向半球 Hemisphere of Directions 立体角 solid angle 透视缩小效应 foreshortening 辐射度 radiance 辐照度 irradiance 亮度 intensity 漫反射表面、Lambertian(朗伯)表面 diffuse surface 镜面 Specular Surfaces 漫反射率 diffuse reflectance 明暗模型 Shading Models 环境光照 ambient illumination 互反射interreflection 反射图Reflectance Map 纹理分析Texture Analysis 元素 elements---------------------------------------------------------------最新资料推荐------------------------------------------------------ 基元 primitives 纹理分类 texture classification 从纹理中恢复图像 shape from texture 纹理合成 synthetic 图形绘制 graph rendering 图像压缩 image compression 统计方法 statistical methods 结构方法 structural methods 基于模型的方法 model based methods 分形fractal 自相关性函数autocorrelation function 熵entropy 能量energy 对比度contrast 均匀度homogeneity5/ 10相关性 correlation 上下文约束 contextual constraints Gibbs 随机场吉布斯随机场边缘检测、跟踪、连接 Detection、Tracking、Linking LoG 边缘检测算法(墨西哥草帽算子)LoG=Laplacian of Gaussian 霍夫变化 Hough Transform 链码 chain code B-样条B-spline 有理 B-样条 Rational B-spline 非均匀有理 B-样条Non-Uniform Rational B-Spline 控制点control points 节点knot points 基函数 basis function 控制点权值 weights 曲线拟合 curve fitting---------------------------------------------------------------最新资料推荐------------------------------------------------------ 内插 interpolation 逼近 approximation 回归 Regression 主动轮廓Active Contour Model or Snake 图像二值化Image thresholding 连通成分connected component 数学形态学mathematical morphology 结构元structuring elements 膨胀Dilation 腐蚀 Erosion 开运算 opening 闭运算 closing 聚类clustering 分裂合并方法 split-and-merge 区域邻接图 region adjacency graphs7/ 10四叉树quad tree 区域生长Region Growing 过分割over-segmentation 分水岭watered 金字塔pyramid 亚采样sub-sampling 尺度空间 Scale Space 局部特征 Local Features 背景混淆clutter 遮挡occlusion 角点corners 强纹理区域strongly textured areas 二阶矩阵 Second moment matrix 视觉词袋 bag-of-visual-words 类内差异 intra-class variability---------------------------------------------------------------最新资料推荐------------------------------------------------------ 类间相似性inter-class similarity 生成学习Generative learning 判别学习discriminative learning 人脸检测Face detection 弱分类器weak learners 集成分类器ensemble classifier 被动测距传感passive sensing 多视点Multiple Views 稠密深度图 dense depth 稀疏深度图 sparse depth 视差disparity 外极epipolar 外极几何Epipolor Geometry 校正Rectification 归一化相关 NCC Normalized Cross Correlation9/ 10平方差的和 SSD Sum of Squared Differences 绝对值差的和 SAD Sum of Absolute Difference 俯仰角 pitch 偏航角 yaw 扭转角twist 高斯混合模型Gaussian Mixture Model 运动场motion field 光流 optical flow 贝叶斯跟踪 Bayesian tracking 粒子滤波 Particle Filters 颜色直方图 color histogram 尺度不变特征转换 SIFT scale invariant feature transform 孔径问题 Aperture problem。

Approximation by quadrilateral finite elements

Approximation by quadrilateral finite elements

a rX iv:mat h /536v1[mat h.NA ]3Ma y2APPROXIMATION BY QUADRILATERAL FINITE ELEMENTS DOUGLAS N.ARNOLD,DANIELE BOFFI,AND RICHARD S.FALK Abstract.We consider the approximation properties of finite element spaces on quadri-lateral meshes.The finite element spaces are constructed starting with a given finite di-mensional space of functions on a square reference element,which is then transformed to a space of functions on each convex quadrilateral element via a bilinear isomorphism of the square onto the element.It is known that for affine isomorphisms,a necessary and suffi-cient condition for approximation of order r +1in L 2and order r in H 1is that the given space of functions on the reference element contain all polynomial functions of total degree at most r .In the case of bilinear isomorphisms,it is known that the same estimates hold if the function space contains all polynomial functions of separate degree r .We show,by means of a counterexample,that this latter condition is also necessary.As applications we demonstrate degradation of the convergence order on quadrilateral meshes as compared to rectangular meshes for serendipity finite elements and for various mixed and nonconforming finite elements. 1.Introduction Finite element spaces are often constructed starting with a finite dimensional space ˆV of shape functions given on a reference element ˆK and a class S of isomorphic mappings of the reference element.If F ∈S we obtain a space of functions V F (K )on the image element K =F (ˆK )as the compositions of functions in ˆV with F −1.Then,given a partition T of a domain Ωinto images of ˆK under mappings in S ,we obtain a finite element space as a subspace 1of the space V T of all functions on Ωwhich restrict to an element of V F (K )on each K ∈T .For example,if the reference element ˆK is the unit triangle,and the reference space ˆV is the space P r (ˆK )of polynomials of degree at most r on ˆK ,and the mapping class S isthe space Aff(ˆK )of affine isomorphisms of ˆK into R 2,then V T is the familiar space of all piecewisepolynomials of degree at most r on an arbitrary triangular mesh T .WhenS =Aff(ˆK ),as in this case,we speak of affine finite elements .If the reference element ˆKis the unit square,then it is often useful to take S equal to a larger space than Aff(ˆK ),namely the space Bil (ˆK )of all bilinear isomorphisms of ˆK into R 2.Indeed,if we allow only affine images of the unit square,then we obtain only parallelograms,and we are quite limited as to the domains that we can mesh (e.g.,it is not possible to mesh2DOUGLAS N.ARNOLD,DANIELE BOFFI,AND RICHARD S.F ALKa triangle with parallelograms).On the other hand,with bilinear images of the square we obtain arbitrary convex quadrilaterals,which can be used to mesh arbitrary polygons. The above framework is also well suited to studying the approximation properties offinite element spaces.See,e.g.,[2]and[1].A fundamental result holds in the case of affinefinite elements:S=Aff(ˆK).Under the assumption that the reference spaceˆV⊇P r(ˆK),the following result is well known:if T1,T2,...is any shape-regular sequence of triangulations of a domainΩand u is any smooth function onΩ,then the L2error in the best approximation of u by functions in V T n is O(h r+1)and the piecewise H1error is O(h r),where h=h(T n)is the maximum element diameter.It is also true,even if less well-known,that the condition thatˆV⊇P r(ˆK)is necessary if these estimates are to hold.The above result does not restrict the choice of reference elementˆK,so it applies to rectangular and parallelogram meshes by takingˆK to be the unit square.But it does not apply to general quadrilateral meshes,since to obtain them we must choose S=Bil(ˆK), and the result only applies to affinefinite elements.In this case there is a standard result analogous to the positive result in the previous paragraph,[2],[1],[4,Section I.A.2].Namely, ifˆV⊇Q r(ˆK),then for any shape-regular sequence of quadrilateral partitions of a domainΩand any smooth function u onΩ,we again obtain that the error in the best approximation of u by functions in V T n is O(h r+1)in L2and O(h r)in(piecewise)H1.It turns out,as we shall show in this paper,that the hypothesis thatˆV⊇Q r(ˆK)is strictly necessary for these estimates to hold.In particular,ifˆV⊇P r(ˆK)butˆV Q r(ˆK),then the rate of approximation achieved on general shape-regular quadrilateral meshes will be strictly lower than is obtained using meshes of rectangles or parallelograms.More precisely,we shall exhibit in Section3a domainΩand a sequence,T1,T2,...of quadrilateral meshes of it,and prove that whenever V(ˆK) Q r(ˆK),then there is a function u onΩsuch thatu−v L2(Ω)=o(h r),infv∈V T n(and so,a fortiori,is=O(h r+1)).A similar result holds for H1approximation.The counterexample is far from pathological.Indeed,the domainΩis as simple as possible, namely a square;the mesh sequence T n is simple and as shape-regular as possible in that all elements at all mesh levels are similar to a singlefixed trapezoid;and the function u is as smooth as possible,namely a polynomial of degree r.The use of a reference space which contains P r(ˆK)but not Q r(ˆK)is not unusual,but the degradation of convergence order that this implies on general quadrilateral meshes in comparison to rectangular(or parallelogram)meshes is not widely appreciated.It has been observed in special cases,often as a result of numerical experiments,cf.[7,Section8.7]. Wefinish this introduction by considering some examples.Henceforth we shall always useˆK to denote the unit square.First,considerfinite elements with the simple polynomial spaces as shape functions:ˆV=P r(ˆK).These of course yield O(h r+1)approximation in L2 for rectangular meshes.However,since P r(ˆK)⊇Q⌊r/2⌋(ˆK)but P r(ˆK) Q⌊r/2⌋+1(ˆK),on general quadrilateral meshes they only afford O(h⌊r/2⌋+1)approximation.A similar situation holds for serendipityfinite element spaces,which have been popular in engineering computation for thirty years.These spaces are constructed using as reference shape functions the space S r(ˆK)which is the span of P r(ˆK)together with the two monomialsAPPROXIMATION BY QUADRILATERAL FINITE ELEMENTS3ˆx rˆy andˆyˆx r.(The purpose of the additional two functions is to allow local degrees of freedom which can be used to ensure interelement continuity.)For r=1,S1(ˆK)=Q1(ˆK),but for r>1the situation is similar to that for P r(ˆK),namely S r(ˆK)⊇Q⌊r/2⌋(ˆK)but S r(ˆK) Q⌊r/2⌋+1(ˆK).So,again,the asymptotic accuracy achieved for general quadrilateral meshes is only about half that achieved for rectangular meshes:O(h⌊r/2⌋+1)in L2and O(h⌊r/2⌋)in H1.In Section4we illustrate this with a numerical example.While the serendipity elements are commonly used for solving second order differential equations,the pure polynomial spaces P r can only be used on quadrilaterals when interele-ment continuity is not required.This is the case in several mixed methods.For example,a popular element choice to solve the stationary Stokes equations is bilinearly mapped piece-wise continuous Q2elements for the two components of velocity,and discontinuous piecewise linear elements for the pressure.Typically the pressure space is taken to be functions which belong to P1(K)on each element K.This is known to be a stable mixed method and gives second order convergence in H1for the velocity and L2for the pressure.If one were to define the pressure space instead by using the construction discussed above,namely by composing linear functions on reference square with bilinear mappings,then the approximation prop-erties of mapped P1discussed above would imply that method could be at mostfirst order accurate,at least for the pressures.Hence,although the use of mapped P1as an alternative to unmapped P1pressure elements is sometimes proposed[6],it is probably not advisable. Another place where mapped P r spaces arise is for approximating the scalar variable in mixedfinite element methods for second order elliptic equations.Although the scalar variable is discontinuous,in order to prove stability it is generally necessary to define the space for approximating it by composition with the mapping to the reference element(while the space for the vector variable is defined by a contravariant mapping associated with the mapping to the reference element).In the case of the Raviart–Thomas rectangular elements,the scalar space on the reference square is Q r(ˆK),which maintains full O(h r+1)approximation properties under bilinear mappings.By contrast,the scalar space used with the Brezzi-Douglas-Marini and the Brezzi-Douglas-Fortin-Marini spaces is P r(ˆK).This necessarily results in a loss of approximation order when mapped to quadrilaterals by bilinear mappings. Another type of element which shares this difficulty is the simplest nonconforming quadri-lateral element,which generalizes to quadrilaterals the well-known piecewise linear non-conforming element on triangles,with degrees of freedom at the midpoints of edges.On the square,a bilinear function is not well-defined by giving its value at the midpoint of edges(or its average on edges),because these quantities do not comprise a unisolvent set of degrees of freedom(the function(ˆx−1/2)(ˆy−1/2)vanishes at the four midpoints of the edges of the unit square).Hence,various definitions of nonconforming elements on rectangles replace the basis functionˆxˆy by some other function such asˆx2−ˆy2.Consequently,the reference space contains P1(ˆK),but does not contain Q1(ˆK),and so there is a degradation of convergence on quadrilateral meshes.This is discussed and analyzed in the context of the Stokes problem in[5].As afinal application,we remark that many of thefinite element methods proposed for the Reissner-Mindlin plate problem are based on mixed methods for the Stokes equations and/or for second order elliptic problems.As a result,many of them suffer from the same4DOUGLAS N.ARNOLD,DANIELE BOFFI,AND RICHARD S.F ALKsort of degradation of convergence on quadrilateral meshes.An analysis of a variety of these elements will appear in forthcoming work by the present authors.In Section3,we prove our main result,the necessity of the condition that the reference space contain Q r(ˆK)in order to obtain O(h r+1)approximation on quadrilateral meshes.The proof relies on an analogous result for affine approximation on rectangular meshes,where the space P r(ˆK)enters rather than Q r(ˆK).While this is a special case of known results, for the convenience of the reader we include an elementary proof in Section2.In thefinal section we illustrate the results with numerical computations.2.Approximation theory of rectangular elementsIn this section we prove some results concerning approximation by rectangular elements which will be needed in the next section where the main results are proved.The results in this section are essentially known,and many are true in far greater generality than stated here.If K is any square with edges parallel to the axes,then K=F K(ˆK)where F K(ˆx):= x K+h Kˆx with x K∈R2and h K>0the side length.For any function u∈L2(K),we define ˆu K=u◦F K∈L2(ˆK),i.e.,ˆu K(ˆx)=u(x K+h Kˆx).Given a subspaceˆS of L2(ˆK)we define the associated subspace on an arbitrary square K byS(K)={u:K→R|ˆu K∈ˆS}.Finally,letΩdenote the unit cube(ΩandˆK both denote the unit cube,but we use the notationΩwhen we think of it as afixed domain,while we useˆK when we think of it as a reference element).For n=1,2,...,let T h be the uniform mesh ofΩinto n d subcubes when h=1/n,and defineS h={u:Ω→R|u|K∈S(K)for all K∈T h}.In this definition,when we write u|K∈S(K)we mean only that u|K agrees with a function in S K almost everywhere,and so do not impose any interelement continuity.The following theorem gives a set of equivalent conditions for optimal order approximation of a smooth function u by elements of S h.Theorem1.LetˆS be afinite dimensional subspace of L2(ˆK),r a non-negative integer. The following conditions are equivalent:1.There is a constant C such that infv∈S hu−v L2(Ω)≤Ch r+1|u|r+1for all u∈H r+1(Ω).2.infv∈S hu−v L2(Ω)=o(h r)for all u∈P r(Ω).3.P r(ˆK)⊂ˆS.Proof.Thefirst condition implies that inf v∈Shu−v L2(Ω)=0for u∈P r(Ω),and so implies the second condition.The fact that the third condition implies thefirst is a well-known consequence of the Bramble–Hilbert lemma.So we need only show that the second condition implies the third.The proof is by induction on r.First consider the case r=0.We haveinf v∈S h u−v 2L2(Ω)= K∈T h inf v K∈S(K) u−v K 2L2(K)=h2 K∈T h inf w∈ˆS ˆu K−w 2L2(ˆK),(1)APPROXIMATION BY QUADRILATERAL FINITE ELEMENTS 5where we have made the change of variable w =ˆv K in the last step.In particular,for u ≡1on Ω,ˆu K ≡1on ˆKfor all K ,so the quantity c :=inf w ∈ˆSˆu K −w 2L 2(ˆK )is independent of K .Thusinf v ∈S h u −v 2L 2(Ω)=h 2 K ∈T h c =cThe hypothesis that this quantity is o (1)implies that c =0,i.e.,that the constant function belongs to ˆS.Now we consider the case r >0.We again apply (1),this time for u an arbitrary homogeneous polynomial of degree r .Thenˆu K (ˆx )=u (x K +h ˆx )=u (h ˆx )+p (ˆx )=h r u (ˆx )+p (ˆx ),(2)where p ∈P r −1(ˆK ).Substituting in (1),and invoking the inductive hypothesis that ˆS ⊇P r −1(ˆK ),we get that inf v ∈S h u −v 2L 2(Ω)=h 2+2r K ∈T h inf w ∈ˆS u −w 2L 2(ˆK )=h 2r inf w ∈ˆSu −w 2L 2(ˆK ).Again the last infimum is independent of K so we immediately deduce that u belongs to ˆS .Thus ˆScontains all homogeneous polynomials of degree r and all polynomials of degree less than r (by induction),so it indeed contains all polynomials of degree at most r .A similar theorem holds for gradient approximation.Since the finite elements are not necessarily continuous we write ∇h for the gradient operator applied piecewise on each element.Theorem 2.Let ˆSbe a finite dimensional subspace of L 2(ˆK ),r a non-negative integer.The following conditions are equivalent:1.There is a constant C such that inf v ∈S h ∇h (u −v ) L 2(Ω)≤Ch r |u |r +1for all u ∈H r +1(Ω).2.inf v ∈S h ∇h (u −v ) L 2(Ω)=o (hr −1)for all u ∈P r (Ω).3.P r (ˆK)⊂P 0(ˆK )+ˆS .Proof.Again,we need only prove that the second condition implies the third.In analogy to(1),we have inf v ∈S h K ∈T h ∇(u −v ) 2L 2(K )= K ∈T h inf v K ∈S (K )∇(u −v K ) 2L 2(K )=K ∈T h inf w ∈ˆS ∇(ˆu K −w ) 2L 2(ˆK),(3)where we have made the change of variable w =ˆv K in the last step.The proof proceeds by induction on r ,the case r =0being trivial.For r >0,apply(3)with u an arbitrary homogeneous polynomial of degree r .Substituting (2)in (3),and6DOUGLAS N.ARNOLD,DANIELE BOFFI,AND RICHARD S.F ALKinvoking the inductive hypothesis that P0(ˆK)+ˆS⊇P r−1(ˆK),we get thatinfv∈S h∇h(u−v) 2L2(Ω)=h2r K∈T h inf w∈ˆS ∇(u−w) 2L2(ˆK)=h2r−2inf w∈ˆS ∇(u−w) 2L2(ˆK).Since we assume that this quantity is o(h2r−2),the last infimum must be0,so u differs from an elementˆS by a constant.Thus P0(ˆK)+ˆS contains all homogeneous polynomials of degree r and all polynomials of degree less than r(by induction),so it indeed contains all polynomials of degree at most r.Remarks.1.IfˆS contains P0(ˆK),which is usually the case,then the third condition of Theorem2reduces to that of Theorem1.2.A similar result holds for higher derivatives(replace∇h by∇m h in thefirst two conditions, and P0(ˆK)by P m−1(ˆK)in the third).3.Approximation theory of quadrilateral elementsIn this,the main section of the paper,we consider the approximation properties offinite element spaces defined with respect to quadrilateral meshes using bilinear mappings starting from a givenfinite dimensional space of polynomialsˆV on the unit squareˆK=[0,1]×[0,1]. For simplicity we assume thatˆV⊇P0(ˆK).For exampleˆV might be the space P r(ˆK)of polynomials of degree at most r,or the space Q r(ˆK)of polynomials of degree at most r in each variable separately,or the serendipity space S r(ˆK)spanned by P r(ˆK)together with the monomialsˆx r1ˆx2andˆx1ˆx r2.Let F be a bilinear isomorphism ofˆK onto a convex quadrilateral K=F(ˆK).Then for u∈L2(K)we defineˆu F∈L2(ˆK)byˆu F(ˆx)=u(Fˆx),and setV F(K)={u:K→R|ˆu F∈ˆV}.(Note that the definition of this space depends on the particular choice of the bilinear iso-morphism F ofˆK onto K,but whenever the spaceˆV is invariant under the symmetries of the square,which is usually the case in practice,this will not be so.)We also note that the functions in V F(K)need not be polynomials if F is not affine,i.e.,if K is not a parallelogram. Given a quadrilateral mesh T of some domain,Ω,we can then construct the space of functions V T consisting of functions on the domain which when restricted to a quadrilateral K∈T belong to V FK(K)where F K is a bilinear isomorphism ofˆK onto K.(Again,ifˆV is not invariant under the symmetries of the square,the space V T will depend on the specific choice of the maps F K.)It follows from the results of the previous section that if we consider the sequence of meshes of the unit square into congruent subsquares of side length h=1/n,then each of the approximation estimatesinf v∈V T h u−v L2(Ω)≤Ch r+1|u|r+1for all u∈H r+1(Ω),(4)inf v∈V T h ∇h(u−v) L2(Ω)≤Ch r|u|r+1for all u∈H r+1(Ω),(5)holds if and only P r(ˆK)⊂ˆV.It is not hard to extend these estimates to shape-regular sequences of parallelogram meshes as well.However,in this section we show that for theseAPPROXIMATION BY QUADRILATERAL FINITE ELEMENTS7 estimates to hold for more general quadrilateral mesh sequences,a stronger condition onˆV is required,namely thatˆV⊇Q r(ˆK).The positive result,that whenˆV⊇Q r(ˆK),then the estimates(4)and(5)hold for any shape regular sequence of quadrilateral meshes T h,is known.See,e.g.,[2],[1],or[4,Section I.A.2].We wish to show the necessity of the conditionˆV⊇Q r(ˆK).As afirst step we show that the condition V F(K)⊇P r(K)is necessary and sufficient to have thatˆV⊇Q r(ˆK)whenever F is a bilinear isomorphism ofˆK onto a convex quadrilateral. This is proven in the following two theorems.Theorem3.Suppose thatˆV⊇Q r(ˆK).Let F be any bilinear isomorphism ofˆK onto a convex quadrilateral.Then V F(K)⊇P r(K).Proof.The components of F(ˆx,ˆy)are linear functions ofˆx andˆy,so if p is a polynomial of degree at most r,then p(F(ˆx,ˆy))is of degree at most r inˆx andˆy,i.e.,p◦F∈Q r(ˆK)⊂ˆV. Therefore p∈V F(K).The reverse implication holds even under the weaker assumption that V F(K)contains P r(K)just for the two specific bilinear isomorphism˜F(ˆx,ˆy)=(ˆx,ˆy(ˆx+1)),¯F(ˆx,ˆy)=(ˆy,ˆx(ˆy+1)),both of which mapˆK isomorphically onto the quadrilateral K′with vertices(0,0),(1,0), (0,1),and(1,2).This fact is established below.Theorem4.LetˆV be a vectorspace of functions onˆK.Suppose that Q r(ˆK) ˆV.Then either V˜F(K′) P r(K′)or V¯F(K′) P r(K′).Remark.If the spaceˆV is invariant under the symmetries of the square,then V˜F (K′)=V¯F(K′)so neither contains P r(K′).Proof.Assume to the contrary that V˜F(K′)⊇P r(K′)and V¯F(K′)⊇P r(K′).We prove thatˆV⊇Q r(ˆK)by induction on r.The case r=0being true by assumption,we consider r>0and show that the monomialsˆx rˆy s andˆx sˆy r belong toˆV for s=0,1,...,r.From the identityˆx rˆy s=ˆx r−s[ˆy(ˆx+1)]s−st=1 s t ˆx r−tˆy s=˜F1(ˆx,ˆy)r−s˜F2(ˆx,ˆy)s−s t=1 s t ˆx r−tˆy s,(6)we see that for0≤s<r,the monomialˆx rˆy s is the sum of a polynomial which clearly belongs toˆV(since˜F1(ˆx,ˆy)r−s˜F2(ˆx,ˆy)s=x r−s y s∈P r(K′)⊂V˜F(K′))and a polynomial in Q r−1(ˆK),which belongs toˆV by induction.Thus each of the monomialsˆx rˆy s with0≤s<r belongs toˆV,and,using¯F,we similarly see that all the monomialsˆx sˆy r,0≤s<r belong toˆV.Finally,from(6)with s=r,we see thatˆx rˆy r is a linear combination of an element of ˆV and monomialsˆx sˆy r with s<r,so it too belongs toˆV.We now combine this result with the those of the previous section to show the necessity of the conditionˆV⊇Q r(ˆK)for optimal order approximation.LetˆV be somefixedfinite dimensional subspace of L2(ˆK)which does not include Q r(ˆK).Consider the specific division of the unit squareˆK into four quadrilaterals shown on the left in Figure1.For definiteness we8place the the midpoints of theA meshThe four specific quadri-lateral K′4,of the four quadrilaterals K′′shown in Figure1an isomorphism F′′from the unit square so that V F′′(K′′) P r(K′′).If we letˆS be the subspace of L2(ˆK)consisting of functions which restrict to elements of V F′′(K′′)on each of the four quadrilaterals K′′,then certainlyˆS does not contain P r(ˆK),since even its restriction to any one of the quadrilaterals K′′does not contain P r(K′′).Next,for n=1,2,...consider the mesh T′h of the unit squareΩshown in Figure1b, obtained byfirst dividing it into a uniform n×n mesh of subsquares,n=1/h,and then dividing each subsquare as in Figure1a.Then the space of functions u onΩwhose restrictions on each subsquare K∈T h satisfyˆu K(ˆx)=u(x K+hˆx)withˆu K∈ˆS is precisely the same as the space V(T′h)constructed from the initial spaceˆV and the mesh T′h.In view of Theorems1 and2and the fact thatˆS P r(ˆK),the estimates(4)and(5)do not hold.In fact,neither of the estimatesinfv∈V(T h)u−v L2(Ω)=o(h r),norinfv∈V(T h)∇(u−v) L2(Ω)=o(h r−1),holds,even for u∈P r(Ω).While the conditionˆV⊇Q r(ˆK)is necessary for O(h r+1)on general quadrilateral meshes, the conditionsˆV⊇P r(ˆK)suffices for meshes of parallelograms.Naturally,the same is true for meshes whose elements are sufficiently close to parallelograms.We conclude this section with a precise statement of this result and a sketch of the proof.IfˆV⊇P r(ˆK)and K=F(ˆK)with F∈Bil(ˆK),then by standard arguments,as in[1],we getv−πv L2(K)≤C J F 1/2L∞(ˆK)|v◦F|H r+1(ˆK),where J F is the Jacobian determinant of F.Now,using the formula for the derivative of a composition(as in,e.g.,[3,p.222]),and the fact that F is quadratic,and so its third andAPPROXIMATION BY QUADRILATERAL FINITE ELEMENTS9 higher derivatives vanish,we get that|v◦F|H r+1(ˆK)≤C J F−1 1/2L∞(K)v H r+1(K)⌊(r+1)/2⌋i=0|F|r+1−2i W1∞(ˆK)|F|i W2∞(ˆK).Now,J FL∞(ˆK)≤Ch2K, J F−1L∞(ˆK)≤Ch−2K,|F|W1∞(ˆK)≤Ch K,where h K is the diameter of K and C depends only on the shape-regularity of K.We thus getv−πv L2(K)≤C v H r+1(K) i h r+1−2i K|F|i W2∞(ˆK).It follows that if|F|W2∞(ˆK)=O(h2K),we get the desired estimatev−πv L2(K)≤Ch r+1Kv H r+1(K).Following[5],we measure the deviation of a quadrilateral from a parallelogram,by the quantityσK:=max(|π−θ1|,|π−θ2|),whereθ1is the angle between the outward normals of two opposite sides of K andθ2is the angle between the outward normals of the other two sides.Thus0≤σK<π,withσK=0if and only if K is a parallelogram.As pointed out in[5],|F|W2∞(ˆK)≤Ch K(h K+σK).This motivates the definition that a family of quadrilateralmeshes is asymptotically parallelogram ifσK=O(h K),i.e.,ifσK/h K is uniformly bounded for all the elements in all the meshes.From the foregoing considerations,if the reference space contains P r(ˆK)we obtain O(h r+1)convergence for asymptotically parallelogram,shape regular meshes.As afinal note,we remark that any polygon can be meshed by an asymptotically paral-lelogram,shape regular family of meshes with mesh size tending to zero.Indeed,if we begin with any mesh of convex quadrilaterals,and refine it by dividing each quadrilateral in four by connecting the midpoints of the opposite edges,and continue in this fashion,as in the last row of Figure2,the resulting mesh is asymptotically parallelogram and shape regular.4.Numerical resultsIn this section we report on results from a numerical study of the behavior of piecewise continuous mapped biquadratic and serendipityfinite elements on quadrilateral meshes(i.e., thefinite element spaces are constructed starting from the spaces Q2(ˆK)and S2(ˆK)on the reference square,and then imposing continuity).We present the results of two test problems. In both we solve the Dirichlet problem for Poisson’s equation−∆u=f inΩ,u=g on∂Ω,(7)where the domainΩis the unit square.In thefirst problem,f and g are taken so that the exact solution is the quartic polynomialu(x,y)=x3+5y2−10y3+y4.Table1shows results for both types of elements using meshes from each of thefirst two mesh sequences shown in Figure2.Thefirst sequence of meshes consists of uniform square subdi-visions of the domain into n×n subsquares,n=2,4,8,....Meshes in the second sequence10DOUGLAS N.ARNOLD,DANIELE BOFFI,AND RICHARD S.F ALKare partitions of the domain into n×n congruent trapezoids,all similar to the trapezoid with vertices(0,0),(1/2,0),(1/2,2/3),and(0,1/3).In Table1we report the errors in L2 for thefinite element solution and its gradient both in absolute terms and as a percentage of the L2rate ofrates ofcon-verge is true for the above, for the is reducedandAs a choosing theAs seen on the veryfine meshes(and very high precision computation)would be required to see thefinal asymptotic orders.Finally we return to thefirst test problem,and consider the behavior of the serendipity elements on the third mesh sequence shown in Figure2.This mesh sequence begins with the same mesh of four quadrilaterals as in previous case,and continues with systematic refinement as described at the end of the last section,and so is asymptotically parallelogram. Therefore,as explained there,the rate of convergence for serendipity elements is the same as for affine meshes.This is clearly illustrated in Table3.While the asymptotic rates predicted by the theory are confirmed in these examples,it is worth noting that in absolute terms the effect of the degraded convergence rate is not very pronounced.For thefirst example,on a moderatelyfine mesh of16×16trapezoids,theAPPROXIMATION BY QUADRILATERAL FINITE ELEMENTS11 Table1.Errors and rates of convergence for the test problem with polyno-mial solution.Mapped biquadratic elementssquare meshes trapezoidal meshesu−u h L2 u−u h L2err.%rate err.%rate2 4.5e−0137.253 5.9e−0148.5764 1.1e−019.333 2.0 1.5e−0112.082 2.08 2.8e−02 2.329 2.0 3.7e−02 3.017 2.0167.1e−030.583 2.09.2e−030.753 2.032 1.8e−030.146 2.0 2.3e−030.188 2.064 4.4e−040.036 2.0 5.7e−040.047 2.0∇(u−u h) L2 ∇(u−u h) L2 n err.%rate err.%rate3.5e−02 2.877 5.0e−024.0664.4e−030.360 3.0 6.7e−030.548 2.95.5e−040.045 3.09.7e−040.080 2.86.9e−050.006 3.0 1.6e−040.013 2.68.6e−060.001 3.0 3.3e−050.003 2.31.1e−060.000 3.07.4e−060.0012.1solution error with serendipity elements exceeds that of mapped biquadratic elements by a factor of about2,and the gradient error by a factor of2.5.Even on thefinest mesh shown, with64×64elements,the factors are only about5.5and8.5,respectively.Of course,if we were to compute onfiner andfiner meshes with sufficiently high precision,these factors would tend to infinity.Indeed,on any quadrilateral mesh which contains a non-parallelogram element,the analogous factors can be made as large as desired by choosing a problem in which the exact solution is sufficiently close to—or even equal to—a quadratic function,which the mapped biquadratic elements capture exactly,while the serendipity elements do not(such a quadratic function always exists).However,it is not unusual that the serendipity elements perform almost as well as the mapped biquadratic elements for reasonable,and even for quite small,levels of error.This,together with their optimal convergence on asymptotically parallelogram meshes,provides an explanation of why the lower rates of convergence have not been widely noted.References1.P.G.Ciarlet,Thefinite element method for elliptic problems,North-Holland,Amsterdam,1978.2.P.G.Ciarlet and P.-A.Raviart,Interpolation theory over curved elements with applications tofiite elementmethods,Comput.Methods Appl.Mech.Engrg.1(1972),217–249.3.H.Federer,Geometric measure theory,Springer-Verlag,New York,1969.4.V.Girault and P.-A.Raviart,Finite element methods for Navier-Stokes equations,Springer-Verlag,NewYork,1986.。

Color quantization

Color quantization

Color quantizationAn example image in 24-bit RGB colorThe same image reduced to a palette of 16 colorsspecifically chosen to best represent the image;the selected palette is shown by the squares above In computer graphics, color quantization or color imagequantization is a process that reduces the number of distinct colorsused in an image, usually with the intention that the new image shouldbe as visually similar as possible to the original image. Computeralgorithms to perform color quantization on bitmaps have been studiedsince the 1970s. Color quantization is critical for displaying imageswith many colors on devices that can only display a limited number ofcolors, usually due to memory limitations, and enables efficientcompression of certain types of images.The name "color quantization" is primarily used in computer graphicsresearch literature; in applications, terms such as optimized palettegeneration , optimal palette generation , or decreasing color depth areused. Some of these are misleading, as the palettes generated bystandard algorithms are not necessarily the best possible.AlgorithmsMost standard techniques treat color quantization as a problem ofclustering points in three-dimensional space, where the points representcolors found in the original image and the three axes represent thethree color channels. Almost any three-dimensional clusteringalgorithm can be applied to color quantization, and vice versa. After the clusters are located, typically the points in each cluster areaveraged to obtain the representative color that all colors in that clusterare mapped to. The three color channels are usually red, green, andblue, but another popular choice is the Lab color space, in which Euclidean distance is more consistent with perceptual difference.The most popular algorithm by far for color quantization, invented by Paul Heckbert in 1980, is the median cut algorithm. Many variations on this scheme are in use. Before this time, most color quantization was done using the population algorithm or population method , which essentially constructs a histogram of equal-sized ranges and assigns colors to the ranges containing the most points. A more modern popular method is clustering using octrees,first conceived by Gervautz and Purgathofer and improved by Xerox PARC researcher Dan Bloomberg.A small photograph that has had its blue channelremoved. This means all of its pixel colors lie in atwo-dimensional plane in the color cube.The color space of the photograph to the left, along witha 16-color optimized palette produced by Photoshop. TheVoronoi regions of each palette entry are shown.If the palette is fixed, as is often the case in real-time color quantization systems such as those used in operating systems, color quantization is usually done using the "straight-line distance" or "nearest color" algorithm, which simply takes each color in the original image and finds the closest palette entry, where distance is determined by the distance between the two corresponding points in three-dimensional space. In other words, if the colors are and , we want to minimize the Euclidean distance:This effectively decomposes the color cube into a Voronoi diagram, where the palette entries are the points and a cell contains all colors mapping to a single palette entry. There are efficient algorithms from computational geometry for computing Voronoi diagrams and determining which region a given point falls in; in practice, indexed palettes are so small that these are usually overkill.A colorful image reduced to 4 colors usingspatial color quantization.Color quantization is frequently combined with dithering, which caneliminate unpleasant artifacts such as banding that appear whenquantizing smooth gradients and give the appearance of a largernumber of colors. Some modern schemes for color quantization attemptto combine palette selection with dithering in one stage, rather thanperform them independently.A number of other much less frequently used methods have beeninvented that use entirely different approaches. The Local K-meansalgorithm, conceived by Oleg Verevka in 1995, is designed for use inwindowing systems where a core set of "reserved colors" is fixed foruse by the system and many images with different color schemes mightbe displayed simultaneously. It is a post-clustering scheme that makesan initial guess at the palette and then iteratively refines it.The high quality but slow NeuQuant algorithm reduces images to 256colors by training a Kohonen neural network "which self-organisesthrough learning to match the distribution of colours in an input image.Taking the position in RGB-space of each neuron gives a high-qualitycolour map in which adjacent colours are similar." [1] It is particularlyadvantageous for images with gradients.Finally, one of the most promising new methods is spatial colorquantization , conceived by Puzicha, Held, Ketterer, Buhmann, andFellner of the University of Bonn, which combines dithering with palette generation and a simplified model of human perception to produce visually impressive results even for very small numbers ofcolors. It does not treat palette selection strictly as a clustering problem,in that the colors of nearby pixels in the original image also affect the color of a pixel. See sample images [2].History and applicationsIn the early days of PCs, it was common for video adapters to support only 2, 4, 16, or (eventually) 256 colors due to video memory limitations; they preferred to dedicate the video memory to having more pixels (higher resolution)rather than more colors. Color quantization helped to justify this tradeoff by making it possible to display many high color images in 16- and 256-color modes with limited visual degradation. The Windows operating system and many other operating systems automatically perform quantization and dithering when viewing high color images in a 256color video mode, which was important when video devices limited to 256 color modes were dominant. Modern computers can now display millions of colors at once, far more than can be distinguished by the human eye, limiting this application primarily to mobile devices and legacy hardware.Nowadays, color quantization is mainly used in GIF and PNG images. GIF, for a long time the most popular lossless and animated bitmap format on the World Wide Web, only supports up to 256 colors, necessitating quantization for many images. Some early web browsers constrained images to use a specific palette known as the web colors,leading to severe degradation in quality compared to optimized palettes. PNG images support 24-bit color, but can often be made much smaller in filesize without much visual degradation by application of color quantization, since PNG files use fewer bits per pixel for palettized images.The infinite number of colors available through the lens of a camera is impossible to display on a computer screen;thus converting any photograph to a digital representation necessarily involves some quantization. In practice, 24-bit color is sufficiently rich to represent almost all colors perceivable by humans with sufficiently small error as to bevisually identical (if presented faithfully) .With the few colors available on early computers, different quantization algorithms produced very different-looking output images. As a result, a lot of time was spent on writing sophisticated algorithms to be more lifelike.Editor supportMany bitmap graphics editors contain built-in support for color quantization, and will automatically perform it when converting an image with many colors to an image format with fewer colors. Most of these implementations allow the user to set exactly the number of desired colors. Examples of such support include:•Photoshop's Mode→Indexed Color function, supplies a number of quantization algorithms ranging from the fixed Windows system and Web palettes to the proprietary Local and Global algorithms for generating palettes suited toa particular image or images.•Paint Shop Pro, in its Colors→Decrease Color Depth dialog, supplies three standard color quantization algorithms: median cut, octree, and the fixed standard "web safe" palette.•The GIMP's Generate Optimal Palette with 256 Colours option, known to use the median cut algorithm. There has been some discussion in the developer community of adding support for spatial color quantization.[3]Color quantization is also used to create posterization effects, although posterization has the slightly different goal of minimizing the number of colors used within the same color space, and typically uses a fixed palette.Some vector graphics editors also utilize color quantization, especially for raster-to-vector techniques that create tracings of bitmap images with the help of edge detection.•Inkscape's Path→Trace Bitmap: Multiple Scans: Color function uses octree quantization to create color traces.[4]References[1].au/~dekker/NEUQUANT.HTML[2]/~dcoetzee/downloads/scolorq/#sampleimages[3]/lists/gimp-user/2000-April/001024.html, /lists/gimp-developer/2000-April/012205.html[4]Bah, Tavmjong (2007-07-23). "Inkscape » Tracing Bitmaps » Multiple Scans" (http://tavmjong.free.fr/INKSCAPE/MANUAL/html/Trace-Multi.html). . Retrieved 2008-02-23.Further reading•Paul S. Heckbert. Color Image Quantization for Frame Buffer Display (/web/ 20050606233131//heckbert80color.html). ACM SIGGRAPH '82 Proceedings. First publication of the median cut algorithm.•Dan Bloomberg. Color quantization using octrees (/papers/colorquant.pdf).Leptonica.•Oleg Verevka. Color Image Quantization in Windows Systems with Local K-means Algorithm (http://citeseer./100440.html). Proceedings of the Western Computer Graphics Symposium '95.•J. Puzicha, M. Held, J. Ketterer, J. M. Buhmann, and D. Fellner. On Spatial Quantization of Color Images (http:// www.iai.uni-bonn.de/III/forschung/publikationen/tr/abstracts/IAI-TR-98-1.abstract-en.html). ( full text.ps.gz (rmatik.uni-bonn.de/III/forschung/publikationen/tr/reports/IAI-TR-98-1.ps.gz))Technical Report IAI-TR-98-1, University of Bonn. 1998.Article Sources and Contributors5Article Sources and ContributorsColor quantization Source: /w/index.php?oldid=511881716 Contributors: Andreas Kaufmann, Bobblehead, CountingPine, Davidhorman, Dcoetzee, ExcaliburZX,Giftlite, HairyWombat, Ioannes Pragensis, Kuefi, Loadmaster, M.O.X, MEmreCelebi, Master Bigode, Nahum Reduta, Quasimondo, Ricardo Cancho Niemietz, Rjwilmsi, Trlkly, Venkadachalam, Wingman4l7, Xrchz, 37 anonymous editsImage Sources, Licenses and ContributorsFile:Dithering example undithered.png Source: /w/index.php?title=File:Dithering_example_undithered.png License: GNU Free Documentation License Contributors: Andyzweb, Jamelan, Jorva, Kri, Origamiemensch, 1 anonymous editsFile:Dithering example undithered 16color palette.png Source: /w/index.php?title=File:Dithering_example_undithered_16color_palette.png License: GNU FreeDocumentation License Contributors: CountingPine, DcoetzeeFile:Rosa Gold Glow 2 small noblue.png Source: /w/index.php?title=File:Rosa_Gold_Glow_2_small_noblue.png License: GNU Free Documentation LicenseContributors: DcoetzeeFile:Rosa Gold Glow 2 small noblue color space.png Source: /w/index.php?title=File:Rosa_Gold_Glow_2_small_noblue_color_space.png License: Public DomainContributors: User:DcoetzeeFile:Spatial color quantization - rainbow, 4 colors.png Source: /w/index.php?title=File:Spatial_color_quantization_-_rainbow,_4_colors.png License: CreativeCommons Zero Contributors: DcoetzeeLicenseCreative Commons Attribution-Share Alike 3.0 Unported///licenses/by-sa/3.0/。

  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
contours and level sets provide a solid formal framework for image segmention. The problem, stated as the minimization of a functional containing terms of conformity to data and regularization, is solved by curve evolution implemented via level set partial differential equations (PDEs). The purpose of this study is to investigate approximation by basis functions as a model for image representation in segmentation by level set PDEs. This model is mathematically yielding, affords more generality than current piecewise constant and Gaussian models, and can be just as efficient as the most general piecewise smooth model. We state the problem using this model to measure conformity of segmentation to data. The resulting functional is minimized via level set evolution PDEs. Experimental results are shown to demonstrate the formulation. 1. INTRODUCTION Image segmentation is a fundamental problem with numerous useful applications in image processing and computer vision. Active contours [1] and level sets [2] have led to a new class of algorithms which have succeeded in segmenting difficult images. Several algorithms have been developped for image segmentation [3–8]. These algorithms follow a well posed problem stated according to a variational formulation where assumptions and constraints are transparent. The solution follows the minimization by curve evolution of an objective functional containing terms of conformity to data and regularization. Curve evolution is realized via level set PDEs. A level set implementation has several well known advantages over the classical snake active contour implementation. In particular, it is numerically stable and accounts for variations in the topology of evolving curves. Most studies on image segmentation by level set PDEs have used general formulations of the problem to concentrate on issues such as evolution equations that lead to a partition of the image domain, implementation efficiency, and application to inputs other than grey-level images such as disparity fields, motion fields, vectorial images, or images of a particular domain of application, such as medical imaging, to incorporate specific, a priori information about the data. The problem of image representation has been secondary in these studies, understandably so, although a good repThis work was supported by the Natural Sciences and Engineering Research Council of Canada under grant OGP004234
resentation is just as important as a good formulation. The image models most often used are the piecewise constant model [4] and the Gaussian model [7]. More representative models have been used in two recent studies, namely the piecewise smooth model in a two-region level set implementation of the Mumford-Shah algorithm [9] and histograms computed over the regions of segmentation [10]. The purpose of this study is to investigate approximation by basis functions as a model for image representation, and its use in segmentation by level set PDEs. This general parametric model is most relevant for several reasons. It is mathematically yielding, affords more generality than current piecewise constant and Gaussian models, is more economical and can be just as efficient as the most general piecewise smooth model. Also, it can serve better important applications of segmentation where parametric models are efficient representations, in, for instance, region-based image reconstruction, region-based coding, and motion-based image partitioning. We state the problem using this model to measure conformity of segmentation to data. The resulting functional is minimized by a greedy algorithm via level set evolution PDEs. Examples are given to demonstrate the algorithm. The remainder of the paper is organized as follows. In Section 2 we present a formal description of the problem for two-region segmentation. The approximation and regularization terms of the energy functional are presented and the evolution equations are derived. Section 3 presents an extension to multiple regions. In Section 4 we show experimental results that validate the approach. Section 5 contains a conclusion. 2. TWO-REGION SEGMENTATION 2.1. Formulation For simplicity, we first formulate the two-region segmentation problem. A generalization to multiple regions will be given in Section 3. Let I : Ω → R be an image function defined on Ω ⊂ R2 . Let γ : [0, 1] → Ω be a simple closed planar curve parameterized by arc parameter s ∈ [0, 1], Rγ the region enclosed by γ , and R = c {R1 = Rγ , R2 = Rγ } the corresponding partition of the image domain. Finally, let Θ = span{θi (x)} be the space spanned by basis functions {θi (x)}i=1,...,M . The goal is to determine γ and α1 = (α11 , ..., α1M )T , α2 = (α21 , ..., α2M )T that minimize the
相关文档
最新文档