Hyperspectral Image Segmentation Using a New Bayesian Approach With Active Learning

合集下载

【最新精选】《数字图像处理》结课小论文题目汇总及要求

【最新精选】《数字图像处理》结课小论文题目汇总及要求

《数字图像处理》结课测试题目题目的路径:首先在Matlab的Command window中键入“demo”,进入demo 窗口。

然后在树形选择框中选择“Toolboxes\Image Processing”和“Blocksets\ Video and Image Processing”。

最后逐个查看并选择自己感兴趣的题目。

所有题目汇总如下:图像去模糊1. Deblurring Images Using the Blind Deconvolution Algorithm基于盲解卷算法的图像去模糊2. Deblurring Images Using the Lucy-Richardson Algorithm使用LR算法进行图像去模糊3. Deblurring Images Using a Regularized Filter使用正则滤波器进行图像去模糊4. Deblurring Images Using the Wiener Filter使用维纳滤波器进行图像去模糊图像增强5. Contrast Enhancement Techniques图像对比度增强技术6. Correcting Nonuniform Illumination如何对不均匀光照进行校正7. Enhancing Multispectral Color Composite Images多谱(卫星遥感) 图像增强技术图像配准8. Finding the Rotation and Scale of a Distorted Image计算失真图像的旋转参数和尺度参数9. Registering an Aerial Photo to an Orthophoto基于控制点的多幅航拍图像的配准10. Registering an Image Using Normalized Cross-Correlation使用归一化交叉相关法来配准图像图像分割11. Batch Processing Image Files Using Distributed Computing分布式计算对图像序列进行批处理12. Color-Based Segmentation Using the L*a*b* Color Space基于Lab色彩空间的彩色图像分割13. Color-Based Segmentation Using K-Means Clustering 基于K-均值聚类的彩色图像分割14. Detecting a Cell Using Image Segmentation使用图像分割技术来检测细胞15. Finding V egetation in a Multispectral Image多谱图像(卫星遥感)上的农作物区域分割16. Marker-Controlled Watershed Segmentation基于标记控制的分水岭分割算法17. Texture Segmentation Using Texture Filters基于纹理滤波器的纹理图像分割图像几何变换18. Creating a Gallery of Transformed Images常见的图像几何变换简介19. Exploring a Conformal Mapping图像的保角变换(共形映射)20. Extracting Slices from a 3-Dimensional MRI Data Set 如何从3维MRI数据集中提取切片图21. Padding and Shearing an Image Simultaneously图像的剪切变换和填充操作图像的测量22. Finding the Length of a Pendulum in Motion从单摆图像序列中计算摆长23. Granulometry of Snowflakes使用形态学方法对雪花的颗粒度进行测量24. Identifying Round Objects在图像中计算物体的“似圆度”25. Measuring Angle of Intersection在图像中计算钢梁的交叉角度26. Measuring the Radius of a Roll of Tape如何用图像方法测量胶带的半径图像的Radon变换27. Reconstructing an Image from Projection Data基于拉东(Radon)变换的CT图像重建视频检测和跟踪28. Abandoned Object Detection遗弃物体检测技术29. Motion Detection基于SAD的运动检测系统30. Lane Departure Warning System车道偏离预警系统31. Lane Detection and Tracking基于Hough变换的车道检测和跟踪32. Traffic Warning Sign Recognition交通警示牌自动识别技术33. People Tracking基于背景差分的行人检测技术34. Color Segmentation基于色彩分割的人体检测35. Tracking Cars Using Background Estimation 基于背景估计的汽车检测36. Tracking Cars Using Optical Flow基于光流法的汽车检测37. Surveillance Recording基于主帧检测的监控记录技术38. Pattern Matching基于模板匹配的PCB检测系统压缩技术39. V ideo Compression基于DCT变换的视频压缩技术40. Image Compression基于DCT变换的图像压缩技术视频分析技术41. Histogram Display图像直方图的实时显示42. Concentricity Inspection光纤的同心性检测系统43. Edge Detection边缘检测技术简介44. V ideo Focus Assessment视频自动聚焦参量计算视频增强45. V ideo Stabilization基于模板的电子稳像技术46. Periodic Noise Reduction针对周期噪声的图像降噪算法47. Histogram Equalization基于直方图均衡的图像增强48. Rotation Correction基于Hough变换的旋转图像校正基于形态学的视频分割技术49. Cell Counting细胞自动计数系统50. Feature Extraction如何自动计算视频中扇形的数目51. Object Counting如何自动计算订书钉的数目52. Object Extraction and Replacement视频目标的实时提取和替换视频回放处理53. Continuous Image Rotation图像连续旋转效果的实现54. Projecting Videos onto a Rotating Cube 如何将视频投影到旋转的立方体上55. V isual Effects图像浮雕效果的实现56. Picture in Picture画中画效果的实现57. Panorama Creation全景照片技术58. Bouncing Balls如何在图像上叠加动画《数字图像处理》结课测试报告规范1.内容要求(1)本报告(论文)的名字,系统功能、实现了什么结果。

A Fast and Accurate Plane Detection Algorithm for Large Noisy Point Clouds Using Filtered Normals

A Fast and Accurate Plane Detection Algorithm for Large Noisy Point Clouds Using Filtered Normals

A Fast and Accurate Plane Detection Algorithm for Large Noisy Point CloudsUsing Filtered Normals and Voxel GrowingJean-Emmanuel DeschaudFranc¸ois GouletteMines ParisTech,CAOR-Centre de Robotique,Math´e matiques et Syst`e mes60Boulevard Saint-Michel75272Paris Cedex06jean-emmanuel.deschaud@mines-paristech.fr francois.goulette@mines-paristech.frAbstractWith the improvement of3D scanners,we produce point clouds with more and more points often exceeding millions of points.Then we need a fast and accurate plane detection algorithm to reduce data size.In this article,we present a fast and accurate algorithm to detect planes in unorganized point clouds usingfiltered normals and voxel growing.Our work is based on afirst step in estimating better normals at the data points,even in the presence of noise.In a second step,we compute a score of local plane in each point.Then, we select the best local seed plane and in a third step start a fast and robust region growing by voxels we call voxel growing.We have evaluated and tested our algorithm on different kinds of point cloud and compared its performance to other algorithms.1.IntroductionWith the growing availability of3D scanners,we are now able to produce large datasets with millions of points.It is necessary to reduce data size,to decrease the noise and at same time to increase the quality of the model.It is in-teresting to model planar regions of these point clouds by planes.In fact,plane detection is generally afirst step of segmentation but it can be used for many applications.It is useful in computer graphics to model the environnement with basic geometry.It is used for example in modeling to detect building facades before classification.Robots do Si-multaneous Localization and Mapping(SLAM)by detect-ing planes of the environment.In our laboratory,we wanted to detect small and large building planes in point clouds of urban environments with millions of points for modeling. As mentioned in[6],the accuracy of the plane detection is important for after-steps of the modeling pipeline.We also want to be fast to be able to process point clouds with mil-lions of points.We present a novel algorithm based on re-gion growing with improvements in normal estimation and growing process.For our method,we are generic to work on different kinds of data like point clouds fromfixed scan-ner or from Mobile Mapping Systems(MMS).We also aim at detecting building facades in urban point clouds or little planes like doors,even in very large data sets.Our input is an unorganized noisy point cloud and with only three”in-tuitive”parameters,we generate a set of connected compo-nents of planar regions.We evaluate our method as well as explain and analyse the significance of each parameter. 2.Previous WorksAlthough there are many methods of segmentation in range images like in[10]or in[3],three have been thor-oughly studied for3D point clouds:region-growing, hough-transform from[14]and Random Sample Consen-sus(RANSAC)from[9].The application of recognising structures in urban laser point clouds is frequent in literature.Bauer in[4]and Boulaassal in[5]detect facades in dense3D point cloud by a RANSAC algorithm.V osselman in[23]reviews sur-face growing and3D hough transform techniques to de-tect geometric shapes.Tarsh-Kurdi in[22]detect roof planes in3D building point cloud by comparing results on hough-transform and RANSAC algorithm.They found that RANSAC is more efficient than thefirst one.Chao Chen in[6]and Yu in[25]present algorithms of segmentation in range images for the same application of detecting planar regions in an urban scene.The method in[6]is based on a region growing algorithm in range images and merges re-sults in one labelled3D point cloud.[25]uses a method different from the three we have cited:they extract a hi-erarchical subdivision of the input image built like a graph where leaf nodes represent planar regions.There are also other methods like bayesian techniques. In[16]and[8],they obtain smoothed surface from noisy point clouds with objects modeled by probability distribu-tions and it seems possible to extend this idea to point cloud segmentation.But techniques based on bayesian statistics need to optimize global statistical model and then it is diffi-cult to process points cloud larger than one million points.We present below an analysis of the two main methods used in literature:RANSAC and region-growing.Hough-transform algorithm is too time consuming for our applica-tion.To compare the complexity of the algorithm,we take a point cloud of size N with only one plane P of size n.We suppose that we want to detect this plane P and we define n min the minimum size of the plane we want to detect.The size of a plane is the area of the plane.If the data density is uniform in the point cloud then the size of a plane can be specified by its number of points.2.1.RANSACRANSAC is an algorithm initially developped by Fis-chler and Bolles in[9]that allows thefitting of models with-out trying all possibilities.RANSAC is based on the prob-ability to detect a model using the minimal set required to estimate the model.To detect a plane with RANSAC,we choose3random points(enough to estimate a plane).We compute the plane parameters with these3points.Then a score function is used to determine how the model is good for the remaining ually,the score is the number of points belonging to the plane.With noise,a point belongs to a plane if the distance from the point to the plane is less than a parameter γ.In the end,we keep the plane with the best score.Theprobability of getting the plane in thefirst trial is p=(nN )3.Therefore the probability to get it in T trials is p=1−(1−(nN )3)ing equation1and supposing n minN1,we know the number T min of minimal trials to have a probability p t to get planes of size at least n min:T min=log(1−p t)log(1−(n minN))≈log(11−p t)(Nn min)3.(1)For each trial,we test all data points to compute the score of a plane.The RANSAC algorithm complexity lies inO(N(Nn min )3)when n minN1and T min→0whenn min→N.Then RANSAC is very efficient in detecting large planes in noisy point clouds i.e.when the ratio n minN is 1but very slow to detect small planes in large pointclouds i.e.when n minN 1.After selecting the best model,another step is to extract the largest connected component of each plane.Connnected components mean that the min-imum distance between each point of the plane and others points is smaller(for distance)than afixed parameter.Schnabel et al.[20]bring two optimizations to RANSAC:the points selection is done locally and the score function has been improved.An octree isfirst created from point cloud.Points used to estimate plane parameters are chosen locally at a random depth of the octree.The score function is also different from RANSAC:instead of testing all points for one model,they test only a random subset and find the score by interpolation.The algorithm complexity lies in O(Nr4Ndn min)where r is the number of random subsets for the score function and d is the maximum octree depth. Their algorithm improves the planes detection speed but its complexity lies in O(N2)and it becomes slow on large data sets.And again we have to extract the largest connected component of each plane.2.2.Region GrowingRegion Growing algorithms work well in range images like in[18].The principle of region growing is to start with a seed region and to grow it by neighborhood when the neighbors satisfy some conditions.In range images,we have the neighbors of each point with pixel coordinates.In case of unorganized3D data,there is no information about the neighborhood in the data structure.The most common method to compute neighbors in3D is to compute a Kd-tree to search k nearest neighbors.The creation of a Kd-tree lies in O(NlogN)and the search of k nearest neighbors of one point lies in O(logN).The advantage of these region growing methods is that they are fast when there are many planes to extract,robust to noise and extract the largest con-nected component immediately.But they only use the dis-tance from point to plane to extract planes and like we will see later,it is not accurate enough to detect correct planar regions.Rabbani et al.[19]developped a method of smooth area detection that can be used for plane detection.Theyfirst estimate the normal of each point like in[13].The point with the minimum residual starts the region growing.They test k nearest neighbors of the last point added:if the an-gle between the normal of the point and the current normal of the plane is smaller than a parameterαthen they add this point to the smooth region.With Kd-tree for k nearest neighbors,the algorithm complexity is in O(N+nlogN). The complexity seems to be low but in worst case,when nN1,example for facade detection in point clouds,the complexity becomes O(NlogN).3.Voxel Growing3.1.OverviewIn this article,we present a new algorithm adapted to large data sets of unorganized3D points and optimized to be accurate and fast.Our plane detection method works in three steps.In thefirst part,we compute a better esti-mation of the normal in each point by afiltered weighted planefitting.In a second step,we compute the score of lo-cal planarity in each point.We select the best seed point that represents a good seed plane and in the third part,we grow this seed plane by adding all points close to the plane.Thegrowing step is based on a voxel growing algorithm.The filtered normals,the score function and the voxel growing are innovative contributions of our method.As an input,we need dense point clouds related to the level of detail we want to detect.As an output,we produce connected components of planes in the point cloud.This notion of connected components is linked to the data den-sity.With our method,the connected components of planes detected are linked to the parameter d of the voxel grid.Our method has 3”intuitive”parameters :d ,area min and γ.”intuitive”because there are linked to physical mea-surements.d is the voxel size used in voxel growing and also represents the connectivity of points in detected planes.γis the maximum distance between the point of a plane and the plane model,represents the plane thickness and is linked to the point cloud noise.area min represents the minimum area of planes we want to keep.3.2.Details3.2.1Local Density of Point CloudsIn a first step,we compute the local density of point clouds like in [17].For that,we find the radius r i of the sphere containing the k nearest neighbors of point i .Then we cal-culate ρi =kπr 2i.In our experiments,we find that k =50is a good number of neighbors.It is important to know the lo-cal density because many laser point clouds are made with a fixed resolution angle scanner and are therefore not evenly distributed.We use the local density in section 3.2.3for the score calculation.3.2.2Filtered Normal EstimationNormal estimation is an important part of our algorithm.The paper [7]presents and compares three normal estima-tion methods.They conclude that the weighted plane fit-ting or WPF is the fastest and the most accurate for large point clouds.WPF is an idea of Pauly and al.in [17]that the fitting plane of a point p must take into consider-ation the nearby points more than other distant ones.The normal least square is explained in [21]and is the mini-mum of ki =1(n p ·p i +d )2.The WPF is the minimum of ki =1ωi (n p ·p i +d )2where ωi =θ( p i −p )and θ(r )=e −2r 2r2i .For solving n p ,we compute the eigenvec-tor corresponding to the smallest eigenvalue of the weightedcovariance matrix C w = ki =1ωi t (p i −b w )(p i −b w )where b w is the weighted barycenter.For the three methods ex-plained in [7],we get a good approximation of normals in smooth area but we have errors in sharp corners.In fig-ure 1,we have tested the weighted normal estimation on two planes with uniform noise and forming an angle of 90˚.We can see that the normal is not correct on the corners of the planes and in the red circle.To improve the normal calculation,that improves the plane detection especially on borders of planes,we propose a filtering process in two phases.In a first step,we com-pute the weighted normals (WPF)of each point like we de-scribed it above by minimizing ki =1ωi (n p ·p i +d )2.In a second step,we compute the filtered normal by us-ing an adaptive local neighborhood.We compute the new weighted normal with the same sum minimization but keep-ing only points of the neighborhood whose normals from the first step satisfy |n p ·n i |>cos (α).With this filtering step,we have the same results in smooth areas and better results in sharp corners.We called our normal estimation filtered weighted plane fitting(FWPF).Figure 1.Weighted normal estimation of two planes with uniform noise and with 90˚angle between them.We have tested our normal estimation by computing nor-mals on synthetic data with two planes and different angles between them and with different values of the parameter α.We can see in figure 2the mean error on normal estimation for WPF and FWPF with α=20˚,30˚,40˚and 90˚.Us-ing α=90˚is the same as not doing the filtering step.We see on Figure 2that α=20˚gives smaller error in normal estimation when angles between planes is smaller than 60˚and α=30˚gives best results when angle between planes is greater than 60˚.We have considered the value α=30˚as the best results because it gives the smaller mean error in normal estimation when angle between planes vary from 20˚to 90˚.Figure 3shows the normals of the planes with 90˚angle and better results in the red circle (normals are 90˚with the plane).3.2.3The score of local planarityIn many region growing algorithms,the criteria used for the score of the local fitting plane is the residual,like in [18]or [19],i.e.the sum of the square of distance from points to the plane.We have a different score function to estimate local planarity.For that,we first compute the neighbors N i of a point p with points i whose normals n i are close toFigure parison of mean error in normal estimation of two planes with α=20˚,30˚,40˚and 90˚(=Nofiltering).Figure 3.Filtered Weighted normal estimation of two planes with uniform noise and with 90˚angle between them (α=30˚).the normal n p .More precisely,we compute N i ={p in k neighbors of i/|n i ·n p |>cos (α)}.It is a way to keep only the points which are probably on the local plane before the least square fitting.Then,we compute the local plane fitting of point p with N i neighbors by least squares like in [21].The set N i is a subset of N i of points belonging to the plane,i.e.the points for which the distance to the local plane is smaller than the parameter γ(to consider the noise).The score s of the local plane is the area of the local plane,i.e.the number of points ”in”the plane divided by the localdensity ρi (seen in section 3.2.1):the score s =card (N i)ρi.We take into consideration the area of the local plane as the score function and not the number of points or the residual in order to be more robust to the sampling distribution.3.2.4Voxel decompositionWe use a data structure that is the core of our region growing method.It is a voxel grid that speeds up the plane detection process.V oxels are small cubes of length d that partition the point cloud space.Every point of data belongs to a voxel and a voxel contains a list of points.We use the Octree Class Template in [2]to compute an Octree of the point cloud.The leaf nodes of the graph built are voxels of size d .Once the voxel grid has been computed,we start the plane detection algorithm.3.2.5Voxel GrowingWith the estimator of local planarity,we take the point p with the best score,i.e.the point with the maximum area of local plane.We have the model parameters of this best seed plane and we start with an empty set E of points belonging to the plane.The initial point p is in a voxel v 0.All the points in the initial voxel v 0for which the distance from the seed plane is less than γare added to the set E .Then,we compute new plane parameters by least square refitting with set E .Instead of growing with k nearest neighbors,we grow with voxels.Hence we test points in 26voxel neigh-bors.This is a way to search the neighborhood in con-stant time instead of O (logN )for each neighbor like with Kd-tree.In a neighbor voxel,we add to E the points for which the distance to the current plane is smaller than γand the angle between the normal computed in each point and the normal of the plane is smaller than a parameter α:|cos (n p ,n P )|>cos (α)where n p is the normal of the point p and n P is the normal of the plane P .We have tested different values of αand we empirically found that 30˚is a good value for all point clouds.If we added at least one point in E for this voxel,we compute new plane parameters from E by least square fitting and we test its 26voxel neigh-bors.It is important to perform plane least square fitting in each voxel adding because the seed plane model is not good enough with noise to be used in all voxel growing,but only in surrounding voxels.This growing process is faster than classical region growing because we do not compute least square for each point added but only for each voxel added.The least square fitting step must be computed very fast.We use the same method as explained in [18]with incre-mental update of the barycenter b and covariance matrix C like equation 2.We know with [21]that the barycen-ter b belongs to the least square plane and that the normal of the least square plane n P is the eigenvector of the smallest eigenvalue of C .b0=03x1C0=03x3.b n+1=1n+1(nb n+p n+1).C n+1=C n+nn+1t(pn+1−b n)(p n+1−b n).(2)where C n is the covariance matrix of a set of n points,b n is the barycenter vector of a set of n points and p n+1is the (n+1)point vector added to the set.This voxel growing method leads to a connected com-ponent set E because the points have been added by con-nected voxels.In our case,the minimum distance between one point and E is less than parameter d of our voxel grid. That is why the parameter d also represents the connectivity of points in detected planes.3.2.6Plane DetectionTo get all planes with an area of at least area min in the point cloud,we repeat these steps(best local seed plane choice and voxel growing)with all points by descending order of their score.Once we have a set E,whose area is bigger than area min,we keep it and classify all points in E.4.Results and Discussion4.1.Benchmark analysisTo test the improvements of our method,we have em-ployed the comparative framework of[12]based on range images.For that,we have converted all images into3D point clouds.All Point Clouds created have260k points. After our segmentation,we project labelled points on a seg-mented image and compare with the ground truth image. We have chosen our three parameters d,area min andγby optimizing the result of the10perceptron training image segmentation(the perceptron is portable scanner that pro-duces a range image of its environment).Bests results have been obtained with area min=200,γ=5and d=8 (units are not provided in the benchmark).We show the re-sults of the30perceptron images segmentation in table1. GT Regions are the mean number of ground truth planes over the30ground truth range images.Correct detection, over-segmentation,under-segmentation,missed and noise are the mean number of correct,over,under,missed and noised planes detected by methods.The tolerance80%is the minimum percentage of points we must have detected comparing to the ground truth to have a correct detection. More details are in[12].UE is a method from[12],UFPR is a method from[10]. It is important to notice that UE and UFPR are range image methods and our method is not well suited for range images but3D Point Cloud.Nevertheless,it is a good benchmark for comparison and we see in table1that the accuracy of our method is very close to the state of the art in range image segmentation.To evaluate the different improvements of our algorithm, we have tested different variants of our method.We have tested our method without normals(only with distance from points to plane),without voxel growing(with a classical region growing by k neighbors),without our FWPF nor-mal estimation(with WPF normal estimation),without our score function(with residual score function).The compari-son is visible on table2.We can see the difference of time computing between region growing and voxel growing.We have tested our algorithm with and without normals and we found that the accuracy cannot be achieved whithout normal computation.There is also a big difference in the correct de-tection between WPF and our FWPF normal estimation as we can see in thefigure4.Our FWPF normal brings a real improvement in border estimation of planes.Black points in thefigure are non classifiedpoints.Figure5.Correct Detection of our segmentation algorithm when the voxel size d changes.We would like to discuss the influence of parameters on our algorithm.We have three parameters:area min,which represents the minimum area of the plane we want to keep,γ,which represents the thickness of the plane(it is gener-aly closely tied to the noise in the point cloud and espe-cially the standard deviationσof the noise)and d,which is the minimum distance from a point to the rest of the plane. These three parameters depend on the point cloud features and the desired segmentation.For example,if we have a lot of noise,we must choose a highγvalue.If we want to detect only large planes,we set a large area min value.We also focus our analysis on the robustess of the voxel size d in our algorithm,i.e.the ratio of points vs voxels.We can see infigure5the variation of the correct detection when we change the value of d.The method seems to be robust when d is between4and10but the quality decreases when d is over10.It is due to the fact that for a large voxel size d,some planes from different objects are merged into one plane.GT Regions Correct Over-Under-Missed Noise Duration(in s)detection segmentation segmentationUE14.610.00.20.3 3.8 2.1-UFPR14.611.00.30.1 3.0 2.5-Our method14.610.90.20.1 3.30.7308Table1.Average results of different segmenters at80%compare tolerance.GT Regions Correct Over-Under-Missed Noise Duration(in s) Our method detection segmentation segmentationwithout normals14.6 5.670.10.19.4 6.570 without voxel growing14.610.70.20.1 3.40.8605 without FWPF14.69.30.20.1 5.0 1.9195 without our score function14.610.30.20.1 3.9 1.2308 with all improvements14.610.90.20.1 3.30.7308 Table2.Average results of variants of our segmenter at80%compare tolerance.4.1.1Large scale dataWe have tested our method on different kinds of data.We have segmented urban data infigure6from our Mobile Mapping System(MMS)described in[11].The mobile sys-tem generates10k pts/s with a density of50pts/m2and very noisy data(σ=0.3m).For this point cloud,we want to de-tect building facades.We have chosen area min=10m2, d=1m to have large connected components andγ=0.3m to cope with the noise.We have tested our method on point cloud from the Trim-ble VX scanner infigure7.It is a point cloud of size40k points with only20pts/m2with less noise because it is a fixed scanner(σ=0.2m).In that case,we also wanted to detect building facades and keep the same parameters ex-ceptγ=0.2m because we had less noise.We see infig-ure7that we have detected two facades.By setting a larger voxel size d value like d=10m,we detect only one plane. We choose d like area min andγaccording to the desired segmentation and to the level of detail we want to extract from the point cloud.We also tested our algorithm on the point cloud from the LEICA Cyrax scanner infigure8.This point cloud has been taken from AIM@SHAPE repository[1].It is a very dense point cloud from multiplefixed position of scanner with about400pts/m2and very little noise(σ=0.02m). In this case,we wanted to detect all the little planes to model the church in planar regions.That is why we have chosen d=0.2m,area min=1m2andγ=0.02m.Infigures6,7and8,we have,on the left,input point cloud and on the right,we only keep points detected in a plane(planes are in random colors).The red points in thesefigures are seed plane points.We can see in thesefig-ures that planes are very well detected even with high noise. Table3show the information on point clouds,results with number of planes detected and duration of the algorithm.The time includes the computation of the FWPF normalsof the point cloud.We can see in table3that our algo-rithm performs linearly in time with respect to the numberof points.The choice of parameters will have little influence on time computing.The computation time is about one mil-lisecond per point whatever the size of the point cloud(we used a PC with QuadCore Q9300and2Go of RAM).The algorithm has been implented using only one thread andin-core processing.Our goal is to compare the improve-ment of plane detection between classical region growing and our region growing with better normals for more ac-curate planes and voxel growing for faster detection.Our method seems to be compatible with out-of-core implemen-tation like described in[24]or in[15].MMS Street VX Street Church Size(points)398k42k7.6MMean Density50pts/m220pts/m2400pts/m2 Number of Planes202142Total Duration452s33s6900sTime/point 1ms 1ms 1msTable3.Results on different data.5.ConclusionIn this article,we have proposed a new method of plane detection that is fast and accurate even in presence of noise. We demonstrate its efficiency with different kinds of data and its speed in large data sets with millions of points.Our voxel growing method has a complexity of O(N)and it is able to detect large and small planes in very large data sets and can extract them directly in connected components.Figure 4.Ground truth,Our Segmentation without and with filterednormals.Figure 6.Planes detection in street point cloud generated by MMS (d =1m,area min =10m 2,γ=0.3m ).References[1]Aim@shape repository /.6[2]Octree class template /code/octree.html.4[3] A.Bab-Hadiashar and N.Gheissari.Range image segmen-tation using surface selection criterion.2006.IEEE Trans-actions on Image Processing.1[4]J.Bauer,K.Karner,K.Schindler,A.Klaus,and C.Zach.Segmentation of building models from dense 3d point-clouds.2003.Workshop of the Austrian Association for Pattern Recognition.1[5]H.Boulaassal,ndes,P.Grussenmeyer,and F.Tarsha-Kurdi.Automatic segmentation of building facades using terrestrial laser data.2007.ISPRS Workshop on Laser Scan-ning.1[6] C.C.Chen and I.Stamos.Range image segmentationfor modeling and object detection in urban scenes.2007.3DIM2007.1[7]T.K.Dey,G.Li,and J.Sun.Normal estimation for pointclouds:A comparison study for a voronoi based method.2005.Eurographics on Symposium on Point-Based Graph-ics.3[8]J.R.Diebel,S.Thrun,and M.Brunig.A bayesian methodfor probable surface reconstruction and decimation.2006.ACM Transactions on Graphics (TOG).1[9]M.A.Fischler and R.C.Bolles.Random sample consen-sus:A paradigm for model fitting with applications to image analysis and automated munications of the ACM.1,2[10]P.F.U.Gotardo,O.R.P.Bellon,and L.Silva.Range imagesegmentation by surface extraction using an improved robust estimator.2003.Proceedings of Computer Vision and Pat-tern Recognition.1,5[11] F.Goulette,F.Nashashibi,I.Abuhadrous,S.Ammoun,andurgeau.An integrated on-board laser range sensing sys-tem for on-the-way city and road modelling.2007.Interna-tional Archives of the Photogrammetry,Remote Sensing and Spacial Information Sciences.6[12] A.Hoover,G.Jean-Baptiste,and al.An experimental com-parison of range image segmentation algorithms.1996.IEEE Transactions on Pattern Analysis and Machine Intelligence.5[13]H.Hoppe,T.DeRose,T.Duchamp,J.McDonald,andW.Stuetzle.Surface reconstruction from unorganized points.1992.International Conference on Computer Graphics and Interactive Techniques.2[14]P.Hough.Method and means for recognizing complex pat-terns.1962.In US Patent.1[15]M.Isenburg,P.Lindstrom,S.Gumhold,and J.Snoeyink.Large mesh simplification using processing sequences.2003.。

基于u-net的“高分五号”卫星高光谱图像土地类型分类

基于u-net的“高分五号”卫星高光谱图像土地类型分类

第40卷第6期航天返回与遥感2019年12月SPACECRAFT RECOVERY & REMOTE SENSING99基于U-net的“高分五号”卫星高光谱图像土地类型分类孙晓敏1,2,3郑利娟4吴军5陈前6徐崇斌1马杨1陈震1(1 北京空间机电研究所,北京 100094)(2 北京航天创智科技有限公司,北京 100076)(3 北京市航空智能遥感装备工程技术研究中心,北京 100094)(4 自然资源部国土卫星遥感应用中心,北京 100048)(5 国网湖北省电力有限公司直流运检公司,宜昌 443000)(6 中国资源卫星应用中心,北京 100094)摘要“高分五号”卫星是世界首颗实现对大气和陆地综合观测的全谱段高光谱卫星,对于土地利用类型分类具有重要的应用价值,如何利用深度学习技术开展高光谱图像分类是当前研究的热点问题。

深度学习中的语义分割方法在地面场景的图像中已经获得较好的应用,但是对于高光谱遥感图像的精度和适用性较差,无法准确获得精确的分类结果。

文章采用U-net模型开展高光谱土地利用类型分类研究,首先基于“高分五号”卫星高光谱数据,构建样本数据集,然后训练分类模型,进行土地利用类型分类,探讨语义分割方法在高分五号高光谱数据上的应用能力。

结果表明,采用深度学习中的语义分割方法能够有效提高精度水平,U-net模型的整体分类精度为0.9357,Kappa系数达到0.92,均高于SVM方法和CNN方法。

采用深度学习中的语义分割方法,可以为“高分五号”高光谱数据的土地利用分类提供技术支撑,有效提升“高分五号”卫星的应用能力。

关键词U-网络模型深度学习高光谱图像土地利用分类高分五号卫星应用中图分类号: TP75文献标志码: A 文章编号: 1009-8518(2019)06-0099-08DOI: 10.3969/j.issn.1009-8518.2019.06.012Land Classification of GF-5 Satellite HyperspectralImages Using U-net ModelSUN Xiaomin1,2,3 ZHENG Lijuan4 WU Jun5 CHEN Qian6 XU Chongbin1 MA Yang1CHEN Zhen1(1 Beijing Institute of Space Mechanics & Electricity, Beijing, 100094, China)(2 Beijing Aerospace Innovative Intelligence Science and Technology Co., Ltd, Beijing, 100076, China)(3 Beijing Engineering Technology Research Center of Aerial Intelligence Remote Sensing Equipments, Beijing, 100094, China)(4 Land Satellite Remote Sensing Application Center, Ministry of Natural Resources of P R China, Beijing, 100048, China)(5 State Grid Hubei DC Operation & Mainteance Company, Yichang, 443000, China)(6 China Centre for Resources Satellite Data and Application, Beijing, 100094, China)Abstract GF-5 satellite is the world's first full-spectrum hyper-spectral imagery satellite to observe the收稿日期:2019-07-12引用格式:孙晓敏, 郑利娟, 吴军, 等. 基于U-net的“高分五号”卫星高光谱图像土地类型分类[J]. 航天返回与遥感, 2019, 40(6): 99-106.SUN Xiaomin, ZHENG Lijuan, WU Jun, et al. Land Classification of GF-5 Satellite Hyperspectral Images Using U-net Model[J]. Spacecraft Recovery & Remote Sensing, 2019, 40(6): 99-106. (in Chinese)100航天返回与遥感2019年第40卷Earth’s atmosphere and surface. It is important for the classification of land use types. How to use deep learning technology to carry out hyperspectral image classification is a hot issue in current research. Semantic segmentation method in depth learning has been well applied in the image of ground scene, but the accuracy and applicability of hyperspectral remote sensing images are relatively poor, and the accurate classification results are difficult to be obtained. In this paper, the U-net model is used to study the classification of hyperspectral land use types. Firstly, based on the hyperspectral data of GF-5 satellite, the sample data set is constructed, then the classification model is trained, the land use type classification is carried out, and the application ability of semantic segmentation method on hyperspectral data of GF-5 satellite is discussed. The results show that the semantic segmentation method in deep learning can effectively improve the accuracy level. The overall classification accuracy of the U-net model is 0.9357, and the Kappa coefficient is 0.92, which is higher than the SVM method and CNN method. Using the semantic segmentation method in deep learning, it can provide technical support for land use classification of GF-5 satellite hyperspectral data, and effectively improve the application ability of GF-5 satellite.Keywords U-net model; deep learning; land use classification; hyper-spectral images; GF-5; satellite applications0引言“高分五号”(GF-5)卫星是我国高分辨率地球观测系统的最重要的遥感卫星之一,是我国实现高光谱分辨率对地观测能力的重要标志[1]。

基于高光谱成像的苹果病害无损检测方法

基于高光谱成像的苹果病害无损检测方法

基于高光谱成像的苹果病害无损检测方法刘思伽;田有文;冯迪;张芳;崔博【摘要】Disease is easy to occur in apple fruit. Traditional detection of apple disease is not adapted to the requirement of apple grading on-line detection. In order to achieve the fast, effective online detection for the disease apple, hyperspectral imaging was adopted to study the nondestructive detection of the anthracnose, bitter pox disease and black fruit rot and leaf spot disease in Hanfu apple. According to the relative reflectance spectrum difference between disease area and normal area, the improved manifold distance method was proposed. The total improved manifold distance L value was comprehensive calculated by the relative reflectance spectra of the disease and normal area, disease with stem/calyx area, normal and stem/calyx area. So three feature wavelengths were selected respectively from the whole band wavelength, 700, 765, 904nm. In order to get the mask image, the image of the characteristic wave band at 700 nm was threshold segmented. The interested area was extracted after secondary threshold segmentation of the mask image. The relative reflectance spectra of the three characteristic wave bands were combined, respectively, as the BP neural network input vector, to detect whether apple fruit was diseased. Finally, the relative reflectance spectra under 700 nm to 904 nm band were selected as the best combination by comparing the detection results. A recognition rate of the normal apples and diseased apples respectively were 96.25%. Results showed that the twocharacteristics of band obtained by hyperspectral imaging technology can effectively detect disease for apple and provide the reference for the development of multispectral imaging of appleˊs quality detection and classification system.%苹果果实易发生病害,传统的苹果病害的检测不适应苹果分级在线检测的要求。

HSI Classification by Exploiting the Spectral-Spatial Correlations in the Sparse Coefficients

HSI Classification by Exploiting the Spectral-Spatial Correlations in the Sparse Coefficients

Hyperspectral Image Classification by Exploiting the Spectral-Spatial Correlations in the Sparse CoefficientsDan Li u, Sh u tao Li, and Ley u an Fan gColle g e of Electrical and Information En g ineerin g,H u nan University, Chan g sha, 410012, China{liudan1,shutao_li,leyuan_fang}@ Abstract. This paper proposes a novel hyperspectral ima g e (HSI) classificationmethod based on sparse model, which incorporates the spectral and spatial in-formation of the sparse coefficient. Firstly, a sparse dictionary is b u ilt by u sin gthe trainin g sampl es and the sparse coefficient is obtained thro ug h the sparserepresentation method. Secondly, a probability map for each class is establishedby s u mmin g the sparse coefficients of each class. Thirdly, the mean filterin g isapplied on each probability map to exploit the spatial information. Finally, wecompare the probabil ity map to find the maxim u m probabil ity for each pixeland then determine the class label of each pixel. Experimental res u l ts demon-strate the effectiveness of the proposed method.Keywords: Hyperspectral ima g e classification, sparse representation, spectral-spatial information, mean filter.1IntroductionHyperspectral ima g e (HSI) is formed by tens to h u ndreds of contin u o u s and s u bdivided spectral bands while reflectin g interested tar g et areas sim u ltaneo u sl y. In HSI, different materials have different spectral information, which can be u sed for classification.Many m u ltispectral ima g e classification methods, s u ch as s u pport vector machines (SVMs) [1], [2], ne u ral network [3], and adaptive artificial imm u ne network [4], have been applied to HSI classification. Generally, these methods have obtained g ood per-formance.Researchers show that HSI contains rich spatial information and the pixel s in a small nei g hborhood have similar spectral characteristics. If the pixels are in a small nei g hbor, they sho u ld belon g to the same material. Therefore, Some methods [5], [6], [7] have combined spectral information and spatial information, and the classification acc u racy has been improved. In partic u l ar, the se g mentation based method [8] first se g ment the HSI into many local re g ion with similar spectral characteristics and then cl assify each re g ion. After u sin g the spatial information, the cl assifiers can obtain improved performance.Recentl y, sparse representation has become a powerf u l tool to solve some prob-lems, s u ch as face reco g nition [9], tar g et detection [10], [11], remote sensin g ima g e S. Li et al. (Eds.): CC P R 2014, P art I, CCIS 483, pp. 151–158, 2014.© Sprin g er-Verla g Berlin Heidelber g 2014152 D. Li u , S. Li, and L. Fan gf u sion [12] and medical imag e reconstr u ction [13], [14]. Recently, the sparse repre-sentation method has also been extended to HSI classification [7], [15], [16]. Basical-l y, the previo u s sparse representation based HSI cl assification methods u til ize the reconstr u ction error for the classification. In this paper, we propose a novel method that can combines the spatial information and spectral information in the sparse coef-ficients for the cl assification. Firstl y, we u se the trainin g sampl es to constr u ct the trainin g dictionary and then u til ize the sim u ltaneo u s ortho g onal matchin g p u rs u it (SOM P ) to obtain the sparse coefficient of each spectral pixel. Differ from other sparse representation based methods which u ses the resid u al to determine the pixel ’s class, the proposed method first employs the coefficients to constr u ct several proba-bility maps. S u bseq u ently, we exploit the spatial information by filterin g every map and g ain a probabil ity map for each cl ass. Final l y, we can determine the pixel ’s class by comparin g the probability maps.The rest of this paper is constr u cted as follows. Section 2 introd u ces the proposed cl assification method. Section 3 shows the experimental res u lts and concl u sions are g iven in the section 4.2 The Proposed Classification MethodFi g . 1 shows the schematic of the proposed classification method. It is constr u cted by fo u r steps: Firstl y, the sparse representation method is adopted to obtain the sparse coefficients. Then, the coefficients bel on g in g to each cl ass are s u mmed to obtain probability map for each pixel. S u bseq u entl y, a mean filterin g is cond u cted on each probabil ity map to expl oit the spatial information. Final l y, cl assification is accom-plished by comparin g the maps. The details of each step are ill ustrated in the follows.x x x 4x x }}}Fig. 1. The scheme of the proposed classification methodHyperspectra l Ima g e Classification by Exploitin g the Spectral-Spatial Correlations 153 Step 1: In HSI, every spectral pixel can be re g arded as a vector i x and the trainin g pixels constr u ct a matrix =12n D [d ,d ,...,d ]which is called dictionary. Every pixel canbe represented by the dictionary.1212...n i i i n i i ααα=+++=x d d d D α (1)In the eq u ation (1), 12,,...,n d d d is cal l ed atom and 12[,,...,]n i i i i ααα=αis cal ed sparse coefficient vector. The sparse coefficient vector can be obtained by solvin g the optimization problem.200ˆar g min s u bject to i i i i K =−≤αx A αα (2)where 0K is the maxim u m val u e of the sparsity level. This optimization problem is aN P -hard and cannot be sol ved directl y. However, it can be sol ved by g reedy al g o-rithms approximately, s u ch as s u bspace p u rs u it (S P ) [17], ortho g onal matchin g p u r-s u it (OM P ) [18] and Sim u ltaneo u s OM P (SOM P ) [7]. In this paper, the SOM P isadopted to obtain the sparse coefficient vector ˆi αfor each spectral pixel i x . Step 2: In the sparse coefficient vector ˆi α, there are onl y a few nonzero sparse coefficients. The lar g er the nonzero coefficients val u es in one specific class, the more probability the test pixel belon g s to this class. We denote the nonzero coefficients in one cl ass as the ,i m α, where {1,2,...,}m M ∈, and M is the total n u mber of cl asses.Then, we s u m the nonzero coefficients ,i m αfor each class of each spectral pixel,(),,s u m ,{1,2,...,},and {1,2,...,}s um i m i m m M i N =∈∈αα (3) where N is the total n u mber of spectral pixels in the HSI. In each class, the s u mmed coefficients ,s um i m αfor al l the spectral pixel s in the HSI can constr u ct one probabil itymap m z .Step 3: As disc u ssed above, one coefficient in a class probability map m z can be re-g arded as the l ikel ihood for the correspondin g pixel bel on g in g to this cl ass. If the probability map m z is directly u sed for determinin g the class of each pixel, the spatialinformation in the probability map is not exploited. To exploit the spatial information, a mean filterin g operation is cond u cted on each m z ,()meanfilterin g ,{1,2,...,}m e an f m m z z m M =∈ (4)where the window for mean operation is selected to 3×3.Step 4: the cl ass l abel of each pixel i x is obtained by comparin g the coefficients in the fil tered probabil ity maps,,1,...,ˆmax (),{1,2,...,}m e an f i m i i m Mm z i N ==∈x (5) where max is the operation to comp u te the max coefficient amon g different maps.154 D.Li u, S. Li, and L. Fan g3Experimental ResultsThis section tests the effectiveness of the proposed classification method on two real HSIs (Indian pines and Salinas scene). The classification res u lts of the proposed me-thod are compared with those obtained by SVM [19], SVM-CK [20], OM P [7] and SOM P [7]. SVM [19] is desi g ned for the classification of the spectral pixel witho u t u tilizin g the spatial information. SVM-CK [20] is a method that incorporates spatial information via a composite kernel. OM P and SOM P are two sparse representation based methods.In o u r first experiment, we u sed Airborne Visible/Infrared Ima g in g Spectrometer (AVIRIS) ima g e Indian pines as testin g HSI. This ima g e is a widely u sed data set and was taken over Indiana’s Indian P ine test site in J u ne 1992. The Indian P ines has a size of 145×145×220, with 220 spectral bands. Beca u se 20 bands is water absorp-tion, these bands are removed. There are 16 g ro u nd-tr u th classes and the size is from 20 to 2455 pixels (the total pixels are 10249).We chose 10% of the samples for each class as trainin g sample and the remainder as testin g sampl es. For each method, we did five experiments and avera g ed the re-s u l ts. The n u mber of the trainin g sample and the testin g sample is presented in Table 1.In this table, we can see the overall acc u racy (OA), avera g e acc u racy (AA) and the kappa coefficient by u sin g different methods (the SOM P-P is denoted as o u r method). Table 1. Trainin g sets, testin g sets and cl assification acc u racy (%) obtained from different methods for the Indian P ines ima g el ass Train Test SVM SVM-CK OM P SOM P SOM P-P Cl fa 6 40 77.73 91.25 55.1292.26 95.04Al faCorn-N 144 1284 77.35 92.79 61.60 93.46 97.77Corn-M 84 746 78.56 93.98 58.62 90.22 97.4042.2187.32 95.1168.75Corn 24 21387.28Grass-M 50 433 88.87 94.90 87.29 95.20 94.04Grass-T 75 655 89.12 99.51 95.30 96.12 96.57 Grass-P 3 25 95.37 85.20 85.20 87.10 87.14Hay-W 49 429 95.09 99.91 96.44 99.10 99.8767.65Oats 2 1883.33 36.67 55.78 0Soybean-N 97 875 78.64 90.33 71.10 93.45 93.47Soybean-M 247 2208 81.19 96.25 74.11 95.10 99.20Soybean-C 62 531 79.74 89.04 51.05 87.49 97.61Wheat 22 183 92.26 99.07 96.85 88.20 97.76 Woods 130 1135 92.72 98.63 91.85 99.00 100B u ildin g s 38 348 69.79 92.64 41.67 83.05 97.7291.51 99.35stone 10 83 97.96 90.24 91.9093.6673.3894.82OA - -82.9197.4983.17AA - -92.77 71.06 89.83 91.010.6960.9310.8050.941k - -0.971Hyperspectra l Ima g e Classification by Exploitin g the Spectral-Spatial Correlations 155 The Table 1 shows the trainin g sets, testin g sets and classification maps obtained by SVM, SVM-CK, OM P, SOM P and SOM P-P and the res u lt is the avera g e of five experiments. From the Tabl e 1, we can see that o u r al g orithm has the best perfor-mance in terms of overall acc u racy and kappa coefficient. As for its avera g e acc u racy, it is only a little worse than the classifier SVM-CK.(a)(b)(c)(d)(e)(f)(g)Fig. 2. Indian P ines: (a) Train samples, (b)Test samples, and the classification res u lts obtainedby (c) SVM, (d) SVM-CK, (e) OM P, (f) SOM P, (g) SOM P-PTable 2. Trainin g sets, testin g sets and cl assification acc u racy (%)obtained from different methods for the Salinas scene ima g eCl ass TrainTestSVMOM P SOM P SOM P-P Weed_1 20198999.8898.68100 100 Weed_2 37368998.5298.7899.7299.95Fall ow 20195692.4894.5598.70 98.41Fallow plow 14 1380 97.46 99.35 96.93 99.69Fallow smooth 27 2651 97.19 93.26 97.45 99.24 St u bbl e 40391999.9899.7299.97100 Cel ery 36354398.1499.4099.55100 Grapes 1131115876.1172.8384.5094.12Soil 62614198.6397.4199.37100 Corn 33324589.2988.1495.2498.04Lett u ce 4wk 11 1057 92.82 96.18 99.26 100Lett u ce 5wk 19 1908 96.16 99.77 96.73 99.73 Lett u ce 6wk 9 907 94.99 98.05 92.53 99.15Lett u ce 7wk 11 1059 94.85 90.87 97.40 99.43 Vineyard u ntrained 73 7195 71.90 57.77 85.24 83.15Vineyard trellis 18 1789 98.87 95.06 98.91 98.92 OA --89.1686.4893.4796.21AA --93.5992.5396.1398.11k --0.8900.8490.92740.958156 D.Li u, S. Li, and L. Fan gIn the Fi g. 3, (a) and (b) are an example of the trainin g and testin g samples. (c) is the classification map obtained from SVM, similarly, (d), (e), (f) are the classification maps of SVM-CK, OM P, SOM P and SOM P-P respectively.In o u r second experiment, we u se the HSI Salinas scene which was collected by 224-band over Sal inas Val ey and Cal ifornia. The size of the Sal inas ima g e is 512×217×224.Al so, beca u se 20 bands is water absorption which is the same as Indian P ines, the n u mber of bands is red u ced to 204. There are 16 g ro u nd-tr u th classes containin g ve g etables, bare soils, and vineyard fields and the size is from 916 to11271 pixels (the total pixels are 54129).We chose 1% of the samples for each class as trainin g sample and the rest as test-in g sample. The n u mber of the trainin g sample and the testin g sample is presented in Table 2. In this table, we can see the overall acc u racy (OA), avera g e acc u racy (AA) and the kappa coefficient by u sin g different methods (the SOM P-P is o u r method). It is easy to see that the performance of the proposed methods is fine. The Fi g. 2 shows the classification maps.Fig. 3. Salinas scene: (a) Train samples, (b) Test samples, and the classification res u lts obtained by(c) SVM, (d) OM P, (e) SOM P, (f) SOM P-P4ConclusionsIn this paper, we have proposed a novel HSI cl assification method base on sparse representation. Differ from other traditional sparse classification technolo g ies which expl oit the sparse coefficient and resid u al to cl assify directl y, this method u ses the sparse coefficient to constr u ct probability maps and then exploits the spatial informa-tion in the maps for classification. Experimental res u lts show that the proposed me-thod has better performance than several well-known classifiers.Hyperspectra l Ima g e Classification by Exploitin g the Spectral-Spatial Correlations 157 Acknowledgement. This work was s u pported in part by the National Nat u ral Science Fo u ndation of China u nder Grant No. 61172161, the National Nat u ral Science Fo u n-dation for Distin gu ished Yo u n g scholars of China u nder Grant No. 61325007.References1.G u altieri, J.A., Cromp, R.F.: S u pport Vector Machines for Hyperspectral Remote Sensin gClassification. In: P roc. S P IE, vol. 3584, pp. 221–232 (1998)2.Mel g ani, F., Br u zzone, L.: Cl assification of Hyperspectral Remote Sensin g Ima g e withS u pport Vector Machines. IEEE. Trans. Geosci. Remote Sens. 42(8), 1778–1790 (2004) 3.Ratle, F., Camps, G.V., Weston, J.: Semis u pervised Ne u ral Networks for Efficient Hyper-spectral Ima g e Classification. IEEE Trans. Geosci. Remote Sens. 48(5), 2271–2282 (2010) 4.Zhon g, Y., Zhan g, L.: An Adaptive Artificial Imm u ne Network for S u pervised Classifica-tion of M u ti-/Hyperspectra Remote Sensin g Ima g ery. IEEE Trans. Geosci. Remote Sens. 50(3), 894–909 (2012)5.Rand, R.S., Keenan, D.M.: Spatial l y smooth partitionin g of hyperspectral ima g ery u sin gspectral/spatial meas u res of disparity. IEEE Trans. Geosci. Remote Sens. 41(6), 1479–1490 (2003)6.Kan g, X., Li, S., Fan g, L.: Extended Random Walker-Based Classification of Hyperspec-tral Ima g es. IEEE Trans. Geosci. Remote Sens., 1–10 (May 2014)7.Chen, Y., Nasrabadi, N.M., Tran, T.D.: Hyperspectral Ima g e Classification Usin g Dictio-nary-Based sparse Representation. IEEE Trans. Geosci. Remote Sens. 49(10), 3973–3985 (2011)8.Driesen, J., Thoonen, G., Sche u nders, P.: Spatial Hyperspectral Ima g e Cl assification byP rior Se g mentation. In: IEEE Geosci. Remote Sens. Symp., vol. 3, pp. 709–712 (2009)9.John, W., Yan g, A.Y., Arvind, G., Sastry, S.S., Ma, Y.: Rob u st Face Reco g nition viaSparse Representation. IEEE Trans. P attern Anal. 31(2), 210–227 (2009)10.Chen, Y., Nasrabadi, N.M., Tran, T.D.: Sparse Representation for Tar g et Detection inHyperspectral Ima g ery. IEEE Jo u rnal of Selected Topics in Si g nal P rocessin g 5(3), 629–640 (2011)11.Fan g, L., Li, S., H u, J.: M u ltitemporal ima g e chan g e detection with compressed sparse re-presentation. IEEE Ima g e P rocessin g, 2673–2676 (2011)12.Li, S., Yin, H., Fan g, L.: Remote Sensin g Ima g e F u sion via Sparse Representations OverLearned Dictionaries. IEEE Trans. Geosci. Remote Sens. 51(9), 4779–4789 (2013)13.Fan g, L., Li, S., Kan g, X., Benediktsson, J.A.: Spectral-Spatial Hyperspectral Ima g e Clas-sification via M u l tiscal e Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens., 1–12 (2014)14.Fan g, L., Li, S., Ryan, M., Qin g, N., Anthony, K.: Fast Acq u isition and Reconstr u ction ofOptical Coherence Tomo g raphy Ima g e via Sparse Representation. IEEE Trans. Med. Im-a g. 32(11), 2034–2049 (2013)15.Fan g, L., Li, S., Kan g, X., Benediktsson, J.: Spectral-Spatial Hyperspectral Ima g e Classifi-cation via M u ltiscale Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens., 1–12 (2014)16.Fan g, L., Li, S., Kan g, X.: Spectral-Spatial Hyperspectral Ima g e Classification via M u ltis-cal e Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens. 52(12), 7738–7749 (2014)17.Dai, W., Mil enkovic, O.: S u bspace Pu rs u it for Compressive Sensin g Si g nal Reconstr u c-tion. IEEE Trans. Inf. Theory 55(5), 2230–2249 (2009)158 D.Li u, S. Li, and L. Fan g18.Tropp, J., Gilbert, A.: Si g nal recovery from random meas u rements via ortho g onal match-in g p u rs u it. IEEE Trans. Inf. Theory 53(12), 4655–4666 (2007)19.G u altieri, J.A., Cromp, R.F.: S u pport Vector machines for Hyperspectral Remote Sensin gClassification. In: P roc. S P IE, vol. 3584, pp. 221–232 (1998)20.Fa u vel, M., Chan u ssot, J., Benediktsson, J.A.: Adaptive P ixel Nei g hborhood Definition forthe Classification of Hyperspectral Ima g es with S u pport Vector Machines and Composite Kernel. In: P roc. IEEE Int. Conf. Ima g e P rocess., pp. 1884–1887 (2008)。

遥感erdas界面翻译

遥感erdas界面翻译

一、主页(一)Information1、ContentsContentsRetrieverGeocoder2、MetadataView/Edit Image Metadata View/Edit Point Cloud Metadata View/Edit Vector Metadata View/Edit Annotation Metadata View/Edit NITF MetadataView/Edit IMAGINE HFAEdit Image Metadata3、SelectSelectSelect by BoxSelect by LineSelect by EllipseSelect by PolygonFollow HyperlinksSelector PropertiesPick Properties4、InquireInquire BoxInquire ShapeInquire Color5、MeasureMeasure(二)Edit1、Cut2、Copy3、Paste4、Delete5、Undo6、Paste From Selected Object (三)Extent1、Fit to Frame2、ResetReset (一)信息内容:内容检索地理编码元数据:查看/编辑影像元数据查看/编辑点云元数据查看/编辑矢量元数据查看/编辑注记元数据查看NITF元数据查看IMAGINE HFA文件内容元数据编辑选择:选择通过拉框选择通过画线选择通过画椭圆选择通过画多边形选择跟踪超链接选择器属性采集器属性查询:查询框查询光标形状查询光标颜色测量:测量(二)编辑剪切复制粘贴删除撤销由选中内容粘贴(三)内容全景显示重置:重置(七)Roam1、HorizontalHorizontalVerticalUser-Defined2、Speed Down3、Speed4、Speed UpSpeed UpSpeed Reset5、Go to Start6、Step Backwards7、Reverse8、Stop9、Start/Pause10、Step Forwards11、Go to End12、Snail TrailSnail TrailMerge Snail Trails13、Roam Properties(七)漫游水平:水平垂直自定义减速速度加速:加速重置速度从头开始快退倒退停止开始/暂停快进到最后追踪:追踪合并追踪漫游属性二、Manage Data管理数据(一)Catalog1、Hexagon ContentHexagon ContentImage Catalog(二)Conversion1、Coordinate Calculator2、Import Data3、Export Data4、GeoMedia ToolsShapefiles to WarehouseWarehouse to ShapefilesGeoMedia Utilities5、Pixels To ASC||6、ASC|| to Pixels7、Graphical Importer(一)目录海克斯康内容:海克斯康内容影像目录(二)转换坐标计算器数据导入数据导出空间数据仓库工具:Shp转仓库仓库转Shp空间数据仓库工具栅格转文本文本转栅格GeoRaster导入器8、GeoRaster Manager9、Imagizer Data Prep(三)VectorizeRaster to Shape to Shape to Annotation (四)Rasterize1、Vector to RasterVector to RasterAnnotation to Raster(五)Image1、Edit Image Metadata2、Pyramids &StatisticsCompute Pyramids and Statistics Create ThumbnailsProcess Footprints and RSETS3、Compare Images4、Create New Image5、Create ECW Transparency(六)NITF/NSIF1、NITFView NITF MetadataExtract Shape LASDPPDB WorkstationMake RPF TOCCIB JPEG 2000 ExporterRPF Frame SearchMake ECRG/ECIB TOCECRG/ECIB Frame Search(七)Office Tools1、Send to PowerPoint2、Send to Word3、Send to GeoPDF4、Send to JPEG GeoRaster管理器IMAGIZER数据预处理(三)矢量化栅格转shp:栅格转shp栅格转注记(四)栅格化矢量转栅格:矢量转栅格注记转栅格(五)影像编辑元数据金字塔和统计:计算金字塔和统计创建缩略图处理范围线和金字塔影像比较创建新影像创建ECW透明度(六)NITF/NSIFNITF:查看NITF元数据提取Shp文件提取LAS文件合并数字式目标定位数据库工作站生成影像目录CIB JPEG 2000输出RPF帧搜索制作ECRG/ECIB目录表ECRG/ECIB帧搜索(七)Office工具发送到PPT发送到Word发送到GeoPDF发送到JPEG三、Raster栅格(一)Resolution (一)分辨率·Two Layer Union Operators Zonal AttributesMatrix UnionSummary Report of Matrix Overlay by Min or Max Index by Weighted Sum (七)Scientific1、FunctionsTwo Image Functions Single Image Functions2、Fourier Analysis Fourier TransformFourier Transform Editor Inverse Fourier Transform Fourier Magnitude ·双层联合计算区域分析矩阵分析归纳分析叠加分析加权分析(七)科学的函数分析:两个影像函数单个影像函数傅里叶分析:傅里叶变换傅里叶变换编辑傅里叶逆变换傅里叶幅值计算四、Vector矢量(一)Manage1、Copy Vector Layer2、Rename Vector Layer3、Delete Vector Layer4、Buffer Analysis5、Attribute to Annotation(二)ER Mapper Vector to Shape 1、Reproject Shape Shape Elevation (三)Raster To VectorRaster to Shapefile (一)管理复制矢量重命名矢量删除矢量缓冲区分析属性转注记ER映射矢量转Shp Shape重投影Shp裁切高程重投影(二)栅格转矢量栅格转Shp五、Terrain地形(一)Manage (一)管理六、Toolbox工具箱(一)Common1、IMAGINE Photogrammetry2、Image Equalizer3、Spatial Model Editor Spatial Model EditorLaunch Spatial Model4、Model MakerModel MakerModel Librarian5、MosaicMosaicProMosaicPro from 2D View Mosaic ExpressUnchip NITF6、AutoSync Workstation AutoSync Workstation Georeferencing wizardEdge Match WizardOpen AutoSync Project7、Stereo AnalystStereo AnalystAuto-Texturize from Block Texel MapperExport 3D shape KML Extended Features to Ground 8、MapsMap Series ToolMap Database ToolEdit Composition Paths9、VirtualGISVirtual World EditorCreate MovieRecord Flight Path with GPS Create TIN Mesh (一)通用图像摄影测量影像匀光器空间模型编辑器:空间模型编辑器发射空间模型空间建模:空间建模模型库管理影像镶嵌:启动专业镶嵌镶嵌视窗显示影像镶嵌快车合并自动配准:自动配准地理参考向导边缘匹配向导打开自动配准工程立体分析:立体分析自动纹理纹理编辑器输出Shp为KML构建地面实体地图工具:图幅地图地图数据库工具修改制图文件路径虚拟GIS:虚拟世界编辑器录像录制通过GPS点定义飞行路径建立不规则三角网七、Help帮助(一)Reference Library1、Help2、About IMAGINE3、Reference booksHexGeoWiki4、WorkflowsCommon WorkflowsSpatial Modeler WorkflowsClassification WorkflowsPhotogrammetry WorkflowsPoint Cloud WorkflowsZonal Change WorkflowsRectification WorkflowsMap Making WorkflowsMosaic WorkflowsVector WorkflowsModel Maker WorkflowsNITF Workflows5、User GuidesAAIC User GuideAutonomous Spectral Image Processing User GuideAutoSync User GuideDeltaCue User GuideHyperspectral User GuideIMAGINE Objective User GuideIMAGIZER Data Prep User GuideIMAGIZER Viewer User Guide Photogrammetry Suite Contents Operational Radar User GuideRader InterferometryStereo Analyst User GuideSubpixel Classifier User GuideVirtual GIS User GuideInstallation and Configuration Guide6、Spatial ModelingSpatial Model EditorModel Maker(Legacy) (一)相关阅览帮助关于IMAGINE参考书:希格维基工作流:常见工作流空间建模器工作流分类工作流摄影测量工作流点云工作流分区变化工作流整流工作流地图制图工作流镶嵌工作流矢量工作流模型制作工作流NITF工作流使用指南:AAIC使用指南自主光谱图像处理用户指南自动配准使用指南变化检测使用指南高光谱使用指南IMAGINE面向对象使用指南IMAGIZER数据准备使用指南IMAGIZER视窗使用指南摄影测量套件目录操作雷达用户指南雷达干涉立体分析使用指南子像元分类使用指南虚拟GIS使用指南安装与配置指南空间建模:空间模型编辑器模型制作者(Legacy)Spatial Modeler Language(Legacy)Graphical Models Reference Guide7、Language ReferenceERDAS Macro Language8、Release NotesERDAS IMAGINE Issues ResolvedAutonomous Spectral Image ProcessingRelease Notes9、ERDAS IMAGINE Release Notes(二)Search Commands1、Search2、Search Box(三)Page1、Previous2、Next空间建模语言(Legacy)图估模型的参考指南建模和定制:ERDAS宏语言发布说明:ERDAS IMAGINE问题解决自主光谱图像处理发布说明ERDAS IMAGINE发行说明(二)搜索命令搜索搜索框(三)页码向前向后八、Multispectral(一)Enhancement1、Adjust RadiometryGeneral ContrastBrightness/ContrastPhotography EnhancementsPiecewise ContrastBreakpointsLoad BreakpointsSave BreakpointsData Scaling2、Discrete DRADiscrete DRADRA Properties(二)Brightness Contrast1、Contrast Down/Up2、Brightness Down/Up(三)Sharpness1、Sharpness Down/Up2、FilteringConvolution FilteringStatistical FilteringReset Convolution(四)Bands1、Sensor Types2、Common Band Combinations (五)View1、Set Resampling Method2、Pixel Transparency(六)Utilities1、Subset & ChipCreate Subset ImageNITF ChipMaskDice ImageImage Slicer2、Spectral Pro Pro Pro Pro Features3、Pyramids & Statistics Compute Pyramids && Statistics Compute Statistics on Window Generate RSETs(七)Transform & Orthocorrect 1、Transform & OrthoOrtho Using Existing ModelOrtho With Model Selection Transform Using Existing Model Create Affine CalibrationPerform Affine Resample Resample Pixel Size2、Control Points3、Single Point4、Check Accuracy(八)Edit1、Fill2、Offset3、Interpolate九、Drawing(一)Edit1、Cut2、Copy3、Paste4、Delete5、Undo6、Paste from Selected Object (二)Insert Geometry1、Point2、Insert Tic3、Arc4、Create Freehand Polyline5、Rectangle6、Polygon7、Ellipse8、Create Concentric Rings9、Text10、Place GeoPoint11、Place GeoPoint Properties12、GrowGrowGrowing Properties13、EasyTrace14、Lock15、Layer Creation Options (三)Modify1、Enable Editing2、SelectSelectSelect by BoxSelect by LineSelect by EllipseSelect by PolygonFollow HyperlinksSelector PropertiesPick Properties3、LineLineReshapeReplace a portion of a lineSplineDensifyGeneralizeJoinSplit4、AreaAreaReshapeSplit polygon with PolylineReplace a portion of a polygon Append to existing polygonInvert Region5、Vector Options(四)Insert Map Element1、Map GridMap GridUTM GridGeographic GridMGRS Grid Zone ElementDeconflict Grid TicmarksGrid Tic Modifier toolGrid Preferences2、Scale Bar3、Legend4、North ArrowNorth ArrowDefault North Arrow Style5、Dynamic ElementsDynamic ElementsDynamic Text EditorConvert to Text(五)Font/Size1、Font Face2、Font/Symbol Unit Type3、Font/Symbol Size4、Font/Symbol Units5、Bold6、Italic7、Underline8、Colors(六)Locking1、Lock Annotation OrientationLock Annotation OrientationReset Annotation Orientation to Screen Reset Annotation Orientation to Map Lock Annotation Orientation Set Default (七)Styles1、Object Style GalleryAdd to GalleryCustomize Styles2、Customize Styles(八)Shape1、Area Fill2、Line Color3、Line StyleLine Thickness1 pt2 pt4 pt6 ptLine PatternSolid LineDotted LineDashed LineDashed Dotted Line OutlineNo OutlineArrowsNo ArrowStart ArrowEnd ArrowBoth Ends(九)Arrange1、ArrangeOrder ObjectsBring to FrontBring ForwardSend To BackSend BackwardGroup ObjectsGroupUngroupPosition ObjectsRotate North十、Format(一)Insert Geometry1、point2、Insert Tic3、Arc4、Create Freehand Polyline5、Rectangle6、Polygon7、Ellipse8、Create Concentric Rings9、Text10、Place GeoPoint11、Place GeoPoint Properties12、GrowGrowGrowing Properties13、EasyTrace14、Lock15、Layer Creation Options (二)Text1、Text GalleryAdd to Gallery(三)Font1、Font Face2、Font Unit Type3、Font Size4、Font Units5、Bold6、Italic7、Underline8、Colors(四)Symbol1、Symbol Size2、Symbol Units3、Symbol Unit Type(五)Locking1、Lock Annotation OrientationLock Annotation OrientationReset Annotation Orientation to Screen Reset Annotation Orientation to Map Lock Annotation Orientation Set Default (六)Styles1、Object Style GalleryAdd to GalleryCustomize Styles2、Customize Styles(七)Shape1、Area Fill2、Line Color3、Line StyleLine Thickness1 pt2 pt4 pt6 ptLine PatternSolid LineDotted LineDashed LineDashed Dotted Line OutlineNo OutlineArrowsNo ArrowStart ArrowEnd ArrowBoth Ends(八)Arrange1、Bring to Front Bring to FrontBring Forward2、Send to Back Send to BackSend Backward3、GroupGroupUngroup4、AlignAlign Horizontal Left Align Horizontal Center Align Horizontal Right Align Vertically Top Align Vertically Center Align Vertically Bottom Distribute Horizontally Distribute Vertically Alignment..5、FlipFlip VerticallyFlip Horizontally6、Rotate North(十一)Table(一)View1、Show AttributesShow AttributesFrom View Attribute2、Switch Table View(二)Drive1、Drive Viewer to first selected item2、Drive to previous selected feature3、Drive to next selected feature4、Drive Viewer to last selected item5、Zoom to Item(三)Column1、Unselect Columns2、Select All Columns3、Invert Column Selection4、Add Class Name5、Add Area(四)Row1、Unselect Rows2、Select All Rows3、Invert Row Selection4、Criteria(五)Query1、Merge2、Colors3、Column Properties(六)Edit1、Edit Column Next2、Edit Row Next。

纹理物体缺陷的视觉检测算法研究--优秀毕业论文

纹理物体缺陷的视觉检测算法研究--优秀毕业论文

摘 要
在竞争激烈的工业自动化生产过程中,机器视觉对产品质量的把关起着举足 轻重的作用,机器视觉在缺陷检测技术方面的应用也逐渐普遍起来。与常规的检 测技术相比,自动化的视觉检测系统更加经济、快捷、高效与 安全。纹理物体在 工业生产中广泛存在,像用于半导体装配和封装底板和发光二极管,现代 化电子 系统中的印制电路板,以及纺织行业中的布匹和织物等都可认为是含有纹理特征 的物体。本论文主要致力于纹理物体的缺陷检测技术研究,为纹理物体的自动化 检测提供高效而可靠的检测算法。 纹理是描述图像内容的重要特征,纹理分析也已经被成功的应用与纹理分割 和纹理分类当中。本研究提出了一种基于纹理分析技术和参考比较方式的缺陷检 测算法。这种算法能容忍物体变形引起的图像配准误差,对纹理的影响也具有鲁 棒性。本算法旨在为检测出的缺陷区域提供丰富而重要的物理意义,如缺陷区域 的大小、形状、亮度对比度及空间分布等。同时,在参考图像可行的情况下,本 算法可用于同质纹理物体和非同质纹理物体的检测,对非纹理物体 的检测也可取 得不错的效果。 在整个检测过程中,我们采用了可调控金字塔的纹理分析和重构技术。与传 统的小波纹理分析技术不同,我们在小波域中加入处理物体变形和纹理影响的容 忍度控制算法,来实现容忍物体变形和对纹理影响鲁棒的目的。最后可调控金字 塔的重构保证了缺陷区域物理意义恢复的准确性。实验阶段,我们检测了一系列 具有实际应用价值的图像。实验结果表明 本文提出的纹理物体缺陷检测算法具有 高效性和易于实现性。 关键字: 缺陷检测;纹理;物体变形;可调控金字塔;重构
Keywords: defect detection, texture, object distortion, steerable pyramid, reconstruction
II

基于分组双阶段双向卷积长短期方法的高光谱图像超分辨率网络

基于分组双阶段双向卷积长短期方法的高光谱图像超分辨率网络

智城实践NO.04 20241智能城市 INTELLIGENT CITY基于分组双阶段双向卷积长短期方法的高光谱图像超分辨率网络林建君1侯钧译2杨翠云2(1.烟台职业学院信息工程系,山东 烟台 264670;2.青岛科技大学信息科学技术学院,山东 青岛 266000)摘要:文章提出基于分组的双阶段Bi-ConvLSTM网络(GDBN),可以充分利用图像的空间和光谱信息,通过使用以波段为单位的分组策略,有效缓解了计算负担,并对光谱信息进行保护。

在编码器的不同阶段,对浅层信息提取模块和深度特征提取模块进行不同层次信息的提取,浅层信息提取模块能够对不同尺度的浅层特征信息进行充分捕捉,深度特征提取模块能够捕捉图像的高频特征信息。

文章还引入通道注意力机制,增强网络对特征的组织能力,并在自然数据集cave上进行大量实验,效果普遍优于目前主流的深度学习方法。

关键词:双向卷积长短期记忆网络;高光谱图像超分辨率;通道注意力;神经网络;深度学习中图分类号:TP391 文献标识码:A 文章编号:2096-1936(2024)04-0001-03DOI:10.19301/ki.zncs.2024.04.001Hyperspectral image super-resolution network based on groupedtwo-stage biconvolution long-term and short-term methodLIN Jian-jun HOU Jun-yi YANG Cui-yunAbstract:In this paper, a two-stage Bi-ConvLSTM network based on grouping (GDBN) is proposed, which can make full use of the spatial and spectral information of images, and effectively relieve the computational burden and protect the spectral information by using the grouping strategy based on band units. At different stages of the encoder, the shallow information extraction module and the depth feature extraction module can extract different levels of information. The shallow information extraction module can fully capture the shallow feature information of different scales, and the depth feature extraction module can capture the high-frequency feature information of the image. The paper also introduces channel attention mechanism to enhance the network's ability to organize features, and conducts a large number of experiments on natural data set cave, and the effect is generally better than the current mainstream deep learning methods.Key words:bidirectional convolution long-term and short-term memory network; hyperspectral image super-resolution; channel attention; neural network; deep learning近年来,基于深度学习[1-2]的单图像超分辨率方法取得了广泛发展。

基于X射线相衬显微CT的肝肿瘤定量分析研究

基于X射线相衬显微CT的肝肿瘤定量分析研究

TECHNOLOGY AND INFORMATION科学与信息化2023年5月下 175基于X射线相衬显微CT的肝肿瘤定量分析研究*林瑶 路文平 张耀中 郑焕圣 王坤(通讯作者)新疆第二医学院 新疆 克拉玛依 834000摘 要 新生血管对肿瘤提供无限生长的应用,但目前传统成像技术只能实现200μm的血管。

本研究收集乏血型和富血型肝肿瘤标本20例,进行X射线相衬断层图像中可清晰显示肝肿瘤组织微血管的分布特征,与对应病理切片表现基本相吻合;通过灰度直方图和灰度-梯度共生矩阵提取图像特征,肿瘤图像特征统计结果表明两组存在显著差异(P<0.05),富血型肿瘤相比于乏血型肿瘤组织图像灰度分布不均匀、变化不规则、图像复杂度升高、清晰度降低,为进一步对不同类型肝肿瘤的诊断和特征评价提供数据支持和科学依据。

关键词 肝肿瘤;X射线相衬显微CT;灰度直方图;灰度-梯度共生矩阵Quantitative Analysis of Liver Tumors Based on X-ray Phase Contrast Micro-CT Lin Yao, Lu Wen-ping, Zhang Yao-zhong, Zheng Huan-sheng, Wang Kun (corresponding author)Xinjiang Second Medical College, Karamay 834000, Xinjiang Uygur Autonomous Region, ChinaAbstract Neovascularization provides unlimited growth for tumors, but currently only 200μm vessels can be achieved by conventional imaging techniques. In this study, 20 cases of liver tumor specimen with poor blood type and rich blood type are collected. X-ray phase-contrast tomography images could clearly show the distribution characteristics of microvessels in liver tumor tissue, which are basically consistent with the corresponding pathological section manifestations. The image features are extracted by grayscale histogram and grayscale-gradient co-occurrence matrix, and the statistical results of tumor image features show that there are significant differences between the two groups (P<0.05). Compared with poor-blood tumor tissue image, the rich-blood tumor tissue image has uneven gray distribution, irregular changes, increased image complexity, and decreased clarity, which provide data support and scientific basis for further diagnosis and characteristic evaluation of different types of liver tumors.Key words liver tumor; X-ray phase-contrast micro-CT; grayscale histogram; grayscale-gradient co-occurrence matrix引言肝肿瘤是一种常见的恶性肿瘤,2016年肝肿瘤已位居我国常见的恶性肿瘤的第二位,因此对肝肿瘤的早期诊断和治疗成为临床工作中的重要环节[1]。

基于graph cuts 和主动轮廓的纹理感知图像分割

基于graph cuts 和主动轮廓的纹理感知图像分割

图的对应关系。左图图像中的像素看作是右图中的结点,邻接像素之间的相似性看作是边上的权值
。类似图,除了普通结点外,还包含两个称为“终点”的点 s、t。边集 E 中包含两种边,一种是 连接相邻结点之间的边(n-links),一种是连接普通结点和“终点”之间的边(t-links)。
假设整幅图像的标签label为L= {l1,l2,,,, lp },其中li为0(背景)或者1(目标)。那假设图像的 分割为L时,图像的能量可以表示为:
Active contour(主动轮廓)、 Color-texture(彩色纹理)、 Structure tensor(结构张量)
2. 本文的研究意义:由于graphcut 的缺点,本论文在四个方面推广延伸了graphcut 算法:
(1):把纹理考虑到分割过程中,我们设计了一个有效地、可靠的纹理探测器,用探测结果生成 一幅增强图像。然后对增强图像用graphcut算法,这样可以利用graphcu进行纹理感知以及有
决方法,对于图像我们适应它,并融合作为后期处理从而增强分割的平滑性、准确性。
(4):当分割比如伪彩色这样的复杂图像时,我们建议在图像分割过程中包含软约束,这样可以 允许用户去勾勒从而效地指导算法去找寻最满意的结果。
(1)给定输入图像和用户的前景/背景输入,算法分析图像并生成一幅增强图像,其中包含了原始 图像和纹理检测结果。
3. 纹理感知graphcut 分割
这一部分首先简单地介绍了graphcut分割作为我们接下来工作的基础,然后提出了两项研究,
构造了纹理增强图像来代替输入图像用于分割以及把结构张量融入到graphcut模型当中。
3.1 graphcut 分割 graph Cut交互式方法的主要思想是:将对图像的目标和背景进行标记作为硬约束,在满足这

基于SNIC的双时相SAR图像超像素协同分割算法

基于SNIC的双时相SAR图像超像素协同分割算法

Vol43 No5May2021第43卷第5期2021 年5 月系统工程与电子技术SystemsEngineeringandElectronics文章编号!001-506X(2021)05-1198-12网址 !www sys-elecom基于SNIC 的双时相SAR 图像超像素协同分割算法马 倩,邹焕新**,李美霖,成 飞,贺诗甜收稿日期:2020 -05 - 27;修回日期:2020 - 08 - 06;网络优先出版日期:2020 - 12 - 10。

网络优先出版地址:https : 〃kns. cnki. net/kcms/detail/11. 2422. TN. 20201210. 0845. 002. html基金项目:国家自然科学基金(62071474)资助课题* 通讯作者 E-mailhxzou2008@163 com引用格式:马倩,邹焕新,李美霖,等.基于SNIC 的双时相SAR 图像超像素协同分割算法:J ).系统工程与电子技术,2021, 43(5): 1198-1209Reerence#ormat !MA Q , ZOU H X , LIM L , etal Superpixelcooperativesegmentationalgorithmforbi-temporalSARimagebasedonSNIC[J ). Systems Engineering and Electronics , 2021, 43(5) : 1198-1209.(国防科技大学电子科学学院,湖南长沙410073)摘 要:针对面向区域的合成孔径雷达(synthetic aperture radar,SAR)图像变化检测方法中存在的双时相图像边缘和空间对应关系不一致的问题,提出了一种基于简单非迭代聚类(simple non-iterative clustering, SNIC )的双时相SAR 图像超像素协同分割算法°首先,构造一幅包含双时相SAR 图像特征的融合图像,计算待处理像素点到聚类中心的像素强度相似度和空间距离相似度°其次,采用一种高效的多尺度弱边缘检测算法,对双时相SAR 图像分别进行边缘检测并融合边缘检测结果°最后,将像素强度相似度、空间距离相似度和边缘信息进行加权以替代原始SNIC 算法中的距离测度,实现对SAR 融合图像的超像素分割,得到与双时相SAR 图像中真实地 物边缘均贴合的协同分割结果°基于一组仿真和一组实测双时相SAR 图像的超像素协同分割实验结果表明,该算法的边缘贴合率、欠分割误差和可达分割准确率均优于其他7种经典方法°关键词:双时相合成孔径雷达图像&超像素&协同分割&简单非迭代聚类&变化检测中图分类号:TN 957 文献标志码:A DOI :10. 12305/j. issn. 1001-506X. 2021. 05. 06Super pixel cooperative segmentation algorithm for bi-temporalSAR image based on SNICMA Qian , ZOU Huanxin * , LI Meilin , CHENG Fei , HE Shitian(College of Electronic Science and Technology , National University of Defense Technology ,Changsha410073 'China "Abstract : Aiming at the problem of the inconsistency of bi-temporal images * boundaries and spatialcorespondenceinthetaskofregion-basedsyntheticapertureradar (SAR "imagechangedetection 'asuperpixel cosegmentation algorithm based on simple non-iterative clustering (SNIC "forbi-temporalSARimagesisproposed'First 'afusedimagecontainingthefeaturesofthebi-temporalSARimagesisconstructed 'andthepixelintensity similarityandspatialdistancesimilaritybetweenthepixelstobeprocessedandtheclustercenteriscalculated'Second 'acomputationalyeficientmultiscaleedgedetectionalgorithmisadoptedandusedtodetecttheedgesofthebi-temporal SARimagesrespectively 'andtheedgedetectionresultsarefusedtoformanedgemap'Finaly 'thepixelintensitysimilarity 'spatialdistancesimilarityandedge map information are weighted to replace the distance measureinthe originalSNICalgorithmandtheimprovedSNICisutilizedtoperformsuperpixelsegmentationonthefusedimagetoobtain the segmentation result which fits the real terrain edges in the bi-temporalSARimages'The experimentalresultsconductonapairofsimulatedSARimagesandapairofreal-worldbi-temporalSARimagesdemonstratethat theboundaryrecal 'under-segmentationerrorandachievablesegmentationaccuracyoftheproposedmethodarebeterthanthoseofothersevenstate-of-the-artmethods'第5期马倩等:基于SNIC的双时相SAR图像超像素协同分割算法・1199・Keyword:bi-temporalsyntheticapertureradar simple non-iterative clustering;change detectiono引言近年来,遥感平台和传感器技术取得了飞速发展,实现了对全球大部分区域的连续重复观测,积累了海量多源、多时相遥感数据(11$在合成孔径雷达(synthetic aperture ra­dar,SAR)、光学、红外、多光谱等多源数据中,SAR具备全天时、全天候、高分辨、大幅宽等多种优点,是一种良好的信息源$SAR图像广泛应用在目标检测、地物分类、动态监测和变化检测等领域,其中变化检测在国民经济和国防建设上发挥着巨大作用,已经成为遥感数据应用研究中的一个重要内容$尽管不少学者围绕双时相(或者多时相,为行文简便起见,以下统一用双时相来描述)SAR图像变化检测开展了大量工作。

遥感图像场景分类综述

遥感图像场景分类综述

人工智能及识别技术本栏目责任编辑:唐一东遥感图像场景分类综述钱园园,刘进锋*(宁夏大学信息工程学院,宁夏银川750021)摘要:随着科技的进步,遥感图像场景的应用需求逐渐增大,广泛应用于城市监管、资源的勘探以及自然灾害检测等领域中。

作为一种备受关注的基础图像处理手段,近年来众多学者提出各种方法对遥感图像的场景进行分类。

根据遥感场景分类时有无标签参与,本文从监督分类、无监督分类以及半监督分类这三个方面对近年来的研究方法进行介绍。

然后结合遥感图像的特征,分析这三种方法的优缺点,对比它们之间的差异及其在数据集上的性能表现。

最后,对遥感图像场景分类方法面临的问题和挑战进行总结和展望。

关键词:遥感图像场景分类;监督分类;无监督分类;半监督分类中图分类号:TP391文献标识码:A文章编号:1009-3044(2021)15-0187-00开放科学(资源服务)标识码(OSID ):Summary of Remote Sensing Image Scene Classification QIAN Yuan-yuan ,LIU Jin-feng *(School of Information Engineering,Ningxia University,Yinchuan 750021,China)Abstract:With the progress of science and technology,the application demand of remote sensing image scene increases gradually,which is widely used in urban supervision,resource exploration,natural disaster detection and other fields.As a basic image pro⁃cessing method,many scholars have proposed various methods to classify the scene of remote sensing image in recent years.This pa⁃per introduces the research methods in recent years from the three aspects of supervised classification,unsupervised classification and semi-supervised classification.Then,combined with the features of remote sensing images,the advantages and disadvantages of these three methods are analyzed,and the differences between them and their performance performance in the data set are com⁃pared.Finally,the problems and challenges of remote sensing image scene classification are summarized and prospected.Key words:remote sensing image scene classification;Unsupervised classification;Supervise classification;Semi-supervised clas⁃sification遥感图像场景分类,就是通过某种算法对输入的遥感场景图像进行分类,并且判断某幅图像属于哪种类别。

虹膜图像识别处理外文翻译

虹膜图像识别处理外文翻译

外文一:AbstractThe biological features recognition is one kind of basis human body own inherent physiology characteristic and the behavior characteristic distinguishes the status the technology,Namely through the computer and optics, acoustics, the biosensor and the biometrics principle and so on high tech method unifies closely,Carry on individual status using the human body inherent physiology characteristic and the behavior characteristic the appraisal。

The biological features recognition technology has is not easy to forget, the forgery-proof performance good, not easy forge or is robbed, “carries” along with and anytime and anywhere available and so on merits.Iris recognition is a new method for man identification based on the biological features, which has the significant value in the information and security field. Combined with the previous work of other researchers, a discussion is elaborately made on the key techniques concerning the capture of iris images, location of iris circle and some improved and approaches to these problems are put forward. The location of iris recognition is realized which proves efficient.Iris location is a crucial part in the process of iris recognition,thus obtaining the iris localization precisely and fleetly is the prelude of effective iris localization .Iris location of is a kernel procession in an iris recognition system. The speed an accuracy of the iris location decide the performance of the iris recognition system.Take the advantages of the iris image, per-processes the images, decides the pesudo –center of pupil by a method of gray projection .Then the application calculus operator law carries on inside and outside the iris the boundary localization,in this paper ,this algorithm is based on the Daugman algorithm .Finally realizes the localization process in matlab.Keywords: Iris location,Biological features recognition,Calculus operator,Daugman algorithmTable of ContentsThe 1 Chapter Introduction1.1 The research background of iris recognition (6)1.2 The purpose and significance (8)1.3 Domestic and foreign research (9)Chapter 2 of iris recognition technology Introduction2.1 biometric identification technology (14)2.1.1 The status and development (14)2.1.2 Several biometric technology (17)2.2Iris recognition technology (23)2.3 Summary (26)Chapter 3 Research Status of iris location algorithm3.1Several common localization algorithm (27)3.1.1 Hough transform method (27)3.1.2 Geometric features location method (28)3.1.3 Active contour positioning method (29)3.2 Positioning algorithm studied (31)Chapter 4 operator calculus based iris localization algorithm4.1Image preprocessing (34)4.1.1Iris image smoothing (denoising) (36)4.1.2 Sharpen the image (filter)..................37.4.2Coarse positioning the inner edge of the iris (39)4.3 the iris to locate calculus operator law (40)4.4 Summary (41)Chapter 5 Conclusion (41)References (43)The first chapter1.1 The research background of iris recognitionBiometrics is a technology for personal identification using physiological characteristics and behavior characteristics inherent in the human body. Can be used for the biological characteristics of biological recognition, fingerprint, hand type face, iris, retina, pulse, ear etc.. Behavior has the following characteristics: signature, voice, gait, etc.. Based on these characteristics, it has been the development of hand shape recognition, fingerprint recognition, facial recognition, iris recognition, signature recognition and other biometric technology, many techniques have been formed and mature to application of.Biological recognition technology in a , has a long history, the ancient Egyptians throughidentification of each part of the body size measure to carry out identity may be the earliest human based on the earliest history of biometrics. But the modern biological recognition technology began in twentieth Century 70 time metaphase, as biometric devices early is relatively expensive, so only a higher security level atomic test, production base.due to declining cost of microprocessor and various electronic components, precision gradually improve, control device of a biological recognition technology has been gradually applied to commerce authorized, such as access control, attendance management, management system, safety certification field etc..All biometric technology, iris recognition is currently used as a convenient and accurate. Making twenty-first Century is information technology, network technology of the century, is also the human get rid of traditional technology, more and more freedom of the century. In the information, free for the characteristics of the century, biometric authentication technology, high-tech as the end of the twentieth Century began to flourish, will play a more and more important role in social life, fundamentally change the human way of life . Characteristics of the iris, fingerprint, DNA the body itself, will gradually existing password, key, become people lifestyle, instead of at the same time, personal data to ensure maximum safety, maximize the prevention of various types of crime, economic crime.Iris recognition technology, because of its unique in terms of acquisition, accuracy and other advantages, will become the mainstream of biometric authentication technology in the future society. Application of safety control, the customs import and export inspection, e-commerce and other fields in the future, is also inevitable in iris recognition technology as the focus. This trend, now in various applications around the world began to appear in the.1.2 Objective and significance of iris recognitionIris recognition technology rising in recent years, because of its strong advantages and potential commercial value, driven by some international companies and institutions have invested a lot of manpower, financial resources and energy research. The concept of automatic iris identification is first proposed by Frown, then Daugman for the first time in the algorithm becomes feasible.The iris is a colored ring in the pupil in the eye of fabric shape, each iris contains a structure like the one and only based on the crown, crystalline, filaments, spots, structure, concave point, ray, wrinkles and fringe characteristic. The iris is different from the retina, retinal is located in the fundus, difficult to image, iris can be seen directly, biometric identification technology can obtain the image of iris fine with camera equipment based on the following basis: Iris fibrous tissue details is rich and complicated, and the formation and embryonic tissue of iris details the occurrence stage of the environment, have great random the. The inherent characteristics of iris tissue is differ from man to man, even identical twins, there is no real possibility of characteristics of the same.When the iris are fully developed, he changes in people's life and tiny. In the iris outer, with a layer of transparent corneal it is separated from the outside world. So mature iris less susceptible to external damage and change.These characteristics of the iris has the advantages, the iris image acquisition, the human eye is not in direct contact with CCD, CMOS and other light sensor, uses a non technology acquisition invasion. So, as an important biometric identity verification system, iris recognition by virtue of the iris texture information, stability, uniqueness and non aggressive, more and more attention from both academic and industrial circles.1.3 Status and application of domestic and foreign research on iris recognitionIDC (International Data Group) statistics show that: by the end of 2003, the global iris recognition technology and related products market capacity will reach the level of $2000000000. Predicted conserved survey China biometric authentication center: in the next 5 years, only in the Chinese, iris recognition in the market amounted to 4000000000 rmb. With the expansion of application of the iris recognition technology, and the application in the electronic commerce domain, this number will expand to hundreds of billions.The development of iris recognition can be traced back to nineteenth Century 80's.. In 1885, ALPHONSE BERTILLON will use the criminal prison thoughts of the application of biometrics individual in Paris, including biological characteristics for use at the time: the size of the ears, feet in length, iris.In 1987, ARAN SAFIR and LEONARD FLOM Department of Ophthalmology experts first proposed the concept, the use of automatic iris recognition iris image in 1991, USA Los ala Moss National Laboratory JOHNSON realized an automatic iris recognition system.In 1993, JOHN DAUGMAN to achieve a high performance automatic iris recognition system.In 1997, the first patent Chinese iris recognition is approved, the applicant, Wang Jiesheng.In 2005, the Chinese Academy of Sciences Institute of automation, National Laboratory of pattern recognition, because of outstanding achievement "in recognition of" iris image acquisition and aspects, won the two "National Technology Invention Prize", the highest level represents the development of iris recognition technology in china.In 2007 November, "requirements for information security technology in iris recognition system" (GB/T20979-2007) national standards promulgated and implemented, the drafting unit: Beijing arithen Information Technology Co., ltd..Application of safety control, the customs import and export inspection, e-commerce and other fields in the future, is also inevitable in iris recognition technology as the focus. This trend, now in various applications around the world began to appear in the. In foreign countries, iris recognition products have been applied in a wide range.In February 8, 2002, the British Heathrow Airport began to test an advanced security system, the new system can scan the passenger's eyes, instead of to check passports. It is reported, the pilot scheme for a period of five months, a British Airways and virgin Airlines passengers can participate in this test. The International Air Transport Association interested in the results of this study are, they encourage the Heathrow Airport to test, through the iris boarding passengers to determine its identity as a boarding pass.Iris recognition system America "Iriscan" developed has been applied in the three business department of Union Bank of American Texas within. Depositors to be left with nothing whatsoever to banking, no bank card password, no more memories trouble. They get money fromthe A TM, a camera first eye of the user to scan, and then scan the image into digital information and data check, check the user's identity.America Plumsted school in New Jersey has been in the campus installed device of iris recognition for security control of any school, students and staff are no longer use cards and certificates of any kind, as long as they passed in the iris camera before, their location, identity is system identification, all foreign workers must be iris data logging to enter the campus. At the same time, through the central login and access control system to carry on the control to enter the scope of activities. After the installation of the system, various campus in violation of rules and infringement, criminal activity is greatly reduced, greatly reducing the campus management difficulty.In Afghanistan, the United Nations (UN) and the United Nations USA federal agency refugee agency (UNHCR) using iris recognition system identification of refugees, to prevent the same refugee multiple receive relief goods. Use the same system in refugee camps in Pakistan and Afghanistan. A total of more than 2000000 refugees use iris recognition system, this system to a key role for the United Nations for distribution of humanitarian aid from.In March 18, 2003, Abu Zabi (one of the Arabia and the United Arab Emirates) announced the iris recognition technology for expelled foreigners iris tracking and control system based on the borders opened the world's first set of national level, this system began construction from 2001, its purpose is to prevent all expelled by Abu Zabi tourists and other personnel to enter the Abu Zabi. Without this system in the past, due to the unique characteristics of the surface of the Arabs (Hu Xuduo), and the number of the expulsion of the numerous, customs inspection staff is very difficult to distinguish between what is a deported person. By using this system, illegal immigration, all be avoided, the maximum guarantee of national security.Kennedy International Airport in New Jersey state (John F. Kennedy International Airport) of the iris recognition system installed on its international flights fourth boarding port, 300 of all 1300 employees have already started to use the system login control. By using this system, all can enter to the apron personnel must be after the system safety certification of personnel. Unauthorized want to break through, the system will automatically take emergency measures to try to force through personnel closed in the guard space. Using this system, the safety grade Kennedy International Airport rose from B+ to A+ grade. The Kennedy International Airport to travel to other parts of the passengers has increased by 18.7%.Generally speaking, the iris recognition technology has already begun in all walks of life in various forms of application in the world. At the same time, to the application of their units of all had seen and what sorts of social benefits and economic benefits are not see. This trend is to enhance the high speed, the next 10 years will be gradually achieve the comprehensive application of iris recognition in each industry.In China, due to the Chinese embargo and iris technology itself and the difficulty in domestic cannot develop products. So far, there has not been a real application of iris recognition system. However, many domestic units are expressed using strong intention, especially the "9 · 11" later, security anti terrorism consciousness has become the most concerned problems in the field of aviation, finance. Iris recognition system is a major airline companies, major financial institutions and other security mechanisms (such as aerospace bureau) become the focus of attention of object and other key national security agency. As with the trend of development in the world, iris recognition technology will in the near future in application China set off climax.The second chapter of introduction of iris recognition technology2.1 Technology of biological feature recognition based on2.1.1 Present status and development of biological feature recognition“9.11" event is an important turning point in the devel opment of biometric identification technology in the world, the importance of it makes governments more clearly aware of the biological recognition technology. Traditional identity recognition technologies in the face of defect anti terrorism has shown, the government began a large-scale investment in the research and application of biometric technology. At the same time, the public understanding of biological recognition technology with "9.11" exposure rate and greatly improve the.The traditional method of individual identification is the identity of the people with knowledge, identity objects recognition. The so-called identity: knowledge refers to the knowledge and memory system of personal identification, cannot be stolen, and the system is easy to install, but once the identification knowledge stolen or forgotten, the identity of easily being fake or replaced, this method at present in a wide range of applications. For example: the user name and password. The so-called identity items: refers to the person, master items. Although it is stable and reliable, but mainly depend on the outer body, lost or stolen identification items once proof of identity, the identity of easily being fake or replaced, for example: keys, certificates, magnetic card, IC card etc..Biometric identification technology is related to physical characteristics, someone using prior record of behavior, to confirm whether the facts. Biometric identification technology can be widely used in all fields of society. For example: a customer came into the bank, he did not take bank card, also did not remember the password directly drawing, when he was drawing in the drawing machine, a camera to scan on his eyes, and then quickly and accurately complete the user identification and deal with business. This is the application of the iris recognition system of modern biological identification technology. "".America "9.11" after the incident, the anti terrorist activity has become the consensus of governments, it is very important to strengthen the security and defense security at the airport, some airports USA can in the crowd out a face, whether he Is it right? Wanted. This is the application of modern technology in biological feature recognition "facial recognition technology".Compared with the traditional means of identity recognition, biometric identity recognition technology in general has the following advantages:(1) the security performance is good, not easy to counterfeit or stolen.(2) carry, whenever and wherever possible, therefore more safety and security and other identification method.For the biological information of biometric recognition, its basic nature must meet the following three conditions: universality, uniqueness and permanency.The so-called universality, refers to any individual has the. Uniqueness, is in addition to other than himself, other people did not have any, namely different. The so-called permanent, refers to the character does not change over time, namely, life-long.Feature selection of organisms with more than three properties, is the first step of biological recognition.In addition, there are two important indexes in biological recognition technology. The rejection rate and recognition rate. Adjusting the relation of these two values is very important. The reject rate, the so-called false rejection, this value is high, use frequency is low, the errorrecognition, its value is high, safety is relatively reduced. So in the biological identification of any adjustment, the two index is a can not abandon the process. The choice of range size, related to the biological identification is feasible and available .And technology of identity recognition based on iris feature now appears, it is the development of biometric identification technology quickly, due to its uniqueness, stability, convenience and reliability, so the formation of biometric identification technology has the prospects for development.Generally speaking, the biological recognition system consists of 4 steps. The first step, the image acquisition system of collecting biometric image; the second step, the biological characteristics of image preprocessing (location, normalization, image enhancement and so on); the third step, feature information extraction, converted into digital code; the fourth step, the generation of test code and database template code to compare, make identification。

利用高光谱图像技术检测水果轻微损伤

利用高光谱图像技术检测水果轻微损伤

利⽤⾼光谱图像技术检测⽔果轻微损伤⾼光谱图像技术检测苹果轻微损伤摘要传统的近红外光谱分析法和可见光图像技术应⽤于⽔果品质⽆损检测中存在的检测区域⼩、检测时间长、仅能检测表⾯情况等局限性。

提出了利⽤⾼光谱图像技术检测⽔果轻微损伤的⽅法。

试验以苹果为研究对象, 利⽤ 500~ 900nm范围内的⾼光谱图像数据, 通过主成分分析提取 547nm 波长下的特征图像, 然后设计不均匀⼆次差分消除了苹果图像亮度分布不均匀的影响, 最后通过合适的数字图像处理⽅法提取苹果的轻微损伤。

关键词: ⽆损检测苹果⾼光谱图像检测轻微损伤引⾔⽔果在采摘或运输过程中, 因外⼒的作⽤使其表⽪受到机械损伤, 损伤处表⽪未破损, 伤⾯有轻微,⾊稍变暗, ⾁眼难于觉察。

受⽔果⾊泽的影响, 传统的计算机视觉技术不能对轻微损伤加以检测。

但是轻微损伤是⽔果在线检测的主要指标之⼀, 随着时间的延长, 轻微损伤部位逐渐褐变, 最终导致整个果实腐烂并影响其他果实。

因此, ⽔果轻微损伤的快速有效检测是⽬前研究的难点和热点之⼀。

虽然轻微损伤和正常区域在外部特征上呈现出极⼤的相似性, 但是损伤区域的内部组织发⽣⼀定的变化, 这种变化可以通过特定波长下的光谱表现出来。

当前, ⼀种能集成光谱检测和图像检测优点的新技术。

⾼光谱图像技术正好能满⾜⽔果表⾯轻微损伤检测的需要。

⾼光谱图像技术是光谱分析和图像处理在最低层⾯上的融合技术, 可以对研究对象的内外部特征进⾏可视化分析。

在国内, ⾼光谱图像技术在农畜产品品质检测的应⽤还没有相关的⽂献报道; 在国外,近⼏年来有部分学者将该技术应⽤于⾁类和果蔬类的品质检测上。

本⽂采⽤⾼光谱图像技术对⽔果表⾯轻微损伤检测进⾏研究, 并通过合适的数据处理⽅法寻找到最能准确辨别⽔果表⾯损伤的特征波长下的图像, 为实现⾼光谱图像技术对⽔果轻微损伤的在线检测提供依据。

1 ⾼光谱图像基本原理⾼光谱图像是在特定波长范围内由⼀系列波长处的光学图像组成的三维图像块。

Image Segmentation Using Thresholding and Genetic Algorithm

Image Segmentation Using Thresholding and Genetic Algorithm

Image Segmentation Using Thresholding and GeneticAlgorithm#P. Kanungo, P. K. Nanda and U. C. SamalImage Processing and Computer Vision Lab.Department of Electrical EngineeringNational Institute of Technology, Rourkela 769008 priyadarshikanungo@, pknanda@nitrkl.ac.in, umesh.samal@AbstractIn this paper the problem of image segmentation is addressedusing the notion of thresholding. A new approach based on GeneticAlgorithm (GA) is proposed for selection of threshold from thehistogram of images. Specifically GA based crowding algorithm isproposed for determination of the peaks and valleys of thehistogram. Experimental results are provided for histogram withbimodal feature, however, this technique can be extended to multithreshold selection for histograms with multimodal feature.Index TermsSegmentation, Genetic Algorithms (GAs)1IntroductionIt is important in picture processing to select an adequate threshold of gray level for extracting object from there background. Image thresholding is a necessary step in many image analysis applications [1]-[4]. In its simplest form, thresholding means to classify the pixels of a given image into two groups (e.g. objects and background). One including those pixels with their gray values above a certain threshold, and the other including those with grey values equal to and below the threshold. This is called bi-level thresholding. Generally, one can select more than one threshold, and use them to divide the whole range of gray values in to several sub ranges. This process is called multilevel thresholding. Most thresholding techniques [5]-[8] utilize shape information of the histogram of given image while selecting thresholds.In an ideal case, for images having two class, the histogram has a deep and sharp valley between two peaks representing objects and back ground respectively. Thus the threshold can be chosen at the bottom of this valley [5]. However, for most real pictures, it is often difficult to detect the valley precisely, because (i) valley could be flat and broad and (ii) the two peaks could be extremely unequal in height, often producing no traceablevalley. Rosenfeld et. al [6] proposed the valley sharpening techniques which restricts the histogram to the pixels with large absolute values of derivatives, S. Watanable et. al.[9] proposed the difference histogram method, which selects threshold at the gray level with the maximal amount of difference. These utilize information concerning neighboring pixels or edges in the original picture to modify the histogram so as to make it useful for thresholding. Another class of methods deal directly with the grey level histogram by parametric techniques. The histogram is approximated in the least square sense by a sum of Gaussian distributions, and statistical decision procedures are applied [8]. However, such methods are tedious and computationally involved.In our proposed method we used GA to find out the peaks and valley in bimodal class of images. GAs are used for function optimization process and hence determining the global optimal solutions. In last couple of years there were new strategies and algorithms proposed to detect the global as well as local solutions in a nonlinear multimodal function optimization [14]-[18]. Crowding originally proposed by Dejong (1975) helps to maintain multiple peaks (Global as well as Local) in multimodal function optimization problem. We have considered the images whose histogram has two peaks. Crowding method will help us to detect the two peaks. After getting the two peaks we can use GA to find out the valley bottom between these peaks. Here we have considered both types of image having flat valley as well as sharp valley in the histogram. Our discussion is confined to the elementary case of threshold selection where only the gray-level histogram suffices without other a priori knowledge. Our algorithm does not require any valley sharpening techniques.2Problem StatementThe problem considered is to extract objects from their background. Thresholding is a popular tool for segmenting grey level images. The approach is based on the assumptions that object and background pixels in the image can be distinguished by their gray level threshold. The dominant values of object and background intensities the original grey levels image can be transformed in to a segmented image of two classes (for example; one object and the other background). Although the method appears to be simple, it is an important and basic one with wide applicability. We have only considered the histogram without any priory information. For a two class problem the aim is to determine a threshold at the grey value in the valley between the two peaks of the histogram. Determination of these two peaks is not an obvious task. Prewitt et al.[5] proposed the mode method. In which theychose thresholds at the valleys (or antimodes) on the histogram. The automatic selection scheme involved some smoothing of the histogram data, searching for modes, and placing thresholds at the minima between them. Their method relied heavily on the structure of the gray level histogram, which contained peaks and valleys corresponding to gray level subpopulation of the image. Object and background regions (represented by histogram peaks) were assumed to be of fairly constant gray level, and to differ in average gray level. Edges were composed of intermediate grey levels and were less populated than either object or back grounds. Heuristic search method also fails to find the two peaks. Also it is difficult to find the exact threshold point if the valley is flat. However, the bottom of the valley is some thing difficult to locate.Several methods have been proposed for transforming the histogram so that the valley is deepened, or is converted to peak. Thus, correct threshold may be selected efficiently. In our proposed method we applied GA to find the two peaks. Genetic Algorithms are used for getting the global solution. But in this problem we need global as well as local solutions i.e. determining the niches in the multimodal function. To maintain stable sub-population by replacing population with like individuals is known as Crowding method. We have used crowding mechanism to maintain subpopulation at the two peaks. After getting the two peaks we used the GA to determine the valley bottom between the peaks. GA used here to get the global solution. We found that this method works even if the valley is flat. We do not need any valley sharpen method to depend the valley of histogram. Experimental results presented here are only for two class images. The histogram of these images have two distinct peaks. Our proposed method is works efficiently if the histogram of the given image has clearly two modes of any size i.e the image has a two class image. Our proposed algorithm fails to segment properly if there are more than two class in an image (histogram has more than two peaks).3GA Based Class ModelUsually GAs are used for optimization of nonlinear multimodal function and hence determines the global optimal solution. In case of nonlinear multimodal function optimization, the problem of determining the global optimal solution as well as the local solution reduces to determining the niches in the multimodal function. Thus the problem boils down to determining the niches of the multimodal function. Substantial efforts hasbeen directed in this direction for last couples of years [14-18], where new strategies and algorithm are proposed.Method3.1 CrowdingTo maintain stable sub-population by replacing population with like individuals can be broadly called crowding method. Crowding, originally proposed by Dejong in the year 1975 is motivated by analogy with competition for limited resourced among similar member of a natural population. Dissimilar population member often occupy different environmental niches. Older members of the niche will be replaced by the fittest of the younger member. Stochastic replacement error prevents the basic crowding algorithm from maintaining more than two peaks of multimodal fitness.Deterministic crowding eliminates replacement error and maintains multiple peaks. It works by randomly pairing the population to yield n/2 pairs for n individuals in the Population. Each pair of parent yields two children by undergoing crossover and mutation and these two children compete with the parent. In tournament selection, the pair containing the maximally fit element wins.In the deterministic crowding, sampling occurs without replacement. We will assume that an element in a given class is closer to its own class than to elements of other class. Our previous assumption is that a crossover operation between two elements of same class yields two elements of that class, and crossover operation between two elements of different classes yield one element of the both classes. Therefore, if two elements of class-A gets randomly paired, the offspring will also be of class-A, and the resulting tournament will advance two class-A elements to the next generation. The random pairing of two class-B elements will similarly result in no net change to the distribution in the next generation. If an element of class-A gets paired with an element of class-B, one offspring will be form class-A, and the other from class-B. the class-A offspring will compete against the class-A parent, the class-B offspring against the class-B parent. The end result will be that one element of the both classes advances to the next generation, and hence no net change. Since each element receives exactly one trial, the mean and variance for the number of population elements in class-A after one generation are µA=I A and σA=0.3.2 Salient Steps of the Proposed Algorithm.(i) Initialize randomly a population space of size N (each element corresponds to a gray value between 0 to 256) and their classes are determined.(ii) Choose two parents randomly for crossover and mutation operation with crossover probability P C and mutation probability P M. Compute the fitness of parents and off-springs. The fitness function is the normalized histogram function p(g).(iii) The offspring generated complete with the parents based on the concept of tournament selection strategy.(iv) After selection the selected elements are put in their respective classes.(v) Step (ii), (iii) and (iv) are repeated for all elements in the population.(vi) Steps (v) is repeated till the convergence is met i.e. the elements of respective classes are equally fit.(vii) The peaks will be determined from the converged classes of step (vi)(viii) Initialize randomly a population space of size n between the two peaks (i.e. between the two corresponding gray values).(ix) Choose two parents randomly for crossover and mutation operation with crossover probability P C and mutation probability P M. Compute the fitness of parents and off-springs. The fitness function is the histogram function p(g).(x) The fittest two elements between the parents and offspring are selected for the next generation in the selection strategy.(xi)Step (ix), (x) are repeated for all elements in the population.(xii) Step (xi) is repeated till the convergence is met.(xiii) The converged value is the gray value corresponding to the valley between the two peaks. The image is then segmented using this value as threshold.4Results and DiscussionIn simulation, we have considered images whose histograms exhibit bimodal feature. Fig 1.(a) is the original image of size 507x284, whose normalized histogram is shown in Fig. 1(b). From Fig. 1 (b), it is observed that there are two clearly separated peaks having unequal heights. The distributions of the two sets of gray values are not analogous. Selection of a global threshold now reduces to determine a suitable gray value in between the peaks of the histogram. The proposed crowding algorithm is used to determine the two peaks. The parameters of the proposed algorithm are (i) number of population elements “N”is 20 (ii) crossover probability P c =0.9 (iii) mutation probability P m follows a decaying exponential function with starting value 0.05. The valley point between the two peaks in the histogram, in other words the optimal threshold is obtained by searching the minimum gray value in between the gray values corresponding to two peaks. This minimum point is determined by employing Genetic Algorithm (GA). Fig. 1(d) shows the detected peaks and valleys using the proposed algorithm. Here, the peaks are at gray value 91 and 148 and valley is at gray value 117. Using the gray value 117 as threshold value segmentation of the original image is carried out and the segmented image is shown in Fig. 1(d). From Fig 1(d), it can be observed that the object and background is clearly distinguished.Fig 2(a) shows the original image of size 238x238 and the corresponding histogram is shown in Fig. 2(b). From this histogram it can be seen that the two modes are clearly separated by a long valley. Fig 2(c) shows the peaks and valley detected by the proposed algorithm. Peaks are at gray values 2 and 187 and the valley is at gray value 58. The segmented image is shown in Fig. 2(d) by taking gray value 58 as threshold.Fig. 3(b) shows the histogram of another image shown in Fig. 3(a). The peaks and valleys found by our proposed algorithm are shown in Fig. 3(c). Thus the threshold is selected to be 98 and the segmented image using this threshold is shown in Fig. 3(d). Thus in this case also proper segmentation could be achieved.Fig. 4(b) shows the histogram of the image shown in Fig. 4(a). From this histogram we can see that the two peaks are quite uneven in size and they are separated by a flat valley. Fig. 4(c) shows the peaks and valley detected by the proposed algorithm. Taking the valley point gray value 136 as threshold the original image is segmented and the segmented image is shown in Fig. 4(d). Here the object is clearly segmented from the background.Fig 5(a) is the original image of size 255x255 and the corresponding normalized histogram is shown in Fig. 5(b). From the histogram we can observe that there are three peaks. So this image consists of three classes. When the proposed two class algorithm is applied to this image it yields the two peaks at 122 gray value and other at 216 and 195 as the valley point for threshold which is shown in Fig. 5(c). After thresholding at this gray value 195 the image is shown in fig 5(d). It can be seen that there are some misclassification, the black dots in background and white dots in the object. This problem can be solved by modifying the algorithm for more than two classes.Fig. 1(a). Original Image of size 507x284 Fig. 1(b). Normalized Histogram of Fig 1(a)Fig. 1(c). Picks(91,148) and valley(117) detected using GA Fig. 1(d). Segmented Image after puttingThreshold at the valley(117)Fig. 2(a). Original Image of size 238x238 Fig. 2(b). Normalized Histogram of Fig. 2(a)Fig. 2(c). Picks(2,187) and valley(58) detected using GA Fig. 2(d). Segmented Image after putting Threshold at the valley (58)Fig. 3(a). Original Image of size 320x240 Fig. 3(b). Normalized Histogram of Fig. 3(a)Fig. 3(c). Picks (19,203) valley(98) detected using GA Fig. 3(d). Segmented Image after puttingThreshold at the valley (98)Fig. 4(a). Original Image of size 256x256 Fig. 4(b). Normalized Histogram of Fig. 4(a)Fig. 4(c). Picks(25,163) and valley(136) detected using GA Fig. 4(d). Segmented Image after putting Threshold at the valley (136)Fig.5(a). Original Image of size 255x255 Fig. 5(b). Normalized Histogram of Fig. 5(a)Fig. 5(c). Picks(122,216) and valley(195) detected using GA Fig. 5(d). Segmented Image after puttingThreshold at the valley (195)5ConclusionThe problem of separating object from the background in a given image is considered. Hence, the problem boils down to determining the threshold using histogram of the given image. Often, in practice, the histograms do not show two clearly separated classes rather overlapping classes. Many methods have been suggested in the past for such kind of problem but still for overlapping classes, it is hard to determine a global threshold. Hence, attempts have been made by proposing a new approach to determine the global threshold for image segmentation. The algorithm is found to produce satisfactory results for images having histograms with bimodal feature. The algorithm fails for images having histograms with tri-modal features. Currently, attempts are made to address two classimages with noises and images requiring multiple thresholds.Reference[1]K. S. Fu and J. K. Mui, “A Survey on Image segmentation”, Pattern Recognition, vol.13, pp. 3-16, Pergamon Press Ltd, 1981.[2]N. R. Pal and S. K. Pal, “A Review on Image Segmentation Techniques”, PatternRecognition, vol. 26, No. 9, pp. 1277-1294, 1993.[3]P. K. Sahoo, S. Soltani and A. K. C. Wong, “A Survey of Thresholding Techniques”,Computer Vision, Graphics, and Image Processing, vol. 41, 133-260 (1988).[4]M. Sezgin and B. Sankur, Survey over image thresholding techniques and quantitativeperformance evaluation”, Journal of Electronic Imaging, vol. 13(1), pp. 146-165, Jan.2004.[5]J. M. S. Prewitt and M. L. Mendelsohn, “The analysis of cell images”, Ann. N. Y. Acad.Sci., vol. 128, pp. 1035-1053, 1966.[6]J. S. Weszka, R. N. Nagel, and A. Rosenfeld, “A threshold selection technique”, IEEETrans. Comput., vol. C-23, pp. 1322-1326, 1974.[7]N. Otsu, “A Threshold Selection Method from Gray-Level Histograms”, IEEE Trans.Syst., Man, Cybern., vol. SMC-9 (1), pp. 62-66, Jan. 1979.[8]S. Watanable and CYBEST Group, “An automated apparatus for cancer processing:CYBEST”, Comp. Graph. Image processing, vol.3, pp. 350-358, 1974.[9]J. Kittler and J. Illingworth, “Minimum Error Thresholding”, Pattern Recognition, vol.19, No. 1, pp. 41-47, 1986.[10]K. S. Fu and J. K. Mui, “A Survey on Image segmentation”, Pattern Recognition, vol.13, pp. 3-16, Pergamon Press Ltd, 1981.[11]N. R. Pal and S. K. Pal, “A Review on Image Segmentation Techniques”, PatternRecognition, vol. 26, No. 9, pp. 1277-1294, 1993.[12]P. K. Sahoo, S. Soltani and A. K. C. Wong, “A Survey of Thresholding Techniques”,Computer Vision, Graphics, and Image Processing, vol. 41, 133-260 (1988).[13]M. Sezgin and B. Sankur, Survey over image thresholding techniques and quantitativeperformance evaluation”, Journal of Electronic Imaging, vol. 13(1), pp. 146-165, Jan.2004.[14]S. W. Mahfoud, “Simple Analytical Models of Genetic Algorithms for MultimodalFunction Optimization”, Technical Report, Illinois Genetic Algorithm Laboratory, IlliGAL report No. 93001, Feb. 1993[15]S. W. Mahfoud, “Cross over Interaction among Niches”, Technical Report, IllinoisGenetic Algorithm Laboratory, IlliGAL report, April. 1994[16]P. K. Nanda & P. Kanungo,“Parallel Genetic Algorithm Based Class Model forPattern Classification”, Proceedings of the all India Seminar on AES 2002, pp.35-40, February 2002.[17]P. K. Nanda & P. Kanungo, “Parallelized Crowding Scheme Using a NewInterconnection Model”, Proceedings of 2002 AFSS International conference on Fuzzy Systems, (Calcutta, India, February 2002), vol. LNAI 2275, Springer-Verlag, pp. 436-443, (2002).[18]P. K. Nanda & P. Kanungo “Parallel Genetic Algorithm based Crowding SchemeUsing Neighboring Net Topology”, Proceedings of sixth international conference on Information Technology, (Bhubaneswar, India, December 2003), pp. 583-585, (2003).。

加权空-谱局部信息保持极限学习机的高光谱图像分类算法

加权空-谱局部信息保持极限学习机的高光谱图像分类算法

2020年软 件2020, V ol. 41, No. 7基金项目: 辽宁省自然科学基金(批准号:20170540574);辽宁省教育厅科学研究项目(批准号:LJ2019014)作者简介: 邢钰佳(1995–),女,硕士研究生,主要研究方向:机器学习、字典学习、遥感图像分类;闫德勤(1962–),男,教授,主要研究方向:机器学习、字典学习、深度学习、遥感图像分类;刘德山(1970–),男,教授,主要研究方向:智能信息处理,机器学习、模式加权空-谱局部信息保持极限学习机的高光谱图像分类算法邢钰佳,闫德勤,刘德山,王军浩(辽宁师范大学,辽宁 大连 116081)摘 要: 高光谱图像的分类研究是高光谱图像处理与应用的重要环节。

为有效提取高光谱遥感图像的空间信息和光谱信息,本文基于极限学习机提出新的研究。

在模式识别和机器学习领域,极限学习机以其简单、快捷和良好的泛化能力得到越来越多的关注。

但由于在高光谱遥感图像的学习过程中极限学习机缺乏对空间信息和光谱信息的有效提取,无法在分类中提供良好的分类结果。

为此,基于谱局部信息的思想构造本文的研究框架,提出一种加权空-谱局部信息保持极限学习机分类算法。

为验证所提算法的有效性,本文在两组常用的高光谱数据集Indian Pines 和University of Pavia 上进行实验,通过与传统的分类算法SVM 和目前较为流行的分类算法KELM ,KCRT-CK ,MLR 和LPKELM 相比,本文算法具有较好的分类精度。

关键词: 极限学习机;高光谱遥感图像分类;加权空-谱;局部信息保持中图分类号: TP3 文献标识码: A DOI :10.3969/j.issn.1003-6970.2020.07.023本文著录格式:邢钰佳,闫德勤,刘德山,等. 加权空-谱局部信息保持极限学习机的高光谱图像分类算法[J]. 软件,2020,41(07):113 119+135Hyperspectral Image Classification Algorithm for Weighted Spatial SpectralLocality Information Preserving Extreme Learning MachineXING Yu-jia, YAN De-qin, LIU De-shan, WANG Jun-hao(Department of Computer Science, Liaoning Normal University, Dalian 116081, China )【Abstract 】: Classification of hyperspectral images is an important part of hyperspectral image processing and ap-plication. In order to effectively extract the spatial and spectral information of hyperspectral remote sensing images, this paper proposes a new study based on extreme learning machines. In the field of pattern recognition and machine learning, extreme learning machines have attracted more and more attention due to their simplicity, speed and good generalization capabilities. However, due to the lack of effective extraction of spatial information and spectral in-formation during the learning process of hyperspectral remote sensing images, extreme learning machines cannot provide good classification results in classification. To this end, based on the idea of spectral local information, a research framework for this paper is constructed, and a weighted spatial spectral locality information preserving ex-treme learning machine classification algorithm is proposed. In order to verify the effectiveness of the proposed al-gorithm, this paper performs experiments on two commonly used hyperspectral data sets, Indian Pines and Univer-sity of Pavia, and compares with the traditional classification algorithm SVM and the currently popular classifica-tion algorithms KELM, KCRT-CK, MLR Compared with LPKELM, our algorithm has better classification accu-racy.【Key words 】: Extreme learning machine; Hyperspectral image classification; Weighted spatial spectral; Locality information preserving第41卷 第7期 软件0引言高光谱遥感图像(HSI)是由数百个光谱带组成的3维立体图像,如何从中提取大量信息应用于图像分类是遥感图像领域面临的一项挑战。

一维对象复杂度的灰度图像分割算法

一维对象复杂度的灰度图像分割算法

第18卷 第6期太赫兹科学与电子信息学报Vo1.18,No.6 2020年12月 Journal of Terahertz Science and Electronic Information Technology Dec.,2020文章编号:2095-4980(2020)06-1058-07一维对象复杂度的灰度图像分割算法章 怡,王海峰(江苏理工学院信息中心,江苏常州 213001)摘 要:从图像复杂度的角度,提出一种一维对象复杂度的灰度图像分割算法。

用阈值将灰度图像分为背景与目标2类,统计其对应直方图与总像素个数,并计算对象复杂度;依据图像复杂度分割准则算法公式,遍历每一灰度级对应的图像复杂度值,选取图像复杂度值最小对应的灰度值为最佳分割阈值。

仿真实验结果表明,与经典Otsu算法、信息最大熵算法和最小交叉熵算法相比,本文算法速度快,稳定性和效率最好,是一种通用有效的图像分割算法。

关键词:对象复杂度;图像分割;Otsu算法;最大熵;最小交叉熵中图分类号:TN21;TP393文献标志码:A doi:10.11805/TKYDA2018142Grayscale image segmentation algorithm based on one-dimensional object complexityZHANG Yi,WANG Haifeng(Information Center,Jiangsu University of Technology,Changzhou Jiangsu 213001)Abstract:Inspired by the classical segmentation algorithm, this paper proposes a grayscale image segmentation algorithm based on the image complexity. Firstly, the grayscale image is divided intobackground and target categories by the threshold, the corresponding histogram and total number of pixelsare calculated, as well as the complexity of objects. Secondly, according to the image complexitysegmentation criterion, the image complexity of each gray level is calculated. Finally, the optimalsegmentation threshold is obtained by the minimum value of the object complexity. Compared with theother three classical algorithms, the experimental results show that the proposed image segmentationalgorithm is fast, stable and efficient.Keywords:object complexity;image segmentation;Otsu;maximum entropy;minimum cross entropy图像分割是图像理解与计算机视觉的前提,也是图像处理与分析的基本技术之一。

基于深度可分离卷积网络的皮肤镜图像病灶分割方法

基于深度可分离卷积网络的皮肤镜图像病灶分割方法

Dermoscopic image lesion segmentation method based on deep separable convolutional network
CUI Wencheng, ZHANG Pengxia, SHAO Hong
School of Information Science and Engineering, Shenyang University of Technology, Shenyang 110870, China
若 N 占每行、每列像素点总数的比例超过 0.6,则 将该行、该列标记为黑框。定位所有黑框后,在原
始图像中分别根据锁定的位置进行垂直和水平的
黑框移除。 2.2 毛发移除
毛发、血管等天然噪声是无法通过图像采集技
术避免的,但它们的存在对病灶的精准分割造成了
很大的干扰,因此需对图像做毛发移除处理。
毛发移除算法的主要步骤如下。
通过设计和使用神经网络分割框架实现病灶 分割是当前分割任务的研究热点,相比于传统方 法,该方法的分割效果有明显的改善。但是针对皮 肤镜图像的分割任务,目前仍存在黑框和毛发噪声 妨碍定位病灶、边界模糊和病灶分布不均导致精准 分割难以实现的难点。
本文针对皮肤镜图像存在黑框和毛发噪声这 一难点进行黑框和毛发移除处理,以编解码架构为 基础,设计基于深度可分离卷积的特征提取器,实 现皮肤镜图像分割。
Abstract: Aiming at the problem of the difficulty in locating the lesions in dermoscopic images and achieving precise segmentation of the lesions, a method of lesion segmentation in dermatological images based on deep separable convolutional network was proposed. Firstly, perform the black frame removal and hair removal processing on the dermoscopic image to remove the artificial and natural noise that hinders the location of the lesion in the image. Then the image after the noise reduction process was deformed and rotated to expand the data set. Finally, a encoder-decoder segmentation model based on depth separable convolution and hole convolution was constructed. The coding part extracts the features of the image, and the decoding part fuses the feature maps and restores the image details. Experimental results show that this method can achieve better segmentation results for the problem of skin disease image lesion segmentation. The accuracy of segmenting lesions reaches 95.24%. Compared with the segmentation model U-Net, the accuracy is improved by 6.17%. Key words: dermoscopic image, lesion segmentation, hole convolution, deep separable convolution, encoder-decoder model

深度卷积神经网络在光学影像分类中的应用综述

深度卷积神经网络在光学影像分类中的应用综述
文献综述汇报
深度卷积神经网络在光学影像分类中的应用 综述
同济大学 罗新
2018年1月25日
深度学习遥感应用
AutoEncoder自动编码器(Vincent, et al., 2010) RNN循环神经网络(Mou et al., 2017)
深度学习
Geoffrey Hinton
University of Toronto
1、CNN介绍及实现
经典CNN模型
ImageNet LSVRC图像识别竞赛
LeNet(2层), 1986 AlexNet(8层),2012年,ImageNet竞赛冠军
GoogleNet(22层),2014年,ImageNet竞赛冠军 VGGNet(19层),2014年, ImageNet竞赛亚军 ReNet(152层),2015年,ImageNet竞赛冠军
遥 感 应 用
CNN卷积神 经网络
SAR影 像处理
目标识别(Chen et al., 2016) 影像配准(Wang et al., 2017) 目标识别(Chen et al., 2016)
其 他 领 域 应 用
光学影 像处理
图像分割(Alam et al., 2016) 高分影像分类
(Nogueira et al., 2017)
ImageNet Classification top-5 error: 3.57%
Page 7
1、CNN介绍及实现
主流深度学习框架
各个开源框架在GitHub上的数据统计(2017年1月)
Reference: /zuochao_2013/article/details/56024172
Softmax层
Page 22

高光谱图像处理与信息提取前沿

高光谱图像处理与信息提取前沿

3
3.1 3.1.1
高光谱图像处理与信息提取方法
噪声评估与数据降维方法 噪声评估 典型地物具有的诊断性光谱特征是高光谱遥
感目标探测和精细分类的前提,但是由于成像光 谱仪波段通道较密而造成光成像能量不足,相对 于全色图像,高光谱图像的信噪比提高比较困 难。在图像数据获取过程中,地物光谱特征在噪 声的影响下容易产生“失真”,如对某一吸收特征进 行探测,则要求噪声水平比吸收深度要低至少一 个数量级。因此,噪声的精确估计无论对于遥感 器性能评价,还是对于后续信息提取算法的支 撑,都具有重要意义。
张兵:高光谱图像处理与信息提取前沿
1063
得新的突破。高光谱图像处理与信息提取技术的 研究主要包括数据降维、图像分类、混合像元分 解和目标探测等方向(张兵和高连如,2011)。本文 首先从上述4个方向梳理高光谱图像处理与信息提 取中的关键问题,然后分别针对每个方向,在回 顾相关经典理论和模型方法的基础上,介绍近年 来取得的新的代表性成果、发展趋势和未来的研 究热点。此外,高性能计算技术的发展显著提升 了数据处理与分析的效率,在高光谱图像信息提 取中也得到了广泛而成功的应用,因此本文还将 介绍高光谱图像高性能处理技术的发展状况。
题制图的基础数据,在土地覆盖和资源调查以及 环境监测等领域均有着巨大的应用价值。高光谱 图像分类中主要面临Hughes现象(Hughes,1968)和 维数灾难 (Bellman , 2015) 、特征空间中数据非线 性分布等问题。同时,传统算法多是以像元作为 基本单元进行分类,并未考虑遥感图像的空间域 特征,从而使得算法无法有效处理同物异谱问 题,分类结果中地物内部易出现许多噪点。 (4) 高光谱图像提供的精细光谱特征可以用于 区分存在细微差异的目标,包括那些与自然背景 存在较高相似度的目标。因此,高光谱图像目标 探测技术在公共安全和国防领域中有着巨大的应 用潜力和价值。高光谱图像目标探测要求目标具 有诊断性的光谱特征,在实际应用中受目标光谱 的变异性、背景信息分布与模型假设存在差异、 目标地物尺寸处于亚像元级别等问题影响,有时 存在虚警率过高的问题,需要发展稳定可靠的新 方法。 此外,高光谱遥感观测的目的是获取有用的 目标信息,而不是体量巨大的高维原始数据,传 统图像处理平台和信息提取方式难以满足目标信 息快速获取的需求。尽管高性能处理器件的迅猛 发展,为亟待解决的高光谱图像并行快速处理和 在轨实时信息提取提供了实现途径,但也面临着 一系列的关键技术问题。并行处理和在轨实时处 理都需要对算法架构进行优化,同时要依据处理 硬件的特点考虑编程方面的问题,此外,在轨实时 处理还对硬件在功耗等方面提出了特殊的要求。
  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
Jun Li, José M. Bioucas-Dias, Member, IEEE, and Antonio Plaza, Senior Member, IEEE
Abstract—This paper introduces a new supervised Bayesian approach to hyperspectral image segmentation with active learning, which consists of two main steps. First, we use a multinomial logistic regression (MLR) model to learn the class posterior probability distributions. This is done by using a recently introduced logistic regression via splitting and augmented Lagrangian algorithm. Second, we use the information acquired in the previous step to segment the hyperspectral image using a multilevel logistic prior that encodes the spatial information. In order to reduce the cost of acquiring large training sets, active learning is performed based on the MLR posterior probabilities. Another contribution of this paper is the introduction of a new active sampling approach, called modified breaking ties, which is able to provide an unbiased sampling. Furthermore, we have implemented our proposed method in an efficient way. For instance, in order to obtain the time-consuming maximum a posteriori segmentation, we use the α-expansion min-cut-based integer optimization algorithm. The state-of-the-art performance of the proposed approach is illustrated using both simulated and real hyperspectral data sets in a number of experimental comparisons with recently introduced hyperspectral image analysis methods. Index Terms—Active learning, graph cuts, hyperspectral image segmentation, ill-posed problems, integer optimization, mutual information (MI), sparse multinomial logistic regression (MLR).
I. I NTRODUCTION ITH THE recent developments in remote sensing instruments, hyperspectral images are now widely used in different application domains [1]. The special characteristics of hyperspectral data sets bring difficult processing problems. Obstacles, such as the Hughes phenomenon [2], come out as the data dimensionality increases. These difficulties have fostered the development of new classification methods, which are able to deal with ill-posed classification problems. For instance, several machine learning techniques are applied to extract relevant information from hyperspectral data sets [3]–[5]. However,
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 49, NO. 10, OCTOBER 2011
3947Βιβλιοθήκη Hyperspectral Image Segmentation Using a New Bayesian Approach With Active Learning
although many contributions have been made to this area, the difficulty in learning high-dimensional densities from a limited number of training samples (an ill-posed problem) is still an active area of research. Discriminative approaches, which learn the class distributions in high-dimensional spaces by inferring the boundaries between classes in feature space [6]–[8], effectively tackle the aforementioned difficulties. Specifically, support vector machines (SVMs) [9] are among the state-of-the-art discriminative techniques that can be applied to solve ill-posed classification problems. Due to their ability to deal with large input spaces efficiently and to produce sparse solutions, SVMs have been used successfully for supervised and semisupervised classifications of hyperspectral data using limited training samples [1], [3], [10]–[15]. On the other hand, multinomial logistic regression (MLR) [16] is an alternative approach to deal with ill-posed problems, which has the advantage of learning the class probability distributions themselves. This is crucial in the image segmentation step. As a discriminative classifier, MLR directly models the posterior densities instead of the joint probability distributions. The distinguishing features of discriminative classifiers have been reported in the literature before [7], [8], [17]. For instance, effective sparse MLR (SMLR) methods are available in the literature [18]. These ideas have been applied to hyperspectral image classification [5], [19], [20], yielding good performance. Another well-known difficulty in supervised hyperspectral image classification is the limited availability of training data, which are difficult to obtain in practice as a matter of cost and time. In order to effectively work with limited training samples, several methodologies have been proposed, including feature extraction methods such as principal component analysis (PCA), linear discriminant analysis (LDA), discriminant analysis feature extraction, multiple classifiers, and decision fusion [21], among many others [1]. Active learning, which is another active research topic, has been widely studied in the literature [22]–[28]. These studies are based on different principles, such as the evaluation of the disagreement between a committee of classifiers [25], the use of hierarchical classification frameworks [24], [27], unbiased query by bagging [28], or the exploitation of a local proximity-based data regularization framework [26]. In this paper, we use active learning to construct small training sets with high training utility, with the ultimate goal of systematically achieving noticeable improvements in classification results with regard to those found by randomly selected training sets of the same size. Since active learning is intrinsically biased sampling, an issue to be investigated in our experiments
相关文档
最新文档