Graph-based hierarchical conceptual clustering
图神经网络综述
第47卷第4期Vol.47No.4计算机工程Computer Engineering2021年4月April 2021图神经网络综述王健宗,孔令炜,黄章成,肖京(平安科技(深圳)有限公司联邦学习技术部,广东深圳518063)摘要:随着互联网和计算机信息技术的不断发展,图神经网络已成为人工智能和大数据处理领域的重要研究方向。
图神经网络可对相邻节点间的信息进行有效传播和聚合,并将深度学习理念应用于非欧几里德空间的数据处理中。
简述图计算、图数据库、知识图谱、图神经网络等图结构的相关研究进展,从频域和空间域角度分析与比较基于不同信息聚合方式的图神经网络结构,重点讨论图神经网络与深度学习技术相结合的研究领域,总结归纳图神经网络在动作检测、图系统、文本和图像处理任务中的具体应用,并对图神经网络未来的发展方向进行展望。
关键词:图神经网络;图结构;图计算;深度学习;频域;空间域开放科学(资源服务)标志码(OSID ):中文引用格式:王健宗,孔令炜,黄章成,等.图神经网络综述[J ].计算机工程,2021,47(4):1-12.英文引用格式:WANG Jianzong ,KONG Lingwei ,HUANG Zhangcheng ,et al.Survey of graph neural network [J ].Computer Engineering ,2021,47(4):1-12.Survey of Graph Neural NetworkWANG Jianzong ,KONG Lingwei ,HUANG Zhangcheng ,XIAO Jing(Federated Learning Technology Department ,Ping An Technology (Shenzhen )Co.,Ltd.,Shenzhen ,Guangdong 518063,China )【Abstract 】With the continuous development of the computer and Internet technologies ,graph neural network has become an important research area in artificial intelligence and big data.Graph neural network can effectively transmit and aggregate information between neighboring nodes ,and applies the concept of deep learning to the data processing of non-Euclidean space.This paper briefly introduces the research progress of graph computing ,graph database ,knowledge graph ,graph neural network and other graph-based techniques.It also analyses and compares graph neural network structures based on different information aggregation modes in the spectral and spatial domain.Then the paper discusses research fields that combine graph neural network with deep learning ,and summarizes the specific applications of graph neural networks in action detection ,graph systems ,text and image processing tasks.Finally ,it prospects the future development research directions of graph neural networks.【Key words 】graph neural network ;graph structure ;graph computing ;deep learning ;spectral domain ;spatial domain DOI :10.19678/j.issn.1000-3428.00583820概述近年来,深度学习技术逐渐成为人工智能领域的研究热点和主流发展方向,主要应用于高维特征规则分布的非欧几里德数据处理中,并且在图像处理、语音识别和语义理解[1]等领域取得了显著成果。
Graph based concept learning
Graph Based Concept LearningJesus Gonzalez, Lawrence B. Holder, Diane J. Cook University of Texas at Arlington, Department of Computer Science and EngineeringBox 19015, Arlington, TX 76019-0015{gonzalez,holder,cook}@URL: /subdue/ConceptLearningIntroductionConcept Learning is a Machine Learning technique in which the learning process is driven by providing positive and negative examples to the learner. From those examples, the learner builds a hypothesis (concept) that describes the positive examples and excludes the negative examples. Inductive Logic Programming (ILP) systems have successfully been used as concept learners. Examples of those are Foil [1] and Progol [5]. The main engine of these systems is based in first order logic. In this research we introduce a graph based relational concept learning system called SubdueCL, which through the experiments has shown that it is competitive with ILP systems in different types of domains. SubdueCL is an extension made to the Subdue [2] system, which is an unsupervised graph based learner. The extensions made to Subdue to provide the concept learning capability were mainly related to the ability to accept as input positive and negative examples and the criteria used to choose the rules to be added to the hypothesis to be learned.Subdue Discovery SystemGiven a labeled graph as input, the Subdue system [2] discovers substructures (sub-graphs) that compress the input graph, according to the minimum description length principle [6], and represent structural concepts in the data. Note that the input graph need not be a connected graph, which will be true during concept learning when multiple examples are included in one input graph.The main discovery algorithm is a computationally-constrained beam search. The algorithm begins with the substructure matching a single vertex in the graph. Each iteration the algorithm selects the best substructure and incrementally expands the instances of the substructure. The algorithm searches for the best substructure until all possible substructures have been considered or the total amount of computation exceeds a given limit. Evaluation of each substructure is determined by how well the substructure compresses the description length of the input graph. The best substructure found by Subdue can be used to compress the input graph, which can then be input to another iteration of Subdue. After several iterations, Subdue builds a hierarchical description of the input data where later substructures are defined in terms of substructures discovered on previous iterations.Because instances of a substructure can appear in different forms throughout the data, an inexact graph match is used to identify substructure instances with a bounded amount of variation from the substructure definition. Variation is described in terms of basic transformations such as deletion, insertion, and substitution of vertices, edges and/or labels. The fraction of the size of an instance by which the instance can be different from the pattern definition can be specified with the threshold parameter. For example, when a threshold of 0.2 is specified, Subdue will discover all instances of the pattern that have less than 20% difference from the pattern definition. The inexact graph match is constrained to run in polynomial time. Subdue is capable of handling very large graphs (e.g., millions of vertices and edges) by first partitioning the graph so as to minimize the loss of frequent edges, analyzing each partition in parallel, and collaborating to evaluate each substructure to identify a global best substructure [4]. Subdue has been applied to several domains including image analysis, CAD circuit analysis, Chinese characterdatabases, program source code, chemical reaction chains, protein databases, and artificially-generated databases [3, 7].SubdueCLSince SubdueCL is an extension to Subdue, it uses Subdue’s core functions to perform graph operations, but the learning process is different since it works as a supervised learner by differentiating positive and negative examples using a set-covering approach. The hypothesis found by SubdueCL consists of a set of disjunctions of conjunctions (substructures), i.e., the concept may contain several rules. SubdueCL forms one of these conjunctions (rules) in each iteration. Positive example graphs covered in a previous iteration are removed from the graph for subsequent iterations.The way in which SubdueCL decides if the substructures (or rules) will be part of the concept or not is also different from Subdue. SubdueCL uses an evaluation formula to give a value to all the generated substructures. This formula assigns a value to the substructure according to how well they describe the positive examples (or a subset of the positive examples) without describing the negative examples. Then, positive examples covered by the substructure increase the substructure value while negative examples decrease its value. In this formula the positive examples that are not covered and the negative examples covered by the substructure are considered errors because the ideal substructure would be one covering all the positive examples without covering any negative example. Then, the substructure value is calculated as follows:NegEgsPosEgs red NegEgsCove overed PosEgsNotC value ####1++−= Using this formula, SubdueCL chooses rules that maximize the substructure’s value and in this way it minimizes the number of errors made by the substructures used to form the concept.SubdueCL allows hypotheses where some of the rules may cover negative examples. We are working on a version of SubdueCL that produces only consistent hypotheses; this means that the rules of the hypothesis will only cover positive examples without covering any negative example.ResultsAs part of the experiments SubdueCL is being compared with the two ILP systems Foil [1] and Progol [5]. The type of domain that is being used corresponds to attribute-value databases. Four databases have been used for the comparison: Golf, Diabetes, Credit, and Vote. The comparison process consists of a ten fold cross validation (that is 9/10 of the examples are used for the training phase and 1/10 for testing) that produces 10 independent results and a significance test for each domain.The results show that SubdueCL performed better for the Golf, Credit, and Vote domains and Progol performed better for the Diabetes domain. Foil performed second for the Golf, Credit, and Vote domains and third for the Diabetes domain. Progol performed last in the Golf, Credit, and Vote domains. We also compared SubdueCL with C4.5 and it performed slightly better in non numeric domains (Golf, Credit and Vote).ConclusionThe graph based concept learning system SubdueCL has been shown to be competitive (even better) that the ILP systems Foil and Progol in attribute-value domains. The goal now is to show how SubdueCL performs with relational domains where ILP systems have produced very good results. Some preliminary results using relational domains show that SubdueCL is able to learn the tic-tac-toe domain and produce a perfect output just as Foil and Progol did.References[1] R. M. Cameron-Jones and J. R. Quinlan. Efficient top-down induction of logicprograms. SIGART Bulletin, 5(1):33-42, 1994.[2] D. J. Cook, L. B. Holder. Substructure discovery using minimum description lengthand background knowledge. Journal of Artificial Intelligence Research, 1:231-255, 1994.[3] D. J. Cook, L. B. Holder, and S. Djoko. Scalable discovery of informative structuralconcepts using domain knowledge. IEEE Expert, 11(5):59-68, 1996.[4] G. Galal, D. J. Cook, and L. B. Holder. Improving the scalability in a scientificdiscovery system by exploiting parallelism. In Proceedings of the Third International Conference on Knowledge Discovery and Data Mining, pages 171-174, 1997.[5] S. Muggleton. Inverse entailment and Progol. New Generation Computing, 13:245-286, 1995.[6] J. Rissanen. Stochastic Complexity in Statistical Inquiry. World Scientific PublishingCompany, 1989.[7] S. Su, D. J. Cook, and L. B. Holder. Applications of knowledge discovery tomolecular biology: Identifying structural regularities in proteins. In Proceedings of the Pacific Symposium on Biocomputing, pages 190-201, 1999.。
2019年国家自然科学奖提名公示
2019年国家自然科学奖提名公示项目名称 高维图像类脑化语义信息提取理论与方法提 名 者 工业和信息化部提名意见该项目围绕高维图像类脑化语义信息提取展开研究,是国际学术前沿和国家战略需求中的重大科学问题,取得了具有重要创新性和理论价值的基础研究成果。
受大脑基于部件的认知过程启发,提出了一套语义可解释鉴别化紧致特征提取理论和方法,达到了认知可解释和特征可鉴别的统一。
受大脑协同思维的启发,提出了基于时空协同和内部协同的图像协同处理架构,显著提升了图像识别的准确率。
受大脑演化认知特性启发,提出了基于演化表达的图像信息描述理论和方法,突破了手工设计描述子的局限。
该项目的研究成果得到了中国、美国、英国、加拿大等国院士和IEEE/ACM/OSA/SPIE/IAPR/AAAS Fellow等国际同行的高度认可。
有力推动了本学科及相关学科的发展。
符合国家科学技术奖的推荐条件。
提名该项目为国家自然科学奖二等奖。
项目简介针对语义丰富的高维图像信号,借鉴和发展脑科学、认知科学的成果,对高维图像进行类脑化处理是减少和消除语义鸿沟重大问题的关键,代表了当前信息领域的发展方向之一。
本项目以类脑化处理为手段,建立了语义可解释、协同认知、演化表达三个方面的高维图像语义信息提取理论和方法。
取得了如下科学发现:1.受大脑基于部件的认知过程启发,提出了同时提取隐藏语义和保持内在流形结构的非负矩阵分解理论框架,并进一步将其发展为流形正则化稀疏非负矩阵分解方法并应用到高光谱图像处理和恢复中,解决了非负矩阵分解可解释性强但鉴别能力弱的难题。
还提出了确定性列矩阵分解子空间学习方法。
形成了一套可解释鉴别化语义特征提取理论和方法,一定程度上达到了认知可解释和特征可鉴别的统一。
2.受大脑协同思维的启发,提出了图像及视频的协同处理框架:针对视频内容识别,提出了时空拉普拉斯金字塔协同编码机制,能够同时捕获结构信息和运动信息,克服了传统方法误配准问题。
AlgebraicGraphTheory
Algebraic Graph TheoryChris Godsil(University of Waterloo),Mike Newman(University of Ottawa)April25–291Overview of the FieldAlgebraic graph theory comprises both the study of algebraic objects arising in connection with graphs, for example,automorphism groups of graphs along with the use of algebraic tools to establish interesting properties of combinatorial objects.One of the oldest themes in the area is the investigation of the relation between properties of a graph and the spectrum of its adjacency matrix.A central topic and important source of tools is the theory of association schemes.An association scheme is,roughly speaking,a collection of graphs on a common vertex set whichfit together in a highly regular fashion.These arise regularly in connection with extremal structures:such structures often have an unex-pected degree of regularity and,because of this,often give rise to an association scheme.This in turn leads to a semisimple commutative algebra and the representation theory of this algebra provides useful restrictions on the underlying combinatorial object.Thus in coding theory we look for codes that are as large as possible, since such codes are most effective in transmitting information over noisy channels.The theory of association schemes provides the most effective means for determining just how large is actually possible;this theory rests on Delsarte’s thesis[4],which showed how to use schemes to translate the problem into a question that be solved by linear programming.2Recent Developments and Open ProblemsBrouwer,Haemers and Cioabˇa have recently shown how information on the spectrum of a graph can be used to proved that certain classes of graphs must contain perfect matchings.Brouwer and others have also investigated the connectivity of strongly-regular and distance-regular graphs.This is an old question,but much remains to be done.Recently Brouwer and Koolen[2]proved that the vertex connectivity of a distance-regular graph is equal to its valency.Haemers and Van Dam have worked on extensively on the question of which graphs are characterized by the spectrum of their adjacency matrix.They consider both general graphs and special classes,such as distance-regular graphs.One very significant and unexpected outcome of this work was the construction,by Koolen and Van Dam[10],of a new family of distance-regular graphs with the same parameters as the Grassmann graphs.(The vertices of these graphs are the k-dimensional subspaces of a vector space of dimension v over thefinitefield GF(q);two vertices are adjacent if their intersection has dimension k1.The graphs are q-analog of the Johnson graphs,which play a role in design theory.)These graphs showed that the widely held belief that we knew all distance-regular graphs of“large diameter”was false,and they indicate that the classification of distance-regular graphs will be more complex(and more interesting?)than we expected.1Association schemes have long been applied to problems in extremal set theory and coding theory.In his(very)recent thesis,Vanhove[14]has demonstrated that they can also provide many interesting results in finite geometry.Recent work by Schrijver and others[13]showed how schemes could used in combination with semidef-inite programming to provide significant improvements to the best known bounds.However these methods are difficult to use,we do not yet have a feel for we might most usefully apply them and their underlying theory is imperfectly understood.Work in Quantum Information theory is leading to a wide range of questions which can be successfully studied using ideas and tools from Algebraic Graph Theory.Methods fromfinite geometry provide the most effective means of constructing mutually unbiased bases,which play a role in quantum information theory and in certain cryptographic protocols.One important question is to determine the maximum size of a set of mutually unbiased bases in d-dimensional complex space.If d is a prime power the geometric methods just mentioned provide sets of size d+1,which is the largest possible.But if d is twice an odd integer then in most cases no set larger than three has been found.Whether larger sets exist is an important open problem. 3Presentation HighlightsThe talks mostlyfitted into one of four areas,which we discuss separately.3.1SpectraWillem Haemers spoke on universal adjacency matrices with only two distinct eigenvalues.Such matrices are linear combinations of I,J,D and A(where D is the diagonal matrix of vertex degrees and A the usual adjacency matrix).Any matrix usually considered in spectral graph theory has this form,but Willem is considering these matrices in general.His talk focussed on the graphs for which some universal adjacency matrix has only two eigenvalues.With Omidi he has proved that such a graph must either be strong(its Seidel matrix has only two eigenvalues)or it has exactly two different vertex degrees and the subgraph induced by the vertices of a given degree must be regular.Brouwer formulated a conjecture on the minimum size of a subset S of the vertices of a strongly-regular graph X such that no component of X\S was a single vertex.Cioabˇa spoke on his recent work with Jack Koolen on this conjecture.They proved that it is false,and there are four infinite families of counterexamples.3.2PhysicsAs noted above,algebraic graph theory has many applications and potential applications to problems in quantum computing,although the connection has become apparent only very recently.A number of talks were related to this connection.One important problem in quantum computing is whether there is a quantum algorithm for the graph isomorphism problem that would be faster than the classical approaches.Currently the situation is quite open.Martin Roetteler’s talk described recent work[1]on this problem.For our workshop’s viewpoint,one surprising feature is that the work made use of the Bose-Mesner algebra of a related association scheme; this connection had not been made before.Severini discussed quantum applications of what is known as the Lov´a sz theta-function of a graph.This function can be viewed as an eigenvalue bound and is closely related to both the LP bound of Delsarte and the Delsarte-Hoffman bound on the size of an independent set in a regular graph.Severini’s work shows that Lov´a sz’s theta-function provides a bound on the capacity of a certain channel arising in quantum communication theoryWork in quantum information theory has lead to interest in complex Hadamard matrices—these are d×d complex matrices H such that all entries of H have the same absolute value and HH∗=dI.Both Chan and Sz¨o ll˝o si dealt with these in their talks.Aidan Roy spoke on complex spherical designs.Real spherical designs were much studied by Seidel and his coworkers,because of their many applications in combinatorics and other areas.The complex case languished because there were no apparent applications,but now we have learnt that these manifest them-selves in quantum information theory under acronyms such as MUBs and SIC-POVMs.Roy’s talk focussedon a recent 45page paper with Suda [12],where (among other things)they showed that extremal complex designs gave rise to association schemes.One feature of this work is that the matrices in their schemes are not symmetric,which is surprising because we have very few interesting examples of non-symmetric schemes that do not arise as conjugacy class schemes of finite groups.3.3Extremal Set TheoryCoherent configurations are a non-commutative extension of association schemes.They have played a sig-nificant role in work on the graph isomorphism problem but,in comparison with association schemes,they have provided much less information about interesting extremal structures.The work presented by Hobart and Williford may improve matters,since they have been able to extend and use some of the standard bounds from the theory of schemes.Delsarte [4]showed how association schemes could be used to derive linear programs,whose values provided strong upper bounds on the size of codes.Association schemes have both a combinatorial structure and an algebraic structure and these two structures are in some sense dual to one another.In Delsarte’s work,both the combinatorial and the algebraic structure had a natural linear ordering (the schemes are both metric and cometric)and this played an important role in his work.Martin explained how this linearity constraint could be relaxed.This work is important since it could lead to new bounds,and also provide a better understanding of duality.One of Rick Wilson’s many important contributions to combinatorics was his use of association schemes to prove a sharp form of the Erd˝o s-Ko-Rado theorem [15].The Erd˝o s-Ko-Rado theorem itself ([5])can certainly be called a seminal result,and by now there are many analogs and extensions of it which have been derived by a range of methods.More recently it has been realized that most of these extensions can be derived in a very natural way using the theory of association schemes.Karen Meagher presented recent joint work (with Godsil,and with Spiga,[8,11])on the case where the subsets in the Erd˝o s-Ko-Rado theorem are replaced by permutations.It has long been known that there is an interesting association scheme on permutations,but this scheme is much less manageable than the schemes used by Delsarte and,prior to the work presented by Meagher,no useful combinatorial information had been obtained from it.Chowdhury presented her recent work on a conjecture of Frankl and F¨u redi.This concerns families F of m -subsets of a set X such that any two distinct elements of have exactly λelements in common.Frankl and F¨u redi conjectured that the m -sets in any such family contain at least m 2 pairs of elements of X .Chowdhury verified this conjecture in a number of cases;she used classical combinatorial techniques and it remains to see whether algebraic methods can yield any leverage in problems of this type.3.4Finite GeometryEric Moorhouse spoke on questions concerning automorphism groups of projective planes,focussing on connections between the finite and infinite case.Thus for a group acting on a finite plane,the number of orbits on points must be equal to the number of orbits on lines.It is not known if this must be true for planes of infinite order.Is there an infinite plane such that for each positive integer k ,the automorphism group has only finitely many orbits on k -tuples?This question is open even for k =4.Simeon Ball considered the structure of subsets S of a k -dimensional vector space over a field of order q such that each d -subset of S is a basis.The canonical examples arise by adding a point at infinity to the point set of a rational normal curve.These sets arise in coding theory as maximum distance separable codes and in matroid theory,in the study of the representability of uniform matroids (to mention just two applications).It is conjectured that,if k ≤q −1then |S |≤q +1unless q is even and k =3or k =q −1,in which case |S |≤q +2.Simeon presented a proof of this theorem when q is a prime and commented on the general case.He developed a connection to Segre’s classical characterization of conics in planes of odd order,as sets of q +1points such that no three are collinear.There are many analogs between finite geometry and extremal set theory;questions about the geometry of subspaces can often be viewed as q -analogs of questions in extremal set theory.So the EKR-problem,which concerns characterizations of intersecting families of k -subsets of a fixed set,leads naturally to a study of intersecting families of k -subspaces of a finite vector space.In terms of association schemes this means we move from the Johnson scheme to the Grassmann scheme.This is fairly well understood,with thebasic results obtained by Frankl and Wilson[6].But infinite geometry,polar spaces form an important topic. Roughly speaking the object here is to study the families of subspaces that are isotropic relative to some form, for example the subspaces that lie on a smooth quadric.In group theoretic terms we are now dealing with symplectic,orthogonal and unitary groups.There are related association schemes on the isotropic subspaces of maximum dimension.Vanhove spoke on important work from his Ph.D.thesis,where he investigated the appropriate versions of the EKR problem in these schemes.4Outcome of the MeetingIt is too early to offer much in the way of concrete evidence of impact.Matt DeV os observed that a conjecture of Brouwer on the vertex connectivity of graphs in an association scheme was wrong,in a quite simple way. This indicates that the question is more complex than expected,and quite possibly more interesting.That this observation was made testifies to the scope of the meeting.On a broader level,one of the successes of the meeting was the wide variety of seemingly disparate topics that were able to come together;the ideas of algebraic graph theory touch a number of things that would at first glance seem neither algebraic nor graph theoretical.There was a lively interaction between researchers from different domains.The proportion of post-docs and graduate students was relatively high.This had a positive impact on the level of excitement and interaction at the meeting.The combination of expert and beginning researchers created a lively atmosphere for mathematical discussion.References[1]A.Ambainis,L.Magnin,M.Roetteler,J.Roland.Symmetry-assisted adversaries for quantum state gen-eration,arXiv1012.2112,35pp.[2]A.E.Brouwer,J.H.Koolen.The vertex connectivity of a distance-regular graph.European bina-torics30(2009),668–673.[3]A.E.Brouwer,D.M.Mesner.The connectivity of strongly regular graphs.European binatorics,6(1985),215–216.[4]P.Delsarte.An algebraic approach to the association schemes of coding theory.Philips Res.Rep.Suppl.,(10):vi+97,1973.[5]P.Erd˝o s,C.Ko,R.Rado.Intersection theorems for systems offinite sets.Quart.J.Math.Oxford Ser.(2),12(1961),313–320.[6]P.Frankl,R.M.Wilson.The Erd˝o s-Ko-Rado theorem for vector binatorial Theory,SeriesA,43(1986),228–236.[7]D.Gijswijt,A.Schrijver,H.Tanaka.New upper bounds for nonbinary codes based on the Terwilligeralgebra and semidefinite binatorial Theory,Series A,113(2006),1719–1731. [8]C.D.Godsil,K.Meagher.A new proof of the Erd˝o s-Ko-Rado theorem for intersecting families of per-mutations.arXiv0710.2109,18pp.[9]C.D.Godsil,G.F.Royle.Algebraic Graph Theory,Springer-Verlag,(New York),2001.[10]J.H.Koolen,E.R.van Dam.A new family of distance-regular graphs with unbounded diameter.Inven-tiones Mathematicae,162(2005),189-193.[11]K.Meagher,P.Spiga.An Erdos-Ko-Rado theorem for the derangement graph of PGL(2,q)acting onthe projective line.arXiv0910.3193,17pp.[12]A.P.Roy,plex spherical Codes and designs,(2011),arXiv1104.4692,45pp.[13]A.Schrijver.New code upper bounds from the Terwilliger algebra and semidefinite programming.IEEETransactions on Information Theory51(2005),2859–2866.[14]F.Vanhove.Incidence geometry from an algebraic graph theory point of view.Ph.D.Thesis,Gent2011.[15]R.M.Wilson.The exact bound in the Erds-Ko-Rado binatorica,4(1984),247–257.。
知识图谱综述表示、构建、推理与知识超图理论
知识图谱综述表示、构建、推理与知识超图理论一、本文概述知识图谱作为一种结构化的知识库,集成了来自多个来源的信息,通过实体、概念和关系来组织和呈现现实世界中的复杂知识。
近年来,随着大数据和技术的快速发展,知识图谱已成为信息抽取、自然语言处理、智能问答、推荐系统等多个领域的研究热点。
本文旨在全面综述知识图谱的表示、构建、推理及其与知识超图理论的联系。
文章首先介绍知识图谱的基本概念和应用背景,随后深入探讨其表示方法、构建技术和推理算法,并在此基础上分析知识超图理论与知识图谱的内在关联。
本文的综述将为相关领域的研究者提供全面、深入的理论支持和实践指导。
二、知识图谱的表示知识图谱的表示是知识图谱构建和应用的关键环节。
它涉及到如何将现实世界中的实体、属性、关系等复杂的信息结构转化为计算机可以理解和处理的数据结构。
知识图谱的表示方式主要分为两大类:符号表示和分布式表示。
符号表示:符号表示是传统的知识表示方法,主要包括一阶谓词逻辑、描述逻辑、语义网络等。
这种表示方式能够清晰地描述实体间的复杂关系,易于人类理解和解释。
然而,符号表示在处理大规模知识图谱时存在效率低下的问题,难以处理模糊和不确定的知识。
分布式表示:为了克服符号表示的不足,近年来分布式表示方法逐渐受到关注。
分布式表示方法将实体和关系表示为低维稠密的向量,通过向量运算来模拟实体间的关系推理。
这种方法能够有效地处理大规模知识图谱,并且在处理模糊和不确定知识方面具有一定的优势。
其中,TransE、TransH、TransR等翻译模型是分布式表示中的代表性方法,它们在链接预测、实体分类等任务中取得了显著的效果。
随着深度学习技术的发展,基于神经网络的知识表示方法也逐渐兴起。
这类方法通过训练神经网络来学习实体和关系的表示,能够捕获更丰富的语义信息。
例如,卷积神经网络(CNN)和循环神经网络(RNN)等模型被应用于知识图谱的表示学习中,取得了良好的效果。
知识图谱的表示方法经历了从符号表示到分布式表示再到基于神经网络的知识表示的发展历程。
基于个性化本体的图像语义标注和检索
w l a es ma t i lr yme s r a e n c n e ts t n n e—ea in hp ”I— ” h smo e s lme t d a d el s t e n i smi i a u e b s d o o c p e d i trr lt s i h c at a o sA .T i d l Wa i e n e mp n s me r lt d e p r n s w r a r d o t T e e p r n a e u t v i ae h er v l a c r c s h g s 8 . % . o e ae x e me t e e c r e u . h x e me tl rs l a d t te rt e a c u a y a ih a 8 6 i i i s l i
i g e e ta td h i g e e a i faurs a e o te o c p s t n u e i trr lto s p ”I— ” o n oo t ma , xr c e t e ma s m nt e t e b s d n h c n e t e a d s d ne —e ain hi c sA f o t lg o y me s r h i l rt t e h m a e n e y Th f c ly le n te pe s n lz d e outo ft o no o , a a u e t e smiaiy bewe n t e i g s a d qu r . e di ut isi h ro aie v l i n o he tp o tlg i y s
国内外遥感核心期刊
国内外遥感核心期刊中国科技论文统计源期刊-中国科技核心期刊:科技部中信所评价期刊学术质量和影响得出,用于科研绩效评估。
中国科学引文索引数据库:中科院编制,偏重于基础科学领域的期刊中文核心期刊:北京大学图书馆编制,指导图书馆的文献采购中国核心期刊遴选数据库:万方数据公司制作的科技期刊资源数据库,不用于评价中国学术期刊(光盘版)/中国期刊全文数据库/中国学术期刊综合评价数据库:清华大学制作的科技期刊资源数据库,不用于评价中文核心期刊:------------------------------------------------------------------------------- 遥感学报地理与地理信息科学地理研究计算机工程与应用微计算机信息计算机应用研究中国图像图形学报计算机应用与软件测绘学报武汉大学学报信息科学版测绘通报地图遥感学报大地测量与地球动力学测绘科学测绘学院学报安徽农业科学中国科技核心期刊:---------------------------------------------------------遥感技术与应用遥感信息地球科学信息世界地质国土资源遥感环境保护科学测绘工程普通期刊:-------------------------------------------------------------地理空间信息国外遥感类相关杂志与投稿1. 期刊名称:GPS SOLUTIONSISSN: 1080-5370出版频率: Quarterly出版社: SPRINGER HEIDELBERG, TIERGARTENSTRASSE 17, HEIDELBERG, GERMANY, D-69121 影响因子:主题范畴: REMOTE SENSING变更情况: 2005New2. 期刊名称:SURVEY REVIEWISSN: 0039-6265出版频率: Quarterly出版社: COMMONWEALTH ASSOC SURVEYING LAND ECONOMY, C A S L E, UNIV WEST ENGLAND,C/O FACULTY BUILT ENVIRONMENT,FRENCHAY CAMPUS, COLDHARBOUR LBRISTOL, ENGLAND, BS16 1QY 期刊网址:/影响因子: 0.102(2002)主题范畴: GEOSCIENCES, MULTIDISCIPLINARY; REMOTE SENSING; ENGINEERING, CIVIL 3. 期刊名称:PHOTOGRAMMETRIC RECORDISSN: 0031-868X出版频率: Quarterly出版社: PHOTOGRAMMETRIC SOC, UNIV COLL LONDON, DEPT GEOMATIC ENGINEERING,GOWER ST, LONDON, ENGLAND, WC1E 6BT出版社网址:/期刊网址:/publications/publicationsFrameset.htm影响因子: 0.353(2001); 0.633(2002)主题范畴: GEOGRAPHY, PHYSICAL; GEOSCIENCES, MULTIDISCIPLINARY; REMOTE SENSING; PATHOLOGY4. 期刊名称:INTERNATIONAL JOURNAL OF REMOTE SENSINGISSN: 0143-1161版本: SCI-CDE出版频率: Semimonthly出版社: TAYLOR & FRANCIS LTD, 4 PARK SQUARE, MILTON PARK, ABINGDON, ENGLAND, OX14 4RN出版社网址:/期刊网址:/journals/tf/01431161.html影响因子: 0.827(2001),1.154(2002)主题范畴: REMOTE SENSING; IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY5. 期刊名称:PHOTOGRAMMETRIC ENGINEERING AND REMOTE SENSINGISSN: 0099-1112版本: SCI-CDE出版频率: Monthly出版社: AMER SOC PHOTOGRAMMETRY, 5410 GROSVENOR LANE, SUITE 210, BETHESDA, MD, 20814-2160出版社网址:/期刊网址:/publications.html影响因子: 0.841(2001);1.176(2002)主题范畴: GEOGRAPHY, PHYSICAL; GEOSCIENCES, MULTIDISCIPLINARY; REMOTE SENSING; PATHOLOGY6. 期刊名称:JOURNAL OF GEODESYISSN: 0949-7714版本: SCI-CDE出版频率: Monthly出版社: SPRINGER-VERLAG, 175 FIFTH AVE, NEW YORK, NY, 10010出版社网址:/期刊网址:/app/ ... gpublicationresults,id:100435,1 影响因子: 0.960(2001),0.726(2002)主题范畴: GEOCHEMISTRY & GEOPHYSICS; REMOTE SENSING7. 期刊名称:ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSINGISSN: 0924-2716出版频率: Quarterly出版社: ELSEVIER SCIENCE BV, PO BOX 211, AMSTERDAM, NETHERLANDS, 1000 AE出版社网址:http://www.elsevier.nl期刊网址:http://www.elsevier.nl/locate/isprsjprs影响因子: 0.963(2001),0.389(2002)主题范畴: GEOGRAPHY, PHYSICAL; GEOSCIENCES, MULTIDISCIPLINARY; REMOTE SENSING; IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY8. 期刊名称:RADIO SCIENCEISSN: 0048-6604版本: SCI-CDE出版频率: Bimonthly出版社: AMER GEOPHYSICAL UNION, 2000 FLORIDA AVE NW, WASHINGTON, DC, 20009出版社网址:/期刊网址:/journals/rs/影响因子: 1.139(2001),0.796(2002)主题范畴: GEOCHEMISTRY & GEOPHYSICS; METEOROLOGY & ATMOSPHERIC SCIENCES; REMOTE SENSING; TELECOMMUNICATIONS; INSTRUMENTS & INSTRUMENTATION9. 期刊名称:IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSINGISSN: 0196-2892版本: SCI-CDE出版频率: Bimonthly出版社: IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 345 E 47TH ST, NEW YORK, NY, 10017-2394出版社网址:/portal/index.jsp期刊网址:/soc/grss/tgars.html影响因子: 1.605(2001),1.603(2002)主题范畴: GEOCHEMISTRY & GEOPHYSICS; REMOTE SENSING; ENGINEERING, ELECTRICAL & ELECTRONIC10. 期刊名称:REMOTE SENSING OF ENVIRONMENTISSN: 0034-4257版本: SCI-CDE出版频率: Monthly出版社: ELSEVIER SCIENCE INC, 360 PARK AVE SOUTH, NEW YORK, NY, 10010-1710出版社网址:http://www.elsevier.nl/期刊网址:http://www.elsevier.nl/inca/publications/store/5/0/5/7/3/3/index.htt 影响因子: 1.697(2001),1.992(2002)主题范畴: REMOTE SENSING; ENVIRONMENTAL SCIENCES; IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY11. 期刊名称:CANADIAN JOURNAL OF REMOTE SENSINGISSN: 0703-8992出版频率: Bimonthly出版社: CANADIAN AERONAUTICS SPACE INST, 1685 RUSSELL RD, UNIT 1-R, OTTAWA, CANADA, K1G 0N1出版社网址:http://www.casi.ca/期刊网址:http://www.casi.ca/index.php?pg=cjrs影响因子: no主题范畴: REMOTE SENSING12. 期刊名称:IEE Proceedings -- Radar, Sonar & Navigation (已经更名:IET Radar, Sonar & Navigation)ISSN: 1350-2395版本: SCIE出版频率: Bimonthly出版社: IEE-INST ELEC ENG, MICHAEL FARADAY HOUSE SIX HILLS WAY STEVENAGE, HERTFORD, ENGLAND, SG1 2AY出版社网址:期刊网址:/IP-RSN影响因子: no主题范畴: radar, radio location, radio navigation and surveillance purposes. Examples of the fields of application include radar, sonar, electronic warfare, avionic and navigation systems. Processing directed towards the above application areas includes advances in matched filters and wideband signal correlation for radar and sonar systems; algorithms and processor designs for adaptive array; bearing estimation; range/Doppler radar and acoustic image processing operations for SAR, sonar, target identification functions, etc13. 期刊名称:IEEE Transactions on Image ProcessingISSN: 1057-7149版本: SCI出版频率: Monthly出版社: IEEE Signal Processing Society出版社网址:期刊网址:/servlet/opac?punumber=83影响因子: no主题范畴: Signal-processing aspects of image processing, imaging systems, and image scanning, display, and printing. Includes theory, algorithms, andarchitectures for image coding, filtering, enhancement, restoration, segmentation, and motion estimation; image formation in tomography, radar, sonar, geophysics, astronomy, microscopy, and crystallography; image scanning, digital half-toning and display, andcolor reproduction.14. 期刊名称:Geophysical Research LettersISSN: 0094-8276版本: SCI出版频率: Semimonthly出版社: AMER GEOPHYSICAL UNION, 2000 FLORIDA AVE NW, WASHINGTON, USA, DC, 20009 出版社网址:期刊网址:/journals/gl/影响因子: 2.491(2005)主题范畴: focus on a specific discipline or apply broadly to the geophysical science community15. 期刊名称: IEEE Transactions on Geoscience and Remote Sensing Letter ISSN: 0196-2892版本:出版频率:出版社: TGARS Manuscript Reivew Assistant, GEOSCIENCE AND REMOTE SENSING LETTERS,IEEE Periodicals,445 Hoes Lane Piscataway, NJ 08855 USA出版社网址:期刊网址:/menu.taf?menu=publications&detail=GRSL影响因子:主题范畴: GEOCHEMISTRY & GEOPHYSICS; REMOTE SENSING; ENGINEERING, ELECTRICAL & ELECTRONIC16. 期刊名称:IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMSISSN: 0018-9251版本: sci出版频率: Quarterly出版社: Aerospace & Electronic Systems Society出版社网址:期刊网址:/servlet/opac?punumber=7影响因子:主题范畴: the equipment, procedures, and techniques applicable to the organization, installation, and operation of functional systems designed to meet the high performance requirements of earth and space systems17. 期刊名称:Pattern Recognition LettersISSN: 0167-8655版本: SCIE出版频率: Subscriptions for the year 2007, Volume 28, 16 issues出版社: ELSEVIER SCIENCE BV, PO BOX 211, AMSTERDAM, NETHERLANDS, 1000 AE出版社网址:期刊网址:/wps/find/journaldescription.cws_home/505619/descr iption#description影响因子: 2005: 1.138主题范畴: ? statistical, structural, syntactic pattern recognition;? neural networks, machine learning, data mining;? discrete geometry, algebraic, graph-based techniques for pattern recognition;? signal analysis, image coding and processing, shape and texture analysis;? computer vision, robotics, remote sensing;? document processing, text and graphics recognition, digital libraries;? speech recognition, music analysis, multimedia systems;? natural language analysis, information retri;? biometrics, biomedical pattern analysis and information systems;? scientific, engineering, social and economical applications of pattern recognition;? special hardware architectures, software packages for pattern recognition.18. 期刊名称:Multidimensional Systems and Signal ProcessingISSN: 0923-6082 (Print) 1573-0824 (Online)版本:出版频率: Monthly出版社: Springer Netherlands出版社网址:期刊网址:/site/catalog/Journal/1582.jsp?top=2&mid=3&bottom=7&su bsection=12影响因子: 0.722 (2005)主题范畴: While the subject of multidimensional systems is concerned with mathematical issues designed to tackle a broad range of models, its applications in signal processing have been known to cover spatial and temporal signals of diverse physical origin. The current problem faced, due to the widely scattered nature of publications in this area, will be circumvented through the unity of theme in thisjournal, so that research is facilitated and expected with much reduced duplication of effort and much enhanced communication.19. 期刊名称:International Journal of Applied Earth Observation and Geoinformation ISSN: 0303-2434版本: SCIE出版频率: Quarterly出版社: ELSEVIER出版社网址:期刊网址:/wps/find/journaldescription.cws_home/622741/descr iption#description影响因子:主题范畴: The International Journal of Applied Earth Observation and Geoinformation publishes original papers that apply earth observation data to inventory and management of natural resources and the environment. In this context, earth observation data are normally those acquired from remote sensing platforms such as satellites and aircraft, complemented and supplemented by surface and subsurface measurements and mapping. Natural resources include forests, agricultural land, soils, water resources, mineral deposits, and land itself as a foundation for infrastructure and housing. Environmental issues include biodiversity, land degradation, industrial pollution and natural hazards such as earthquakes, floods and landslides. The focus, which can be either conceptual or data driven, includes all major themes in geoinformation, like capturing, databasing, visualization and interpretation of data, but also issues of data quality and spatialuncertainty. Since the scope is large, contributions should be of the highest quality. Some will convey important recommendations for environmental management and policy, and we encourage 'Discussion' articles that stimulate dialogue between earth observation studies and managers in a statistically sound way. Papers addressing these topics in the context of the social fabric and economic constraints of developing countries are particularly welcome.20. 期刊名称:Computers & GeosciencesISSN: 0098-3004版本: SCIE出版频率: Subscriptions for the year 2007, Volume 33, 10 issues出版社: ELSEVIER出版社网址:期刊网址:/wps/find/journaldescription.cws_home/398/descript ion#description影响因子: 2005: 0.779主题范畴: spatial analysis, geomathematics, modelling, simulation, statistical and artificial intelligence methods, e-geoscience, geoinformatics, geomatics, geocomputation, image analysis, remote sensing, and geographical information science.21. 期刊名称:SIGNAL PROCESSINGISSN: 0165-1684版本: SCIE出版频率: Monthly出版社: ELSEVIER出版社网址:期刊网址:/wps/find/journaldescription.cws_home/505662/descr iption#description影响因子: 2005: 0.694主题范畴: Signal Theory; Stochastic Processes; Detection and Estimation; Spectral Analysis; Filtering; Signal Processing Systems; Software Developments; Image Processing; Pattern Recognition; Optical Signal Processing; Digital Signal Processing; Multi-dimensional Signal Processing; Communication Signal Processing; Biomedical Signal Processing; Geophysical and Astrophysical Signal Processing; Earth Resources Signal Processing; Acoustic and Vibration Signal Processing; Data Processing; Remote Sensing; Signal Processing Technology; Speech Processing; Radar Signal Processing; Sonar Signal Processing; Industrial Applications; New Applications22. 期刊名称:Journal of Quantitative Spectroscopy & Radiative TransferISSN: 0022-4073版本: SCIE出版频率: Subscriptions for the year 2007, Volumes 103-108, 18 issues出版社: ELSEVIER出版社网址:期刊网址:/wps/find/journaldescription.cws_home/272/description#description影响因子: 2005: 1.685主题范畴:· Theoretical and experimental aspects of the spectra of atoms, molecules, ions, and plasmas. · Spectral lineshape studies including models and computational algorithms. · Atmospheric spectroscopy. · Theoretical and experimental aspects of light scattering. · Application of light scattering in particle characterization and remote sensing. · Application of light scattering in biological sciences and medicine. · Radiative transfer in absorbing, emitting, and scattering media. · Radiative transfer in stochasticmedia. · Electromagnetic energy transfer with near-field, nano-scale, and coherent effects. · Planetary, atmospheric, and environmental radiative transfer. · Radiative transfer in high-temperature environments, combustion systems, and fires. · Radiant energy emission from plasmas.。
R L 算 法 介 绍 及 比 较
人工智能常见算法简介人工智能的三大基石—算法、数据和计算能力,算法作为其中之一,是非常重要的,那么人工智能都会涉及哪些算法呢?不同算法适用于哪些场景呢?一、按照模型训练方式不同可以分为监督学习(Supervised Learning),无监督学习(Unsupervised Learning)、半监督学习(Semi-supervised Learning)和强化学习(Reinforcement Learning)四大类。
常见的监督学习算法包含以下几类:(1)人工神经网络(Artificial Neural Network)类:反向传播(Backpropagation)、波尔兹曼机(Boltzmann Machine)、卷积神经网络(Convolutional Neural Network)、Hopfield网络(hopfield Network)、多层感知器(Multilyer Perceptron)、径向基函数网络(Radial Basis Function Network,RBFN)、受限波尔兹曼机(Restricted Boltzmann Machine)、回归神经网络(Recurrent Neural Network,RNN)、自组织映射(Self-organizing Map,SOM)、尖峰神经网络(Spiking Neural Network)等。
(2)贝叶斯类(Bayesin):朴素贝叶斯(Naive Bayes)、高斯贝叶斯(Gaussian Naive Bayes)、多项朴素贝叶斯(Multinomial Naive Bayes)、平均-依赖性评估(Averaged One-Dependence Estimators,AODE)贝叶斯信念网络(Bayesian Belief Network,BBN)、贝叶斯网络(Bayesian Network,BN)等。
(3)决策树(Decision Tree)类:分类和回归树(Classification and Regression Tree,CART)、迭代Dichotomiser3(Iterative Dichotomiser 3, ID3),C4.5算法(C4.5 Algorithm)、C5.0算法(C5.0 Algorithm)、卡方自动交互检测(Chi-squared Automatic Interaction Detection,CHAID)、决策残端(Decision Stump)、ID3算法(ID3 Algorithm)、随机森林(Random Forest)、SLIQ(Supervised Learning in Quest)等。
基于高层语义的图像检索
Levels 2 and 3 together are referred to as semantic image retrieval, and the gap between Levels 1 and 2 as the semantic gap .
3.The low-level image feature
In CBIR,images are indexed by their visual content, such as color,texture, shapes. A pioneering work was published by Chang in 1984, in which the author presented a picture indexing and abstraction approach for pictorial database retrieval . The pictorial database consists of picture objects and picture relations.To construct picture indexes, abstraction operations are formulated to perform picture object clustering and classification.
通信系统之英语单词
通信系统之英语单词1, inoperative [in'?p?r?tiv]a. 不活动的,无效力的,不发生效力的2, synchronization [,si?kr?nai'zei??n]同步化3, synchronize ['si?kr?naiz]v. 使同时,同时发生4, luminance ['lju:min?ns]亮度5, conventional [k?n'ven??nl]a. 普通的,常见的,习惯的,常规的6, monochrome ['m?n?ukr?um]n. 单色画,单色画法a. 单色的,黑白的[计算机] 单色的7, chrominance ['kr?umin?ns][物]色度( 任意一种颜色与亮度相同的一个指定的参考色之间的差异,如彩色电视采用白色为参考色)8, facsimile [f?k'simili]n. 复写,传真v. 传真9, binarya. 二进位的,二元的10, punctuation [p??ktj?'ei?(?)n]n. 标点11, appendv. 附加,添加,悬挂n.[计算机] DOS 命令建立数据文件的子目录搜索路经12, violate ['vai?leit]v. 违犯,亵渎,干扰13, designate ['dezigneit]v. 指定,标示14, terminology [,t?:mi'n?l?d?i] n. 用辞,术语15, asynchronous [ei'si?kr?n?s] adj. 异步的;不同时的;不同期的16, compression [k?m'pre?(?)n]n. 压缩,压榨,缩小17, reversible [ri'v?:s?bl]a. 可逆的(双向的,回行的,可反转的)18, data compaction【计算机】数据精简19, intrinsicallyad. 内在地(固有地,本征地,实质地)20, clinical ['klinik?l]a. 临床的21, degraded [di'greidid]a. 被贬低的,堕落的22, cater ['keit?]v. 备办食物,投合,迎合23, coefficient [k?ui'fi??nt]n. 系数24, uncorrelated [?n'k?rileitid] adj. 不相关的25, quantization [,kw?ntai'zei??n] 量子化,数字化26, discard [dis'kɑ:d]n. 丢弃,扔掉v. 丢弃,抛弃27, compact ['k?mp?kt]a. 紧凑的,紧密的,简洁的v. 使装满,使简洁28, hierarchical [,hai?'rɑ:kik?l] 分层的,体系的29, prior ['prai?]a. 更重要的,较早的,在先的ad. 居先30, segment ['segm?nt]n. 部分v. 部分,段,分割vt. 分割31, exceeding [ik'si:di?]a. 超越的,非常的32, reassemble ['ri:?'sembl] 重装,再汇编33, proceed [pr?'si:d]v. 著手进行,继续进行34, encryption [-'krip?Ln]加密术35, acronym ['?kr?nim]n. 缩写字,字头语36, paradigm ['p?r?daim, -dim] n. 例,模范,词形变化表37, subnet ['s?b,net]子[分支]网络38, reside [ri'zaid]v. 住,居留,属于39, pertinent ['p?:tin?nt]a. 相关的,中肯的,切题的40, recipient [ri'sipi?nt]n. 接受者41, gigantic [d?ai'g?ntik]a. 巨大的42, variation [,v??ri'ei??n] n. 变化,变动,变种,变异43, interval ['int?v?l]n. 间隔44, interleave [,int?(:)'li:v]交错,隔行扫描,插入纸45, substitute ['s?bstitju:t]n. 代替者,代用品v. 代替vi. (for)代替vt. 用...代替,代以46, ultimate ['?ltimit]n. 终极,根本a. 终极的,根本的,极限的,最后的47, transducer [tr?nz'dju:s?]转换器,传感器48, acoustic [?'ku:stik]a. 美声的,听觉的,音感教育,音响学的49, compactlyad. 细密地,简洁地50, restriction [ris'trik??n]n. 限制,约束vi. 结果,(作为结果)发生51, articulationn. 关节,接合,清晰发音52, derivative [di'riv?tiv]a. 引出的,系出的n. 引出之物,系出物,衍生字n. 导数, 微商53, encasedadj. 包装的;被包住的v. 盖住;包起;装箱(encase的过去分词)54, polyvinyl chloriden. 聚氯乙烯55, proximity [pr?k'simiti]n. 接近,亲近56, susceptible [s?'sept?bl] n. 易受影响者,易受感染者a. 易受影响的,易感动的,容许57, mitigate ['mitiɡeit]vt. 使缓和,使减轻vi. 减轻,缓和下来58, coaxial [k?u'?ks?l]同轴的59, dielectric [,daii'lektrik] n. 电介体a. 诱电性的,非传导性的60, immunity [i'mju:niti]n. 免疫,免疫性,免除61, feasible ['fi:z?bl]a. 可行的,可能的62, comfinessn. comfy的变形63, confine ['k?nfain]v. 限制,闭居[计算机] 限制64, subscriber [s?bs'kraib?] n. 签署者,捐献者,订户65, confined [k?n'faind]a. 被限制的,狭窄的,在分娩中的66, cladding ['kl?di?]n. 金属包层法(表面处理,路面) 电镀67, silica ['silik?]矽土68, theoretical [θi?'retik?l] a. 理论上的69, inherent [in'hi?r?nt] a. 内在的,固有的70, diameter [dai'?mit?]n. 直径71, ruggedness ['r?gidnis]n. 强度(坚固性,耐久性,崎岖) 72, exemplify [iɡ'zemplifai] vt. 例证;例示73, tensile ['tensail]a. 可拉长的,可伸长的74, fabricatedv. 制造(装配,伪造)a. 制造好的(装配式的)75, metallic [mi 't?lik]a. 金属的76, scarce [sk??s]a. 缺乏的,不足的,稀少的,罕见的77, fuelledv. 刺激(fuel的过去式);加燃料78, excite [ik'sait]v. 刺激,使...兴奋,激励79, afield [?'fi:ld]ad. 远离着,在远处80, diffract [di'fr?kt]v. 使分散,绕射,曲折而分散81, albeitconj. 虽然(即使)82, incurredv. 招致(遭受)83, converting [k?n'v?:ti?]n. 转换(吹炼)84, intermediate [,int?'mi:dj?t] a. 中级的,中间的n. 中间体,媒介物85, detector [di'tekt?]n. 探测器86, mobility [m?u'biliti]n. 可动性,变动性,情感不定87, terrestrial [ti'restri?l]n. 地球上的人a. 地球的,地上的88, encounter [in'kaunt?]n. 意外的相见,遭遇v. 遇到,偶然碰到,遭遇89, encounteredn. 遇到90, scattering ['sk?t?ri?]n. 散布vbl. 散布,散落,散播91, statistical [st?'tistik?l]a. 统计的,统计学的92, invaluable [in'v?lju?bl]a. 无价的93, coverage ['k?v?rid?]n. 涉及范围(作用距离,覆盖厚度,涂层)94, distinct [dis'ti?kt]a. (from)独特的,不同的,明显的,清楚的95, geosynchronous [,d?i:??'si?kr?n?s] 与地球的相对位置不变的96, orbit ['?:bit]n. 轨道,常轨,眼眶v. 绕轨道而行,进入轨道,盘旋97, considerable [k?n'sid?r?bl]a. 相当的,可观的,重要的98, platform ['pl?tf?:m]n. 平台,月台,讲台,坛99, uplink ['?p,li?k][电信]向上传输,上行线, 卫星上行链路100, amplifiedA V C 放大式自动音量控制101, transponder [tr?n'sp?nd?(r)] n. 异频雷达收发机102, circuitry ['s?:kitri]n. 电路,线路103, repeater [ri'pi:t?]中继器,转发器104, attributen. 属性,特征v. 归于,属于105, attenuation衰减106, insignificant [,insig'nifik?nt] a. 无关紧要的,可忽略的107, emission [i'mi??n]n. 发射,射出,发行108, galactic [g?'l?ktik]a. 星系的,银河系的109, summarize ['s?m?raiz]v. 概述,摘要而言110, properties道具111, diagram ['dai?gr?m]n. 图解,图表v. 用图解法表示112, recognizable ['rek?gnaiz?bl] a. 可认识的,可承认的,可辨别的113, modify ['m?difai]v. 修改,更正,修饰114, distortion [dis't?:??n]n. 扭曲,变形,曲解115, scheme [ski:m]n. 方案,计划,阴谋v. 计画,设计,体系,结构,图谋116, sinusoidal [,sain?'s?idLl]a. 正弦曲线,窦状隙,窦状小管117, instantaneous [,inst?n'teinj?s] a. 瞬间的,即刻的ad. 瞬间,霎时118, rectangular [rek't??gjul?]n. 矩形119, counterpart ['kaunt?pɑ:t]n. 相似之物(正副二份中之一,对应物)120, quantized量子化的121, roundoffvt. 舍入(修整)122, prescribeda. 规定的123, discernible [di's?:n?bl,-'z?:-] a. 可辨别的124, emerge [i'm?:d?]v. 浮现,(由某种状态)脱出,(事实)显现出来125, regeneration [ri,d?en?'rei??n] n. 再生,重建126, multiplexing ['m?ltipleksi?] n. 多路技术127, simultaneous [,sim?l'teinj?s] a. 同时发生的128, slot [sl?t]n. 水沟,细长的孔,狭缝,硬币投币口v. 留细长的孔129, distinctive [dis'ti?ktiv]a. 有特色的,出众的130, overlap ['?uv?'l?p]n. 重叠,重复v. 重叠,重复[计算机] 重叠131, hence [hens]ad. 今后,从此,因此,所以132, crosstalk ['kr?s,t?:k]n. 串话干扰(交扰,交调失真,顶嘴,相声) 133, interaction [,int?r'?k??n]n. 相互作用,相互影响134, timeshare ['taim?e?(r)]adj. 分时享用度假别墅的n. 分时享用度假别墅所有权135, concurrent [k?n'k?r?nt]a. 同时发生的136, portion ['p?:??n]n. 部分,份,命运v. 将...分配,分配137, reciprocal [ri'sipr?k?l]a. 相互的,互惠的[数]n. 倒数138, sink [si?k]n. 接收端,沟渠,污水槽v. 下沉,使...低落,陷于139, specifiedadj. 指定的140, guidlinen. 指南;指导原则141, estimate ['estimeit]n. 估计,估价v. 估计,估价,评价142, option ['?p??n]n. 选择143, block [bl?k]n. 街区,木块,石块v. 阻塞144, rational ['rnl]a. 合理的,理性的n. 有理数145, functional ['f??k??nl]a. 功能的146, redundant [ri'd?nd?nt]a. 多余的147, virtue ['v?:tju:]n. 美德,优点148, finite ['fainait]a. 有限的149, waveform ['weivf?:m]波形150, conceptual [k?n'sept?u?l, -tju?l] a. 概念上的151, robust [r?'b?st]a. 强壮的,强健的152, mechanical [mi'k?nikl]a. 机械的,力学的,呆板的153, vibration [vai'brei??n]n. 震动,颤动154, stringent ['strind??nt]a. 迫切的,严厉的155, linearity [,lini'?riti]线性,直线性156, superficial [sju:p?'fi??l]a. 表面的,肤浅的157, tailor ['teil?]n. 裁缝师,成衣匠v. 缝制,做裁缝,使...适应158, impairment [im'p??m?nt]n. 损伤(毁损)159, solely ['s?u(l)li]ad. 独自地,单独地160, ever-increasing demandever-increasing demand: 日益增长的需求ever-increasing: 不断增加的|不断增长的|不断增长的,持续增长的161, affinity [?'finiti]n. 密切关系,吸引力162, consume [k?n'sju:m]v. 消耗,消费,饮食;毁灭;消磨;枯萎163, dictate [dik'teit]v. 听写,口述,口授n. 命令,指挥,指令164, implementation [,implimen'tei??n]n. 安装启用,实行,履行工具165, configure [k?n'fig?]v. 配置166, objective [?b'd?ektiv]a. 客观的n. 目标,目的167, constraint [k?n'streint]约束, 强制,约束条件[计算机] 限制168, allowable [?'lau?b?l]容许的,承认的169, emphatic [im'f?tik]a. 语调强的,用力的,强调了的170, embodiedv. 呈现(embody的过去式及过去分词形式);具体表达171, celebrated ['selibreitid]a. 著名的172, theorem ['θi?r?m]n. 定理173, denote [di'n?ut]v. 指示,表示174, boundn. 跳跃,界限,范围a. 受约束的,装订的,有义务的v. 跳跃175, unity ['ju:niti]n. 个体,一致,结合176, equation [i'kwei??n]n. 方程(式),等式177, idealizedadj. 理想化的v. 把…理想化;以理想的形式表现事物(idealize 的过去分词)178, framework ['freimw?:k]n. 结构,构架,框架179, second-ordern. 二阶180, partial ['pɑ:??l] a. 部分的,偏袒的,偏爱的181, substituent [s?b'stitju?nt]n. 取代182, ensemble [?:n'sɑ:mbl]n. 全体,合唱曲,女人的全套服装183, Rayleigh ['reili]n. 雷利(极光和夜天光的发光强度的单位) 184, probabilistic [,pr?b?bi'listik] 概率性的185, appendix [?'pendiks]n. 附录186, component [k?m'p?un?nt]n. 元件,组件,成份a. 组成的,构成的187, vicinity [vi'siniti]n. 邻近,附近188, thermal ['θ?:m?l]a. 热的,热量的n. 上升的热气流189, graph [grɑ:f]n. 图表,曲线图v. 图表,用图表表示vt. 以(曲线)图表示190, suppress [s?'pres]v. 镇压,使...止住,禁止191, formally ['f?:m?li]ad. 正式地,拘泥礼仪,形式上192, initiatedv. 开始,发起;开创(initiate的过去式)193, exhibit [ig'zibit]n. 展览品,陈列品,展览v. 展现,陈列,展览[计算机] 呈现194, profound [pr?'faund]a. 极深的,深奥的,深厚的195, implication [,impli'kei??n]n. 暗示,含意196, depict [di'pikt]v. 描述197, spatial ['spei??l]a. 空间的198, joint [d??int]a. 联合的,共同的n. 关节,接合处v. 连接,贴合199, henceforth [hens'f?:θ]ad. 今后200, subclass ['s?bklɑ:s]子类,小分类,子集合201, fraction ['fr?k??n]n. 分数,小部分,破片202, distortionlessadj. 无畸变的;不失真的;无失真203, additive ['?ditiv]n. 附加物(加法)a. 加添的204, correlator [k?ri'leit?]相关因子,相关器205, thr eshold ['θre?h?uld]n. 极限,门槛,入口,开端206, invokingv. 援引;祈求;唤起(invoke的ing形式)207, invoke [in'v?uk][计算机] 调用208, inverse ['in'v?:s]a. 相反的,倒转的209, proportional [pr?'p?:??nl]a. 成比例的,相称的n. [数]比例项210, scaling ['skeili?]缩放比例211, justification [d??stifi'kei?(?)n] n. 辩护,证明正当,释罪[计算机] 对齐212, variable ['v??ri?bl]a. 可变的,易变的n. 变量213, conserve [k?n's?:v]n. 蜜饯,果酱v. 保存,保全214, incomplete [,ink?m'pli:t]a. 不完全的,不完整的。
知识图谱表示模型方法选择评估
知识图谱表示模型方法选择评估在知识图谱中,表示模型方法是为了有效处理和表示知识图谱中的实体和关系而提出的。
这些方法帮助我们更好地理解和利用知识图谱中的信息,对于许多应用领域都具有重要的意义。
然而,随着不断涌现的表示模型方法,选择合适的方法成为了一个关键的挑战。
本文将介绍一些常见的知识图谱表示模型方法,并探讨如何评估这些方法的性能。
一、常见的知识图谱表示模型方法1. TransE模型TransE是最早提出的知识图谱表示模型之一。
该模型基于平移操作来表示实体和关系之间的语义关联,通过最小化三元组中头实体、关系和尾实体之间的距离来优化模型参数。
TransE方法简单有效,常被用于处理知识图谱中的实体关系预测任务。
2. TransH模型TransH在TransE的基础上进行了改进,通过引入关系特定的映射矩阵来解决实体和关系之间的语义关联问题。
该模型克服了TransE中的一些限制,并在一些特定的知识图谱任务中表现出更好的性能。
3. TransR模型TransR模型进一步发展了TransE和TransH模型的思想,通过引入矩阵关系来表示实体和关系之间的语义关联。
该模型能够更准确地捕捉实体和关系之间的语义关系,并在一些复杂的知识图谱任务中表现出很好的性能。
二、评估知识图谱表示模型方法的选择在选择合适的知识图谱表示模型方法时,我们需要考虑几个重要的评估指标:1. 准确率准确率是评估模型预测结果的重要指标之一。
我们可以通过比较模型在测试数据集上的预测结果和实际标签之间的一致性来评估模型的准确率。
2. 召回率召回率是评估模型对于真实标签的覆盖能力。
在知识图谱中,我们关心的是模型能够捕捉到尽可能多的实体和关系之间的语义关联。
因此,召回率也是一个重要的评估指标。
3. F1值F1值综合了准确率和召回率,可以更全面地评估模型的性能。
F1值的计算公式是F1 = 2 * (precision * recall) / (precision + recall),其中precision表示准确率,recall表示召回率。
基于本体的图像语义标注与检索模型
“ n t i h t g a h tke y Ke , h tl I h sp o o r p a n b n t eIa y f o b l p a e b roBa go i ik ng o t al ly rRo e t g i sk c i h al n sa i m. t e b l i t d u ”
( g ce c n e h oo yL b r tr, c o l f u o t n Not wetr o ye h ia U i es y Xi n7 0 7 ) I ma eS in e dT c n lg a o aoy S h o A t mai , r a o o h s nP ltc nc l n v ri , ’ 1 0 2 e t a
1 概述
如何合理有效地描述 图像是 图像处理技术的难点之一 。 基于关键 词的描述方法在处理大规模 图像时 ,容易出现含义 模糊或矛盾的描述结果。基于图像处理知识、应用图像 低层
2 图像语义描述
2 语义 描述简介 . 1 图像的特点是能表达 复杂的内容含义 ,要合理、完整地
描 述一幅图像 所表达 不同层 次上的复杂语义 ,必须提 出一个 适当的图像语义 标注模 型。 自然语言对一 幅图像进行描述 , 用
I g Q so ate dl el efr di g n ae n ditlg n t ea. ma e wt t e lp r me i ema g me tn el e te iv 1 h h mo w o h n ma a n i rr
[ ywod l d mano tlg ;ma ean tt n i g n gmetitlgn t e a Ke r s o i noo y i g oao ;ma emaa e n;neieter v n i l r i l
基于自然语言处理的知识图谱构建研究
基于自然语言处理的知识图谱构建研究自然语言处理(Natural Language Processing,简称NLP)是人工智能领域中的一个重要分支,它致力于让计算机能够理解、分析和生成人类的自然语言,实现人机之间的交互。
知识图谱(Knowledge Graph)是一种结构化的知识表示模型,它以图的形式存储和展示知识之间的关系,可以帮助计算机理解和推理出复杂的知识关联。
本文将探讨基于自然语言处理的知识图谱构建研究,旨在介绍知识图谱的基础概念和构建方法,并对自然语言处理在知识图谱构建中的应用进行详细讨论。
首先,了解知识图谱的基础概念是构建的关键。
知识图谱是一种结构化、可组织、可扩展的知识表示模型,它由一系列实体(Entity)和它们之间的关系(Relation)构成。
实体可以是人、物、地点等事物的抽象,关系则描述了实体之间的联系和特征。
通过知识图谱,我们可以将分散的知识元素组织起来,形成丰富的语义网络。
那么,如何构建知识图谱呢?一般来说,知识图谱的构建包含三个主要步骤:知识抽取、知识表示和知识链接。
知识抽取是指从文本等非结构化数据中提取实体和关系的过程。
在自然语言处理领域,我们可以利用文本挖掘、实体识别和关系抽取技术,从大规模的文本数据中自动抽取出实体和关系的信息。
例如,从新闻文章中抽取出人物、地点和事件,并建立它们之间的联系。
知识表示是将抽取得到的知识转化为机器可理解的形式。
通常使用语义表示模型,如词嵌入(Word Embedding)和图神经网络(Graph Neural Network),将实体和关系表示为高维向量。
这样可以方便计算机进行语义相似度计算和推理推断。
此外,还可以利用领域本体(Ontology)和语义关系定义来进一步增强知识的语义表达能力。
知识链接是将已有的知识与构建的知识图谱相链接。
这里的已有知识可以是结构化的数据,如数据库和知识库,也可以是非结构化的数据,如互联网上的文本和多媒体内容。
人工智能中知识图谱的使用方法
人工智能中知识图谱的使用方法人工智能(Artificial Intelligence,AI)是指通过模拟、延伸人的智能实现与人类一样的智能行为的科学与技术。
在人工智能的发展过程中,知识图谱作为一种表示和存储知识的方式,发挥着重要的作用。
知识图谱是一种将知识组织在一起的结构化数据,可以帮助机器理解人类语言,支持机器学习和推理,从而实现更高级的智能任务。
本文将介绍人工智能中知识图谱的使用方法。
首先,知识图谱的构建是人工智能应用的关键环节。
构建知识图谱需要通过对大量的文本、数据进行语义解析和知识提取,将提取的知识组织成结构化的图谱形式。
常见的知识图谱构建方法包括规则匹配、模式识别和机器学习等。
规则匹配是通过定义一系列规则,根据文本或数据中的特征进行匹配提取知识。
模式识别是通过训练模型,自动识别文本或数据中的特征,并将其转化为知识图谱。
机器学习是一种通过训练数据,使机器能够从中学习并得出知识的方法。
构建知识图谱时,可以根据具体任务的需求选择合适的方法。
其次,知识图谱在人工智能中的应用非常广泛。
知识图谱可以用于自然语言理解、问答系统、智能推荐等多个领域。
在自然语言理解中,知识图谱可以帮助机器理解语义,从而更准确地理解和生成自然语言。
例如,在对话系统中,机器可以通过与知识图谱进行交互,获取实体、属性和关系的信息,从而更好地响应用户的问题。
在问答系统中,知识图谱可以充当知识库,存储和管理各种领域的知识,帮助机器回答用户的问题。
在智能推荐中,知识图谱可以通过对用户行为和偏好的分析,将用户与相关的实体和属性进行链接,从而提供个性化的推荐服务。
总之,知识图谱为人工智能的应用提供了丰富的知识库和语义理解能力,从而提高了应用的准确性和智能化程度。
第三,知识图谱的更新和维护是保证应用效果的重要手段。
知识图谱中的知识是不断变化的,需要及时更新和维护。
更新和维护知识图谱可以通过监控和分析外部数据源,根据新的数据进行知识更新。
多模态知识图谱表示学习综述
多模态知识图谱表示学习综述多模态知识图谱表示学习综述摘要:随着大数据时代的到来,知识图谱成为了对现实世界进行建模和分析的重要工具。
然而,传统的知识图谱主要基于文本信息进行构建和表示,忽略了其他多模态数据的丰富信息。
针对这个问题,多模态知识图谱表示学习应运而生。
本文将对多模态知识图谱表示学习的研究现状、方法和应用进行综述,以期为相关领域的研究者提供参考和启发。
一、引言知识图谱是一种以图的形式表达的知识库,其中知识以实体、关系和属性的形式存储。
传统的知识图谱以基于文本的方式进行构建和表示,通过对文本进行实体抽取、关系抽取等技术来获得知识。
然而,文本信息属于单模态数据,仅能够提供有限的知识表达能力。
随着多模态数据的快速增长,如图像、音频和视频等,如何将多模态数据融入知识图谱表示学习成为当前研究的热点和挑战。
二、多模态知识图谱表示学习的研究现状多模态知识图谱表示学习旨在利用多模态数据增强知识图谱的表达能力。
已有的研究主要可以分为两类:基于图的方法和基于张量的方法。
基于图的方法使用图神经网络(GNN)来建模并融合多模态数据,利用节点和边的信息进行知识表示学习。
基于张量的方法则将多模态数据表示为高阶张量,通过张量分解等技术进行知识表示学习。
三、多模态知识图谱表示学习的方法多模态知识图谱表示学习的方法多种多样,以下是其中几种常见的方法:1. 卷积神经网络(CNN)和循环神经网络(RNN):这两种方法广泛用于图像和文本数据的表示学习,可以将其应用于多模态知识图谱表示学习中,从而提高知识图谱的表达能力。
2. 图卷积神经网络(GCN):GCN是一种特殊的卷积神经网络,它通过聚合周围节点的信息来更新当前节点的表示,已被广泛应用于多模态知识表示学习中。
3. 张量分解:张量分解可以将多维张量分解为若干低维张量,从而实现对多模态数据的表示学习。
常用的张量分解方法包括SVD、CP分解等。
四、多模态知识图谱表示学习的应用多模态知识图谱表示学习在许多领域中具有广泛的应用前景,以下是其中几个常见的应用:1. 音乐推荐:通过将音乐数据和用户数据融入知识图谱表示学习,可以提高音乐推荐系统的精确度和个性化程度。
Semi-supervised Graph-based Hyperspectral Image Classification
Semi-supervised Graph-basedHyperspectral Image Classification Gustavo Camps-Valls,Senior Member,IEEE,Tatyana V.Bandos,and Dengyong ZhouAbstractThis paper presents a semi-supervised graph-based method for the classification of hyperspectral images.The method is designed to handle the special characteristics of hyperspectral images,namelyhigh input dimension of pixels,low number of labeled samples,and spatial variability of the spectralsignature.To alleviate these problems,the method incorporates three ingredients,respectively.First,being a kernel-based method,it combats the curse of dimensionality efficiently.Second,following asemi-supervised approach,it exploits the wealth of unlabeled samples in the image,and naturally givesrelative importance to the labeled ones through a graph-based methodology.Finally,it incorporatescontextual information through a full family of composite kernels.Noting that the graph method relieson inverting a huge kernel matrix formed by both labeled and unlabeled samples,we originally introducethe Nystr¨o m method in the formulation to speed up the classification process.The presented semi-supervised graph-based method is compared to state-of-the-art support vector machines(SVMs)in the classification of hyperspectral data.The proposed method produces betterclassification maps which capture the intrinsic structure collectively revealed by labeled and unlabeledpoints.Good and stable accuracy is produced in ill-posed classification problems(high dimensionalspaces and low number of labeled samples).Also,the introduction of the composite kernels frameworkdrastically improves results,and the new fast formulation ranks almost linearly in the computationalManuscript received September2006;revised January2007;G.Camps-Valls and T.V.Bandos are with Grup de Processament Digital de Senyals,GPDS.Dept.Enginyeria Electr`o nica. Escola T`e cnica Superior d’Enginyeria.Universitat de Val`e ncia.C/Dr.Moliner,50.46100Burjassot(Val`e ncia)Spain.E-mail: gustavo.camps@uv.es.D.Zhou is with Microsoft Research,One Microsoft Way Redmond,W A.E-mail:Dengyong.Zhou@cost,rather than cubic as in the original method,thus allowing the use of this method in remote sensingapplications.Index TermsHyperspectral image classification,semi-supervised learning,ill-posed problem,composite kernel, graph Laplacian,undirected graph,Nystr¨o m method.I.I NTRODUCTIONThe information contained in hyperspectral images allows the characterization,identification, and classification of the land-covers with improved accuracy and robustness.However,several critical problems should be considered in the classification of hyperspectral data,among which: (i)the high number of spectral channels,(ii)the spatial variability of the spectral signature, (iii)the high cost of true sample labeling,and(iv)the quality of data.In particular,the high number of spectral channels and low number of labeled training samples pose the problem of the curse of dimensionality(i.e.the Hughes phenomenon[1])and,as a consequence,result in the risk of overfitting the training data.For these reasons,desirable properties of hyperspectral image classifiers should be the ability to produce accurate land cover maps when working with high number of features,low-sized training datasets and high levels of spatial variability of the spectral signature[2].In the remote sensing literature,many supervised and unsupervised classifiers have been developed to tackle the multi-and hyperspectral data classification problem[3].Supervised methods,such as artificial neural networks[4]–[6]readily revealed inefficient when dealing with a high number of spectral bands,and thus in the recent years,kernel-based methods in general and support vector machines(SVMs)[7],[8]in particular have been successfully used for hyperspectral image classification[9]–[12].Certainly,kernel-based classifiers are able to handle large input spaces efficiently,and deal with noisy samples in a robust way[13].However,the main difficulty with all supervised methods is that the learning process heavily depends on the quality of the training dataset,which is only useful for simultaneous images,or for images with the same classes taken under the same conditions.Even worse,the training set is frequently not available, or in a very reduced number,given the very high cost of true sample labeling.On the other hand,unsupervised methods have demonstrated good results[14]–[19]in multi and hyperspectralimage classification.Unsupervised methods are not sensitive to the number of labeled samples since they work on the whole image,but the relationship between clusters and classes is not ensured.Moreover,a preliminary feature selection/extraction step is usually undertaken to reduce the high input space dimension,which is time-consuming,scenario-dependent,and needs prior knowledge.In this context,it becomes natural that using semi-supervised classifiers can yield improved performance.In semi-supervised learning(SSL),the algorithm is provided with some available supervised information in addition to the wealth of unlabeled data.The framework of semi-supervised learning is very active and has recently attracted a considerable amount of research [20]–[22].Essentially,three different classes of SSL algorithms are encountered in the literature: 1)Generative models which involve estimating the conditional density p(x|y),such as ex-pectation-maximization(EM)algorithms withfinite mixture models[23],which have been extensively applied in the context of remotely sensed image classification[24].2)Low density separation algorithms,which maximize the margin for labeled and unlabeledsamples simultaneously,such as Transductive SVM[25],which have been recently applied to hyperspectral image classification[26].3)Graph-based methods[27],[28],in which each sample spreads its label information to itsneighbors until a global stable state is achieved on the whole dataset.This paper concentrates on graph-based methods,which have been lately paid attention because of their solid mathematical background,their relationship with kernel methods,sparseness prop-erties,model visualization,and good results in many areas,such as computational biology[29], web mining[30],or text categorization[22].In this paper,we introduce a semi-supervised graph-based method,previously presented in [31],in the context of hyperspectral image classification.The method is then further improved to tackle the problems imposed by the special characteristics of hyperspectral images,namely high input dimension of pixels,low number of labeled samples,and spatial variability of the spectral signature.To this end,the method has the following characteristics:1)Kernel method.Since the proposed method is kernel-based,the high dimensionality ofsamples is treated efficiently[12].2)Semi-supervised method.Being a semi-supervised method,the huge number of unlabeledsamples in the image is exploited to improve performance[22].3)Graph-based method.The method follows a graph-based methodology,and thus relativeimportance to the labeled samples is given in a natural way[31].4)Context-based method.We incorporate contextual information in the classifier through theintroduction of a family of composite kernels,extending the works[32],[33].5)Fast method.Finally,noting that the method relies on inverting large kernel matrices(built with labeled and unlabeled pixels together),we reformulate the algorithm using the Nystr¨o m method to enable a dramatic speed-up of the classification process[34].The method is evaluated in ill-posed classification problems,that is low number of high dimen-sional labeled samples.Evaluation is carried out in terms of accuracy and robustness when low number of labeled samples is available,and by visual inspection of the provided classification maps.Also,special attention is given to the issues of computational cost,and free parameters tuning.The rest of the paper is organized as follows.Section II reviews the main ideas underlying graph methods and the consistency assumption in semi-supervised learning.The latter motivates Section III,in which we present the proposed semi-supervised graph-based composite kernel classification method.Section IV discusses the classification results of this approach compared to standard SVMs in ill-posed problems.Also,in this section we address the problem of free parameters tuning and computational cost.Finally,section V includes some concluding remarks and indications on further work.II.S EMI-SUPERVISED L EARNING WITH G RAPHSThe key issue in semi-supervised learning is the assumption of consistency,which means that:(1)nearby points are likely to have the same label;and(2)points on the same structure (typically referred to as a cluster or a manifold)are likely to have the same label.In our case, nearby points are those pixels spectrally similar and thus the assumption is applied to the high dimensional space of hyperspectral image pixels.This argument is akin to that in[20],[35]–[38]and is often called the cluster assumption[20],[37].Note that thefirst assumption is local, whereas the second one is global.Traditional supervised learning algorithms,such as k-NN,in general depend only on thefirst assumption of local consistency.Fig.1.Classification on the‘two moons’dataset.(a)Toy data set with only two labeled points and many unlabeled samples conforming a structured domain with intuitively discernable clusters(manifolds);(b)classification result given by the SVM with an RBF kernel;(c)k-NN with k=1;and(d)ideal classification(and the one provided by our method).To illustrate the prior assumption of consistency underlying semi-supervised learning,let us consider a toy dataset generated according to a pattern of two intertwining moons(see Fig.1[a]). Every point should be similar to points in its local neighborhood,and furthermore,points in one moon should be more similar to each other than to points in the other moon.The classification results given by the Support Vector Machine(SVM)with an RBF kernel and k-NN are shown in Fig.1[b]and1[c],respectively.According to the assumption of consistency,however,the two moons should be classified as shown in Fig.1[d].The main differences between the various semi-supervised learning algorithms,such as spectral methods[35],[37],[39],random walks[38],[40],graph mincuts[36]and transductive SVM [25],lie in their way of realizing the assumption of consistency.A principled approach to formalize the assumption is to design a classification function which is sufficiently smooth on the intrinsic structure revealed by known labeled and unlabeled points.In order to construct such a smooth function,we propose here a simple iteration algorithm inspired by the work on spreading activation networks[41],[42]and diffusion kernels[43]–[45],recent work on semi-supervised learning and clustering[35],[37],[46],and more specifically by the work of Zhu et al.[38].The keynote of our method is to let every point iteratively spread its label information to its neighbors until a global stable state is achieved.Graph-based methods rely upon the construction of a graph representation,where the vertices are the(labeled and unlabeled)samples,and edges represent the similarity among samples in the dataset(see Fig.2).Typically,graph methods utilize the graph Laplacian,which is defined as follows.Let G=(V,E)be a graph with a set of vertices,V,connected by a set of edges, E.The edge connecting nodes(or samples)i and j has an associated weight,{W ij}.Then, the weight(or affinity)matrix W is constructed among all labeled and unlabeled samples.The (normalized)graph Laplacian is defined asL=I−D−1/2W D−1/2,(1)where D is a diagonal matrix defined by D ii=jW ij.See[22](Ch.11)for more details ondifferent families of graph-based methods.At this point,it is worth noting that prediction consists in labeling the unlabeled nodes,and thus,these are intrinsically transductive classifiers,i.e.the graph only returns the predicted class label for the unlabeled samples,not a decision function defined on the whole domain.This graph-based classification can be viewed as estimating a function F over the graph,which should be in accordance with the smoothness assumption,that is,a good classification function should not change too much between similar points.This smoothness assumption can be reinforced in the problem of hyperspectral image classification through the integration of spatial and contextual information,as will be described in the next section.III.G RAPH-BASED COMPOSITE KERNEL CLASSIFICATIONIn this section,we present the whole formulation of the graph-based method proposed in this paper.We start by presenting the general graph approach,and introduce a full family of composite kernels to integrate the spatial(contextual)and spectral formulation in the method. Finally,noting the high computational cost of the method,we propose to use the Nystr¨o m method (in combination with Woodbury’s formula)which allow us to speed up the solution.(a)(b)-1-1-1-1-1+1+1+1+1+1+1Fig.2.Graph classification on a toy graph.(a)The two shaded circles are the initially labeled vertices (±1),while the white nodes represent unlabeled samples.The thickness of the edges represent the similarity among samples.(b)Graph methods classify the unlabeled samples according to the weighted distance,not just to the shortest path lengths,the latter leading to incorrectly classified samples.The two clusters (shaded)are intuitively correct,even being connected by (thin weak)edges.A.Semi-supervised graph-based method1)Formulation:Given a dataset of pixels in an N -dimensional input space (being N the number of bands or spectral channels),X ={x 1,...,x l ,x l +1,...,x n }⊂R N and a label set L ={1,...,c },the first l points x i (i ≤l )are labeled as y i ∈L and the remaining points x u (l +1≤u ≤n )are unlabeled.The goal in semi-supervised learning is to predict the labels of the unlabeled points.Let F denote the set of n ×c matrices with non-negative entries.A matrix F =[F 1,...,F n ] ∈F corresponds to a classification on the dataset X by labeling each point x i with a label y i =arg max j ≤c F ij .We can understand F as a vectorial function F :X →R c which assigns a vector F i to each point x i .Define an n ×c matrix Y ∈F with Y ij =1if x i is labeled as y i =j and Y ij =0otherwise.Note that Y is consistent with the initial labels assigned according to the decision rule.At each iteration t ,the algorithm can be summarized as follows:1.-Calculate the affinity matrix W ,for instance using the RBF kernel 1:W ij ≡W (x i ,x j )=exp(− x i −x j 2/2σ2),∀i =j (2)1In the kernel and graph-based frameworks,the use of RBF kernels is a common choice because it has less numerical difficulties,and only the Gaussian width (σ)has to be tuned,which is an easy way to control the smoothness of the mapping function and relates the closeness of samples (spectra)in the feature space.In addition,the RBF kernel is a universal kernel and includes other valid kernels as particular cases [13].and make W ii =0to avoid self-similarity.2.-Construct the matrixS =D −1/2W D −1/2(3)in which D is a diagonal matrix with its (i,i )-element equal to the sum of the i -th row of W.Note that this step corresponds to the normalization in feature spaces.Certainly,if we consider a semi-definite kernel matrix formed by the dot products of mapped samples,W = φ(x i ),φ(x j ) =φ(x i ) φ(x j ),the normalized version is given by:ˆW (x i ,x j )= φ(x i ) φ(x i ) ,φ(x j ) φ(x j ) =W (x i ,x j ) W (x i ,x i )W (x j ,x j ).(4)3.-Iterate the following spreading function until convergence:F (t +1)=αSF (t )+(1−α)Y,(5)where αis a parameter in (0,1).These three steps should be iteratively repited until convergence.Now,if F ∗denotes the limit of the sequence {F (t )},the predicted labels for each point x i is done using:y i =arg max j ≤c F ∗ij .(6)However,it is worth noting here that one can demonstrate [31]that in the limit:F ∗=lim t →∞F (t )=(1−α)(I −αS )−1Y,(7)and thus the final estimating function F ∗can be computed directly without iterations.2)Graph interpretation:This algorithm can be understood intuitively in terms of spreading activation networks from experimental psychology [41],[42],and explained as random walks on graphs [47].Basically,the proposed method can be interpreted as a graph G =(V,E )defined on X ,where the vertex set V is just X and the edges E are weighted by W .In the second step,the weight matrix W of G is normalized symmetrically,which is necessary for the convergence of the following iteration.The first two steps are exactly the same as in spectral clustering [46].During the third step,each sample receives the information from its neighbors (first term),and also retains its initial information (second term).With regard to the free parameter α,one can see that it specifies the relative amount of the information from its neighbors and its initial label information.It is worth noting that self-reinforcement is avoided since the diagonal elements of the affinity matrix are set to zero inthefirst step.Moreover,the information is spread symmetrically since S is a symmetric matrix. Finally,the label of each unlabeled point is set to be the class of which it has received most information during the iterative process.B.Spatio-Spectral composite kernelsNote that,in its standard use,the graph-based method proposed before only would take advantage of the spectral information.Here we propose a toolbox of composite kernels ac-counting for the spatial,spectral,and cross-information between spatial and spectral parts.For this purpose,a pixel entity x i∈R N(recall that N represents the number of spectral bands)is redefined simultaneously both in the spectral domain using its spectral content,xωi∈R Nω,and in the spatial domain by applying some feature extraction to its surrounding area,x s i∈R N s, which yields N s spatial(contextual)features.These separated entities lead to two different similarity matrices,which can be easily computed and combined.At this point,one can sum spectral and spatial dedicated affinity matrices(Wωand W s,respectively),and introduce the cross-information between contextual and spectral features(Wωs and W sω)in the formulation. This simple methodology yields a full family of composite methods for hyperspectral image classification,which was originally presented in[32]for supervised SVM-based classification and we now extend and apply it to semi-supervised classification with graphs.1)The stacked features approach:The common approach to introduce spatial or contextual information in a(hyperspectral image)classifier consists of stacking the spectral and spatial features of a given pixel and then feeding the classifiers with them.This simple method has provided good results with neural networks and SVMs,but a main problem is readily identified; as the number of features increases,the curse of dimensionality is more likely to happen.In the context of kernel methods,the stacked approach can be formalized as follows.Let us define the mappingΦas a transformation of the concatenation x∗i≡{x s i,xωi},then the corresponding‘stacked’affinity matrix is:W{s,ω}≡W(x∗i,x∗j)= Φ(x∗i),Φ(x∗j) ,(8) which does not include explicit cross-relations between x s i and xωj.Here the angle brackets indicate inner product in the feature space.2)The direct summation kernel:Another possibility to avoid building very high dimensional samples to be classified is to treat spectral and spatial features separately.Let us assume two nonlinear (vectorial)transformations ϕ1(·)and ϕ2(·)into Hilbert spaces H 1and H 2,respectively.Then,the following transformation can be constructed:Φ(x i )={ϕ1(x s i ),ϕ2(x ωi )}(9)and the corresponding kernel matrix can be obtained by computing the dot product implicitely in the direct summation space H =H 1 H 2(i.e.there is no need to know the expression of the mappings,only their dot products)as follows:W (x i ,x j )= Φ(x i ),Φ(x j )(10)= {ϕ1(x s i ),ϕ2(x ωi )},{ϕ1(x s j ),ϕ2(x ωj )}=W s (x s i ,x s j )+W ω(x ωi ,x ωj )where W s = ϕ1(x s i ),ϕ1(x s j ) and W ω= ϕ2(x ωi ),ϕ2(x ωj ) are the kernel matrices computed using a valid kernel function such as the RBF kernel function in (2)over spatial or spectralfeatures,respectively.Note that dim (x ωi )=N ω,dim (x s i )=N s ,and dim (W )=dim (W s )=dim (W ω)=n ×n .Therefore,the solution is expressed as the sum of positive definite matrices accounting for the spatial and spectral counterparts independently,and thus the number of features is not duplicated by stacking them and to feed one classifier.This has the noticeable advantage of alleviating the curse of dimensionality in the scenario when a low number of labeled samples is available.3)The weighted summation kernel:By exploiting properties of Mercer’s kernels (see Ap-pendix I),a composite kernel that balances the spatial and spectral content can also be created,as follows:W (x i ,x j )=μW s (x s i ,x s j )+(1−μ)W ω(x ωi ,x ωj )(11)where μis a positive real-valued free parameter (0<μ<1),which is tuned in the training process,and constitutes a trade-off between the spatial and spectral information to classify a given pixel.4)The cross-information kernel:The preceding kernel-based classifiers can be conveniently modified to account for the cross relationship between the spatial and spectral information.As-sume a nonlinear(vectorial)mappingϕ(·)to a Hilbert space H and three linear transformations A k from H to H k,for k=1,2,3.Let us construct the following composite vector:Φ(x i)={A1ϕ(x s i),A2ϕ(xωi),A3(ϕ(x s i)+ϕ(xωi))}(12) and compute the dot productW(x i,x j)= Φ(x i),Φ(x j) (13)=Φ(x s i) R1Φ(x s j)+Φ(xωi) R2Φ(xωj)+Φ(x s i) R3Φ(xωj)+Φ(xωi) R3Φ(x s j)where R1=A 1A1+A 3A3,R2=A 2A2+A 3A3,and R3=A 3A3are three independent positive definite matrices.The important trick here is that including linear transformations A i in the definition of the mapping yields as the main subproduct that the induced kernel matrix in(13)takes into account the similarity between contextual and spectral information among samples,while keeping the size of the input space the same size as in the approaches before. Similarly to the direct summation kernel,it can be demonstrated that(13)can be expressed as the sum of positive definite matrices,accounting for the spatial,spectral,and cross-terms between spatial and spectral counterparts:W(x i,x j)=W s(x s i,x s j)+Wω(xωi,xωj)+W sω(x s i,xωj)+Wωs(xωi,x s j)(14)The only restriction for this formulation to be valid is that x s i and xωj need to have the same dimension(Nω=N s).Otherwise,cross kernels(Wωs and W sω)can not be computed.This condition can be easily ensured by extracting one spatial feature per spectral band.5)Kernels for improved versatility:Also note that one can build up a full family of kernel composition to account for cross-information between spatial and spectral features.For instance, one could think of the following combination of kernels for improved versatility:W(x i,x j)=W s(x s i,x s j)+Wω(xωi,xωj)+W{s,ω}(x∗i,x∗j),(15)which combines the summation kernel and the stacked approach.Similarly,another possibility is to construct the kernel:W(x i,x j)=W s(x s i,x s j)+Wω(xωi,xωj)+W sω(x s i,xωj)+Wωs(xωi,x s j)(16)+W{s,ω}(x∗i,x∗j)which combines the cross-information and the stacked vector approach in one similarity matrix.C.Nystr¨o m method formulationThe formulation of the method proposed so far involves three basic steps:firstly building the W matrix according to a composite specification;secondly,normalizing W to obtain S;and finally,solving the inversion problem given by(7).The algorithm in MATLAB code is given below:%Encode outputs in Y(e.g.,Class+1:[01],Class-1:[10],Unlabeled:[00]) %Precompute a W kernel(similarity)matrixW=W-eye(n);%Avoid self-similarityD=diag(1./sqrt(sum(W)));%Diagonal factorS=D*W*D;%Normalize the affinity matrixF=(1-alpha)*inv(eye(N)-alpha*S)*Y;%SolutionNote that direct inversion of(I−αS)induces a high computational cost of O(n3),since matrix size is size n×n,where n is the number of labeled and unlabeled samples.One method to reduce the computational complexity is to retain only thefirst largest p eigenvalues of the eigen-decomposition of the normalized matrix S:S=VΛV (17) where V represents the unitary matrix of eigenvectors andΛis a diagonal matrix containing their associated eigenvalues.There are methods tofind thefirst eigenvalues without explicitly solving the whole eigenproblem[48].However,computational time is drastically reduced only when p n.In order to reduce the computational cost involved,we introduce here the Nystr¨o m method. The Nystr¨o m method is commonly used to produce an approximate matrix˜S by randomlychoosing m rows/columns of the original matrix S and then making˜S n,n=S n,m S−1m,m S m,n, m≤n,where S n,m represents the n×m block of S.As a result,the method simplifies the solution of the problem to computing an approximated eigen-decomposition of the low-rank kernel matrix˜S=˜V˜Λ˜V ,involving O(mn2)computational cost[34].See Appendix II for the full formulation of the Nystr¨o m method.Therefore,if we approximate the normalized matrix S by expanding a small p×p matrix,˜S=˜V˜Λ˜V ,and substitute it into(7),we obtain:F∗=(1−α)(I−α˜V˜Λ˜V )−1Y.(18) Let us recall now the Woodbury formula from linear algebra,which states the identity:(C+AB)−1=C−1−C−1A(I+BC−1A)−1BC−1(19) where C is an invertible n×n matrix,A∈R n×m and B∈R m×n.By using this formula in our problem statement(18),it is straightforward to demonstrate that:F∗=(1−α)(Y−˜V(˜Λ˜V ˜V−α−1I)−1˜Λ˜V Y),(20) which involves inverting a matrix of size p×p(with p≤m≤n)and thus the computational cost is O(p2n),i.e.linear with the number of samples.This method wasfirst applied in the context of Gaussian Processes[34]but readily extended to spectral clustering and normalized-cut method [49].In this paper,we have formally presented the method in the context of semi-supervised graph-based classification.IV.E XPERIMENTAL R ESULTSIn this section,we show the performance of the proposed family of semi-supervised contextual graph-based classifiers for hyperspectral image classification.We also pay attention to the free parameters tuning,and propose a non-exhaustive procedure for this purpose.A.The AVIRIS Indian Pines datasetIn our experiments,we used the familiar A VIRIS image taken over NW Indiana’s Indian Pine test site in June1992.The data set represents a very challenging land-cover classification scenario,in which the primary crops of the area(mainly corn and soybeans)were very early in their growth cycle,with only about5%canopy cover.Discriminating among the majorcrops under these circumstances can be very difficult(in particular,given the moderate spatial resolution of20meters).This fact has made the scene a challenging benchmark to validate classification accuracy of hyperspectral imaging algorithms.The calibrated data is available online(along with detailed ground-truth information)from http://dynamo.ecn.purdue. edu/∼biehl/MultiSpec.Two different data sets were considered in the experiments.Following[9],wefirst used a part of the scene,called the subset scene,consisting of pixels[27−94]×[31−116]for a size of68×86,which contains four labeled classes(the background pixels were not considered for classification purposes).In this subimage,there are four classes with uneven number of labeled samples:‘Corn-notill’(1008),‘Grass/Trees’(732),‘Soybeans-notill’(727),and‘Soybeans-min’(1926).The latter two classes have very similar spectral signatures as they belong to the same super-class‘Soybeans’.Second,we used the whole scene,consisting of the full145×145pixels, which contains16classes,ranging in size from20−2468pixels,and thus constituting a very difficult situation.From the16different land-cover classes available in the original ground-truth, 7were discarded since an insufficient number of training samples were available and thus,this fact would dismiss the planned experimental analysis.Thefinally selected classes were:‘Corn-no till’(1434),‘Corn-min till’(834),‘Grass/Pasture’(497),‘Grass/Trees’(747),‘Hay-windrowed’(489),‘Soybean-no till’(968),‘Soybean-min till’(2468),‘Soybean-clean till’(614),and‘Woods’(1294).In both images,we removed20noisy bands covering the region of water absorption, andfinally worked with200spectral bands.Before training,data was normalized to give zero mean and unit variance.B.Model DevelopmentThe spectral samples xωi are,by definition,the spectral signature of pixels x i.The contextual samples,x s i,were computed as the mean of a3×3window surrounding x i for each band.This simple method is motivated by the local assumption in the spatial domain,which has previously produced good results in the context of SVMs[32].In all cases,we used the RBF kernel to construct the similarity matrices,W(x i,x j)=exp(− x i−x j 2/(2σ2)),and depending on the composite kernel used,a differentσparameter was to be tuned for each counterpart.All RBF kernel widths were tuned in the rangeσ={10−3,...,103},the regularization parameter for SVM was varied in C={100,...,103},and theαparameter for the graph-based method was。
初二下册英语单词精解系列[十二]
初二下册英语单词精解系列[十二]overcome音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[əʊvə’kʌm] 美[,ovɚ’kʌm]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 过去式overcame 过去分词overcome 现在分词overcoming ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ vt. 克服;胜过vi. 克服;得胜短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ overcome pronunciation:克服发音问题Successfully overcome:成功攻克overcome setbacks:战胜挫折overcome bureaucracy:克服官僚主义Is overcome:被战胜overcome oneself:战胜自己overcame overcome:的过去式overcome despair:摆脱绝望例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.V-T If you overcome a problem or a feeling, you successfully deal with it and control it. 克服2.V-T If you are overcome by a feeling or event, it is so strong or has such a strong effect that you cannot think clearly. 困扰3.V-T If you are overcome by smoke or a poisonous gas, you become very ill or die from breathing it in. 熏倒;熏死[usu passive]knowledge音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’nɒlɪdʒ] 美[’nɑlɪdʒ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 知识,学问;知道,认识;学科n. (Knowledge)人名;(南非)诺利奇短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ explicit knowledge:外显知识;显性知识;明确知识;显性学问Knowledge Graph:知识图谱;知识图;常识图谱;知识图表procedural knowledge:程序性知识;knowledge representation:过程性知识;程序知识;程序性的知识Local Knowledge:知识表示;Knowledge market:知识表达;知识表征;知识表现comprehensive knowledge:地方性知识;implicit knowledge:局部知识;地方知识;了解球路情况Conceptual Knowledge:知识市场;线知识市场例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-UNCOUNT Knowledge is information and understanding about a subject which a person has, or which all people have. 知识; 学识2.PHRASE If you say that something is true to your knowledge or to the best of your knowledge , you mean that you believe it to be true but it is possible that you do not know all the facts. 据某人所知national音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’næʃ(ə)n(ə)l] 美[’næʃnəl]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adj. 国家的;国民的;民族的;国立的n. 国民短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ National Day:国庆日;中华人民共和国国庆节;国庆节;新加坡国庆日National Stadium:国家体育场;苏帕查拉赛体育场;新加坡国家体育场;国家体育馆National University:国立大学;中国公学;美国国立大学;国家级大学National Museum:国家博物馆;博物馆;国度博物馆National Mall:国家广场;国家大草坪;国度广场;大草坪National flag:国旗;升国旗;意大利国旗;美国国旗national treatment:国民待遇;国民待遇原则;国家待遇;国民待遇条款National Review:国家评论;国民评论;全国评论;国度评论National Velvet:玉女神驹;玉仙女驹;外文名称;天鹅绒例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.ADJ National means relating to the whole of a country or nation rather than to part of it or to other nations.全国的2.nationally ADV 全国地3.ADJ National means typical of the people or customs of a particular country or nation. 国民的; 民族的[ADJ n]4.N-COUNT You can refer to someone who is legally a citizen of a country as a national of that country. 国民outline音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’aʊtlaɪn] 美[’aʊtlaɪn]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 过去式outlined 过去分词outlined 现在分词outlining ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 轮廓;大纲;概要;略图vt. 概述;略述;描画…轮廓短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Outline dimensions:外形尺寸;外型尺寸;外形圆;轮廓尺寸hierarchical outline:分层大纲;采取层级结构式大纲tooth outline:齿外形;齿廓Outline Stroke:轮廓化描边;描边轮廓化;外框笔画Topic Outline:题目式提纲;主题提纲;标题式提纲;短语形式Document Outline:文档大纲pit outline:露天矿场边界Outline Writing:提纲要旨例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.V-T If you outline an idea or a plan, you explain it in a general way. 概述2.N-COUNT An outline is a general explanation or description of something. 概要[also ’in’ N]3.V-T PASSIVE You say that an object is outlined when you can see its general shape because there is light behind it. 映衬出…的轮廓4.N-COUNT The outline of something is its general shape, especially when it cannot be clearly seen. 轮廓unhappy音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[ʌn’hæpɪ] 美[ʌn’hæpi]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 比较级unhappier 最高级unhappiest ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adj. 不快乐的;不幸福的;不适当的短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Unhappy China:中国不高兴very unhappy:悲惨的;非常不高兴;不欢快;很不爽unhappy yesterday:昨天的不愉快;不满昨日;昨天不满if unhappy:如果不快乐No unhappy:不准不开心Many unhappy:很多的不开心Unhappy when:不如意的时候unhappy duty:不快乐的勤务unhappy player:闹情绪球员例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.ADJ If you are unhappy , you are sad and depressed. 不幸福的2.unhappily ADV 不快乐地3.ADJ If you are unhappy about something, you are not pleased about it or not satisfied with it. 对…不高兴的;对…不满意的[v-link ADJ]4.unhappiness N-UNCOUNT 不高兴; 不满意5.ADJ An unhappy situation or choice is not satisfactory or desirable. 不令人满意的; 不理想的[ADJ n]orchestra音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _英[’ɔːkɪstrə] 美[’ɔrkɪstrə]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _n. 管弦乐队;乐队演奏处短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _Minnesota Orchestra:明尼苏达交响乐团;明尼苏达管弦乐团;明尼苏达乐团;明尼苏达管弦乐团古典类型Orchestra Seats:巴黎不打烊;蒙田大道;贵宾席;正厅前座Chamber Orchestra:室内乐团;室内管弦乐队;室内管弦乐团;室内乐队orchestra shell:音响反射板;乐团壳;乐队壳MSC Orchestra:管乐号Orchestra ver:好像爱你;一天一天;乐团见;见乐团Epic Orchestra:史诗乐团Duke Orchestra:公爵交响乐团Orchestra Hall:管弦乐厅例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT An orchestra is a large group of musicians who play a variety of different instruments together. Orchestras usually play classical music. 管弦乐队2.→ see also symphony orchestra3.N-SING The orchestra or the orchestra seats in a theatre or concert hall are the seats on the first floor directly in front of the stage. 舞台前方一楼座位[美国英语]violently音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[ˈvaɪələntli] 美[’vaɪələntli]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adv. 猛烈地,激烈地;极端地短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ dance violently:剧烈地跳动violently angry:狂怒object violently:强烈地反对plunge violently:猛烈地投入box violently:强烈地拳击rouse violently:强有力地激发tremble violently:猛烈地颤动thunder violently:猛烈地打雷Not Violently:不暴力地gap音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[gæp] 美[ɡæp]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 间隙;缺口;差距;分歧vi. 裂开vt. 使形成缺口短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ financing gap:资金缺口;root gap:财政缺口;融资缺口;资金缺口英语anion gap:根部间隙gap plate:阴离子间隙;阴离子隙;负离子缺额;离子间隙block gap:锯口罩;补隙板discharge gap:块间隙;块间隔;块隙;gap repair:信息组间隙Knowledge Gap:放电间隙;翻译;电火花磨削Macmillan Gap:缺口修复例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT A gap is a space between two things or a hole in the middle of something solid. 缺口; 裂缝2.N-COUNT A gap is a period of time when you are not busy or when you stop doing something that you normally do. 间隔期3.N-COUNT If there is something missing from a situation that prevents it from being complete or satisfactory, you can say that there is a gap . 缺漏4.N-COUNT A gap between two groups of people, things, or sets of ideas is a big difference between them. 差距solution音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[sə’luːʃ(ə)n] 美[sə’luʃən]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 解决方案;溶液;溶解;解答短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ aqueous solution:水溶液;水溶剂;水剂;水性溶液saline solution:生理盐水;食盐水;盐溶液isotonic solution:等张溶液;normal solution:等渗溶液molar solution:当量溶液;规度溶液;approximate solution:规定溶液;solution polymerization:标准溶液physical solution:克分子溶液;摩尔混液;balanced solution:摩尔溶液;例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT A solution to a problem or difficult situation is a way of dealing with it so that the difficulty is removed. 解决办法2.N-COUNT The solution to a puzzle is the answer to it. 谜底3.N-COUNT A solution is a liquid in which a solid substance has been dissolved. 溶液[also ’in’ N]grocery音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’grəʊs(ə)rɪ] 美[’ɡrosəri]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 复数groceries ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 食品杂货店食品杂货短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Green Grocery:蔬菜水果店;蔬菜店水果店;水果蔬菜店;蔬菜生果店grocery shopping:买菜;生活必需品的购买;去杂货店;买食品杂货grocery department:食品杂货部门;食品干货部Grocery choice:食品选择General grocery:通用件杂货GROCERY ZONE:杂货区Grocery Checkout:商店付款系统Lu Grocery:卢氏杂货店例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT A grocery or a grocery store is a small shop that sells foods such as flour, sugar, and canned goods.食品杂货店[美国英语]2.→ see also supermarket3.N-PLURAL Groceries are foods you buy at a grocery or at a supermarket. 食品杂货manage音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’mænɪdʒ] 美[’mænɪdʒ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ vt. 管理;经营;控制;设法vi. 处理;应付过去短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Manage Up:和上司保持一致;向上管理;管理上司Equipment manage:设备管理;装备管理;仪器设备管理to manage:主导;管理;打理;治理preside manage:统辖;负责manage team:管理球队;管理团队Manage Sources:管理来源;管理源Manage Ads:广告管理;管理广告;管理系统;这个photo manage:照片管理器;照片管理;张照片管理;照片金管家Manage Servers:管理服务器;管理伺服器例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.V-T If you manage an organization, business, or system, or the people who work in it, you are responsible for controlling them. 管理; 经营(机构、企业、系统等)2.V-T If you manage time, money, or other resources, you deal with them carefully and do not waste them. 管理(时间、金钱等)3.V-T If you manage to do something, especially something difficult, you succeed in doing it. 设法(做成某事)4.V-I If you manage , you succeed in coping with a difficult situation. 成功应对5.V-T If you say that you can manage an amount of time or money for something, you mean that you can afford to spend that time or money on it. 腾出(时间、金钱等)6.V-T If you say that someone managed a particular response, such as a laugh or a greeting, you mean that it was difficult for them to do it because they were feeling sad or upset. 勉强做出(某种回应)7.CONVENTION You say " I can manage " or " I’ll manage " as a way of refusing someone’s offer of help and insisting on doing something by yourself. 我能应付tradition音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[trə’dɪʃ(ə)n] 美[trə’dɪʃən]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 惯例,传统;传说短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Chinese tradition:中国传统;民族音乐;中华传统;华夏传统educational tradition:教育传统;教育现代化Cartier Tradition:卡地亚古董珍藏系列;卡地亚珍藏Legal Tradition:法律传统;法律文化传统;法制传统;司法传统Gothic tradition:哥特传统Tradition writing:传统写作philosophical tradition:哲学传统;哲学基础literary tradition:文学传统例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 1.N-VAR A tradition is a custom or belief that has existed for a long time. 传统normally音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’nɔːm(ə)lɪ] 美[’nɔrmli]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adv. 正常地;通常地,一般地短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Normally white:常白;模式的常白方式;显示方式;以寻常白normally acidic:通常酸性designed normally:正常设计normally operation:平时运营Normally Luminating:正常照明normally blossom:正常开花结实normally displayed:正常显示Normally Tired:总感觉疲乏Operating Normally:运行正常normally contact:常开触点例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.ADV If you say that something normally happens or that you normally do a particular thing, you mean that it is what usually happens or what you usually do. 通常2.ADV If you do something normally , you do it in the usual or conventional way. 按常规地[ADV after v]care音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[keə] 美[kɛr]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 过去式cared 过去分词cared 现在分词caring ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 关怀;照料;谨慎;忧虑vi. 照顾;关心;喜爱;顾虑vt. 在意;希望或喜欢n. (Care)人名;(英)凯尔;(塞)察蕾短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ friso care:美素佳儿;美素全称美素佳儿Primary Care:医学院;基层医疗;社区医疗Child care:儿童保健;儿童护理;托儿所;儿童看护Critical Care:急救护理;特级护理;重症监护;医院人生Self Care:照顾自己;自我照顾;自我照料;自顾BUICK CARE:别克关怀;别克眷注parental care:亲代抚育;亲代养育;亲代关怀;亲代照顾MAGIC CARE:魔法护理;邪术照顾护士denture care:清洗假牙;假牙的护理;假牙护理Patient Care:病人照护;病患照顾;病人护理;病人照顾例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.V-T/V-I If you care about something, you feel that it is important and are concerned about it. 关注[no cont]2.V-I If you care for someone, you feel a lot of affection for them. 喜爱[no cont] [表赞许] </p>3.caring N-UNCOUNT 体贴4.V-I If you care for someone or something, you look after them and keep them in a good state or condition.照顾5.N-UNCOUNT Care is also a noun. 照顾6.V-T/V-I You can ask someone if they would care for something or if they would care to do something as a polite way of asking if they would like to have or do something. 想要[no cont]7.N-UNCOUNT If you do something with care , you give careful attention to it because you do not want to make any mistakes or cause any damage. 谨慎8.N-COUNT Your cares are your worries, anxieties, or fears. 烦恼9.→ see also caring , day care , intensive care10.PHRASE You can use for all I care to emphasize that it does not matter at all to you what someone does. 不关某人的事[强调]11.PHRA SE If you say that you couldn’t care less about someone or something, you are emphasizing that you are not interested in them or worried about them. You can also say that you could care less , with the same meaning. 不在乎[强调]12.PHRASE If someone sends you a letter or package care of or in care of a particular person or place, they send it to that person or place, and it is then passed on to you. 经某人转交13.PHRASE If you take care of someone or something, you look after them and prevent them from being harmed or damaged. 照顾某人/某事14.PHRASE If you take care to do something, you make sure that you do it. 留意做某事15.PHRASE To take care of a problem, task, or situation means to deal with it. 处理某事16.PHRASE You can say " Who cares? " to emphasize that something does not matter to you at all. 管它呢[强调]collector音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[kə’lektə] 美[kə’lɛktɚ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 收藏家;[电子] 集电极;收税员;征收者短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Garbage Collector:垃圾回收器;垃圾收集器;无用单元收集程序;清洁工air collector:空气收集器;集气罐;储气罐;空气集热器data collector:藏家;数据采集装置;电子收藏家;Resource Collector:数据收集器solar collector:资源收集器;资源采集船;资源搜集器;资源收集船collector lens:太阳能集热器;collector junction:太阳能收集器;太阳能聚集器;太阳集热器cosine collector:聚光透镜;collector shoe:会聚透镜;聚场透境例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT A collector is a person who collects things of a particular type as a hobby. 收藏家2.N-COUNT You can use collector to refer to someone whose job is to take something such as money, tickets, or rubbish from people. For example, a rent collector collects rent from people. (钱款、票据或垃圾的) 收取人make音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[meɪk] 美[mek]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 过去式made 过去分词made 现在分词making ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ vt. 使得;进行;布置,准备,整理;制造;认为;获得;形成;安排;引起;构成vi. 开始;前进;增大;被制造n. 制造;构造;性情n. (Make)人名;(塞、南非)马克短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ make up:弥补;构成;编造;组成make sense:讲得通;有意义;有道理;言之有理make for:走向;有助于;促进;导致make friends:交朋友;友好相处;交友诀窍;交个朋友吧make mistakes:犯错误;学会犯错;出错make headway:取得进展;有进展;前进;取得盼望make good:成功;弥补;旅行诺言;补偿make it:成功;达到;赶上;在面试make clear:澄清;弄清楚;表明;说明intention音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[ɪn’tenʃ(ə)n] 美[ɪn’tɛnʃən]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 意图;目的;意向;愈合短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ paradoxical intention:矛盾意向;矛盾疗法;矛盾意向法;欲擒故纵法purchasing intention:购买意图;购买意向;购买意愿;购买意向强弱Real intention:真实意图;真正目的;真实意向;本意communicative intention:交际意图;交际意向;行为目的Evil Intention:恶意;邪念aesthetic intention:审美意向;审美内涵;审美心象;审美意图Wilful Intention:故意intention definition:意向定义semantic intention:语义意向;语义意图例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-VAR An intention is an idea or plan of what you are going to do. 意图2.PHRASE If you say that you have no intention of doing something, you are emphasizing that you are not going to do it. If you say that you have every intention of doing something, you are emphasizing that you intend to do it. 不打算/打算[强调]silly音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’sɪlɪ] 美[’sɪli]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adj. 愚蠢的;不明事理的;没头脑的n. 傻瓜n. (Silly)人名;(匈)希伊;(法)西利短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Silly Ho:真傻;淘气;糊涂的Silly Sausage:蠢蠢的腊肠;愚蠢的香肠;弹力香肠狗silly season:新闻缺乏时期;傻瓜球季;翻译Silly things:笑红尘Silly Symphony:糊涂交响曲;糊涂交响乐;糊涂交响樂Silly People:愚蠢的人Silly Science:妙趣实验室;小小科学家silly games:无聊的游戏borderline silly:近乎犯傻例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.ADJ If you say that someone or something is silly , you mean that they are foolish, childish, or ridiculous. 愚蠢的; 幼稚的; 荒唐的curious音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’kjʊərɪəs] 美[’kjʊrɪəs]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 比较级more curious 最高级most curious ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adj. 好奇的,有求知欲的;古怪的;爱挑剔的短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Curious George:好奇的乔治;好奇猴乔治;好奇小猴乔治;好奇乔治Stay Curious:保持好奇;德不罗毅;拒绝的魅力Curious Creature:好奇的精灵专辑;好奇的精灵Curious Owl:叫好奇猫头鹰Curious Camel:好奇的骆驼Curious Balls:好奇的小球Something curious:对某物好奇;对某事感到好奇;对什么事情好奇Curious Minds:好奇心;好奇的小脑袋;奇思妙想例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.ADJ If you are curious about something, you are interested in it and want to know more about it. 好奇的2.curiously ADV 好奇地[ADV after v]3.ADJ If you describe something as curious , you mean that it is unusual or difficult to understand. 不寻常的;难以理解的4.curiously ADV 令人不解的地aspect音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’æspekt] 美[’æspɛkt]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 方面;方向;形势;外貌短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ imperfective aspect:未完成体;未完成式;未完成貌aspect card:标号卡片;标记卡;标志卡;progressive aspect:特征卡片aestival aspect:进行体;进行貌;进行态;进行式devourer aspect:夏季相;盛夏景色viewpoint aspect:吞噬者Aspect Mining:视点体Interesting aspect:方面挖掘;侧面挖掘例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT An aspect of something is one of the parts of its character or nature. 方面2.N-COUNT The aspect of a building or window is the direction in which it faces. (建筑物或窗子的) 朝向[正式]truth音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[truːθ] 美[truθ]附加_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [ 复数truths ]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 真理;事实;诚实;实质短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ in truth:事实上;实际上;的确;实在ground truth:地面实况;地面真值;地表实况;路面实况apriori truth:先天的真理Historical truth:历史事实;历史真实;历史真理;历史的真实Conventional Truth:世俗谛Hidden Truth:法网晴天;丹特丽安的书架;隐藏的真实;隐匿真相Deep Truth:深邃蓝Necessary Truth:必然真理truth criterion:真理标准例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-UNCOUNT The truth about something is all the facts about it, rather than things that are imagined or invented. 实情; 真相2.N-UNCOUNT If you say that there is some truth in a statement or story, you mean that it is true, or at least partly true. 真实性3.N-COUNT A truth is something that is believed to be true. 真理4.PHRASE You say to tell you the truth or truth to tell in order to indicate that you are telling someone something in an open and honest way, without trying to hide anything. 跟你说实话conversation音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[kɒnvə’seɪʃ(ə)n] 美[,kɑnvɚ’seʃən]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 交谈,会话;社交;交往,交际;会谈;(人与计算机的)人机对话短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ conversation practice:对话练习;会话练习;会话操练;对话训练Telephone Conversation:电话交谈;打电话;电话英语;与电话沟通Long Conversation:长对话;促膝长谈;长篇对话;长会话Daily conversation:日常会话;日常对话;日常交谈;生活对话Conversation Starters:开始谈话in conversation:交谈中;在谈话中;唱片名Casual conversation:随意交谈;随意性会话;商务话题随意聊General Conversation:一般性对话;一般的对话;普通性对话conversation assembler:对话汇编程序;交谈组译器;会话汇编程序英语例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT If you have a conversation with someone, you talk with them, usually in an informal situation. (非正式的) 交谈scared音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[skeəd] 美[skɛrd]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ adj. 害怕的v. 使害怕(scare的过去分词)短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ SCARED LAND:吓得土地;害怕土地;害怕的土地scared monsters:圣兽舞姬scared cow:神圣不可侵犯的人Scared love:情圣Scared Lake:圣湖Was scared:就被吓到SCARED HALLOWEN:试验摇滚Easily scared:胆小的例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.ADJ If you are scared of someone or something, you are frightened of them. 害怕的2.ADJ If you are scared that something unpleasant might happen, you are nervous and worried because you think that it might happen. 担心的; 害怕的kite音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _英[kaɪt] 美[kaɪt]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _n. 风筝vi. 使用空头支票;像风筝一样飞;轻快地移动vt. 骗钱;涂改(支票)n. (Kite)人名;(塞)基特;(英)凯特短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _kite surfing:风筝冲浪;鹞子冲浪;冲浪风筝;风帆Brahminy Kite:栗鸢;大红鹰;鹰科;嗨抡杪轻揣kite liberator:少女杀手;风筝解放者;放风筝的人;姹儿刺客A Kite:纸鸢;一个风筝;少女杀手;一只风筝box kite:遥控立体风筝;Stunt kite:匣形风筝;箱形风筝图片Kite Festival:运动风筝;特技风筝life kite:风筝节kite drag:救生发报风筝;救生风筝例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _1.N-COUNT A kite is an object, usually used as a toy, which is flown in the air. It consists of a light frame covered with paper or cloth and has a long string attached which you hold while the kite is flying. 风筝2.N any diurnal bird of prey of the genera Milvus, Elanus, etc, typically having a long forked tail and long broad wings and usually preying on small mammals and insects: family Accipitridae (hawks, etc) 鹰科昼行的鸟3.PHRASE If you say that someone is as high as a kite , you mean that they are very excited or that they are greatly affected by alcohol or drugs. 如痴如狂的; 因饮酒或吸毒而飘飘然的steamboat音标_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 英[’stiːmbəʊt] 美[’stimbot]释义_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ n. 汽船,轮船短语_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Steamboat Bill:小比尔号汽船;船长二世;蒸汽船小比尔号steamboat falls:蒸汽船瀑布Steamboat Inn:汽船旅馆;的评论Richie Steamboat:里奇·斯廷博特ELECTRIC STEAMBOAT:电火锅SteamBoat Food:火锅料Steamboat Hollow:街道地址例句_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 1.N-COUNT A steamboat is a boat or ship that has an engine powered by steam. 汽船; 汽艇。
微软发布全新知识图谱和概念标签模型,让AI更像人
微软发布全新知识图谱和概念标签模型,让AI更像人近日,微软亚洲研究院发布全新的、能帮助计算机更好运作的微软知识图谱(Microsoft Concept Graph)和概念标签模型(Microsoft Concept Tagging),机器可以更好地理解人类交流并且进行语义计算。
微软知识图谱(Microsoft Concept Graph)Microsoft Concept Graph 是一个大型的知识图谱系统。
包含的知识来自于数以亿计的网页和多年积累的搜索日志,可以为机器提供文本理解的常识性知识。
研究员称,现在的人工智能可以战胜李世石,但其学习能力以及完成一般任务的能力可能都远不及一个三四岁的儿童。
因此研究员们从人类的学习成长过程开始入手,试图找到逐步实现机器智能的解决方法。
如何逐步实现机器智能,微软亚洲研究院给出的答案是“常识”,常识是理解是万物的基础。
人们从小就开始学习(对世界的认知),学习作为人类与生俱来的本能可让我们上学之前就已知道“西瓜是甜的水果”、“水是液体”等基础概念,随着年龄和经历的增长,这类并不属于某个细分专业领域的开放性常识也在不断增长。
微软亚洲研究院这次做的事情就是让计算机“懂得”这些常识性概念。
为此,经过六年的研究微软为 Microsoft Concept Graph 收集了超过540 万条概念核心知识库。
除了“建筑”、“诗人”等这类通用知识库外,还包含数百万比较生僻的长尾概念,如“抗帕金森治疗”、“名人婚纱设计师”、“基础的水彩技巧”等。
与此同时,Microsoft Concept Graph 同样包含了大量数据空间(每条知识概念都包含一系列的实体或者子概念,例如“太阳系”底下可能就会包括“水星”、“火星”、“地球”等等)。
微软概念标签模型(Microsoft Concept Tagging)概念标签模型(Microsoft Concept Tagging)与微软知识图谱(Microsoft Concept Graph)相辅相成。
基于有向图的双向匹配分词算法及实现
收稿日期:2004-11-14;修订日期:2005-03-12 基金项目:国家863计划项目;国家自然科学基金资助项目(60403050) 作者简介:陈耀东(1978-),男,江西丰城人,硕士研究生,主要研究方向:自然语言处理; 王挺(1970-),男,湖南长沙人,副教授,主要研究方向:自然语言处理、计算机软件.文章编号:1001-9081(2005)06-1442-03基于有向图的双向匹配分词算法及实现陈耀东,王 挺(国防科学技术大学计算机学院,湖南长沙410073)(ydchen0104@yahoo )摘 要:在分析了现有各种汉语分词算法及其优缺点的基础上,提出以句子覆盖率和分词覆盖率作为评价分词方法的指标,详细介绍了基于网络有向图的双向匹配分词算法的设计与实现,该算法对经典的最大匹配分词算法进行了改进,通过带覆盖歧义标志的有向图生成多候选分词序列。
与最大匹配算法和全切分算法的比较实验显示,基于有向图的双向匹配算法以低复杂度实现了高覆率盖。
关键词:句子覆盖率;分词覆盖率;双向最大匹配算法;全切分;网络有向图中图分类号:TP391.1 文献标识码:AUsi n g d i rected graph ba sed BDMM a lgor ithm for Ch i n ese word segm en t a ti onCHEN Yao 2dong,WANG Ting(School of Co m puter Science,N ational U niversity of D efense Technology,Changsha Hunan 410073,China )Abstract:I n this paper,the authors firstly studied current seg mentati on algorith m s,then,modifid the traditi onal M axi m u m Match (MM )algorith m.W ith the considerati on of both word 2coverage rate and sentence 2coverage rate,a character D irected Graph with a mbiguity mark was i m p le mented f or searching all possible seg mentati on sequences .This methodcompared with the classic MM algorith m s and omni 2seg mentati on algorith m and the experi m ent result shows that the D irectedGraph based algorith m can achieve higher coverage rate and l ower comp lexity .Key words:sentence 2coverage rate;word 2coverage rate;B i 2directi onal M axi m u m Match;omni 2seg mentati on;directed graph0 引言凡涉及句法、语义的项目(如中文搜索引擎、机器翻译等)均要用到分词,汉语自动分词是中文信息处理的第一步骤,分词质量将直接影响整个文本处理任务的后续工作和最终的评测结果。
基于词汇树层次语义模型的图像检索算法
基于词汇树层次语义模型的图像检索算法
张月辉;吴健;陆姗姗;崔志明
【期刊名称】《微电子学与计算机》
【年(卷),期】2012(29)11
【摘要】解决语义鸿沟必须建立图像低层特征到高层语义的映射,针对此问题,本文提出了一种基于词汇树层次语义模型的图像检索方法.首先提取图像包含颜色信息的SIFT特征来构造图像库的特征词汇树,生成描述图像视觉信息的视觉词汇.并在此基础上利用Bayesian决策理论实现视觉词汇到语义主题信息的映射,进而构造了一个层次语义模型,并在此模型基础上完成了基于内容的语义图像检索算法.通过检索过程中用户的相关反馈,不仅可以加入正反馈图像扩展图像查询库,同时能够修正高层语义映射.实验结果表明,基于该模型的图像检索算法性能稳定,并且随着反馈次数的增加,检索效果明显提升.
【总页数】5页(P172-176)
【关键词】词汇树;语义主题信息;层次语义模型;语义映射;图像检索
【作者】张月辉;吴健;陆姗姗;崔志明
【作者单位】苏州大学智能信息处理及应用研究所
【正文语种】中文
【中图分类】TP391
【相关文献】
1.基于主题层次树和语义向量空间模型的用户建模 [J], 胡吉明;胡昌平
2.基于多层次概念语义网络结构的中文医学信息语义标引体系和语义检索模型研究[J], 李毅;庞景安
3.基于ISODATA聚类的词汇树图像检索算法 [J], 张婷;戴芳;郭文艳
4.基于多层次视觉语义特征融合的图像检索算法 [J], 张霞;郑逢斌
5.基于语义绑定的分层视觉词汇库的图像检索 [J], 傅光磊;孙锬锋;蒋兴浩
因版权原因,仅展示原文概要,查看原文内容请购买。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
Graph-Based Hierarchical Conceptual Clustering Istvan Jonyer, Lawrence B. Holder and Diane J. CookDepartment of Computer Science and EngineeringUniversity of Texas at ArlingtonBox 19015 (416 Yates St.), Arlington, TX 76019-0015E-mail: {jonyer | holder | cook}@Phone: (817) 272-2596Fax: (817) 272-3784AbstractHierarchical conceptual clustering has been proven to be a useful data mining technique. Graph-based representation of structural information has been shown to be successful in knowledge discovery. The Subdue substructure discovery system provides the advantages of both approaches. In this paper we present Subdue and focus on its clustering capabilities. We use two examples to illustrate the validityof the approach both in structured and unstructured domains, as well as compare Subdue to an earlier clustering algorithm.IntroductionCluster analysis has been studied and developed in many areas for a wide variety of applications. Among these are model fitting, hypothesis testing, hypothesis generation, data exploration, prediction based on groups, data reduction and finding true topologies [Ball 1971]. Clustering techniques have been applied in as diverse fields as analytical chemistry, geology, biology, zoology and archeology, just to mention a few. Many names have been given to this technique, among which are cluster analysis, Q-analysis, typology, grouping, clumping, classification, numerical taxonomy, mode separation and unsupervised pattern recognition, which further signifies the importance of clustering techniques [Everitt 1980].The purpose of applying clustering to a database is to gain better understanding of the data, in many cases through revealing hierarchical topologies. An example of this is the classification of vehicles into groups such as cars, trucks, motorcycles, tricycles, and so on, which are then further subdivided into smaller and smaller groups based on some other traits.In this paper we present Subdue, a structural knowledge discovery system, specifically focusing on its clustering capabilities. After acknowledging some earlier works, we describe Subdue, and present examples to highlight our results.Copyright © 2000, American Association for Artificial Intelligence (). All rights reserved.Related WorkNumerous clustering techniques have been devised in the past, among which are statistical, syntactic, neural and hierarchical approaches. In all cases, clustering is inherently an unsupervised learning paradigm, since it consists of identifying valuable groupings of concepts, or facts, which hopefully reveal previously unknown information. Most techniques have some intrinsic disadvantages, however. Statistical and syntactic approaches have trouble expressing structural information, and neural approaches are greatly limited in representing semantic information [Schalkoff 1992].Nevertheless, many relatively successful clustering systems have been constructed. An example of an incremental approach is Cobweb, which successively considers a set of object descriptions, while constructing a classification tree [Fisher 1987]. Labyrinth [Thompson and Langley 1991], an extension to Cobweb, can represent structured objects using a probabilistic model. AutoClass [Cheeseman at al. 1988] is an example of a bayesian classification system, which has a probabilistic class assignment scheme. It can deal with real, discrete, or missing values. Yet another algorithm, called Snob, uses the Minimum Message Length (MML) principle to do mixture modeling—another synonym for clustering [Wallace 1968].There also exist hierarchical approaches that work on databases containing data in Euclidian space. Among these are agglomerative approaches that merge clusters until an optimal separation of clusters is achieved based on intra-, and inter-cluster distances. Divisive approaches split existing clusters until an optimal clustering is found. These approaches usually have the disadvantage of being applicable only to metric data, which excludes discrete-valued and structured databases. Examples of these are Chameleon [Karypis, Han and Kumar 1999] and Cure [Guha, Rastogi and Shim 1998].Conceptual Clustering Using Subdue Subdue [Holder and Cook 1993] is a knowledge discovery system that can deal with structured data—a veryimportant feature in more and more applications. Subdue expects a graph as its input, hence a database needs to be represented as a graph before passing it to Subdue. This graph representation includes vertex and edge labels, as well as directed and undirected edges, where objects and data usually map to vertices, and relationships and attributes map to edges (see Figure 2 for an example). Subdue’s discovery algorithm discovers interesting, repetitive substructures in the input graph. Our Graph-Based Hierarchical Conceptual Clustering (GBHCC) algorithm begins with an empty lattice and calls Subdue to find a substructure S that maximally compresses the input graph G. If S achieves some compression of G, then S is added to the lattice and used to compress the graph G. The compressed graph is passed again to Subdue to find another substructure. This iterative approach on successively more compressed graphs allows Subdue to find new substructures defined in terms of previously discovered substructures. Therefore, when substructures are added to the lattice, their parents may include other, non-root nodes in the lattice. If a substructure is composed of two of the same previously-discovered substructures, then there will be two links from the parent to the child in the lattice.Subdue’s discovery algorithm discovers substructures in the input graph. Subdue uses a beam search that–starting with single-node subgraphs–incrementally expands the substructures that seem the best thus far. During the expansion process, a substructure is expanded in all possible ways by its neighboring vertices, and all instances of these new substructures are found. This discovery process continues iteratively until all possible subgraphs have been considered, or the algorithm reaches a user-specified limit. After the best substructure is found and the graph is compressed, the process starts all over, finding another best substructure. This search is guided by the Minimum Description Length (MDL) principle, originally developed by Rissanen [Rissanen 1989]. According to the evaluation heuristic, the best substructure is the one that minimizes the description length of the graph when compressed by the substructure. When compressing the graph, all instances of the substructure are replaced by a single vertex, which is a pointer to the substructure’s definition.This approach imposes more and more hierarchy on the database with each successive iteration. The definition of the best substructure after a single iteration yields the description of a cluster. After identifying a cluster, it is inserted into the classification lattice (see Figure 5). Previous works on clustering suggested the use of classification trees, however, in structured domains the strict tree representation is inadequate. We realized that in certain domains a lattice-like structure emerges instead of a tree.Subdue searches the hypothesis space of all classification lattices. During each iteration of the search process, numerous local minima are encountered, where the global minimum tends to be one of the first few minima. For clustering purposes the first local minimum is used as the best partial hypothesis. The reason for this is easy to see. Subdue starts with all the single-vertex instances of all unique substructures, and iteratively expands the best ones by a single vertex. The local minimum encountered first is therefore caused by a smaller substructure with more instances than the next local minimum, which must be larger, and have fewer instances. A smaller substructure is more general than a larger one, and should be a parent node in the classification lattice for any more specific clusters. Even though it is entirely possible to use the global minimum as the best substructure, we found that if the global minimum is not the first local minimum it may produce overlapping clusters. Overlapping clusters are those that include the same information. For example, in a particular clustering of the vehicles domain two clusters may include the information “number of wheels: 4”. This suggests that perhaps a better clustering may be constructed in which this information is part of a cluster at a higher level. Subdue supports biasing the discovery process. Predefined substructures can be provided to Subdue, which will try to find and expand these substructures, this way “jump-starting” the discovery. The inclusion of background knowledge proved to be of great benefits [Djoko, Cook and Holder 1997]. Inexact graph matching is also provided by Subdue to account for slight variations of a substructure. The user is given control of the degree of similarity between the substructures to be considered the same. Subdue also supports supervised learning, where positive and negative examples are provided to the system. Substructures found that are similar to positive examples are given a higher value, while substructures similar to the negative examples are penalized. This way of influencing the discovery process has proven successful, an example of which is the application of Subdue to the chemical toxicity domain [Chittimoori, Holder and Cook 1999].ExperimentsA small experiment devised by Fisher can serve as an example of Subdue’s performance on unstructured data, as well as offer a brief comparison to Cobweb. The database used for the experiment is given in Table 1. Cobweb produces the classification tree shown in Figure 1, as suggested by Fisher [Fisher 1987].The animal domain is represented in Subdue as a graph, where attribute names (like Name and BodyCover) were mapped to edges, and attribute values (like mammal and hair) were mapped to vertices. In unstructured databasesTable 1 Animal DescriptionsName Body Cover Heart Chamber Body Temp. Fertilization mammal hair four regulated internal bird feathers four regulated internal reptile cornified-skin imperfect-four unregulated internal amphibian moist-skin three unregulated external fish scales twounregulatedexternallike this the data set translates to a collection of small, star-like, connected graphs. Intuitively, we would map the “main” attribute–Name in this case–to the center node and all other attributes would be connected to this central vertex with a single edge. We found, however, that a more general representation yields better results. In this representation the center node becomes a very general description of the example. In this case the center node can be animal . Note that the Name attribute becomes just a regular attribute (see Figure 2). In the most general case, the center node could be named entity , or object , since the designation is quite irrelevant to the discovery process—the purpose is proper structural representation.Subdue generated the hierarchical clustering shown in Figure 3. Subdue actually produces a hierarchical graph as its output that can be viewed with a visualization software. This also allows us to include informative details in theclassification lattice. A node in the classification hierarchy includes the description of vertices and edges that form the cluster.Subdue’s result is similar to that of Cobweb’s. The “mammal/bird ” branch is clearly the same. Amphibians and fish are grouped in the same cluster based on their external fertilization, which is done the same way by Cobweb. Subdue, however, incorporates reptiles with amphibians and fish, based on their commonality in unregulated body temperature. This clustering of the animal domain seems better, since Subdue eliminated the overlap between the two clusters (reptile and amphibian/fish ) by creating a common parent for them that describes this common trait.To illustrate Subdue’s main advantage–the ability to work with structured data–we present a task that involves describing a DNA sequence by clustering. A portion of theFigure 1 Hierarchical clustering over animal descriptions by Cobweb.Figure 2 Graph representation of an animal descriptionmammal bird fish amphibianDNA is shown in Figure 4. To represent the DNA as a graph, atoms and small molecules are mapped to vertices, and bonds are represented by undirected edges. The edges are labeled according to the type of bond, single or double. A portion of the classification lattice is shown in Figure 5. For better understanding, we show the chemical compounds the clusters define, rather than the textual description extracted from the graph representation of the DNA (like in Figure 3).The lattice closely resembles a tree, with the exception that two nodes (bottom-left) have two parents. The lattice in Figure 5 describes 71% of the DNA sequence shown in Figure 4. As the figure shows, smaller, more commonly occurring compounds are found first that compose the first level of the lattice. These account for more than 61% of the DNA. Subsequently identified clusters are based onFigure 4 Portion of a DNA moleculeFigure 3 Hierarchical clustering over animal descriptions by Subdue.Figure 5 Partial hierarchical clustering of a DNA sequence.these smaller clusters that are either combined with each other, or with other atoms or molecules to form a new cluster. The second level of the lattice extends the conceptual clustering description such that an additional 7% of the DNA is covered. Evaluation of the domain-relevance of Subdue’s findings will require the assistance of a domain expert.The best clustering is usually the one that has the minimum number of clusters, with minimum number of overlaps between clusters, such that the entire data set is described. Too many clusters can arise if the clustering algorithm fails to generalize enough in the upper levels of the hierarchy, in which case the classification lattice may become shallow with a high branching factor from the root, and a large number of overlaps. On the other extreme, if the algorithm fails to account for the most specific cases, the classification lattice may not describe the data entirely. Experimental results indicate that Subdue finds clusterings that effectively trade off these extremes.ConclusionsMost previous efforts at clustering work with unstructured databases that simply enlist object descriptions. Subdue overcomes this restriction by representing databases using graphs, which allows for the representation of a large number of relationships between items, which is an integral part of defining clusterings.We have demonstrated that Subdue’s performance on unstructured datasets parallels one of the most prominent algorithms so far, perhaps even outperforming it. We also showed Subdue’s applicability to chemical domains—specifically, a conceptual clustering of a DNA sequence. Future work on Subdue will include defining hierarchical clusterings of real-world domains, and comparisons to other clustering systems. Since data mining researchers have to rely on experts’ opinion to evaluate the effectiveness of their clustering algorithm, it would be extremely useful to devise metrics for objective evaluation of clusterings.References[Ball 1971] Ball, G. H. Classification Analysis. Stanford Research Institute SRI Project 5533, 1971. [Cheeseman at al. 1988] Cheeseman, P., Kelly, J., Self, M., Stutz, J., Taylor, W., & Freeman, D. Autoclass: A bayesian classification system. In Proceedings of the Fifth International Workshop on Machine Learning, 1988, pp.54--64. [Chittimoori, Holder and Cook 1999] R. Chittimoori, L.B. Holder, and D.J. Cook. Applying the Subdue Substructure Discovery System to the Chemical Toxicity Domain. In the Proceedings of the Twelfth International Florida AI Research Society Conference, 1999, pages 90-94.[Djoko, Cook and Holder 1997] S. Djoko, D. J. Cook, and L. B. Holder. An Empirical Study of Domain Knowledge and Its Benefits to Substructure Discovery. In the IEEE Transactions on Knowledge and Data Engineering, Volume 9, Number 4, 1997.[Everitt, 1980] Everitt, B. S. Cluster Analysis. Wiley & Sons, New York, 1980.[Fisher 1987] Fisher, D. H. Knowledge Acquisition Via Incremental Conceptual Clustering, Machine Learning. Kluwer, The Netherlands, 1987.[Guha, Rastogi and Shim 1998] S. Guha, R. Rastogi, K. Shim. CURE: An Efficient Clustering Algorithm for Large Databases. ACM SIGMOD International Conference on Management of Data, 1998.[Holder and Cook 1993] L. B. Holder and D. J. Cook. Discovery of Inexact Concepts from Structural Data. In IEEE Transactions on Knowledge and Data Engineering, Volume 5, Number 6, 1993, pages 992-994.[Karypis, Han and Kumar 1999] G. Karipis, E. Han, V. Kumar. Chameleon: Hierarchical Clustering Using Dynamic Modeling. In Computer, August, 1999, pages 68-75.[Rissanen 1989] J. Rissanen. Stochastic Complexity in Statistical Inquiry. World Scientific Company, 1989. [Schalkoff, 1992] R. Schalkoff. Pattern Recognition. Wiley & Sons, New York, 1992.[Thompson and Langley 1991] Concept formation in structured domains. In Fisher, D.H., & Pazzani, M. (Eds.), Concept Formation: Knowledge and Experience in Unsupervised Learning, chap. 5. Morgan Kaufmann Publishers, Inc., 1991.[Wallace 1968] Wallace, C.S. and Boulton, D.M., An Information Measure for Classification, Computer Journal, Vol.11, No.2, 1968, pp 185-194.。