Nonparametric estimation in a nonlinear cointegration type model
Nonuniversal Correlations and Crossover Effects in the Bragg-Glass Phase of Impure Supercon

Institut f¨ ur Theoretische Physik, Universit¨ at zu K¨ oln, Z¨ ulpicher Straße 77, D-50937 K¨ oln, Germany 2 Physics Department, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA (February 1, 2008) The structural correlation functions of a weakly disordered Abrikosov lattice are calculated in a functional RG-expansion in d = 4 − ǫ dimensions. It is shown, that in the asymptotic limit the Abrikosov lattice exhibits still quasi-long-range translational order described by a nonuniversal exponent ηG which depends on the ratio of the renormalized elastic constants κ = c66 /c11 of the flux line (FL) lattice. Our calculations clearly demonstrate three distinct scaling regimes corresponding to the Larkin, the random manifold and the asymptotic Bragg-glass regime. On a wide range of intermediate length scales the FL displacement correlation function increases as a power law with twice the manifold roughness exponent ζRM (κ), which is also nonuniversal. Correlation functions in the asymptotic regime are calculated in their full anisotropic dependencies and various order parameters are examined. Our results, in particular the κ-dependency of the exponents, are in variance with those of the variational treatment with replica symmetry breaking which allows in principle an experimental discrimination between the two approaches. PACS numbers: 74.60.Ge, 05.20.-y I. INTRODUCTION
Non Parametric Statistics

a hypothesis and in their complexity of computation. While it is usually true that statistics that throw away lot of information about the underlying data lose much of the power in testing a hypothesis, a good statistic is associated with the just optimal amount of information required. In fact, over-information can have negative e ect if it brings in irrelevant and possibly misleading information from the data. The steps involved in proposing a test usually include: 1. 2. 3. 4. Identify the above statistic Identify its distribution and its asymptotic distribution (for large samples) Formulate a rejection test for the hypothesis based on this statistic, under the null hypothesis Compute the power and other measures of accuracy of the test
2.1 Coming up with a nonparametric test : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 3.1 Sign Test for Quantile Estimation : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 3.2 2 test : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 3.3 Kolmogorov-Smirnov test : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Sign test for population comparison : : : : Wilcoxon signed rank test : : : : : : : : : : Kolmogorov-Smirnov test for two samples : Lilliefors test for normality : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
非参数统计分析NonparametricTests菜单详解

非参数统计分析――Nonparametric Tests菜单详解非参数统计分析――Nonparametric Tests菜单详解平时我们使用的统计推断方法大多为参数统计方法,它们都是在已知总体分布的条件下,对相应分布的总体参数进行估计和检验。
比如单样本u检验就是假定该样本所在总体服从正态分布,然后推断总体的均数是否和已知的总体均数相同。
本节要讨论的统计方法着眼点不是总体参数,而是总体分布情况,即研究目标总体的分布是否与已知理论分布相同,或者各样本所在的分布位置/形状是否相同。
由于这一类方法不涉及总体参数,因而称为非参数统计方法。
SPSS的的Nonparametric Tests菜单中一共提供了8种非参数分析方法,它们可以被分为两大类:1、分布类型检验方法:亦称拟合优度检验方法。
即检验样本所在总体是否服从已知的理论分布。
具体包括:Chi-square test:用卡方检验来检验二项/多项分类变量的几个取值所占百分比是否和我们期望的比例有没有统计学差异。
Binomial Test:用于检测所给的变量是否符合二项分布,变量可以是两分类的,也可以使连续性变量,然后按你给出的分界点一分为二。
Runs Test:用于检验样本序列随机性。
观察某变量的取值是否是围绕着某个数值随机地上下波动,该数值可以是均数、中位数、众数或人为制定。
一般来说,如果该检验P值有统计学意义,则提示有其他变量对该变量的取值有影响,或该变量存在自相关。
One-Sample Kolmogorov-Smirnov Test:采用柯尔莫哥诺夫-斯米尔诺夫检验来分析变量是否符合某种分布,可以检验的分布有正态分布、均匀分布、Poission分布和指数分布。
2、分布位置检验方法:用于检验样本所在总体的分布位置/形状是否相同。
具体包括:Two-Independent-Samples Tests:即成组设计的两独立样本的秩和检验。
Tests for Several Independent Samples:成组设计的多个独立样本的秩和检验,此处不提供两两比较方法。
非参数检验讲解

非参数检验 (nonparametric test )
配对设计的符号秩和检验—— 配对设计两样本比较
Wilcoxon符号秩和检验(Wilcoxon 配对法或Wilcoxon signed rank test) 是推断其差值是否来自中位数为零的总体的方法,可用于配对设计差值的比较 和单一样本与总体中位数的比较。
也是非参数检验的优点。 多数非参数检验明显地或隐含地利用了秩的性质;但也有一 些非参数方法没有涉及秩的性质。 掌握对数据进行编秩的方法是学习秩和检验的基本要求。
非参数检验 (nonparametric test )
非参数检验的最常用方法——秩和检验( rank test )
A组: - 、、+、+、+、+、++、++、++、 ++、+++、+++ - ± + + + + ++ ++ ++ ++ +++ +++ 1 2 3 4 5 6 7 8 9 10 11 12
1 2 4.5 4.5 4.5 4.5 8.5 8.5 8.5 8.5 11.5 11.5
平均秩次=(3+6)/2=4.5
非参数检验 (nonparametric test )
常用秩和检验方法
非参数检验 (nonparametric test )
配对设计的符号秩和检验—— 配对设计两样本比较
非参数检验又称为任意分布检验(distribution- free test)
,它不考虑研究对象总体分布具体形式,也不对总体参数 进行统计推断,而是通过检验样本所代表的总体分布形式 是否一致来得出统计结论。
Non-Parametric Method of Estimation

4.3 Kernel Regression
33
other words, if m() is sufficiently smooth, then in a small neighborhood aroundxo, m(Xo) will be nearly constant and may be estimated by taking the average of the Y ^ 's that corresponds to those X^ 's nearxQ. The closer the X/s are to the valuex^, the closer the average of the corresponding Y/s will be to m(Xo). This argues for a weighted average of theY/s, where the weights decline as the X/s get further away from x^. This weighted average procedure for estimating m(x) is the essence of smoothing. More formally, for any arbitrary x, a smoothing estimator of m(x) may be expressed as, AW^xi'^tjWY,,
非参数检验

➢ 编秩:数据相等则取平均秩,
➢ 求秩和
➢ 计算检验统计量H值
H 12 N(N 1)
Ri2 3( N 1) ni
出生体重(kg)xij ABCD
相应秩次 Rij A BCD
2.7 2.9 3.3 3.5
3
4
7 11
2.4 3.2 3.6 3.6
2 5.5 12.5 12.5
2.2 3.2 3.4 3.7
χ 2 12
R
2 i
3(N1)
N(N1) ni
χ2
12 14(14 1)
152
4
152 3
37.52 4
37.52 3
3(14
1)
χ 2 9.375
χ
2 c
1
χ2
(t
3 j
t
j
)
n3 n
1
(23
9.375 2) (33 3) (23
143 14
2)
9.50
四、随机区组设计资料的秩和检验 (Friedman test)
正态近似法
如果n1或n2-n1超出附表的范围,可按下式 计算u值:
u | T n1(N 1) / 2 | 0.5 n1n2 (N 1) / 12
在相同秩次较多时,应用下式进行校正:
uC u / C
C 1
(t
3 j
t
j
)
/(N
3
N)
tj为第j组相同秩次的个数
频数表资料(或等级资料)两样本资料比较
xi (2) 86 71 77 68 91 72 77 91 70 71 88 87
12 对双胞胎兄弟心理测试结果
后出生者得分 差 值
Anomaly Detection A Survey(综述)

A modified version of this technical report will appear in ACM Computing Surveys,September2009. Anomaly Detection:A SurveyVARUN CHANDOLAUniversity of MinnesotaARINDAM BANERJEEUniversity of MinnesotaandVIPIN KUMARUniversity of MinnesotaAnomaly detection is an important problem that has been researched within diverse research areas and application domains.Many anomaly detection techniques have been specifically developed for certain application domains,while others are more generic.This survey tries to provide a structured and comprehensive overview of the research on anomaly detection.We have grouped existing techniques into different categories based on the underlying approach adopted by each technique.For each category we have identified key assumptions,which are used by the techniques to differentiate between normal and anomalous behavior.When applying a given technique to a particular domain,these assumptions can be used as guidelines to assess the effectiveness of the technique in that domain.For each category,we provide a basic anomaly detection technique,and then show how the different existing techniques in that category are variants of the basic tech-nique.This template provides an easier and succinct understanding of the techniques belonging to each category.Further,for each category,we identify the advantages and disadvantages of the techniques in that category.We also provide a discussion on the computational complexity of the techniques since it is an important issue in real application domains.We hope that this survey will provide a better understanding of the different directions in which research has been done on this topic,and how techniques developed in one area can be applied in domains for which they were not intended to begin with.Categories and Subject Descriptors:H.2.8[Database Management]:Database Applications—Data MiningGeneral Terms:AlgorithmsAdditional Key Words and Phrases:Anomaly Detection,Outlier Detection1.INTRODUCTIONAnomaly detection refers to the problem offinding patterns in data that do not conform to expected behavior.These non-conforming patterns are often referred to as anomalies,outliers,discordant observations,exceptions,aberrations,surprises, peculiarities or contaminants in different application domains.Of these,anomalies and outliers are two terms used most commonly in the context of anomaly detection; sometimes interchangeably.Anomaly detectionfinds extensive use in a wide variety of applications such as fraud detection for credit cards,insurance or health care, intrusion detection for cyber-security,fault detection in safety critical systems,and military surveillance for enemy activities.The importance of anomaly detection is due to the fact that anomalies in data translate to significant(and often critical)actionable information in a wide variety of application domains.For example,an anomalous traffic pattern in a computerTo Appear in ACM Computing Surveys,092009,Pages1–72.2·Chandola,Banerjee and Kumarnetwork could mean that a hacked computer is sending out sensitive data to an unauthorized destination[Kumar2005].An anomalous MRI image may indicate presence of malignant tumors[Spence et al.2001].Anomalies in credit card trans-action data could indicate credit card or identity theft[Aleskerov et al.1997]or anomalous readings from a space craft sensor could signify a fault in some compo-nent of the space craft[Fujimaki et al.2005].Detecting outliers or anomalies in data has been studied in the statistics commu-nity as early as the19th century[Edgeworth1887].Over time,a variety of anomaly detection techniques have been developed in several research communities.Many of these techniques have been specifically developed for certain application domains, while others are more generic.This survey tries to provide a structured and comprehensive overview of the research on anomaly detection.We hope that it facilitates a better understanding of the different directions in which research has been done on this topic,and how techniques developed in one area can be applied in domains for which they were not intended to begin with.1.1What are anomalies?Anomalies are patterns in data that do not conform to a well defined notion of normal behavior.Figure1illustrates anomalies in a simple2-dimensional data set. The data has two normal regions,N1and N2,since most observations lie in these two regions.Points that are sufficiently far away from the regions,e.g.,points o1 and o2,and points in region O3,are anomalies.Fig.1.A simple example of anomalies in a2-dimensional data set. Anomalies might be induced in the data for a variety of reasons,such as malicious activity,e.g.,credit card fraud,cyber-intrusion,terrorist activity or breakdown of a system,but all of the reasons have a common characteristic that they are interesting to the analyst.The“interestingness”or real life relevance of anomalies is a key feature of anomaly detection.Anomaly detection is related to,but distinct from noise removal[Teng et al. 1990]and noise accommodation[Rousseeuw and Leroy1987],both of which deal To Appear in ACM Computing Surveys,092009.Anomaly Detection:A Survey·3 with unwanted noise in the data.Noise can be defined as a phenomenon in data which is not of interest to the analyst,but acts as a hindrance to data analysis. Noise removal is driven by the need to remove the unwanted objects before any data analysis is performed on the data.Noise accommodation refers to immunizing a statistical model estimation against anomalous observations[Huber1974]. Another topic related to anomaly detection is novelty detection[Markou and Singh2003a;2003b;Saunders and Gero2000]which aims at detecting previously unobserved(emergent,novel)patterns in the data,e.g.,a new topic of discussion in a news group.The distinction between novel patterns and anomalies is that the novel patterns are typically incorporated into the normal model after being detected.It should be noted that solutions for above mentioned related problems are often used for anomaly detection and vice-versa,and hence are discussed in this review as well.1.2ChallengesAt an abstract level,an anomaly is defined as a pattern that does not conform to expected normal behavior.A straightforward anomaly detection approach,there-fore,is to define a region representing normal behavior and declare any observation in the data which does not belong to this normal region as an anomaly.But several factors make this apparently simple approach very challenging:—Defining a normal region which encompasses every possible normal behavior is very difficult.In addition,the boundary between normal and anomalous behavior is often not precise.Thus an anomalous observation which lies close to the boundary can actually be normal,and vice-versa.—When anomalies are the result of malicious actions,the malicious adversaries often adapt themselves to make the anomalous observations appear like normal, thereby making the task of defining normal behavior more difficult.—In many domains normal behavior keeps evolving and a current notion of normal behavior might not be sufficiently representative in the future.—The exact notion of an anomaly is different for different application domains.For example,in the medical domain a small deviation from normal(e.g.,fluctuations in body temperature)might be an anomaly,while similar deviation in the stock market domain(e.g.,fluctuations in the value of a stock)might be considered as normal.Thus applying a technique developed in one domain to another is not straightforward.—Availability of labeled data for training/validation of models used by anomaly detection techniques is usually a major issue.—Often the data contains noise which tends to be similar to the actual anomalies and hence is difficult to distinguish and remove.Due to the above challenges,the anomaly detection problem,in its most general form,is not easy to solve.In fact,most of the existing anomaly detection techniques solve a specific formulation of the problem.The formulation is induced by various factors such as nature of the data,availability of labeled data,type of anomalies to be detected,etc.Often,these factors are determined by the application domain inTo Appear in ACM Computing Surveys,092009.4·Chandola,Banerjee and Kumarwhich the anomalies need to be detected.Researchers have adopted concepts from diverse disciplines such as statistics ,machine learning ,data mining ,information theory ,spectral theory ,and have applied them to specific problem formulations.Figure 2shows the above mentioned key components associated with any anomaly detection technique.Anomaly DetectionTechniqueApplication DomainsMedical InformaticsIntrusion Detection...Fault/Damage DetectionFraud DetectionResearch AreasInformation TheoryMachine LearningSpectral TheoryStatisticsData Mining...Problem CharacteristicsLabels Anomaly Type Nature of Data OutputFig.2.Key components associated with an anomaly detection technique.1.3Related WorkAnomaly detection has been the topic of a number of surveys and review articles,as well as books.Hodge and Austin [2004]provide an extensive survey of anomaly detection techniques developed in machine learning and statistical domains.A broad review of anomaly detection techniques for numeric as well as symbolic data is presented by Agyemang et al.[2006].An extensive review of novelty detection techniques using neural networks and statistical approaches has been presented in Markou and Singh [2003a]and Markou and Singh [2003b],respectively.Patcha and Park [2007]and Snyder [2001]present a survey of anomaly detection techniques To Appear in ACM Computing Surveys,092009.Anomaly Detection:A Survey·5 used specifically for cyber-intrusion detection.A substantial amount of research on outlier detection has been done in statistics and has been reviewed in several books [Rousseeuw and Leroy1987;Barnett and Lewis1994;Hawkins1980]as well as other survey articles[Beckman and Cook1983;Bakar et al.2006].Table I shows the set of techniques and application domains covered by our survey and the various related survey articles mentioned above.12345678TechniquesClassification Based√√√√√Clustering Based√√√√Nearest Neighbor Based√√√√√Statistical√√√√√√√Information Theoretic√Spectral√ApplicationsCyber-Intrusion Detection√√Fraud Detection√Medical Anomaly Detection√Industrial Damage Detection√Image Processing√Textual Anomaly Detection√Sensor Networks√Table parison of our survey to other related survey articles.1-Our survey2-Hodge and Austin[2004],3-Agyemang et al.[2006],4-Markou and Singh[2003a],5-Markou and Singh [2003b],6-Patcha and Park[2007],7-Beckman and Cook[1983],8-Bakar et al[2006]1.4Our ContributionsThis survey is an attempt to provide a structured and a broad overview of extensive research on anomaly detection techniques spanning multiple research areas and application domains.Most of the existing surveys on anomaly detection either focus on a particular application domain or on a single research area.[Agyemang et al.2006]and[Hodge and Austin2004]are two related works that group anomaly detection into multiple categories and discuss techniques under each category.This survey builds upon these two works by significantly expanding the discussion in several directions. We add two more categories of anomaly detection techniques,viz.,information theoretic and spectral techniques,to the four categories discussed in[Agyemang et al.2006]and[Hodge and Austin2004].For each of the six categories,we not only discuss the techniques,but also identify unique assumptions regarding the nature of anomalies made by the techniques in that category.These assumptions are critical for determining when the techniques in that category would be able to detect anomalies,and when they would fail.For each category,we provide a basic anomaly detection technique,and then show how the different existing techniques in that category are variants of the basic technique.This template provides an easier and succinct understanding of the techniques belonging to each category.Further, for each category we identify the advantages and disadvantages of the techniques in that category.We also provide a discussion on the computational complexity of the techniques since it is an important issue in real application domains.To Appear in ACM Computing Surveys,092009.6·Chandola,Banerjee and KumarWhile some of the existing surveys mention the different applications of anomaly detection,we provide a detailed discussion of the application domains where anomaly detection techniques have been used.For each domain we discuss the notion of an anomaly,the different aspects of the anomaly detection problem,and the challenges faced by the anomaly detection techniques.We also provide a list of techniques that have been applied in each application domain.The existing surveys discuss anomaly detection techniques that detect the sim-plest form of anomalies.We distinguish the simple anomalies from complex anoma-lies.The discussion of applications of anomaly detection reveals that for most ap-plication domains,the interesting anomalies are complex in nature,while most of the algorithmic research has focussed on simple anomalies.1.5OrganizationThis survey is organized into three parts and its structure closely follows Figure 2.In Section2we identify the various aspects that determine the formulation of the problem and highlight the richness and complexity associated with anomaly detection.We distinguish simple anomalies from complex anomalies and define two types of complex anomalies,viz.,contextual and collective anomalies.In Section 3we briefly describe the different application domains where anomaly detection has been applied.In subsequent sections we provide a categorization of anomaly detection techniques based on the research area which they belong to.Majority of the techniques can be categorized into classification based(Section4),nearest neighbor based(Section5),clustering based(Section6),and statistical techniques (Section7).Some techniques belong to research areas such as information theory (Section8),and spectral theory(Section9).For each category of techniques we also discuss their computational complexity for training and testing phases.In Section 10we discuss various contextual anomaly detection techniques.We discuss various collective anomaly detection techniques in Section11.We present some discussion on the limitations and relative performance of various existing techniques in Section 12.Section13contains concluding remarks.2.DIFFERENT ASPECTS OF AN ANOMALY DETECTION PROBLEMThis section identifies and discusses the different aspects of anomaly detection.As mentioned earlier,a specific formulation of the problem is determined by several different factors such as the nature of the input data,the availability(or unavailabil-ity)of labels as well as the constraints and requirements induced by the application domain.This section brings forth the richness in the problem domain and justifies the need for the broad spectrum of anomaly detection techniques.2.1Nature of Input DataA key aspect of any anomaly detection technique is the nature of the input data. Input is generally a collection of data instances(also referred as object,record,point, vector,pattern,event,case,sample,observation,entity)[Tan et al.2005,Chapter 2].Each data instance can be described using a set of attributes(also referred to as variable,characteristic,feature,field,dimension).The attributes can be of different types such as binary,categorical or continuous.Each data instance might consist of only one attribute(univariate)or multiple attributes(multivariate).In To Appear in ACM Computing Surveys,092009.Anomaly Detection:A Survey·7 the case of multivariate data instances,all attributes might be of same type or might be a mixture of different data types.The nature of attributes determine the applicability of anomaly detection tech-niques.For example,for statistical techniques different statistical models have to be used for continuous and categorical data.Similarly,for nearest neighbor based techniques,the nature of attributes would determine the distance measure to be used.Often,instead of the actual data,the pairwise distance between instances might be provided in the form of a distance(or similarity)matrix.In such cases, techniques that require original data instances are not applicable,e.g.,many sta-tistical and classification based techniques.Input data can also be categorized based on the relationship present among data instances[Tan et al.2005].Most of the existing anomaly detection techniques deal with record data(or point data),in which no relationship is assumed among the data instances.In general,data instances can be related to each other.Some examples are sequence data,spatial data,and graph data.In sequence data,the data instances are linearly ordered,e.g.,time-series data,genome sequences,protein sequences.In spatial data,each data instance is related to its neighboring instances,e.g.,vehicular traffic data,ecological data.When the spatial data has a temporal(sequential) component it is referred to as spatio-temporal data,e.g.,climate data.In graph data,data instances are represented as vertices in a graph and are connected to other vertices with ter in this section we will discuss situations where such relationship among data instances become relevant for anomaly detection. 2.2Type of AnomalyAn important aspect of an anomaly detection technique is the nature of the desired anomaly.Anomalies can be classified into following three categories:2.2.1Point Anomalies.If an individual data instance can be considered as anomalous with respect to the rest of data,then the instance is termed as a point anomaly.This is the simplest type of anomaly and is the focus of majority of research on anomaly detection.For example,in Figure1,points o1and o2as well as points in region O3lie outside the boundary of the normal regions,and hence are point anomalies since they are different from normal data points.As a real life example,consider credit card fraud detection.Let the data set correspond to an individual’s credit card transactions.For the sake of simplicity, let us assume that the data is defined using only one feature:amount spent.A transaction for which the amount spent is very high compared to the normal range of expenditure for that person will be a point anomaly.2.2.2Contextual Anomalies.If a data instance is anomalous in a specific con-text(but not otherwise),then it is termed as a contextual anomaly(also referred to as conditional anomaly[Song et al.2007]).The notion of a context is induced by the structure in the data set and has to be specified as a part of the problem formulation.Each data instance is defined using following two sets of attributes:To Appear in ACM Computing Surveys,092009.8·Chandola,Banerjee and Kumar(1)Contextual attributes.The contextual attributes are used to determine thecontext(or neighborhood)for that instance.For example,in spatial data sets, the longitude and latitude of a location are the contextual attributes.In time-series data,time is a contextual attribute which determines the position of an instance on the entire sequence.(2)Behavioral attributes.The behavioral attributes define the non-contextual char-acteristics of an instance.For example,in a spatial data set describing the average rainfall of the entire world,the amount of rainfall at any location is a behavioral attribute.The anomalous behavior is determined using the values for the behavioral attributes within a specific context.A data instance might be a contextual anomaly in a given context,but an identical data instance(in terms of behavioral attributes)could be considered normal in a different context.This property is key in identifying contextual and behavioral attributes for a contextual anomaly detection technique.TimeFig.3.Contextual anomaly t2in a temperature time series.Note that the temperature at time t1is same as that at time t2but occurs in a different context and hence is not considered as an anomaly.Contextual anomalies have been most commonly explored in time-series data [Weigend et al.1995;Salvador and Chan2003]and spatial data[Kou et al.2006; Shekhar et al.2001].Figure3shows one such example for a temperature time series which shows the monthly temperature of an area over last few years.A temperature of35F might be normal during the winter(at time t1)at that place,but the same value during summer(at time t2)would be an anomaly.A similar example can be found in the credit card fraud detection domain.A contextual attribute in credit card domain can be the time of purchase.Suppose an individual usually has a weekly shopping bill of$100except during the Christmas week,when it reaches$1000.A new purchase of$1000in a week in July will be considered a contextual anomaly,since it does not conform to the normal behavior of the individual in the context of time(even though the same amount spent during Christmas week will be considered normal).The choice of applying a contextual anomaly detection technique is determined by the meaningfulness of the contextual anomalies in the target application domain. To Appear in ACM Computing Surveys,092009.Anomaly Detection:A Survey·9 Another key factor is the availability of contextual attributes.In several cases defining a context is straightforward,and hence applying a contextual anomaly detection technique makes sense.In other cases,defining a context is not easy, making it difficult to apply such techniques.2.2.3Collective Anomalies.If a collection of related data instances is anomalous with respect to the entire data set,it is termed as a collective anomaly.The indi-vidual data instances in a collective anomaly may not be anomalies by themselves, but their occurrence together as a collection is anomalous.Figure4illustrates an example which shows a human electrocardiogram output[Goldberger et al.2000]. The highlighted region denotes an anomaly because the same low value exists for an abnormally long time(corresponding to an Atrial Premature Contraction).Note that that low value by itself is not an anomaly.Fig.4.Collective anomaly corresponding to an Atrial Premature Contraction in an human elec-trocardiogram output.As an another illustrative example,consider a sequence of actions occurring in a computer as shown below:...http-web,buffer-overflow,http-web,http-web,smtp-mail,ftp,http-web,ssh,smtp-mail,http-web,ssh,buffer-overflow,ftp,http-web,ftp,smtp-mail,http-web...The highlighted sequence of events(buffer-overflow,ssh,ftp)correspond to a typical web based attack by a remote machine followed by copying of data from the host computer to remote destination via ftp.It should be noted that this collection of events is an anomaly but the individual events are not anomalies when they occur in other locations in the sequence.Collective anomalies have been explored for sequence data[Forrest et al.1999; Sun et al.2006],graph data[Noble and Cook2003],and spatial data[Shekhar et al. 2001].To Appear in ACM Computing Surveys,092009.10·Chandola,Banerjee and KumarIt should be noted that while point anomalies can occur in any data set,collective anomalies can occur only in data sets in which data instances are related.In contrast,occurrence of contextual anomalies depends on the availability of context attributes in the data.A point anomaly or a collective anomaly can also be a contextual anomaly if analyzed with respect to a context.Thus a point anomaly detection problem or collective anomaly detection problem can be transformed toa contextual anomaly detection problem by incorporating the context information.2.3Data LabelsThe labels associated with a data instance denote if that instance is normal or anomalous1.It should be noted that obtaining labeled data which is accurate as well as representative of all types of behaviors,is often prohibitively expensive. Labeling is often done manually by a human expert and hence requires substantial effort to obtain the labeled training data set.Typically,getting a labeled set of anomalous data instances which cover all possible type of anomalous behavior is more difficult than getting labels for normal behavior.Moreover,the anomalous behavior is often dynamic in nature,e.g.,new types of anomalies might arise,for which there is no labeled training data.In certain cases,such as air traffic safety, anomalous instances would translate to catastrophic events,and hence will be very rare.Based on the extent to which the labels are available,anomaly detection tech-niques can operate in one of the following three modes:2.3.1Supervised anomaly detection.Techniques trained in supervised mode as-sume the availability of a training data set which has labeled instances for normal as well as anomaly class.Typical approach in such cases is to build a predictive model for normal vs.anomaly classes.Any unseen data instance is compared against the model to determine which class it belongs to.There are two major is-sues that arise in supervised anomaly detection.First,the anomalous instances are far fewer compared to the normal instances in the training data.Issues that arise due to imbalanced class distributions have been addressed in the data mining and machine learning literature[Joshi et al.2001;2002;Chawla et al.2004;Phua et al. 2004;Weiss and Hirsh1998;Vilalta and Ma2002].Second,obtaining accurate and representative labels,especially for the anomaly class is usually challenging.A number of techniques have been proposed that inject artificial anomalies in a normal data set to obtain a labeled training data set[Theiler and Cai2003;Abe et al.2006;Steinwart et al.2005].Other than these two issues,the supervised anomaly detection problem is similar to building predictive models.Hence we will not address this category of techniques in this survey.2.3.2Semi-Supervised anomaly detection.Techniques that operate in a semi-supervised mode,assume that the training data has labeled instances for only the normal class.Since they do not require labels for the anomaly class,they are more widely applicable than supervised techniques.For example,in space craft fault detection[Fujimaki et al.2005],an anomaly scenario would signify an accident, which is not easy to model.The typical approach used in such techniques is to 1Also referred to as normal and anomalous classes.To Appear in ACM Computing Surveys,092009.Anomaly Detection:A Survey·11 build a model for the class corresponding to normal behavior,and use the model to identify anomalies in the test data.A limited set of anomaly detection techniques exist that assume availability of only the anomaly instances for training[Dasgupta and Nino2000;Dasgupta and Majumdar2002;Forrest et al.1996].Such techniques are not commonly used, primarily because it is difficult to obtain a training data set which covers every possible anomalous behavior that can occur in the data.2.3.3Unsupervised anomaly detection.Techniques that operate in unsupervised mode do not require training data,and thus are most widely applicable.The techniques in this category make the implicit assumption that normal instances are far more frequent than anomalies in the test data.If this assumption is not true then such techniques suffer from high false alarm rate.Many semi-supervised techniques can be adapted to operate in an unsupervised mode by using a sample of the unlabeled data set as training data.Such adaptation assumes that the test data contains very few anomalies and the model learnt during training is robust to these few anomalies.2.4Output of Anomaly DetectionAn important aspect for any anomaly detection technique is the manner in which the anomalies are reported.Typically,the outputs produced by anomaly detection techniques are one of the following two types:2.4.1Scores.Scoring techniques assign an anomaly score to each instance in the test data depending on the degree to which that instance is considered an anomaly. Thus the output of such techniques is a ranked list of anomalies.An analyst may choose to either analyze top few anomalies or use a cut-offthreshold to select the anomalies.2.4.2Labels.Techniques in this category assign a label(normal or anomalous) to each test instance.Scoring based anomaly detection techniques allow the analyst to use a domain-specific threshold to select the most relevant anomalies.Techniques that provide binary labels to the test instances do not directly allow the analysts to make such a choice,though this can be controlled indirectly through parameter choices within each technique.3.APPLICATIONS OF ANOMALY DETECTIONIn this section we discuss several applications of anomaly detection.For each ap-plication domain we discuss the following four aspects:—The notion of anomaly.—Nature of the data.—Challenges associated with detecting anomalies.—Existing anomaly detection techniques.To Appear in ACM Computing Surveys,092009.。
非参数统计实验报告 南邮概要

非参数统计实验报告南邮概要英文回答:Nonparametric Statistical Inference Report。
Introduction。
Nonparametric statistical inference is a branch of statistics that makes no assumptions about the distribution of the population from which a sample is drawn. This makes nonparametric methods less sensitive to outliers and other deviations from normality than parametric methods.Methods。
There are a variety of nonparametric statistical tests, each of which is designed to test a specific hypothesis. Some of the most common nonparametric tests include:The chi-square test for independence。
The Mann-Whitney U test for two independent samples。
The Kruskal-Wallis test for multiple independent samples。
The Friedman test for multiple dependent samples。
Results。
The results of a nonparametric statistical test are typically reported in terms of a p-value. The p-value is the probability of obtaining a test statistic as extreme as or more extreme than the one that was observed, assuming that the null hypothesis is true.Conclusion。
9-非参数检验

3 j
tj)
N N
3
【例】某医生分别测定了10名正常人、单纯性肥胖 和皮质醇增多症患者血浆中总皮质醇的含量见下表。 问三组人的血浆总皮质醇含量有无差别
三组人的血浆总皮质醇测定值(g/L)
正常人 测定值 0.4 1.9 2.2 2.5 2.8 3.1 3.7 3.9 4.6 7.0 秩 次 1 4 6 8 9 10.5 12 13 15 18 单纯性肥胖 测定值 0.6 1.2 2.0 2.4 3.1 4.1 5.0 5.9 7.4 13.6 秩 次 2 3 5 7 10.5 14 16 17 19 24 皮质醇增多症 测定值 9.8 10.2 10.6 13.0 14.0 14.8 15.6 15.6 21.6 24.0 秩 次 20 21 22 23 25 26 27 28 29 30
1
【例】测得铅作业与非铅作业工人的血铅值 (mol/L)如表第(1)、(3)栏,问两组工人 的血铅值有无差别?
两组工人血铅值的秩和检验
非铅作业组 (1) 0.24 0.24 0.29 0.34 0.43 0.58 0.62 秩次 (2) 1 2 3 4 5 6 7 铅作业组 (3) 0.82 0.86 0.96 1.20 1.63 2.06 2.11 n1=7 秩次 (4) 9 10.5 12 14 15 16 17 T1=93.5
检验步骤(2)
三组人的血浆总皮质醇测定值(g/L) 正常人 测定值 秩 次 0.4 1 1.9 4 2.2 6 2.5 8 2.8 9 … … Ri 96.5 ni 10 单纯性肥胖 测定值 秩 次 0.6 2 1.2 3 2.0 5 2.4 7 3.1 10.5 … … 117.5 10 皮质醇增多症 测定值 秩 次 9.8 20 10.2 21 10.6 22 13.0 23 14.0 25 … … 251 10
第十二章 非参数检验(Nonparametric test)

(2)求差值、编秩、求秩和并确 定检验统计量:
编秩: 按绝对值大小 差值为0舍去不计 秩次相等取平均秩次
T+=98,T-=22 任取其中之一作为检验的统计量T 本例取T= T- =22。
(3)确定P值并作出推断结论:
根据T值( T+=98 或 T-=22 )查T界值表 ( P208附表12-1 )确定P值
A、用t检验 B、用u检验 C、用Wilcoxon秩和检验 D、用t检验或Wilcoxon秩和检验均可 E、资料符合t检验还是Wilcoxon秩和 检验
2、配对样本差值的Wilcoxon符号秩检 验,确定P值的方法为: A、T越大,P越大 B、T越大,P越小 C、T值在界值范围内,P小于相应的α D、T值在界值范围内,P大于相应的α E、T值即u值,查u界值表
大样本情况:若k > 3或ni > 5时,理论上, H近似服从自由度为k-1的χ2分布,可查 χ2界值表确定P值。
秩和检验的两两比较
方法有: 1、扩展的t检验 2、Nemenyi法检验 3、q检验
几种方法理论上仍存在争议,故SAS、 SPSS等软件没有提供这方面的分析
第四节 配伍组设计的秩和检验
正态近似法
n>50时,T分布近似正态分布可用正 态近似法作u检验:
u T T | T n(n 1) / 4 | 0.5
非参数统计(non-parametricstatistics)又称任意分布检验(.

2
0.05(2)
=5.99
P 0.05
按=0.05水准,拒绝H 0,接受H1,可认为小白鼠接 种三种不同菌型伤寒杆 菌后存活日数有差别。
四、等级资料的比较
适用范围:完全随机设计分组的两个、以及两个以 上样本等级程度比较,目的在于判断两个以及多个总体 分布是否相同。
注意:等级资料对程度的比较不应选检验。
;
T
在上下界值范围外时,则 P 。
n 9
T 的界值范围是5-40 0.05
P 0.05
按=0.05水准,不拒绝 H 0,故不能认为两法测定 空气中 CS 2的含量有差别。
2、正态近似法
当对子数n 50时,计算统计量 u值。
T n(n 1) / 4 0.5 u n(n 1)(2n 1) / 24
2
0.05(2)
=5.99
P 0.05
按=0.05水准,拒绝H 0,接受H1,可认为三组病人 血浆总皮质醇含量有差别(不同或不全同)。
若还希望分析具体哪些组之间有差别,需进一步两两组 间比较。方法见《卫生统计学》第五版P196,《医学统计学》 第二版P183等。
当相同秩次较多(超过25%)时,需进行如下校正。
H 0:血浆总皮质醇含量的三个总体分布相同 H1:血浆总皮质醇含量的三个总体分布不同或不全同 0.05
(二)计算统计量H值 1、编秩
先将各组数据分别由小到大排列,统一编秩,不同组的
相同数据取平均秩次。 2、求各组秩和 R
i 本例 R1=96.5 R2= 117.5 R3=251 3、计算统计量 H 值 2 n 为各组例数 R i 12 i H ( ) 3( N 1) N n N ( N 1) n i i 12 96.52 117.52 2512 H ( ) 3(30 1) 18.12 30(301) 10 10 10
2020托福物理学专业词汇:非线性Nonlinear

2020托福物理学专业词汇:非线性Nonlinear托福物理学学科分类词汇:非线性Nonlinear非线性,Nonlinear英语短句,例句大全非线性,Nonlinear1)Nonlinear[英]['N?N'lini?][美][Nɑn'l?Ni?]非线性1.Dynamic Tuning Of 1-D Nonlinear Photonic Crystals;非线性一维光子晶体特性的动态调制2.Research Of Nonlinear Simulation On The Machine Tool Control System With Matlab;基于Matlab机床控制系统非线性的仿真研究3.Optical Switch And Bistability Based On Nonlinear One-Dimensional Photonic Crystals;非线性一维光子晶体光开关与光双稳英文短句/例句1.Linearization Of Nonlinear System非线性系统的线性化2.Non-Linear Processing(图象)非线性处理3.Non-Linear Non-Ideal Chromatography非线性非理想色谱法4.Nonlinear Lagrange Methods For Solving Nonlinear Optimization Problems;求解非线性优化问题的非线性Lagrange方法5.Some Discussion For Nonlinear Operators And Nonlinear Equations;非线性算子及非线性方程的若干讨论6.Research On Thermal Optical Nonlinearity Induced By Nonlinear Absorption;非线性吸收诱导的热光学非线性研究7.Nonlinear Aerodynamic Characteristic非线性空气动力特性8.Nonlinear Dynamics Of Elastic Bodies弹性体非线性动力学9.Nonlinear Taper非线性电阻分布特性10.Study On The Nonlinear Characteristics Of Josephson JunctionJosephson结的非线性特性研究11.This System Is Of A Small Nonlinearity ContainingStiff Nonlinear And Damper Nonlinear One.它是含有刚度非线性和阻尼非线性的弱非线性系统。
Non-myopic attribute estimation in regression

Parametric and Non-Parametric

8 10 z 1 2 8 10 z 1 2
xx z s
• On the tables for z =1 we get 0.3413. Then z = = 0.6826. This means that the probability of X assuming a value between 8 and 12 or P (8 < X < 12) IS 68.26%
Distributions
Normal Probability Distribution
• Characteristics of the normal probability distribution:
• Bell-shaped, single peak at the centre of the distribution. • The arithmetic mean, median and mode are equal and located in the centre of the distribution. Thus half the area under the normal curve is to the right of this centre point and the other half to the left of it. • It is symmetrical about the mean. • The distribution is asymptotic. The curve gets closer and closer to the X-axis but never actually touches it.
– subtract the mean of the data set from each observation – Divide each by the standard deviation
非参数检验

n
n
利用秩的大小进行推断就避免了不知道背景分布 的困难。这也是大多数非参数检验的优点。 多数非参数检验明显地或隐含地利用了秩的性质; 但也有一些非参数方法没有涉及秩的性质。 常用的非参数检验的方法有:单样本检验、两独 立样本检验、多个独立样本检验、多个相关样本 检验和列联表某一变量各水平比例检验
H0成立,秩和统计量w随机出现在n1*(N+1)/2两 侧附近并且在T=n1(N+1)/2的地方呈对称分布, 在大多数情况下,T与n1 *(N+1)/2的差值较小 (纯属抽样误差),并且当n1和n2都较大时,T近 似服从均数为n1(N+1)/2,方差为 n1n2 ( N + 1) /12 的正态分布。 若H0非真时,大多数情况下,统计量T远离 n1(N+1)/2处并呈偏态分布。因此在H0成立的情 况下T远离它的期望值n1(N+1)/2为小概率事件, 可认为在一次抽样中是不会发生的,故当出现 这种情况时推断拒绝H0。
参数统计
(parametric statistics)
非参数统计
(nonparametric statistics)
如何判别数据分布类型
n
均数、中位数两者关系
n n n
已知总体分布类型,对 未知参数(μ、π)进行 统计推断 依赖于特定分布类 型,比较的是参数
对总体的分布类 型不作任何要求 不受总体参数的影响, 比较分布或分布位置 适用范围广;可用于任何类型 资料(等级资料,或 “>50mg >50mg” )
ti为第i个相同秩号的数据个数
(3)求秩和并确定检验统计量: 分别将试验组和对照组的秩次累加求和,得 TX=145,TY=180。设较小样本的样本例数 为n1,较大样本的样本例数为n2。取小样本 的秩和作为检验统计量。若n1=n2,可任取 一组的秩和作为统计量。本例n1=15,n2= 10,因此检验统计量T=TY=180。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
for the function f in the nonlinear regression model (1.2) Zt = f (Xt ) + Wt ,
where K is a kernel function whose definition and properties are given in Section 2.1, h is the bandwidth, {Wt } is an unobserved stationary process and {Xt } and {Zt } are observed processes which are nonstationary in a sense to be made precise later. At first, {Xt } and {Wt } will be assumed to be independent processes, which is quite a natural assumption in a nonlinear regression context. However, in a cointegration framework, this independence assumption is rather restrictive and is generally not fulfilled for linear cointegration models. In Section 4, dependence is the main subject. It turns out that dependence between {Xt } and {Wt } for fixed t may disappear asymptotically. The reason for this phenomenon is related to restrictions on the type of dependence which is possible between a stationary and a nonstationary process. A stationary process cannot follow a nonstationary process too closely as this will violate the stationarity. Although the connection between (1.2) and the nonlinear cointegration problem is obvious, we would like to point out that the estimation of the function f in the general context we are considering should also be of interest in other areas of application. In a traditional time series regression problem, some sort of mixing condition is often assumed for {Xt } in order to obtain a central limit theorem for f (x). However, mixing assumptions on {Xt } are ruled out in the general situation we consider. A minimal condition for undertaking asymptotic analysis on f (x) is that as the number of observations on {Xt } increases, there must be infinitely many observations in any neighborhood of x. This means that {Xt } must return to a neighborhood of x infinitely often, which, in turn, implies that the framework
The Annals of Statistics 2007, Vol. 35, No. 1, 252–299 DOI: 10.1214/009053606000001181 c Institute of Mathematical Statistics, 2007
arXiv:0708.0503v1 [math.ST] 3 Aug 2007
We derive an asymptotic theory of nonparametric estimation for a time series regression model Zt = f (Xt ) + Wt , where {Xt } and {Zt } are observed nonstationary processes and {Wt } is an unobserved stationary process. In econometrics, this can be interpreted as a nonlinear cointegration type relationship, but we believe that our results are of wider interest. The class of nonstationary processes allowed for {Xt } is a subclass of the class of null recurrent Markov chains. This subclass contains random walk, unit root processes and nonlinear processes. We derive the asymptotics of a nonparametric estimate of f (x) under the assumption that {Wt } is a Markov chain satisfying some mixing conditions. The finite-sample properties of f (x) are studied by means of simulation experiments.
1. Introduction. Two time series {Xt } and {Zt } are said to be linearly cointegrated if they are both nonstationary and of unit root type and if there exists a linear combination aXt + bZt = Wt such that {Wt } is stationary. This means that the series {Xt , Zt } move together when considered over a long period of time. The concept of cointegration was introduced by Granger [10] and further developed by Engle and Granger [6]. Since its introduction, there have been numerous papers in econometrics exploring its various aspects. Some of the main results are given in Johansen [19]. The long term relationships between two economic time series may not necessarily be linear, however, and the processes {Xt } and {Zt } may not be linearly generated unit root processes. This has led to a search for nonlinear cointegration type relationships such as Zt = f (Xt ) + Wt , for some
2
H. A. KARLSEN, T. MYKLEBUST AND D. TJØSTHEIM
nonlinear function f and some possibly nonlinearly generated input process {Xt }. Indeed, functional relationships of this type have been fitted to economic data (see, e.g., [8, 12]), but to our knowledge, the properties of the resulting nonparametric estimates have not been established (see [27] for a consistency property in a simplified situation, though). A brief discussion of the relationship between our work and recent contributions to the theory of nonlinear cointegration occurs in Section 6. There are at least two difficulties (cf. [11] and others): which class of processes should be chosen as a basic class of nonstationary processes and how should an estimation theory for an estimate of f be constructed? The main goal of this paper is to try to answer these questions, that is, we wish to establish a nonparametric estimation theory of the kernel estimator (1.1) f (x) =