Belief functions

合集下载

《人工智能》测试题答案

《人工智能》测试题答案

《人工智能》测试题答案测试题——人工智能原理一、填空题1.人工智能作为一门学科,它研究的对象是______,而研究的近期目标是____________ _______;远期目标是___________________。

2.人工智能应用的主要领域有_________,_________,_________,_________,_______和__________。

3.知识表示的方法主要有_________,_________,_________,_________和________。

4.产生式系统由三个部分所组成,即___________,___________和___________。

5.用归结反演方法进行定理证明时,可采取的归结策略有___________、___________、_________、_________、_________和_________。

6.宽度优先搜索对应的数据结构是___________________;深度优先搜索是________________。

7.不确定知识处理的基本方法有__________、__________、__________和__________。

8.AI研究的主要途径有三大学派,它们是________学派、________学派和________学派。

9.专家系统的瓶颈是________________________;它来自于两个阶段,第一阶段是,第二阶段是。

10.确定因子法中函数MB是描述________________________、而函数MD是描述________________________。

11.人工智能研究的主要领域有_________、_________、_________、_________、_______和__________。

12.一阶谓词逻辑可以使用的连接词有______、_______、_______和_______。

believe的宾语从句

believe的宾语从句

believe的宾语从句I. IntroductionBelieve is a verb that expresses one's conviction or trust in something or someone. It is a word that holds great power and significance in shaping our thoughts, actions, and beliefs. As human beings, we constantly encounter situations that challenge our beliefs and force us to evaluate the basis upon which we hold them. The use of the verb "believe" is often accompanied by an object or a complement in the form of a noun clause, also known as an indirect statement or a reported speech clause. This essay aims to explore the various types and functions of belief as an object complement and how it impacts our understanding of reality and influences our decision-making processes.II. Types of Belief1. Direct Belief PlacementThe most straightforward type of belief is one where the direct object of "believe" is a statement or a proposition. For example, "I believe that the Earth is round" or "She believes he will win the race." In these cases, the belief is centered around a particular fact or assertion. The object clause serves as a direct complement to the verb "believe" and contains the belief itself.2. Indirect Belief PlacementAnother type of belief is one where the belief is reported or indirectly stated. This type of construction often involves verbslike "think," "claim," or "suggest." For instance, "He believes that she is a skilled pianist" or "They claim that they have found a cure for cancer." In these instances, the beliefs are not expressed directly by the speaker but rather reported through an indirect statement. The object clause indicates the content of the belief and may or may not align with the speaker's personal belief.III. Functions of the Belief Object Clause1. Assertion and ConfirmationThe primary function of the belief object clause is to assert or confirm a particular belief. It allows us to convey our thoughts, theories, or opinions about a specific subject matter. For instance, "I believe that recycling is essential for the environment" or "We think that technology is transforming the way we live." By using the object clause, we provide evidence or justification for our beliefs, making our assertions more convincing and credible.2. Uncertainty and DoubtThe belief object clause can also be used to express uncertainty or doubt. In these cases, the speaker may use phrases such as "I'm not sure," "I doubt," or "I can't believe." For example, "I can't believe that he cheated on his exams" or "She doubts that he will keep his promise." These statements indicate a lack of conviction or complete faith in the belief being expressed.3. Persuasion and InfluenceAnother significant function of the belief object clause is persuasion and influence. By stating our beliefs, we attempt to convince or persuade others to adopt the same viewpoint. For instance, "I believe that everyone should have access to quality education" or "They claim that their product is the best in the market." These statements aim to influence the thoughts and actions of others, leading them to align their beliefs with our own.4. Emotional Expression and EmpathyBeliefs are not solely based on logic or factual evidence but are often rooted in personal experiences and emotions. The belief object clause allows us to express our emotional connection to a particular belief and empathize with others who share the same sentiment. For example, "I believe that love conquers all" or "We feel that everyone deserves to be treated with respect." These statements convey our emotional connection and understanding of people's experiences, creating a sense of unity and connection.IV. The Impact of Belief1. Perception of RealityBeliefs play a crucial role in shaping our perception of reality. They act as filters through which we interpret and make sense of the world around us. Our beliefs influence how we perceive events, situations, and interactions. For example, if we believe that the world is a hostile place, we may interpret neutral situations as threatening. Conversely, if we believe in the inherent goodness of people, we may view their actions in a positive light. Beliefs shapeour reality by determining the significance and meaning we assign to different events and experiences.2. Decision-Making ProcessBeliefs also impact our decision-making processes. They serve as guiding principles that inform our choices and actions. Our beliefs influence the goals we set for ourselves, the risks we are willing to take, and the paths we choose to follow. For instance, if we believe that hard work leads to success, we are more likely to persevere in our pursuits. On the other hand, if we hold the belief that luck determines our fate, we may be more inclined to rely on chance.3. Social Identity and CohesionBeliefs are not just individual constructs, but they also play a significant role in shaping our social identity and group cohesion. Shared beliefs create a sense of belonging and connection among individuals, leading to the formation of communities, religious groups, and cultural identities. Beliefs serve as a common ground that unites people with similar values and aspirations. They provide individuals with a framework for understanding their place in society and their interactions with others.V. ConclusionBeliefs are deeply ingrained in human nature and play a fundamental role in our cognitive, emotional, and social lives. Whether expressed directly or indirectly, belief object clauses allow us to articulate and share our convictions, doubts, andemotions with others. The functions of these clauses vary from asserting our beliefs and confirming our convictions to expressing uncertainty and influencing others. The impact of our beliefs extends beyond our individual experiences, shaping our perception of reality, guiding our decision-making processes, and fostering social cohesion. By understanding the diverse types and functions of belief object clauses, we can navigate the complexity of human beliefs and foster meaningful connections with others.。

证据推理理论(Introduction to Evidential Reasoning)

证据推理理论(Introduction to Evidential Reasoning)

2. plausibility function pl : 2 → [0, 1] is defined as: Shafer : bel is ‘normalized’ => closed world assumption=> bel(Ω)=1, pl(Ω)=1,m(∅)
Observations in belief functions
Belief functions
Ω the frame of discernment(elements of the set Ω are called ‘worlds’) One “actual world” ω0. But which? An agent can only express the strength of his/her opinion (called belief) degree of belief that the actual world belongs to this or that subset of Ω. [0, Shafer belief function bel : 2 → [0 1] bel(A) denotes the strength of Agent’s belief that ω0∈A. bel satisfies the following inequalities:
Not supporting any strictly more specific propositions A basic belief mass given to a set A supports also that the actual world is in every subsets that contains A. The degree of belief bel(A) for A quantifies the total amount of justified specific support given to A. We say justified because we include in bel(A) only the basic belief masses given to subsets of A. m({x,y}) given to {x,y} could support x if further information indicates this.However given the available information the basic belief mass can only be given to {x,y}. We say specific because the basic belief mass m(Ø) is not included in bel(A) as it is given to the subset Ø.

2024届山东省实验中学高三下学期一模英语试题

2024届山东省实验中学高三下学期一模英语试题

绝密★启用并使用完毕前山东省实验中学2024届高三第一次模拟考试英语试题2024.04(本试卷共10页, 共三部分: 全卷满分120分, 考试用时100分钟)注意事项:1. 答卷前, 先将自己的姓名、准考证号填写在试卷和答题纸上。

2. 选择题的作答: 每小题选出答案后, 用2B铅笔把答题卡上对应题目的答案标号涂黑。

如需改动, 用橡皮擦干净后, 再选涂其他答案标号。

3. 非选择题的作答: 用0.5mm黑色签字笔直接答在答题卡上对应的答题区域内, 写在试卷、草稿纸和答题卡上的非答题区域均无效。

第一部分阅读理解(共两节, 满分50分)第一节(共15小题; 每小题2.5分, 满分37.5分)阅读下列短文, 从每题所给的A、B、C、D四个选项中选出最佳选项。

AIntroduction to Drama ExamsOur exams inspire and enable learners across the globe to be confident communicators. Exams are open to anyone looking to gain confidence and experience in speech, communication and performance. There are no age restrictions. As one of the UK's oldest and most respected drama schools and awarding organizations, we examine over 100,000candidates and deliver exams both online and in person in many countries across the globe.Now we are pleased to offer free, online "Introduction to Examinations" information session. Booking is now opening for events until Summer 2024.The 1.5-hour session will begin with an Introduction to Examinations, their history and the format of assessment. Work will then focus on the subjects available to take, and will end with a Q&A phase where participants will be invited to write in their questions to the host organizer.Ifyouhaveanyquestionsregardingthis,********************************.ukandwewillbehappytohelp. Looking forward to seeing you online at this event.1. What is an advantage of the drama exam?A. It is free of charge.B. It offers flexible schedules.C. It suits a wide range of people.D. It puts restrictions on nationality.2. What is required to register for the sessions?A. Payment in advance.B. Contact information.C. Education background.D. Performance experience.3. What should you do if you have a question during the online session?A. Email it to the drama school.B. Write it down before the session.C. Propose it at the beginning of the session.D. Send it to the host organizer in Q&A phase.BCafeterias have been filled with challenges—right from planning, purchasing, and preparing, to reducing waste, staying on budget, managing goods, and training staff. Through the tedious process, restaurateurs lacked a unified platform for efficient management. To bring consistency to the unorganised catering(餐饮)industry, childhood friends Arjun Subramanian and Raj Jain, who shared a passion for innovation, decided to partner in 2019 to explore opportunities in the cafeteria industry.In May 2020, they co-founded Platos, a one-stop solution for restaurants with a custom technology kit to streamline all aspects of cafeteria management. The company offers end-to-end cafeteria management, staff selection and food trials to ensure smooth operations and consistent service. "We believe startups solve real problems and Platos is our shot at making daily workplace food enjoyable again. We aim to simplify the dining experience, providing a convenient and efficient solution that benefits both restaurateurs and customers and creating a connected ecosystem, "says Subramanian, CEO and co-founder.Platos guarantees that a technology-driven cafeteria allows customers to order, pay, pick up, and provide ratings and feedback. It also offers goods and menu management to effectively perform daily operations. Additionally, its applications connect all shareholders for a smart cafeteria experience. "We help businesses that are into catering on condition that they have access to an industrial kitchen setup where they' re making food according to certain standards," Jain states.Since the beginning, Platos claims to have transformed 45 cafeterias across eight cities in the country. Currently, it has over 45,000 monthly users placing more than 200,000 orders. Despite facing challenges in launching cafeterias across major cities in the initial stages, Platos has experienced a 15% increase in its month-over-month profits.As for future plans, the startup is looking to raise $1 million from investors as strategic partners, bringing in capital, expertise, and networks. "Finding the right lead investor is the compass that points your startup toward success," Subramanian says.4. What does the underlined word "tedious" in Paragraph 1 mean?A. Time-consuming.B. Breath-taking.C. Heart-breaking.D. Energy-saving.5. What is the purpose of founding Platos?A. To connect customers with a greener ecosystem.B. To ensure food security and variety in cafeterias.C. To improve cafeteria management with technology.D. To make staff selection more efficient and enjoyable.6. What can we learn from the statistics in Paragraph 4?A. Platos has achieved its ultimate financial goal.B. Platos has gained impressive marketing progress.C. Challenges in food industry can be easily overcome.D. Tech-driven cafeterias have covered most urban areas.7. What is Subramanian's future plan for Platos?A. To reduce costs.B. To increase profits.C. To seek investment.D. To innovate technology.CWith a brain the size of a pinhead, insects possess a great sense of direction. They manage to locate themselves and move through small openings. How do they do this with their limited brain power? Understanding the inner workings of an insect's brain can help us in our search towards energy-efficient computing, physicist Elisabetta Chicca of the University of Groningen shows with her most recent result: a robot that acts like an insect.It's not easy to make use of the images that come in through your eyes when deciding what your feet or wings should do. A key aspect here is the apparent motion of things as you move. "Like when you're on a train,” Chicca explains. "The trees nearby appear to move faster than the houses far away." Insects use this information to infer how far away things are. This works well when moving in a straight line, but reality is not that simple. To keep things manageable for their limited brain power, they adjust their behaviour: they fly in a straight line, make a turn, then make another straight line.In search of the neural mechanism(神经机制)that drives insect behaviour, PhD student Thorben Schoepe developed a model of its neuronal activity and a small robot that uses this model to find the position. His model is based on one main principle: always head towards the area with the least apparent motion. He had his robot drive through a long passage consisting of two walls and the robot centred in the middle of the passage, as insects tend to do. In other virtual environments, such as a space with small openings, his model also showed similar behaviour to insects.The fact that a robot can find its position in a realistic environment is not new. Rather, the model gives insight into how insects do the job, and how they manage to do things so efficiently. In a similar way, you could make computers more efficient.In the future, Chicca hopes to apply this specific insect behaviour to a chip as well. "Instead of using a general-purpose computer with all its possibilities, you can build specific hardware; a tiny chip that does the job, keeping things much smaller and energy-efficient." She comments.8. Why is "a train" mentioned in Paragraph 2?A. To illustrate the principle of train motion.B. To highlight why human vision is limited.C. To explain how insects perceive distances.D. To compare the movement of trees and houses.9. What does Paragraph 3 mainly talk about concerning Schoepe's model?A. Its novel design.B. Its theoretical basis.C. Its possible application.D. Its working mechanism.10. What do the researchers think of the finding?A. Amusing.B. Discouraging.C. Promising.D. Contradictory.11. What will Chicca's follow-up study focus on?A. Inventing insect-like chips.B. Studying general-purpose robots.C. Creating insect-inspired computers.D. Developing energy-efficient hardware.DWith the help from an artificial language(AL)model, MIT neuroscientists have discovered what kind of sentences are most likely to fire up the brain's key language processing centers. The new study reveals that sentences that are more complex, because of either unusual grammar or unexpected meaning, generate stronger responses in these language processing centers. Sentences that are very straightforward barely engage these regions, and meaningless orders of words don't do much for them either.In this study, the researchers focused on language-processing regions found in the left hemisphere(半球)of the brain. By collecting a set of 1,000 sentences from various sources, the researchers measured the brain activity of participants using functional magnetic resonance imaging(fMRI)while they read the sentences. The same sentences were also fed into a large language model, similar to ChatGPT, to measure the model's activation patterns. Once the researchers had all of those data, they trained the model to predict how the human language network would respond to any new sentence based on how the artificial language network responded to these 1,000 sentences.The researchers then used the model to determine 500 new sentences that would drive highest brain activity and sentences that would make the brain less active, and their findings were confirmed in subsequent human participants. To understand why certain sentences generate stronger brain responses, the model examined the sentences based on 11 different language characteristics. The analysis revealed that sentences that were more surprising resulted in greater brain activity. Another linguistic(语言的)aspect that correlated with the brain's language network responses was the complexity of the sentences, which was determined by how well they followed English grammar rules and bow logically they linked with each other.The researchers now plan to see if they can extend these findings in speakers of languages other than English. They also hope to explore what type of stimuli may activate language processing regions in the brain's right hemisphere.12. What sentences make our brain work harder?A. Lengthy.B. Logical.C. Straightforward.D. Complicated.13. What is the function of the AL model in the research?A. To examine language network.B. To reduce language complexity.C. To locate language processing area.D. To identify language characteristics.14. How did the researchers carry out their study?A. By conducting interviews.B. By collecting questionnaires.C. By analyzing experiment data.D. By reviewing previous studies.15. Which of the following is a suitable title for the text?A. AL Model Stimulates Brain ActivitiesB. AL Model Speeds Up Language LearningC. AL Model Reveals the Secrets of Brain ActivationD. AL Model Enhances Brain Processing Capacity第二节(共5小题; 每小题2.5分, 满分12.5分)根据短文内容, 从短文后的选项中选出能填入空白处的最佳选项。

Dempster-Shafer Theory

Dempster-Shafer Theory

Dempster-Shafer TheoryGlenn Shafer1The Dempster-Shafer theory, also known as the theory of belief functions, is a generalization of the Bayesian theory of subjective probability. Whereas the Bayesian theory requires probabilities for each question of interest, belief functions allow us to base degrees of belief for one question on probabilities for a related question. These degrees of belief may or may not have the mathematical properties of probabilities; how much they differ from probabilities will depend on how closely the two questions are related.The Dempster-Shafer theory owes its name to work by A. P. Dempster (1968) and Glenn Shafer (1976), but the kind of reasoning the theory uses can be found as far back as the seventeenth century. The theory came to the attention of AI researchers in the early 1980s, when they were trying to adapt probability theory to expert systems. Dempster-Shafer degrees of belief resemble the certainty factors in MYCIN, and this resemblance suggested that they might combine the rigor of probability theory with the flexibility of rule-based systems. Subsequent work has made clear that the management of uncertainty inherently requires more structure than is available in simple rule-based systems, but the Dempster-Shafer theory remains attractive because of its relative flexibility.The Dempster-Shafer theory is based on two ideas: the idea of obtaining degrees of belief for one question from subjective probabilities for a related question, and Dempster's rule for combining such degrees of belief when they are based on independent items of evidence.To illustrate the idea of obtaining degrees of belief for one question from subjective probabilities for another, suppose I have subjective probabilities for the reliability of my friend Betty. My probability that she is reliable is 0.9, and my probability that she is unreliable is 0.1. Suppose she tells me a limb fell on my car. This statement, which must true if she is reliable, is not necessarily false if she is unreliable. So her testimony alone justifies a 0.9 degree of belief that a limb fell on my car, but only a zero degree of belief (not a 0.1 degree of belief) that no limb fell on my car. This zero does not mean that I am sure that no limb fell on my car, as a zero probability would; it merely means that Betty's testimony gives me no reason to believe that no limb fell on my car. The 0.9 and the zero together constitute a belief function.To illustrate Dempster's rule for combining degrees of belief, suppose I also have a 0.9 subjective probability for the reliability of Sally, and suppose she too testifies, independently of Betty, that a limb fell on my car. The event that Betty is reliable is independent of the event that Sally is reliable, and we may multiply the probabilities of these events; the probability that both are reliable is 0.9x0.9 = 0.81, the probability that neither is reliable is 0.1x0.1 = 0.01, and the probability that at least one is reliable is 1 -0.01 = 0.99. Since they both said that a limb fell on my car, at least of them being reliable implies that a limb did fall on my car, and hence I may assign this event a degree of belief of 0.99.Suppose, on the other hand, that Betty and Sally contradict each other—Betty says that a limb fell on my car, and Sally says no limb fell on my car. In this case, they cannot both be right and hence cannot both be reliable—only one is reliable, or neither is reliable. The prior probabilities that only Betty is reliable, only Sally is reliable, and that neither is reliable are 0.09, 0.09, and 0.01, respectively, and the posteriorprobabilities (given that not both are reliable) are 919 , 919, and 119, respectively. Hence we have a 919degree ofbelief that a limb did fall on my car (because Betty is reliable) and a 919degree of belief that no limb fell on my car (because Sally is reliable).In summary, we obtain degrees of belief for one question (Did a limb fall on my car?) from probabilities for another question (Is the witness reliable?). Dempster's rule begins with the assumption that the questions for which we have probabilities are independent with respect to our subjective probability judgments, but this independence is only a priori; it disappears when conflict is discerned between the different items of evidence.Implementing the Dempster-Shafer theory in a specific problem generally involves solving two related problems. First, we must sort the uncertainties in the problem into a priori independent items of evidence. Second, we must carry out Dempster's rule computationally. These two problems and and their solutions are closely related. Sorting the uncertainties into independent items leads to a structure involving items of evidence that bear on different but related questions, and this structure can be used to make computations1Ronald G. Harper Distinguished Professor of Business, School of Business, Summerfield Hall, University of Kansas, Lawrence, Kansas 66045.feasible. Suppose, for example, that Betty and Sally testify independently that they heard a burglar enter my house. They might both have mistaken the noise of a dog for that of a burglar, and because of this common uncertainty, I cannot combine degrees of belief based on their evidence directly by Dempster's rule. But if I consider explicitly the possibility of a dog's presence, then I can identify three independent items of evidence: my other evidence for or against the presence of a dog, my evidence for Betty's reliability, and my evidence for Sally's reliability. I can combine these items of evidence by Dempster's rule and the computations are facilitated by the structure that relates the different questions involved.For more information, see Shafer (1990) and the articles on the belief functions in Shafer and Pearl (1990).ReferencesDempster, A.P. (1968). A generalization of Bayesian inference. Journal of the Royal Statistical Society, Series B30 205-247.Shafer, Glenn (1976). A Mathematical Theory of Evidence. Princeton University Press.Shafer, Glenn (1990). Perspectives on the theory and practice of belief functions. International Journal of Approximate Reasoning3 1-40.Shafer, Glenn, and Judea Pearl, eds. (1990). Readings in Uncertain Reasoning. Morgan Kaufmann.2。

薛定谔—麦克斯韦尔方程径向解的存在性和多重性(英文)

薛定谔—麦克斯韦尔方程径向解的存在性和多重性(英文)

In 1887, the German physicist Erwin Schrödinger proposed a radial solution to the Maxwell-Schrödinger equation. This equation describes the behavior of an electron in an atom and is used to calculate its energy levels. The radial solution was found to be valid for all values of angular momentum quantum number l, which means that it can describe any type of atomic orbital.The existence and multiplicity of this radial solution has been studied extensively since then. It has been shown that there are infinitely many solutions for each value of l, with each one corresponding to a different energy level. Furthermore, these solutions can be divided into two categories: bound states and scattering states. Bound states have negative energies and correspond to electrons that are trapped within the atom; scattering states have positive energies and correspond to electrons that escape from the atom after being excited by external radiation or collisions with other particles.The existence and multiplicity of these solutions is important because they provide insight into how atoms interact with their environment through electromagnetic radiation or collisions with other particles. They also help us understand why certain elements form molecules when combined together, as well as why some elements remain stable while others decay over time due to radioactive processes such as alpha decay or beta decay.。

中国时政英语词汇

中国时政英语词汇

按照客观规律和科学规律办事act in compliance with objective and scientific laws八个坚持、八个反对eight do’s and eight don’ts八项主张eight-point proposal保持昂扬向上的精神状态be filled with an enterprising spirit保证中央的政令畅通ensure the Central Committee’s decisions are carried out without fail标本兼治address both the symptoms and root causes不确定因素uncertainties参政议政participation in and deliberation of state affairs长期共存、互相监督、肝胆相照、荣辱与共long-term coexistence, mutual supervision, treating each other with all sincerity and sharing weal and woe长治久安maintain prolonged stability崇尚科学respect and promote science传播先进文化spread advanced culture传统安全威胁traditional threats to security从严治军the army must be strict with itself党的领导方式the Party's style of leadership党的民族政策the Party's policy toward ethnic minorities党的侨务政策the Party's policy toward overseas Chinese affairs党的宗教信仰自由政策the Party's policy toward the freedom of religious belief党风廉政建设责任制responsibility system for improving the Party's work style and building clean government党内情况通报制度、情况反映制度和重大决策征求意见制度inner-Party information sharing and reporting systems and the system of soliciting opinions concerning major policy decisions党要管党、从严治党the Party exercises self-discipline and is strict with its members党员管理工作management of Party membership党政机关Party and government organs党政领导干部职务任期制、辞职制和用人失察失误责任追究制the system of fixed tenures, the system of resignation and the system of accountability for neglect of supervisory duty or the use of the wrong person with regard to leading cadres of the Party and government党总揽全局、协调各方的原则principle that the Party commands the overall situation and coordinates the efforts of all quarters电子政务e-government独立负责、步调一致地开展工作assume one’s responsibilities independently and make concerted efforts in one’s work独立公正地行使审判权和检察权exercise adjudicative and procuratorial powers independently and impartially多重多头执法duplicate law enforcement多重多头执法duplicate law enforcement发展民主团结、生动活泼、安定和谐的政治局面develop the political situation characterized by democracy, solidarity, liveliness, stability and harmony发展平等团结互助的社会主义民族关系enhance socialist ethnic relations of equality, solidarity and mutual assistance法定职能legal functions法律援助legal aid法制观念awareness of law防卫作战能力defense capabilities非传统安全威胁non-traditional threats to security丰富民主形式develop diverse forms of democracy干部人事制度cadre and personnel system干部双重管理体制system of dual control over cadres高知识群体prominent intellectuals公共事务public affairs公务员制度system of public servants公益事业programs for public good马列主义、毛泽东思想、邓小平理论、“三个代表”重要思想Marxism-Leninism, Mao Zedong Thought, Deng Xiao-ping Theory, Jiang Zemin “ThreeRepresent’s”important Thought新民主主义革命new-democratic revolution民族独立和人民解放national independence and the liberation of the people经济体制改革和政治体制改革reforms in the economic and political structure社会主义制度socialist system社会变革social transformation建设有中国特色的社会主义事业the cause of building socialism with Chinese characteristics中华民族的伟大复兴the great rejuvenation of the Chinese nation党在社会主义初级阶段的基本理论、基本路线、基本纲领the basic theory, line and program of our Party in the primary stage of socialism改革开放政策the policies of reform and opening to the outside中国共产党十一届三中全会The Third Plenary Session of the 11th Central Committee of the Communist Party of China马克思主义政党Marxist political Party党的第一(第二、第三)代中央领导集体the collective leadership of the Party Central Committee of the first (second/third)generation人民民主专政the p eople’s democratic dictatorship国民经济体系national economic system综合国力aggregate national strength国内生产总值the annual gross domestic product(GDP)独立自主的和平外交政策an independent foreign policy of peace马克思主义基本原理同中国具体实际相结合the fundamental principles of Marxism with the specific situation in China加强和改进党的建设,不断增强党的创造力、凝聚力和战斗力,永葆党的生机与活力strengthen and improve Party building, continuously enhance the creativity,rallying power and combat capability of the Party, and always maintain its vigor and vitality“三个代表”就是必须代表中国先进生产力的发展要求,代表中国先进文化的前进方向,代表中国最广大人民的根本利益,是我们党的立党之本、执政之基、力量之源,是我们党始终站在时代前列,保持先进性的根本体现和根本要求。

Combining Belief Fuctions When Evidence Conflicts

Combining Belief Fuctions When Evidence Conflicts

Ž.Decision Support Systems2920001–9 r locate r dsw Combining belief functions when evidence conflictsCatherine K.MurphyPenn State York,York,PA17403,USAAccepted6December1999AbstractThe use of belief functions to represent and to manipulate uncertainty in expert systems has been advocated by some practitioners and researchers.Others have provided examples of counter-intuitive results produced by Dempster’s rule for combining belief functions and have proposed several alternatives to this rule.This paper presents another problem,the failure to balance multiple evidence,then illustrates the proposed solutions and describes their limitations.Of the proposed methods,averaging best solves the normalization problems,but it does not offer convergence toward certainty,nor a probabilistic basis.To achieve convergence,this research suggests incorporating average belief into the combining rule. q2000Elsevier Science B.V.All rights reserved.Keywords:Belief functions;Expert systems;Decision analysis;Uncertain reasoning;Dempster–Shafer theory1.IntroductionA complicating factor in developing decision sup-port and expert systems is the handling of uncertain judgments.An expert may be unable to provide a definite answer to the question of interest.A number of basic approaches to the management of uncer-tainty in these systems have been developed.These include certainty factors,developed during the work w x w x on MYCIN3;Bayesian theory,fuzzy logic19, and belief functions,also known as Dempster–Shafer Ž.w xD-S theory5,11.Currently,there is no universally accepted method for uncertainty management.Each method has advantages and weaknesses associated w xwith it8,16.Two reasons for considering belief functions for the combination of evidence are their flexibility andŽ.E-mail address:cxm53@ C.K.Murphy.ease of use by the decision maker.Belief functionsw xfit into valuation-based systems17.Belief can be assigned to sets,not just to individual elements.In contrast with Bayesian decision modeling,belief functions do not require the expert to provide a set of prior probabilities.Also,belief in a hypothesis and its negation need not sum to1;some belief can be assigned to the base set,as a measure of uncertainty. This approach is similar to the evidence-gathering process observed when people reason at differentw x w xlevels of abstraction7.Pearl10agrees that this approach is well suited to knowledge elicitation:‘‘an expert may feel more comfortable describing the impact of an evidence in terms of weight assignment to classes rather than to individual points.’’Offsetting this ease of use are problems associated with the combination of belief bina-tion may yield conclusions different from what we expect or consider reasonable.In this paper,we0167-9236r00r$-see front matter q2000Elsevier Science B.V.All rights reserved.Ž.PII:S0167-92369900084-6()C.K.Murphy r Decision Support Systems2920001–9 2examine the combining of contradictory evidence using belief-functions.Our aim is to develop a useful alternative to the combination rule.We will compare the answers that proposed alternatives provide for dilemmas,both previously and newly identified.In the next section,we review the belief function theory and some alternative approaches for combin-ing evidence.Following that,we evaluate the perfor-mance of those alternatives in the context of a new problem,as well as a familiar one.The results of the comparison lead to the recommendation of an ap-proach for dealing with contradictory evidence.2.Belief functionsThe base set,or frame of discernment,u,for belief functions consists of a set of mutually exclu-sive and exhaustive hypotheses.The mass,m,of belief in an element of u can range from0to1, representing how strongly the evidence supports the hypothesis,without supporting a more specific hy-pothesis.This mass of belief is analogous to the weight in favor of the hypothesis.The mass,or basic probability assignment,measures the amount of be-lief attributed directly to an element;it does not include the mass attributed to any subsets of the element.In contrast,the belief function,Bel,of a set does include mass that has been specifically assigned to the element’s subsets.A subset is called a focal element of belief if its mass is greater than zero.The sum of the masses of belief in the base set,u,and its subsets is1since the base set includes the answer toŽ. the question of interest.The mass of the nullØset is defined as zero in the Dempster–Shafer frame-w xwork.In contrast,Smets14argues that,under an open-world assumption,the mass of the null set may be nonzero if the base set does not contain the correct answer.Belief can be assigned to a general conclusion, such as bacteria,as well as to singleton elements, such as pneumococcus.In addition,all belief need not be assigned to specific sets.The remaining mass, which represents uncertainty,or ignorance,is as-signed to the base set,u.This mass could possibly belong to any of the subsets.Its existence makesw possible a range of belief for every subset mass,xmass q ignorance.The range,or plausibility,of a subset extends from its mass to the mass assigned to all sets which include it.bining eÕidenceIn an expert system with IF–THEN rules,eitherŽ.the premises evidence or conclusions may be un-certain.Rules are triggered when the evidence,i.e., test results,referenced in their premises becomes available.When D-S theory is used in a rule-based expert system,applying a rule results in the assign-ment of masses of belief to the elements of the rule’sw xconclusion.Dempster’s5rule of combination han-dles the successive application of rules.Given two rules,based on independent evidence,the orthogonal sum,[,of their mass functions computes the degree of belief for the combined rules.Dempster’s rule,Ž.shown in Eq.1,varies the distributions,X and Y, from the two rules over all subsets of u and com-bines those elements where X l Y s Z.The set inter-sections represent areas where the conclusion of one rule agrees with that of the rule being combined with it.1m[m Z s m X m YŽ.Ž.Ž.Ý12121y kX l Y s Zk s m X m Y1Ž.Ž.Ž.Ý12X l Y s Bk is the mass that the combination assigned to the null subset.It represents contradictory evidence.Di-viding the other mass functions by1y k normalizes them.The process reapportions among the other subsets the belief that was originally assigned to the null set.When k s1,i.e.,when none of the combin-ing masses intersect,the function is undefined.Attractive features of the combining function are w x13:1.Concordant items of evidence reinforce eachother.2.Conflicting items of evidence erode each other.3.A chain of reasoning is weaker than its weakestlink.The first feature is accomplished by reassigning mass in the null set to the focal elements.When the mass in the null set is very large,as occurs when()C.K.Murphy r Decision Support Systems2920001–93conclusions disagree,reassigning it can cause prob-lems.Let us look at the operation of the combining function when the combining rules are in substantialÄ4 agreement.Suppose that a rule confirms A to0.4,Ä4Ä4A orB to0.2,andC to0.4.An expert systemÄ4 combines that rule with a second which confirms A to0.7and assigns the remaining mass to ignorance. The D-S combination of the two rules consists of the products of the masses from the rules’conclusions. The mass products are assigned to the intersection of the combining sets.Table1indicates the set assign-ments of the combined masses.The underlined values represent conflicting con-clusions,where the combining masses have no inter-section.The combined mass of belief in A is0.54 before normalization and0.75after normalization,inŽwhich all evidential masses were divided by1yŽ..mass of null set0.28.The mass assigned to A after normalization is greater than that assigned by either combining rule;normalization produces con-vergence toward the dominant opinion.The igno-rance interval disappeared in the combination.2.2.Problems and solutions in combining eÕidenceThe rules in Table1assigned similar plausibilitiesÄ4to the major element A.In the opposite situation, when evidential rules differ substantially in their conclusions,the combining rule of D-S theory may produce answers that disagree with the evidence. Two problems cited in the literature are the follow-ing.Ž.1D-S combination can assign100%certainty tow xa minority opinion20.Table4illustrates this prob-lem.Ž.2The‘‘ignorance’’interval disappears forever whenever a single piece of evidence imparts all its weight to a proposition and its negation,giving the false impression that precise probabilistic informa-w xtion underlies the belief10.We can see the latter effect in Table 1.The conclusion of rule2contained an interval which represented ignorance,the mass assigned to the base set.After combination with rule1,which assigned all its weight to specific sets,the resultant mass assigned to the base set u was zero.Ž.3Elements of sets with larger cardinality canw xgain a disproportionate share of belief15.Section 3.5presents this example.Because of these problems,several alternatives to the normalization process have been proposed.Ž.1Allow mass in the null set and,thus,eliminatew xthe need for normalization6.This eliminates divi-Ž.sion by1y k from Eq.1,Dempster’s rule.Ž.2Assign the mass in the null set to the base set w xu18.Since the correct destination of the conflict-ing evidence is unknown,distribute it among all the elements,rather than just the elements which happen to be intersections of the combining masses.Ž.3Average the masses assigned to a subset Z tow xŽ. determine its belief function,Bel18.Eq.2covers the case where two rules are combined.1Bel Z s m X q m Y2Ž.Ž.Ž.Ž.ÝÝ122X;Z Y;ZClarke suggests that classical estimation theory may provide the best solution in the form of the meansTable1Products of combining two mass functions using Dempster’s ruleÄ4Ä4Ä4Ä4Ä4 Conclusion A A or B C u B Rule1™m10.400.200.4000 Rule2™m2Ä4Ä4Ä4A0.7A0.28A0.140.280Ä4Ä4Ä4u0.3A0.12A or B0.06C0.120Combined mass0.540.060.1200.28Ž.Combined mass normalized0.750.080.1700 Underlined values:conflicting masses to be assigned to null set.Combined mass:sum of products whose intersection is the element in the column.Ž.Combined mass normalized:masses of focal elements after allocation of mass in null set.()C.K.Murphy r Decision Support Systems 2920001–94Table 2The silent majorityÄ4Ä4Ä4Ä4Ä4Ä4Ä4A A or B B C A or C u B First two rules 0.750.0800.17000Rule 3000.50.5000Combined mass000.040.085000.875()Combined mass normalized 000.320.68000Ž.No normalization r three rules lower bound 000.030.06000.91Ž.Mass ™Union r three rules upper bound 00.21000.280.510()Average three rules 0.3670.0670.1670.30.1and standard deviations of the beliefs.Furthermore,an awareness of the problems caused by assigning zero,rather than very small beliefs,could reduce the w x incidence of paradoxes in D-S combination 4.Ž.w x 4Oblow 9proposed that intersection and union operators constitute a lower and upper bound,respec-tively,for the mass,m ,of the combined informa-c tion.The intersection operator is Dempster’s combi-nation rule without normalization.m Z sm X m Y Lower boundŽ.Ž.Ž.Ýc 12X l Y s Z3Ž.The union operator is defined as:m Z sm X m Y Upper bound 4Ž.Ž.Ž.Ž.Ýc 12X j Y s ZŽ.The union operator of Eq.4assigns mass prod-ucts to the union of the subset of the elements being considered,rather than the intersection .The preceding four alternatives to Dempster’s rule have been proposed because combining conflicting evidence may produce counterintuitive results.In addition,several authors have noted that it may beuseful to track mass in the null set because it repre-w x sents conflicting evidence.In particular,Zeigler 21,in the context of describing a propositional evidence accumulator,suggested that the amount of contradic-tion be used to trigger a warning to the user when it exceeds a preset value.Although not designed to handle conflicting evi-dence,the iterative assignment method proposed by w x Baldwin 1,2represents another approach for com-bining belief functions.Rather than dividing all the intersection products by a single normalizing con-Ž.stant 1y k in Eq.1,the method calculates a separate normalizing constant for each column that has an intersection product corresponding to mass in the null set.That mass is set to zero and reallocated to the focal elements in the column.The result is that the focal elements in each column sum to the mass that the rule assigned to that column.For example in Table 1,the normalizing constant for the third col-umn would be 0.3;dividing the mass assigned to Ä4Ä4C ,0.12,by 0.3gives C a normalized mass of 0.4.This assignment represents only the first iteration as this method repeatedly combines the evidence until the solution converges.Table 3Effect of one strongly confirming ruleÄ4Ä4Ä4Ä4Ä4A A or B B u B Ž.Rule 1result of four rules 0.500.500Rule 50.9000.10Combined mass0.500.0500.45()Combined mass normalized 0.9100.0900Ž.Combined mass w r o normalization for five rules lower bound 0.06300.00600.931Ž.Mass ™Union for five rules upper bound 0.0560.84400.10Average0.5800.40.02Combined mass based on average0.8560.144()C.K.Murphy r Decision Support Systems2920001–95Table4Normalization alternatives for100%certainty assigned to a minority opinionÄ4Ä4Ä4Ä4Ä4Ä4Ä4Ä4A B C AB AC BC u BRule10.900.100000Rule200.90.100000 ()D-S combined mass normalized00 1.0000000Ž.D-S w r o normalization lower bound000.0100000.99Ž.Mass™Union upper bound000.010.810.090.0900 Average0.450.450.100000If the evidence is incompatible,as is the case for the examples in Tables1–4,the solution still con-verges,but it does not retain all the evidence.The iterative assignment method provides a measure of compatibility,the support pair for each focal ele-ment.The support interval for an element is brack-eted by its mass and its plausibility.When two rules are combined,the necessary support for an element is defined as its maximum mass in the combining rules.Its possible support is its minimum plausibility in the rules.For the rules in Table1,the support pair Ä4w xfor A is0.7,0.6,indicating that one expert as-Ä4signs a higher mass to A than the second expert considers plausible.When the evidence conflicts,as in this example,the iterative assignment method will lose evidence in its combination.In the next three examples,we will analyze only the four normaliza-tion alternatives that are suitable for conflicting evi-dence.In Section3.5,the cardinality problem,we will include the solution provided by the iterative assignment method.3.Problems and normalization alternativesThe set of problems identified with the combining rule implies that the following properties are ex-pected in any alternative method for combining con-flicting evidence.1.Assign the preponderance of belief to a majorityopinion,not a minority one.2.Indicate,if possible,an appropriate level of igno-rance.3.Provide a combined belief that reflects the rela-tive strengths and frequencies of the individual estimates.A further requirement is that the method satisfy commutative and associative properties so that the order and grouping of evidence do not affect the result.In considering what have been described as the limitations of the combining rule,we will distinguish Ž.between:1combinations which lack some desir-Ž.able feature and2combinations which produce errors.An example of the former is the disappear-ance of the ignorance interval,as shown in Table1. This is not a misleading result.We can see that the estimate is uncertain despite the absence of mass in the base set because the combined mass resides in two distinct subsets.One could argue that the real value of the ignorance interval is its simplification of knowledge acquisition,and that the presence of an ignorance interval in the result is of secondary im-portance.In this paper,we will examine solutions for those cases which represent errors in the combina-tion of evidence.Of the proposed alternatives,we eliminate the one that assigns mass in the null set to the base set.It gives plausible answers,but its operation is not associative.Changing the grouping of a series of Ž.rules the order in which they fire would change the final mass assignments.In addition,assigning a large mass to the base set,or environment,paves the way for another problem.The next rule which fired would be in a position to establish certainty,prematurely. Because of these flaws,we will not consider further the alternative of assigning conflicting mass to the base set.The other alternatives are both associative and commutative in their operations,as is Dempster’s combination rule.In the next sections,we examine problems associ-ated with multiple rule combinations and tabulate the solutions provided by the proposed normalization()C.K.Murphy r Decision Support Systems2920001–9 6alternatives.These problems have not been described in the literature.The first problem is the loss of a majority opinion because of a single dissenting rule which assigns no plausibility to the majority’s sub-set.The table of alternative solutions includes the mass in the null set so that we can judge its effec-tiveness in warning of possible errors.3.1.Loss of the majority opinion in multiple combi-nationsThe combination rule works when the combining rules are in substantial agreement,as in the example from Table1.Suppose we combine the result with another rule,rule3which assigns all of its mass to the minority sets:0.50to B and0.50to C.The result of the combination,as well as the results produced by alternative methods,is shown in Table2.The combined mass in Table2shows no mass in Ä4A which contained the majority of the mass in the previous two rules.Moreover,no amount of corrobo-rating evidence can resurrect the belief in a majority opinion if the system ever uses a rule which assigns all mass to sets which contradict the majority.The system in Table3assigns68%of its belief to C yet its average plausibility is only40%.In Table2,the amount of mass in the null set before normalization,0.875,warns that the new evi-dence is in conflict with an established pattern; however,the upper and lower bounds are not useful in this example.The lower bound is near zero for all the subsets.The minimum combining mass would be a more informative lower bound for the major ele-ments than Dempster’s rule without normalization. In contrast,the weighted average describes the distri-butions;and its disagreement with the normalized masses also functions as a warning.The average shows that the accumulated evidence does not justify convergence to any conclusion.3.2.Attaining certainty in multiple combinationsThe converse of the loss of a majority opinion is the rapid movement toward certainty when a body of conflicting evidence is overridden by a single piece of strongly confirming evidence,as illustrated in Table3.Suppose that a series of four rules have divided belief between A and B equally;however, the next combining rule confirms A to0.9.Ä4The certainty expressed for A in rule5increases in the combination despite the presence of50% disconfirming evidence in the previous four rules. The failure of belief functions to weight conclusions by the number of contributing rules becomes a prob-lem when most of the rules divide belief fairly equally among competing propositions.Although the conclusion may be correct,the decision maker de-serves a warning that the level of certainty is not so high as the combined mass indicates.Whether this turns out to be a problem depends on which rule is representative of the evidence.With only45%con-flict,this example demonstrates the difficulty of establishing a warning level for the mass in the null set that would be appropriate for all situations.The weighted average gives a clearer picture of the rules’Ä4 disagreement.The average mass in A is high Ž.enough58%to motivate the question of how to reach a conclusion based on averages.The final line in Table3,discussed in the next section,addresses that question.3.3.AchieÕing certainty with aÕeragesAveraging provides an accurate record of con-tributing beliefs,but it lacks convergence.Unlike Dempster’s rule,it does not increase the measure of belief in the dominant subset.To provide conver-gence with an averaging method,we insert the aver-age values of the masses in the combining rule.IfŽ. there are n rules,or pieces of evidence,use Eq.1 to combine the weighted averages of the masses n y1times.We can,thus,avoid overdependence on a single piece of conflicting evidence,such as one causing the disappearance of the majority opinion. This use of the average in Table3leads to a combined belief in A of0.856,compared with0.91 from the individual application of the rules.In gen-eral,combining an average gives a less extreme answer.Others have noted that combining belief functions leads to a paradox—the greater is the conflict between pieces of evidence,the greater is the‘‘certainty’’on ing an average value reduces this effect.The calculation in Table3assumes that rule1 represents five rules.If rule1were a single rule,one combination of the average masses from rules1andŽ5would yield a mass of0.861for A also less than()C.K.Murphy r Decision Support Systems2920001–97.the original0.91.This result is typical of the com-bining rule:combining two identical rules with0.7Ä4mass in A yields a lower value than combining two rules with the differing masses of0.5and0.9.This property,in itself,may constitute another incongruity of the normalization process.3.4.SolÕing a classic problemIn this section,we tabulate the results of applying the D-S combining rule and the proposed normaliza-tion alternatives to a problem involving only twow xrules:Zadeh’s20example of100%certainty as-signed to a minority belief.Ä4Ä4Ä4The headings AB,AC,and BC in Table4 refer to unions of two singleton sets;these headings were included to illustrate the union operator.The mass in the null set before normalization is under-lined to highlight its importance as a warning.All of the alternative approaches are adequate for this extreme example of a normalization problem. The alternative of using the average and the alterna-tive of assigning the mass to the union also furnish a record of the conflicting evidence.The latter opera-Ä4Ä4tion gives an upper bound for A and B equal to their maximum mass in the two combining rules;Ä4however,the upper bound for C,0.19,is higher than its maximum,0.1.Before normalization,the mass in the null set is0.99,a definite warning of a false conclusion,given that the combined mass does not reflect the divergence in the combining distribu-tions.The equal division between A and B in the average shows that the combining evidence does not yield a definite conclusion.3.5.Unearned beliefw xVoorbraak15identified a weakness in a feature usually considered to be a strength of belief func-tions:probability assigned to a set is not divided among its elements but remains with all the ele-ments.In combination with other evidence,this can result in an element of the multi-element set receiv-ing a larger belief than seems justified.He provides this example.Ä4Ä4ÄRule1:A s0.5s B or C;Rule2:A or 4Ä4B s0.5s C.Ä4Ä4 Result after combining rules1and2:A s B s Ä4C s1r3.The equal distribution of belief is counterintuitiveÄ4Ä4because both A and C had individually assignedÄ4Ä4 mass,as well as a share with B;but B had only two shared masses.Because the problem occurs in the intersection operation,omitting the normalization step doesn’t change the relative assignments.However,averaging the masses yields:m A s m C s m A or B s m B or C s0.25.Ž.Ž.Ž.Ž. Averaging followed by D-S combination gives:m A s m C s0.3,m B s0.2,Ž.Ž.Ž.m AB s m BC s0.1.Ž.Ž.Ä4Ä4 This result assigns a higher mass to A and C than Ä4to B although it assigns the same plausibility,0.4,Ä4Ä4Ä4to A,B,and C.The iterative assignment method converges to Ž.Ž.Ž.m A s m C s0.5,m B s0.0,given the supportŽ.w xŽ.Ž. pairs for the rules:S A s0.5,0.5s S C;S B s w x0.0,0.5.Both this method and averaging achieve a satisfactory solution to this problem.4.Evaluation of normalization alternativesAll of the normalization alternatives avoid the counter-intuitive results produced by Dempster’s rule when conflicting evidence is present;however,aver-aging lacks correspondence with Bayesian condition-ing.In addition,each of the alternative approaches has other drawbacks.Mass in the null set warns of problems;however, setting an appropriate level is problematic.We haveŽ.seen an example Table3where a warning is indicated,but the mass in the null set is only45%. One of the combinations in Shafer and Tversky’s w x12anthropological example contains75%mass in the null set;however,the final result is compatible with the evidence.At extreme levels,greater than 90%,null mass is a reliable indicator of problems.The method,that does not normalize,allocates a decreasing mass to the focal elements in successive ing the intersection operator with-out normalization to represent a lower bound on the()C.K.Murphy r Decision Support Systems2920001–9 8belief results in a belief which approaches zero as the number of combinations increases.Coupling this with an upper bound supplied by the union operator results in a constantly increasing gap between belief and plausibility as combinations continue.Thus,there is the paradox that as evidence accumulates,igno-rance increases.As the number of combinations increases,the belief assigned to ignorance ap-proaches one.Upper and lower bounds,calculated in this manner,provide little information about the more weighty focal elements.The maximum and minimum functions are more informative bounds.Similarly,if mass were allowed in the null set with the assumption that there are unknown proposi-tions,these unknown elements would accumulate an increasingly larger share of the belief as evidence combination proceeded.This would occur even if the evidence supported one or more of the known propo-sitions.The average offers an account of the evidence; therefore,it is useful as a reference even when it is not the primary combination method.Substantial differences between the average and the combined belief indicate problems in the combination.The averaging method does not converge to a conclusion; however,we can incorporate convergence by using the combining rule repeatedly with the average val-ues.Moreover,in the averaging process,the decision maker can easily weight evidence by its perceived importance or reliability.Of the proposed methods, only averaging preserves a record of uncertainty Ž.mass assigned to the base set and the relative frequencies of beliefs.From the alternatives,we can construct two prac-tical methods.Ž.1If all the evidence is available at the same time,average the masses,and calculate the combined masses by combining the average values multiple times.Ž.2If decisions are made as evidence accumu-lates,using Dempster’s rule with the individual evi-dential masses is simpler than recomputing averages.Ž.Use guidelines based on either a mass in the null Ž.set or b average masses to warn of possible prob-lems caused by conflicting evidence.A guideline based on average masses summarizes the evidence and is more informative than one based on mass in the null set.5.ConclusionWhen conflicting evidence is present,Dempster’s rule for combining beliefs often produces results that do not reflect the actual distribution of beliefs.Three types of problems were previously identified;two other types are associated with the successive combi-Ž.nation of rules:1a single rule forcing certainty and Ž.2a single rule overruling a majority opinion.Of the alternative methods that address the problems, averaging identifies combination problems,shows the distribution of belief,and preserves a record of Ž.ignorance unassigned belief.In selecting an uncertainty management system for an expert system,two characteristics to avoid are:Ž.1assigning misleading certainty to a recommenda-Ž.tion and2failing to attain certainty when it is justified.Analysis of examples in four problem areas showed that the method of averaging avoids the first error and accounts for all combining rules;moreover, an averaging system can intensify belief in the domi-nant subset by using average values in the combining rule.Averaging also easily incorporates the assign-ment of higher weights to more reliable evidence.Using actual belief functions,rather than averages in the combining rule offers correspondence to Bayesian theory,as well as convergence.This method is computationally simpler when making decisions with incremental evidence.In this situation,Demp-ster–Shafer combination is recommended as the pri-mary uncertainty management system,accompanied by weighted averages to track the accumulated mass and warn of possible errors.Referencesw x1J.F.Baldwin,A calculus for mass assignments in evidentialŽ.reasoning,in:R.Yager,J.Kacprzyk,M.Fedrizzi Eds., Advances in the Dempster–Shafer Theory of Evidence,Wi-ley,New York,1994,pp.513–531.w x2J.F.Baldwin,Combining evidences for evidential reasoning,Ž.International Journal of Intelligent Systems61991569–616.w x3 B.J.Buchanan,E.H.Shortcliffe,A model of inexact reason-Ž.ing in medicine,Mathematical Biosciences231975351–379.w x4M.R.B.Clarke,Discussion of belief functions,by P.Smets,Ž.in:P.Smets, A.Mamdani, D.Dubois,H.Prade Eds.,。

DS证据理论 _浙大

DS证据理论 _浙大
[18] Yaghlane, B. B., et al. Belief function independence: II. The conditional case. International Journal of Approximate Reasoning, 2002, 31: 31-75.
本章的主要参考文献(续4)
浙江大学研究生《人工智能》课件
第五章 D-S证据理论
(Chapter5 D-S Evidential Theory )
徐从富(Congfu Xu) PhD, Associate Professor
Email: xucongfu@ Institute of Artificial Intelligence, College of Computer Science, Zhejiang University, Hangzhou 310027, P.R. China
March 10, 2002第一稿 September 25, 2006第四次修改稿
Outline



本章的主要参考文献 证据理论的发展简况 经典证据理论 关于证据理论的理论模型解释 证据理论的实现途径 基于DS理论的不确定性推理 计算举例
本章的主要参考文献
[1] Dempster, A. P. Upper and lower probabilities induced by a multivalued mapping. Annals of Mathematical Statistics, 1967, 38(2): 325-339. 【提出 证据理论的第一篇文献】 [2] Dempster, A. P. Generalization of Bayesian Inference. Journal of the Royal Statistical Society. Series B 30, 1968:205-247. [3] Shafer, G. A Mathematical Theory of Evidence. Princeton University Press, 1976. 【证据理论的第一本专著,标志其正式成为一门理论】 [4] Barnett, J. A. Computational methods for a mathematical theory of evidence. In: Proceedings of 7th International Joint Conference on Artificial Intelligence(IJCAI-81), Vancouver, B. C., Canada, Vol. II, 1981: 868-875. 【第一篇将证据理论引入AI领域的标志性论文】

证据理论

证据理论

适用策分析,等等。
4、证据理论的局限性
要求证据必须是独立的,而这有时不易满足
证据合成规则没有非常坚固的理论支持,其合理 性和有效性还存在较大的争议
计算上存在着潜在的指数爆炸问题
5、证据理论的发展概况
“Zadeh悖论”:对证据理论的合成公式的合理性进行 质疑。 例子:利用Dempster证据合成规则对两个目击证人 (W1, W2)判断某宗“谋杀案” 的三个犯罪嫌疑人(Peter, Paul, Mary)中究竟谁是真正的凶手,得到的结果(认定Paul 是凶手)却违背了人的常识推理结果,Zadeh认为这样的结果 无法接受。 m1() Peter Paul Mary 0.99 0.01 0.00 m2() 0.00 0.01 0.99 m12() 0.00 1.00 0.00
[12] Dubois, D, Prade, H. Consonant approximations of belief functions. International Journal of Approximate Reasoning, 1990, 4: 279-283.
[13] Tessem, B. Approximations for efficient computation in the theory of evidence. Artificial Intelligence, 1993, 61:315-329. 【注:文献10-12均为证 据理论近似计算方法】 [14] Simard, M. A., et al. Data fusion of multiple sensors attribute information for target identity estimation using a Dempster-Shafer evidential combination algorithm. In: Proceedings of SPIE-International Society for Optical Engineering, 1996, Vol.2759: 577-588. 【提出了一种实现证据理论的“修 剪算法”】

D-S证据理论

D-S证据理论
精品文档
本章的主要参考文献(续3)
[15] Josang, A. The consensus operator for combining beliefs. Artificial Intelligence, 2002, 141(1-2): 157-170. [16] Yang, Jian-Bo, Xu, Dong-Ling. On the evidential reasoning algorithm for multiple attribute decision analysis under uncertainty. IEEE Transaction on Systems, Man, and Cybernetics – Part A: Systems and Humans, 2002, 32(3): 289-304. [17] Yaghlane, B. B., et al. Belief function independence: I. The marginal case. International Journal of Approximate Reasoning, 2002, 29(1): 47-70. [18] Yaghlane, B. B., et al. Belief function independence: II. The conditional case. International Journal of Approximate Reasoning, 2002, 31: 31-75.
[8] Voorbraak, F. On the justification of Dempster’s rule of combination. Artificial Intelligence, 1991, 48:171-197.

DS证据理论

DS证据理论

适用领域:信息融合、专家系统、情报分析、法律 案件分析、多属性决策分析,等等。
4、证据理论的局限性
要求证据必须是独立的,而这有时不易满足
证据合成规则没有非常坚固的理论支持,其合理 性和有效性还存在较大的争议
计算上存在着潜在的指数爆炸问题
5、证据理论的发展概况
“Zadeh悖论”:对证据理论的合成公式的合理性进行 质疑。 例子:利用Dempster证据合成规则对两个目击证人 (W1, W2)判断某宗“谋杀案” 的三个犯罪嫌疑人(Peter, Paul, Mary)中究竟谁是真正的凶手,得到的结果(认定Paul 是凶手)却违背了人的常识推理结果,Zadeh认为这样的结果 无法接受。 m1() Peter Paul Mary 0.99 0.01 0.00 m2() 0.00 0.01 0.99 m12() 0.00 1.00 0.00
5.1
证据理论的发展简况
证据理论(Evidential Theory) Dempster-Shafer理论 Dempster-Shafer证据理论 DS (或D-S)理论
1、证据理论的名称
其它叫法:
Dempster规则 Dempster合成规则 Dempster证据合成规则
March 10, 2002第一稿 September 25, 2006第四次修改稿
Outline



本章的主要参考文献 证据理论的发展简况 经典证据理论 关于证据理论的理论模型解释 证据理论的实现途径 基于DS理论的不确定性推理 计算举例
本章的主要参考文献
[1] Dempster, A. P. Upper and lower probabilities induced by a multivalued mapping. Annals of Mathematical Statistics, 1967, 38(2): 325-339. 【提出 证据理论的第一篇文献】 [2] Dempster, A. P. Generalization of Bayesian Inference. Journal of the Royal Statistical Society. Series B 30, 1968:205-247. [3] Shafer, G. A Mathematical Theory of Evidence. Princeton University Press, 1976. 【证据理论的第一本专著,标志其正式成为一门理论】 [4] Barnett, J. A. Computational methods for a mathematical theory of evidence. In: Proceedings of 7th International Joint Conference on Artificial Intelligence(IJCAI-81), Vancouver, B. C., Canada, Vol. II, 1981: 868-875. 【第一篇将证据理论引入AI领域的标志性论文】

On Certainty

On Certainty

On Certainty*Wolfgang SpohnCertainty is an epistemic quality or, as philosophers are used to say, an epistemic modality. It is not easily accounted for as such; but things get even more complicated due to the fact that certainty is often confused with other epistemic modalities. Since I cannot discuss here the quite sophisticated technical treatments of certainty, I focus on disentangling certainty from and relating it to other epistemic modalities; this is the more important business for getting a hold on certainty. In doing so, I am not presenting any original view or thesis; my view or evaluation implicitly shows just in my selection from material which is familiar in more formally orientied analytic philosophy, but is less known, I think, outside philosophy (and AI).1Certainty is most appropriately equated with unrevisability or infallibility. This relation will be one main focus of my talk. But a more important role in the history of philosophy was played by two other modalities, apriority and analyticity, the first of which is clearly epistemic and the second of which may be so taken. Certainty must be strictly distinguished from both which in turn only partly capture the various uses of the notion of necessity. A prominent instance of confusion is Quine's famous attack on the notion of analyticity which ended up with the conclusion, and confusion in my view, that the only feasible sense which can be given to it is centrality, meaning something like "hardly revisable".2So, my other main focus is to address these distinctions.* Vortrag auf dem 13. Weltkongreß für Soziologie …Contested Boundaries and Shifting Solidarities“ in Bielefeld im Juli 1994.1 In Artificial Intelligence as well great efforts are devoted to the modelling of epistemic states; some references are given below. In fact, philosophy has one of its most fruitful interdisciplinary exchanges in this area.2 Cf. Quine (1951), p. 39ff.Let me start with two basic observations. First, the logical form of the notion of certainty is that it is at least a three-place relation: a is certain of p at t - where a is a subject capable of epistemic atttitudes, t is a time, and p is a proposition or whatever you prefer to take as an object of belief; the nature of these objects is utterly problematic and beyond the scope of my talk.3Secondly, certainty obviously comes in degrees; people are more or less certain and more certain of some things than of others. Thus, certainty is in fact a four-place relation: a is certain of p at t to the degree r. Thus stated, it is tantamount to the most basic epistemological notion, namely: a believes p at t to the degree r. If "certain" behaves like most adjectives, then p is plainly certain for a subject if p is more certain than most other things; compare this with "x is a tall elephant just if x is an elephant taller than most elephants". However, this is not the plain certainty philosophers always talk about; their interest was in absolute certainty, as it was usually emphasized, or in maximal certainty, to express it in terms of degrees.So, the primary theoretical task is to account for these degrees of certainty and belief. One important theoretical connection is that these degrees manifest themselves in our decisions and actions; we base our actions more firmly on our firmer beliefs and less on our lesser certainties. For our present concern, however, this is the less important connection. One reason for this is that this is, so to speak, an impure manifestation of the degrees of belief; they therein mesh with the degrees of our desires or volitional attitudes in general in a very complicated fashion. The other reason is that the only well-working theoretical model of this complicated meshing is decision theory according to which degrees of belief are probabilities.4 Other models of degrees of belief thus drop out of focus because they are not well embedded in the theory of practical reasoning.But such other models exist, as will become clear when we look at the other theoretical connection: degrees of belief play a crucial role in the dynamics of belief. We continuously revise our beliefs or, more generally, change our epistemic state in the light of new evidence or information; this continuous change is described in the3 Philosophers have burdened propositions with multiple roles, as truth bearers, as sentence meanings, as objects of propositional attitudes. The clearer it became that no entity can play all these roles, the unclearer it became how to characterize for each role the approriate entities.4 The most widely used version of decision theory was developed by Savage (1954). Philosophers became acquainted to decision theory by Jeffrey (1965). For a brief comparison of basic decision theoretic models cf. Spohn (1978), ch. 2.dynamics of belief. However, there is no workable account of this dynamics in terms of ungraded belief5; only if one takes belief as graded, one can state reasonable general laws of epistemic change. So, what is changed in the light of evidence is in fact the assignment of degrees of belief or certainty to the various propositions.The most prominent model of epistemic change, of course, is the probabilistic one. Here, epistemic states are represented as probability measures (in the mathematical sense).The crucial point is that conditional probabilities can be defined relative to a probability measure. The dynamic law then basically takes the form of a rule of conditionalization: my new probability for a given proposition is just my old probability for it conditional on the evidence gathered in between.6 Indeed, each account of epistemic change must provide analogous notions of conditional epistemic states and of conditionalization.In the meantime, there exist various alternatives to the probabilistic model and a vast amount of literature about them. The so-called AGM approach to belief revision is perhaps the most carefully worked out one7; there only ordinal degrees of belief are assumed. A bolder and more powerful approach is given by the theory of so-called ranking functions or natural (or ordinal) conditional functions.8In AI two other theories are even more prominent: the Dempster-Shafer theory of belief functions9and fuzzy logic10 (though the latter is in my view misapplied to the epistemic matters under discussion). And more could be mentioned; the field is, despite the enormous amount of work, still in an experimental state, so to speak.115 Cf. Spohn (1988), sect. 2.6 More general rules of probabilistic belief change have been developed. The two most prominent ones are Jeffrey's generalized conditionalization (cf. Jeffrey 1965, ch. 11) and the rule of maximizing entropy or of minimizing relative entropy (cf. Hunter 1991).7 Cf. Alchourrón et al. (1985), Gärdenfors (1988), and Gärdenfors and Rott (1994).8 Cf. Spohn (1988) and (1990) and Goldszmidt and Pearl (1992).9 Cf. Shafer (1976), (1990), and (1992).10 Cf. Dubois and Prade (1988).11 There have been predecessors, of course. Dempster (1967), Shackle (1969), Rescher (1976), and Cohen (1977) perhaps deserve most to be mentioned. Shafer (1978) even mentions J. Bernoulli and J.H. Lambert as early predecessors in the 18th century.Revisability or fallibility12are clearly dynamical notions which are explicable precisely by such theories. Certainty, however, is no less a dynamical notion. A belief is the more certain, the harder it is to revise; it is certain in the vague and loose everyday sense if it is hard to revise; and it is certain in the strict philosophical sense if it is not revisable at all. All these explanations adopt a precise meaning only within the theories of epistemic change referred to, and a different meaning, at that, in different theories. The vague notion of hard revisability allows of a spectrum of exactifications; but the strict notion of unrevisability or absolute certainty receives a unique explication in each theory of epistemic change. Consider the most familiar theory, probability theory: there, a proposition is absolutely certain or unrevisable if and only if it has probability 1, because according to each probabilistic rule of epistemic change a proposition keeps probability 1 forever once it has received probability 1. Thus, a theory of certainty can be developed within probability theory; it is in fact quite simple. Analogous assertions hold for the other theories of epistemic change.So, which propositions are certain? Well, I said that certainty is a subject-relative notion; propositions which are certain for you may be uncertain for me and vice versa; this depends on the subjective epistemic states and their dynamics. So far, the only propositions which turn out to be certain for everyone according to the above-mentioned theories of epistemic change are the logically true ones. Therefore the question arises whether there are more propositions which everyone should or may reasonably take as certain. But note that we are entering a new field with this question. We are no longer explicating certainty, as we have done so far, but we are looking for further rationality constraints on certainty.A first attempt to answer the question may be to say that the propositions to be taken as certain are the necessary propositions. But this answer is not good enough, since there are many kinds of necessity. There is logical necessity, the strictest kind of necessity, for which the answer is true. There is mathematical necessity the nature of which is much discussed - does it reduce to logical necessity, is it a kind of linguistic necessity or a kind of necessity sui generis? - to which the answer applies as well.13 But there are also various kinds of material necessity: causal necessity,12 Here I would like to mention the markedly different approach by Levi (1980) who strictly di-stinguishes between fallibility and revisability.13 By thus declaring mathematical propositions as certain or indubitable, I do not want to deny that there is mathematical doubt. But mathematical doubt is not only beyond the scope of thephysical necessity, historical necessity14,and so on, and the strictest of them, metaphysical necessity15. For each of them the answer is certainly not true. We may doubt or even disbelieve propositions which are necessary in one of these senses, we may discover them, we may find reasons for rejecting them again, and so on. So, these kinds of necessary propositions should not be taken as certain.Hence, the first attempt was not yet specific enough. But it headed into the right direction. Certainty itself may be viewed as a kind of epistemic necessity; thus the kind of necessity which is characteristic of propositions reasonably to be held to be certain is presumably an epistemic one as well. Which kinds of epistemic necessity are there?There are mainly two candidates. Since Kant has made vital use of the analy-tic/synthetic and of the a priori/a posteriori distinction in his epistemological turn of metaphysics, these distinctions have remained in the center of theoretical philosophy and have caused a lot of concern and confusion.To take up analyticity first: The common explanation is that analytic sentences are sentences which are true only by virtue of the meaning of the expressions and the syntactic constructions from which they are built. Analytic truths are thus known by fully competent speakers simply in virtue of their knowledge of language. And come what may, they cannot turn out false. The meanings may change, of course, and the syntactic forms when associated with the new meanings may yield falsehoods. But then you have, in a way, different sentences in a new language; it is not the old analyticities which would thereby turn false. In this sense, analytic sentences are epistemically necessary.The explanation given is so common because it is vague. What is truth in virtue of meaning alone? In order to render this precise, nothing less than a fulltheories I am referring to; I know of no theory at all which would be able to adequately cope with it.14 The simplest account fo these necessities is this: Something is causally (physically, historically) necessary just if it is logically entailed by the causal (physical, historic) laws. Thus, insofar there are no historic laws, there are no historical necessities (except the logical ones). Whether this simple account is adequate in each case is doubtful. It may be better, for instance, to proceed conversely and explain causal laws in terms of causal necessity; cf., e.g., von Fraassen (1989) and Spohn (1993).15 Which has been forcefully reintroduced into the current philosophical discussion by Kripke (1972), among others.meaning theory is required. Indeed, the search for an explication of analyticity was an important motive in developing various and ever more sophisticated meaning theories. Thus, in a way, there are today as many concepts of analyticity as there are theories of meaning.16The state of the notion of apriority is still worse. The common explanation is that a proposition is a priori known by a subject if the subject knows it prior to any experience. From this cautious, subject-relative explanation it needs a substantive step to argue that the same propositions are a priori known by all subjects; these propositions may then be called a priori by themselves. If propositions a priori are known prior to any experience, then no experience can prove them to be false; they are to be believed come what may. In this sense, propositions a priori are again epistemically necessary.This explanation of apriority is still vaguer than that of analyticity. Certainly, the philosophical community was also misled by the logical empiricists' forcefully doing away with apriority by simply identifying it, contra Kant, with analyticity. This misunderstanding has been cleared up cleared up for over 20 years17, and since the notion of apriority is widely and freely used again. However, its use is, to my knowledge, hardly backed up by any theory and stays on a rather insecure informal level.18This is why I said it would be worse off than the notion of analyticity.There is, however, a theoretical framework which in my view improves upon the situation. The part of the theory of meaning which is relevant to an account of analyticity is referential semantics, i.e. that part which is concerned with reference and truth. It has received its most powerful and up-to-date format in the so-called character theory of David Kaplan.19 Kaplan wanted it to keep separate from epistemology, but Robert Stalnaker has given it an explicitly epistemological16 Quine (1960) tries to satisfy us with ersatz concepts like stimulus meaning and stimulus analyticity. Putnam had influential, though changing views on the matter; see Putnam (1975), ch.2 and 12. Lewis (1969) is ultimately an attempt to reestablish the notion of analyticity. And so on; the list could be continued almost indefinitely.17 Due to Kripke (1972) who made very clear that analyticity, metaphysical necessity, and apriority are three different notions and that the latter two are in fact independent, since there are clear cases of necessities a posteriori and of contingencies a priori.18 Cf., e.g., Putnam (1983), ch. 6 and 7, Kitcher (1980) or Casullo (1988).19 Cf. Kaplan (1977) and (1989) and Lewis (1980).reinterpretation.20Given this reinterpretation, the theory is capable not only of explicating metaphysical necessity and analyticity, but also of analyzing apriority.21 It thus provides a framework for studying not only apriority, but also its relation to other central modal notions. In this it is unrivaled, as far as I know, and that is why I am mentioning it here.According to these explications, analyticity is a stronger notion than apriority.22There is, however, a further difference. Analyticity is a communal notion applying to a given language as spoken by a given linguistic community, whereas apriority is rather a subjective notion applying to the propositions or thought contents entertained by a given epistemic subject.23 This difference is relevant here because the truths which are analytic on the communal level need not be a priori known on the subjective level. This is so because knowledge of analytic truths requires full semantic competence in a strong sense which we may well fail to satisfy without ceasing to count as members of our linguistic community.24Thus, analytic truths are not necessarily subjectively certain; only a priori truths taken at the subjective level are certain.This brings us back to our topic. I said which propositions a subject takes as certain is up to her or him and depends on her or his dynamics of epistemic states. But it is certainly a rationality postulate on this dynamics that all and only propositions a priori are taken to be certain, i.e.unrevisable. On the one hand, there will never arise a need to revise the belief in an a priori proposition because it is20 Cf. Stalnaker (1978) and (1987). However, Stalnaker did not conceive of himself as reinterpreting Kaplan; he intended just different things with a formal apparatus similar to Kaplan's. The relation between Kaplan and Stalnaker is reconstructed in Haas-Spohn (1994).21 Technically speaking, metaphysical necessity is truth at the actual context and all indices, apriority is truth at all contexts, and analyticity is truth at all contexts and indices. Thus, analyticity is the strongest notion and comes to a priori necessity, as Kripke (1972), p. 264, has already claimed. The significance of these explications is, however, revealed only by studying the whole framework. Cf. Haas-Spohn (1994), sect. 1.2.22 The standard example for a sentence which is a priori, but not analytic is "I exist now". Whether any of Kant's arguments for his synthetic truths a priori can be made good is questionable.23 One may also make sense of a linguistic or communal a priori, but the subjective one is certainly the primary notion. For a way to account for the communal as well as for the subjective level within the framework of Kaplan and Stalnaker see Haas-Spohn (1994), section 3.9. Of course, my earlier claim about the relative strength of analyticity and apriority holds only when both notions are taken at the same level.24 This is forcefully argued by Burge (1979) who builds his far-reaching doctrine of anti-individualism on this fact.independent from any information, evidence, or experience. On the other hand, there may always arise the need to revise the belief in an a posteriori proposition because it is dependent on information and adopted only after some experience which may have been misleading, which may be amended or superseded by further evidence; it may therefore be disadvantageous to stick to it come what may. Indeed, this postulate is just a generalized version of the regularity axiom of Carnap's inductive logic25 which is widely held to be reasonable. But note that it is only this rationality postulate which positively connects the two parts of my talk. As far as the analysis of certainty is concerned, the second part was only negatively connected with the first part, namely by the warning that the notions discussed in the second part be not confused with certainty.Much more could be said about certainty, but hardly more in 20 minutes. I hope what I said was not already too dense for this brief time. Thank you for your attention!ReferencesAlchourrón, C.E., P. Gärdenfors, D. Makinson (1985), "On the Logic of Theory Change: Partial Meet Functions for Contraction and Revision", Journal of Symbolic Logic 50, 510-530 Burge, T. (1979), "Individualism and the Mental", in: P.A. French, T.E. Uehling jr., H.K.Wettstein (eds.), Midwest Studies in Philosophy IV, Metaphysics, University of Minnesota Press, Minneapolis, pp. 73-121Carnap, R. (1971), "A Basic System of Inductive Logic, Part I", in: R. Carnap, R.C. Jeffrey (eds.), Studies in Inductive Logic and Probability , Vol. I, University of California Press, Berkeley, pp. 33-165Casullo, A. (1988), "Revisability, Reliabilism, and A Priori Knowledge", Philosophy and Penomenological Research 49, 187-213Cohen, L.J., (1977), The Probable and the Provable, Clarendon Press, OxfordDempster, A.P. (1967), "Upper and Lower Probabilities Induced by a Multivalued Mapping", Annals of Mathematical Statistics 38,325-33925 Cf. Carnap (1971), sect. 7. Since Carnap's axiom referred to logically true or analytic propositions, whereas my postulate refers only to a priori propositions, I have slightly generalized Carnap's axiom, in a way Carnap would have agreed to if he had had an independent notion of apriority.Dubois, D., H. Prade (1988), Possibility Theory. An Approach to the Computerized Processing of Uncertainty, Plenum Press, New Yorkvan Fraassen, B.C. (1989), Laws and Symmetry, Oxford University Press, OxfordGärdenfors, P. (1988), Knowledge in Flux, MIT Press, Cambridge, Mass.Gärdenfors, P., H. Rott (1994), "Belief Revision", in: D.M. Gabbay, C.J. Hogger, J.A. Robinson (eds.), Handbook of Logic in AI and Logic Programming. Vol. IV: Epistemic and Temproal Reasoning, Oxford University Press, forthcomingGoldszmidt, M., J. Pearl (1992), "Rank-Based Systems: A Simple Approach to Belief Revision, Belief Update, and Reasoning About Evidence and Actions", Proceedings of the Third International Conference on Principles of Knowledge Representation and Reasoning, Cambridge, Mass.Haas-Spohn, U. (1994), Versteckte Indexikalität und subjektive Bedeutung, Dissertation, TübingenHunter, D. (1991), "Maximum Entropy Updating and Conditionalization", in: W. Spohn, B.C. van Fraassen, B. Skyrms (eds.), Existence and Explanation. Essays in Honor of Karel Lambert, Kluwer, Dordrecht, pp. 45-57Jeffrey, R.C. (1965), The Logic of Decision, University of Chicago Press, Chicago, 2. Aufl. 1983Kaplan, D. (1977), "Demonstratives", in: J. Almog, J. Perry, H. Wettstein (eds.), Themes from Kaplan, Oxford University Press, Oxford 1989, pp. 481-563Kaplan, D. (1989), "Afterthoughts", in: J. Almog, J. Perry, H. Wettstein (eds.), Themes from Kaplan, Oxford University Press, Oxford, pp. 565-614Kitcher, P. (1980), "A Priori Knowledge", The Philosophical Review 89, 3-23Kripke, S.A. (1972), "Naming and Necessity", in: D. Davidson, G. Harman (eds.), Semantics of Natural Language, Reidel, Dordrecht, pp. 253-355, 763-769Levi, I. (1980), The Enterprise of Knowledge. An Essay on Knowledge, Credal Probability, and Chance, MIT Press, Cambridge, Mass.Lewis, D. (1969), Convention: A Philosophical Study, Harvard University Press, Cambridge, Mass.Lewis, D. (1980), "A Subjectivist's Guide to Objective Chance", in R.C. Jeffrey (ed.), Studies in Inductive Logic and Probability, Vol. II, University of California Press,Los Angeles, pp. 263-293Putnam, H. (1975), Mind, Language, and Reality. Philosophical Papers, Vol. 2, Cambridge University Press, CambridgePutnam, H. (1983), Realism and Reason. Philosophical Papers, Vol. 3, Cambridge University Press, CambridgeQuine, W.V.O. (1951), "Two Dogmas of Empiricism", Philosophical Review 60, 20-43Quine, W.V.O. (1960), Word and Object, MIT Press, Cambridge, Mass.Rescher, N. (1976), Plausible Reasoning, Van Gorcum, AssenSavage, L.J. (1954), The Foundations of Statistics, Wiley, New York10Shackle, G.L.S. (1969), Decision, Order, and Time in Human Affairs, Cambridge University Press, CambridgeShafer, G. (1976), A Mathematical Theory of Evidence, Princeton University Press, Princeton: Shafer, G. (1978), "Non-Additive Probabilities in the Work of Bernoulli and Lambert", Archive for the History of Exact Sciences 19, 309-370Shafer, G. (1990), "Perspectives on the Theory and Practice of Belief Functions", International Journal of Approximate Reasoning 4, 323-362Shafer, G. (1992), "Rejoinders to Comments on 'Perspectives on the Theory and Practice of Belief Functions'", International Journal of Approximate Reasoning 6, 445-480Spohn, W. (1978), Grundlagen der Entscheidungstheorie, Scriptor, Kronberg/Ts.Spohn, W. (1988), "Ordinal Conditional Functions. A Dynamic Theory of Epistemic States", in: W.L. Harper, B. Skyrms (eds.), Causation in Decision, Belief Change, and Statistics, Vol.II, Kluwer, Dordrecht, pp. 105-134Spohn, W. (1990), "A General Non-Probabilistic Theory of Inductive Reasoning", in: R.D.Shachter, T.S. Levitt, J. Lemmer, L.N. Kanal (eds..), Uncertainty in Artificial Intelligence 4, Elsevier, Amsterdam, pp. 149-158Spohn, W. (1993), "Causal Laws are Objectifications of Inductive Schemes", in: J. Dubucs (Hg.), Philosophy of Probability, Kluwer, Dordrecht 1993, pp. 223-252Stalnaker, R.C. (1978), "Assertion", in: P. Cole (ed.), Syntax and Semantics Vol. 9: Pragmatics, Academic Press, New York, pp. 315-332Stalnaker, Robert C. (1987), "Belief Attribution and Context", in: R. Grimm, D. Merrill (eds.), Contents of Thought, Tucson, pp. 140-156。

Author to whom correspondence should be addressed

Author to whom correspondence should be addressed

Keywords: dependency bounds, Dempster-Shafer belief functions, interval probabilities, uncertainty propagation
2
Introduction In this paper we address the general problem of performing convolutions of real-valued continuous random variables under binary operations, when there exists variation and uncertainty in the constituent random variables and uncertainty in the dependency between them. Suppose we wish to add, subtract, multiply or divide two continuous real-valued random variables, A and B, to produce a new random variable. Denoting the binary operation by *, we wish to describe the distribution of A*B from the distributions of A and B. The general formulation of the distribution for A*B, given marginal distributions for A and B and assuming A and B are independent, can be described by the following convolution:

D-S证据理论

D-S证据理论
[2] Dempster, A. P. Generalization of Bayesian Inference. Journal of the Royal Statistical Society. Series B 30, 1968:205-247.
[3] Shafer, G. A Mathematical Theory of Evidence. Princeton University Press, 1976. 【证据理论的第一本专著,标志其正式成为一门理论】
适用领域:信息融合、专家系统、情报分析、法律 案件分析、多属性决策分析,等等。
4、证据理论的局限性
要求证据必须是独立的,而这有时不易满足
证据合成规则没有非常坚固的理论支持,其合理 性和有效性还存在较大的争议
计算上存在着潜在的指数爆炸问题
5、证据理论的发展概况
“Zadeh悖论”:对证据理主要参考文献(续3)
[15] Josang, A. The consensus operator for combining beliefs. Artificial Intelligence, 2002, 141(1-2): 157-170.
[16] Yang, Jian-Bo, Xu, Dong-Ling. On the evidential reasoning algorithm for multiple attribute decision analysis under uncertainty. IEEE Transaction on Systems, Man, and Cybernetics – Part A: Systems and Humans, 2002, 32(3): 289-304.
0.99

心理学专业术语

心理学专业术语

感觉记忆(SM)—sensory memory短期记忆(STM)—short-term M.长期记忆(LTM)—long-term memory复诵——rehearsal预示(激发)——priming童年失忆症——childhood amnesia视觉编码(表征)——visual code(representation)听觉编码—acoustic code运作记忆——working memory语意性知识—semantic knowledgeji忆扫瞄程序—memory scanning procedure竭尽式扫瞄程序-exhaustive S.P.自我终止式扫瞄—self-terminated S.程序性知识—procedural knowledge命题(陈述)性知识——propositional(declarative)knowledge 情节(轶事)性知识—episodic K.讯息处理深度—depth of processing精致化处理—elaboration登录特殊性—coding specificity记忆术—mnemonic位置记忆法—method of loci字钩法—peg word(线)探索(测)(激发)字—prime关键词——key word命题思考——propositional thought心像思考——imaginal thought行动思考——motoric thought概念——concept原型——prototype属性——property特征——feature范例策略——exemplar strategy语言相对性(假说)—linguistic relativity th.音素——phoneme词素——morpheme(字词的)外延与内涵意义—denotative & connotative meaning (句子的)表层与深层结构—surface & deep structure语意分析法——semantic differential全句语言—holophrastic speech过度延伸——over-extension电报式语言—telegraphic speech关键期——critical period差异减缩法——difference reduction方法目的分析——means-ends analysis倒推——working backward动机——motive自由意志——free will决定论——determinism本能——instinct种属特有行为——species specific驱力——drive诱因——incentive驱力减低说——drive reduction th.恒定状态(作用)—homeostasis原级与次级动机—primary & secondary M.功能独立—functional autonomy下视丘侧部(LH)—lateral hypothalamus脂肪细胞说——fat-cell theory.下视丘腹中部(VMH)—ventromedial H定点论——set point th.CCK───胆囊调节激素第一性征——primary sex characteristic第二性征——secondary sex characteristic自我效能期望—self-efficiency expectancy内在(发)动机—intrinsic motive外在(衍)动机—extrinsic motive成就需求——N. achievement需求层级—hierarchy of needs自我实现——self actualization冲突——conflict多项仪——polygraph肤电反应——GSR(认知)评估——(cognitive appraisal)脸部回馈假说——facial feedback hypothesis(生理)激发——arousal挫折-攻击假说——frustration-aggression hy.替代学习——vicarious learning发展——development先天——nature后天——nurture成熟——maturation(视觉)偏好法——preferential method习惯法——habituation视觉悬崖——visual cliff剥夺或丰富(环境)——deprivation or enrichment of env. 基模——schema同化——assimilation调适——accommodation平衡——equilibrium感觉动作期——sensorimotor stage物体永久性——objective permanence运思前期——preoperational st.保留概念——conservation道德现实主义——moral realism具体运思期——concrete operational形式运思期——formal operational st.前俗例道德——pre-conventional moral俗例道德——conventional moral超俗例道德——post-conventional moral气质——temperament依附——attachment性别认定——gender identity性别配合——sex typing性蕾期——phallic stage恋亲冲突—Oedipal conflict认同——identification社会学习——social learning情结——complex性别恒定——gender constancy青年期——adolescence青春期—— -puberty第二性征——secondary sex characteristics 认同危机——identity crisis定向统合——identity achievement早闭型统合——foreclosure未定型统合——moratorium迷失型统合——identity diffusion传承——generativity心理动力——psycho-dynamics心理分析——psychoanalysis行为论——behaviorism心理生物观——psycho-biological perspective 认知——cognition临床心理学家-clinical psychologist谘商——counseling人因工程——human factor engineering组织——organization潜意识——unconsciousness完形心理学——Gestalt psychology感觉——sensation知觉——perception实验法——experimental method独变项——independent variable依变项——dependent V.控制变项——control V.生理——physiology条件化——conditioning学习——learning比较心理学——comparative psy.发展——development社会心理学——social psy.人格——personality心理计量学—psychometrics受试(者)——subject 实验者预期效应—experimenter expectancy effect 双盲法——double—blind实地实验——field experiment相关——correlation调查——survey访谈——interview个案研究——case study观察——observation心理测验——psychological test纹理递变度——texture gradient注意——attention物体的组群——grouping of object型态辨识—pattern recognition形象-背景——figure-ground接近律——proximity相似律——similarity闭合律——closure连续律——continuity对称律——symmetry错觉——illusion幻觉——delusion恒常性——constancy大小——size形状——shape位置—— location单眼线索——monocular cue线性透视——linear- perspective双眼线索——binocular cue深度——depth调节作用——accommodation重迭——superposition双眼融合——binocular fusion辐辏作用——convergence双眼像差——binocular disparity向度—— dimension自动效应——autokinetic effect运动视差—— motion parallax诱发运动—— induced motion闪光运动—— stroboscopic motion上下文﹑脉络-context人工智能——artificial intelligence A.I. 脉络关系作用-context effect模板匹配——template matching整合分析法——analysis-by-synthesis丰富性——redundancy选择性——selective无yi识的推论-unconscious inferences运动后效——motion aftereffect特征侦测器—feature detector激发性——excitatory抑制性——inhibitory几何子——geons由上而下处理—up-down process由下而上处理——bottom-up process连结者模式——connectionist model联结失识症——associative agnosia脸孔辨识困难症——prosopagnosia意识——conscious(ness)意识改变状态——altered states of consciousness无意识——unconsciousness前意识——preconsciousness内省法——introspection边缘注意——peripheral attention多重人格——multiple personality午餐排队(鸡尾酒会)效应—lunch line(cocktail party)effect 自动化历程——automatic process解离——dissociate解离认同失常——dissociative identity disorder快速眼动睡眠——REM dream非快速眼动睡眠—NREM dream神志清醒的梦——lucid dreaming失眠——insomnia显性与隐性梦——manifest & latern content心理活动性psychoactive冥想——meditation抗药性——tolerance戒断——withdrawal感觉剥夺——sensory deprivation物质滥用——substance abuse成瘾——physical addiction物质依赖——sub. dependence戒断症状——withdrawal symptom兴奋剂——stimulant幻觉(迷幻)剂——hallucinogen镇定剂——sedative抑制剂——depressant酒精中毒引起谵妄—delirium tremens麻醉剂——narcotic催眠——hypnosis催眠后暗示——posthypnotic suggestion 催眠后失忆posthypnotic amnesia超心理学——parapsychology超感知觉extrasensory perception ESP 心电感应——telepathy超感视——clairvoyance预知——precognition心理动力—psycokinesis PK受纳器——receptor绝对阈——absolute threshold差异阈——difference threshold恰辨差——-JND韦伯律——Weber''s law心理物理——psychophysical费雪纳定律——Fechner''s law频率——frequency振幅——amplitude音频——pitch基音——fundamental tone倍音——overtone和谐音——harmonic音色——timbre白色噪音——white noise鼓膜——eardrum耳蜗——cochlea卵形窗—oval window圆形窗——round window前庭——vestibular sacs半规管——semicircular canal角膜——cornea水晶体——lens虹膜——iris瞳孔——pupil网膜——retina睫状肌——ciliary muscle调节作用——accommodation脊髓——spinal cord反射弧——reflex arc脑干——brain stem计算机轴性线断层扫描——CAT或CT PET——正子放射断层摄影MRI——磁共振显影延脑——medulla桥脑——pons小脑——cerebellum网状结构——reticular formation RAS——网状活化系统视丘——thalamus下视丘——hypothalamus大脑——cerebrum脑(下)垂体(腺)—pituitary gland脑半球——cerebral hemisphere皮质——cortex胼胝体——corpus callosum边缘系统——limbic system海马体——hippocampus杏仁核——amygdala中央沟——central fissure侧沟——lateral fissure脑叶——lobe同卵双生子——identical twins异卵双生子—fraternal twins古典制约——classical conditioning操作制约——operant conditioning非制约刺激—(US unconditioned stimulus 非制约反应—(UR)unconditioned R.制约刺激——(CS)conditioned S.制约反应——(CR)conditioned R.习(获)得——acquisition增强作用——reinforcementxiao除(弱)——extinction自(发性)然恢复——spontaneous recovery 前行制约—forward conditioning同时制约——simultaneous conditioning回溯制约——backward cond.痕迹制约——trace conditioning延宕制约—delay conditioning类化(梯度)——generalization(gradient)区辨——discrimination(次级)增强物——(secondary)reinforcer嫌恶刺激——aversive stimulus试误学习——trial and error learning效果率——law of effect正(负)性增强物—positive(negative)rei.行为塑造—behavior shaping循序渐进——successive approximation自行塑造—autoshaping部分(连续)增强—partial(continuous)R定比(时)时制—fixed ratio(interval)schedule FR或FI变化比率(时距)时制—variable ratio(interval)schedule VR或VI逃离反应——escape R.回避反应—avoidance response习得无助——learned helplessness顿悟——insight学习心向—learning set隐内(潜在)学习——latent learning认知地图——cognitive map生理回馈——biofeedback敏感递减法-systematic desensitization普里迈克原则—Premack''s principle洪水法——flooding观察学习——observational learning动物行为学——ethology敏感化—sensitization习惯化——habituation联结——association认知学习——cognitional L.观察学习——observational L.登录﹑编码——encoding保留﹑储存——retention提取——retrieval回忆——(free recall全现心像﹑照相式记忆——eidetic imagery﹑photographic memory .舌尖现象(TOT)—tip of tongue再认——recognition再学习——relearning节省分数——savings外显与内隐记忆——explicit & implicit memory记忆广度——memory span组集——chunk序列位置效应——serial position effect起始效应——primacy effect新近效应——recency effect心(情)境依赖学习——state-dependent L.无意义音节—nonsense syllable顺向干扰——proactive interference逆向干扰——retroactive interference闪光灯记忆——flashbulb memory动机性遗忘——motivated forgetting器质性失忆症—organic amnesia阿兹海默症——Alzheimer''s disease近事(顺向)失忆症—anterograde amnesia旧事(逆向)失忆—retrograde A.高沙可夫症候群—korsakoff''s syndrome凝固理论—consolidation。

高中英语名词性从句作同位语单选题40题

高中英语名词性从句作同位语单选题40题

高中英语名词性从句作同位语单选题40题1.He received the news that his friend had become a famous scientist.A.whichB.whatC.thatD.who答案:C。

本题考查同位语从句。

that 引导同位语从句,解释说明news 的具体内容,在从句中不充当任何成分。

选项A which 通常在定语从句中使用;选项B what 在名词性从句中要充当成分;选项D who 通常用来指人。

2.The fact that she is a talented singer is known to everyone.A.whichB.whatC.thatD.who答案:C。

that 引导同位语从句,解释说明fact 的内容,在从句中不充当成分。

选项 A which 在定语从句中使用;选项B what 在名词性从句中充当成分;选项D who 指人。

3.We were excited at the news that our team had won the championship.A.whichB.whatC.thatD.who答案:C。

that 引导同位语从句,说明news 的具体内容,在从句中不充当成分。

选项 A which 用于定语从句;选项B what 在名词性从句中充当成分;选项D who 指人。

4.The report that he was seriously ill was not true.A.whichB.whatC.thatD.who答案:C。

that 引导同位语从句,解释report 的内容,在从句中不充当成分。

选项A which 在定语从句中使用;选项B what 在名词性从句中充当成分;选项D who 指人。

5.The rumor that she is going to marry a rich man spread quickly.A.whichB.whatC.thatD.who答案:C。

井冈山时期毛泽东革命话语的认知、实践、信仰功能

井冈山时期毛泽东革命话语的认知、实践、信仰功能

The Cognitive,Practice and Belief Functions of Mao Zedong’s Revolutionary Discourse in the Jinggang
Mountains Period
作者: 杨帆[1]
作者机构: [1]井冈山大学井冈山研究中心,江西吉安343009
出版物刊名: 理论月刊
页码: 12-20页
年卷期: 2020年 第1期
主题词: 井冈山时期;革命话语;毛泽东;认知;实践;革命信仰
摘要:井冈山斗争时期,毛泽东以革命话语作为媒介,引领话语受众认知、实践与信仰中共革命。

通过革命话语认知功能,毛泽东建构了国民党反革命的形象,树立了推翻其统治的正当性与合法性;通过革命话语实践功能,毛泽东引导话语受众积极参与革命实践;通过革命话语信仰功能,毛
泽东塑造话语受众革命价值观,进而形成革命信仰。

在毛泽东革命话语的训诫与规训之下,其革命话语建构了话语受众的话语想象,最终将中共革命理念导入话语受众的精神世界。

一种新的基于证据距离的证据合成算法

一种新的基于证据距离的证据合成算法

一种新的基于证据距离的证据合成算法翟海天;李辉;潘凯【摘要】针对D-S证据理论对于高度冲突证据合成时出现与常理相悖结论的问题,提出了一种基于证据距离的加权证据合成方法,并给出了具体算法.新算法仍然使用D-S合成公式,但需考虑融合系统中证据源的修正问题.利用证据总体的期望值将Jousselme提出的证据距离修正,从而得出证据权,通过证据权对源证据的基本可信度分配函数进行修正,并通过算例比较,说明新的算法具有更有效的合成结果.%Since Dempster-Shafer (DS) theory involves counter-intuitive behaviors when evidences highly conflict, a new approach of combination of weighted belief functions based on evidence distance is proposed. The new approach still uses DS combination formula, but considers the revision means which fuses the evidence source of the system. Jousselme evidence distance is revised by making use of expectation of all evidences, and then obtains the evidence weight. The basic beliability partition function of source evidence is revised through evidence weight. The numerical examples are given to illustrate the effectiveness and the good performance of the new approach in case of high conflict between evidences.【期刊名称】《现代电子技术》【年(卷),期】2011(034)015【总页数】4页(P11-14)【关键词】证据合成;证据距离;证据权;冲突证据【作者】翟海天;李辉;潘凯【作者单位】西北工业大学电子信息学院,陕西西安 710072;西北工业大学电子信息学院,陕西西安 710072;西北工业大学电子信息学院,陕西西安 710072【正文语种】中文【中图分类】TN911.7-340 引言D-S证据理论[1-2]是一种不确定性推理方法,近年来倍受关注。

相关主题
  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

CHAPTER 7BELIEF FUNCTIONSINTRODUCTIONbyGlen n ShaferThe theory of belief functions provides a non-Bayesian way of using mathematical probability to quantify subjective judgements. Whereas a Bayesian assesses probabilities directly for the answer to a question of interest, a belief-function user assesses probabilities for related questions and then considers the implications of these probabilities for the question of interest.Though antecedents for belief functions can be found in the seventeenth and eighteenth centuries, the theory in its present form is due to the work of A.P. Dempster in the 1960s and my own work in the 1970s. For this reason, it is sometimes called the “Dempster-Shafer theory.” It came to the attention of workers in artificial intelligence in the 1980s, in part because of its resemblance to the less systematic calculus of certainty factors developed for MYCIN by Edward Shortliffe and Bruce Buchanan (see Chapter 5).My 1976 book, A Mathematical Theory of Evidence, remains the most comprehensive single reference for the mathematical theory of belief functions, but it has been followed by a large literature on interpretation, application, and computation. This chapter presents a selection from that literature.The first two articles in the chapter represent recent thinking by Dempster, myself, and our co-workers. The article by Rajendra Srivastava and me introduces the belief-function formalism and compares it with the Bayesian formalism in the context of an important application, financial auditing. The article by Dempster and Augustine Kong illustrates how belief functions can be used in belief networks.The third and fourth articles, the article by Jean Gordon and Edward Shortliffe, and the article by Judea Pearl, provide alternative introductions to belief functions. Gordon and Shortliffe present the theory in the context of a medical application, and they compare it with the framework and rules of combination of MYCIN. Pearl's article takes a more critical look at belief functions.The fifth article, by Prakash Shenoy and me, extends to belief functions the methods of local computation in belief networks that are presented for Bayesian networks in Chapter 6. Shenoy and I show that these methods apply to a broad class of formalisms, including the Bayesian formalism, belief functions, Wolfgang Spohn's natural conditional functions, and any other formalism that has operations of combination and marginalization that satisfy certain simple axioms.The last two articles describe consultation systems that help users represent and combine belief functions. The article by John Lowrance, Thomas Garvey, and Thomas Strat describes an early version of the Gister system, which helps users build up belief networks graphically. This article is significant not only because of Gister itself, but also because Lowrance and his co-workers at SRI International were influential in bringing belief functions into artificial intelligence. The article by Debra Zarley, Yen-Teh Hsia, and me describes DELIEF, a system developed at the University of Kansas. This system goes beyond the early versions of Gister by using the mathematics of belief networks to carry out the combination automatically.This introduction surveys some of the issues in the belief-function literature and provides references for further reading. For a fuller survey of current issues, see Shafer (1990).1. The Basics of Be lie f-Function The oryThe theory of belief functions is based on two ideas: the idea of obtaining degrees of belief for one question from subjective probabilities for a related question, and Dempster's rule for combining such degrees of belief when they are based on independent items of evidence.These ideas are illustrated by example in several of the articles in this chapter. The simplest, perhaps, is the example of testimony discussed by Srivastava and me. We can derive degrees of belief for statements made by witnesses from subjective probabilities for the reliability of these witnesses.Degrees of belief obtained in this way differ from probabilities in that they may fail to add to 100%. Suppose, for example, that Betty tells me a tree limb fell on my car. My subjective probability that Betty is reliable is 90%; my subjective probability that she is unreliable is 10%. Since they are probabilities, these numbers add to 100%. But Betty's statement, which must be true if she is reliable, is not necessarily false if she is unreliable. From her testimony alone, I can justify a 90% degree of belief that a limb fell on my car, but only a 0% (not 10%) degree of belief that no limb fell on my car. (This 0% does not mean that I am sure that no limb fell on my car, as a 0% probability would; it merely means that Betty's testimony gives me no reason to believe that no limb fell on my car.) The 90% and the 0%, which do not add to 100%, together constitute a “belief function.”In this example, we are dealing with a question that has only two answers (Did a limb fall on my car? Yes or no.). Belief functions can also be derived for questions for which there are more than two answers. In this case, we will have a degree of belief for each answer and for each set of answers. If the number of answers (or the size of the “frame”) is large, the belief function may be very complex.Dempster's rule is based on the standard idea of probabilistic independence, applied to the questions for which we have subjective probabilities. I can use the rule to combine evidence from two witnesses if I consider the first witness's reliability subjectively independent (before I take account of what the witnesses say) of the second's reliability. (This means that finding out whether one witness is reliable would not change my subjective probability for whether the other is reliable.) The rule uses this subjective independence to determine joint probabilities for various possibilities as to which of the two are reliable.Though it begins with an initial judgment of independence, Dempster's rule goes beyond this independence. After using independence to compute joint probabilities for who is reliable, I must check whether some possibilities are ruled out by what the witnesses say. (If Betty says that a tree limb fell on my car, and Sally says nothing fell on my car, then they cannot both be reliable.) If so, I renormalize the probabilities of the remaining possibilities so they add to one. This is an example of probabilistic conditioning, and it may destroy the initial independence. (After I notice that Betty and Sally have contradicted each other, their reliabilities are no longer subjectively independent for me. Now finding out that one is reliable would tell me that the other is not.) Then I determine what each possibility for the reliabilities implies about the truth of what the witnesses said, and I use the renormalized probabilities to get new degrees of belief.The net effect of Dempster's rule is that concordant items of evidence reinforce each other (two independent witnesses for a limb falling on my car make me believe it more than either alone), conflicting items of evidence erode each other, and a chain of reasoning is weaker than its weakest link. Section 2.3 of the article by Srivastava and me illustrates all these points by example.Bayesian probability measures qualify as belief functions; they represent, in effect, the special case where the unreliable witness always lies. Another special case is that of categorical, completely non-probabilistic information. This is the case where we are 100% confident in the reliability of a witness or other evidence, yet this evidence does not completely answer our question. I might have conclusive evidence, for example, that I lost my wallet in one of three places, without any clue as to which one. This calls for a belief function that assigns a degree of belief of 100% to the three places as a set, but a degree of belief of 0% to each of the three individually.As Dempster and Kong emphasize in their article in this chapter, the ability to represent both probabilistic and categorical information makes belief functions a bridge between Bayesian and categorical reasoning. Bayesian conditioning itself can be understood in terms of this bridge. Conditioning a Bayesian probability measure on given information is equivalent to combining it, by Dempster's rule, with a categorical belief function representing that information (Shafer 1976, pp. 66-67).2. Belief Functions Do Not Express Lower Bounds on True but Unknown ProbabilitiesMathematically, the degrees of belief given by a single belief function can be related to lower bounds on probabilities, but conceptually they must be sharply distinguished from such lower bounds. If we make up numbers by thinking of them as lower bounds on true probabilities, and we then combine these numbers by Dempster's rule, we are likely to obtain erroneous and misleading results.It is easy to see the temptation to interpret belief-function degrees of belief as lower bounds on unknown true probabilities. Consider again my 90% belief that a limb fell on my car, and my 0% belief that no limb fell on my car. These degrees of belief were derived from my 90% and 10%subjective probabilities for Betty being reliable or unreliable. Suppose these subjective probabilities were based on my knowledge of the frequency with which witnesses like Betty are reliable. Then I might think that the 10% of witnesses like Betty who are not reliable make true statements a definite (though unknown) proportion of the time and false statements the rest of the time. Were this the case, I could think in terms of a large population of statements made by witnesses like Betty. In this population, 90% of the statements would be true statements by reliable witnesses, x% would be true statements by unreliable witnesses, and (10-x)% would be false statements by unreliable witnesses, where x is an unknown number between 0 and 10. The total chance of getting a true statement from this population would be (90+x)%, and the total chance of getting a false statement would be(10-x)%. My degrees of belief of 90% and 0% are lower bounds on these chances; since x is anything between 0 and 10, 90% is the lower bound for (90+x)%, and 0% is the lower bound for (10-x)%.As this example suggests, there is a sense in which a single belief function can always be interpreted as a consistent system of probability bounds. It is always possible to find a probability distribution such that each probability is greater than the corresponding degree of belief given by the belief function.The fallaciousness of the probability-bound interpretation of belief functions becomes clear, however, when we consider two or more belief functions addressing the same question but representing different and possibly conflicting items of evidence. The disagreements that such belief functions represent are not disagreements about the values of true probabilities. When Betty says a limb fell on my car, and Sally says nothing fell on my car, they are disagreeing about whether something fell on my car, not about the true probability of something having fallen on my car.Were we to insist on a probability-bound interpretation of belief functions, then we would only be interested in groups of belief functions whose degrees of belief, when interpreted as probability bounds, can be satisfied simultaneously. When belief functions are given their proper interpretation, however, it is of no particular significance whether there exist probabilities that simultaneously satisfy the bounds defined by a whole group of belief functions. Consider two cases that might arise when we use belief functions to represent contradictory evidence from Betty and Sally:Case 1. Before hearing their testimony, we think highly of the reliability of both Betty and Sally. We represent Betty's evidence by a belief function that gives a95% degree of belief to a limb having fallen on my car, and we represent Sally'sevidence by a belief function that gives a 95% degree of belief to nothing havingfallen on my car. In this case, the two belief functions are contradictory asprobability bounds; if the true probability of a limb having fallen on my car is greaterthan 95%, then the true probability of nothing having fallen on my car cannot also begreater than 95%.Case 2. Before hearing their testimony, we think that both Betty and Sally are fairly unreliable. So in both belief functions, we assign a 35% degree of belief ratherthan a 95% degree of belief. In this case, the two belief functions define consistentprobability bounds; the true probability of a limb having fallen on my car and ofnothing having fallen on my car can both be greater than 35%.From the belief-function point of view, there is no conceptual difference between these two cases. In both cases, we can combine the two belief functions by Dempster's rule. In both cases, there is conflict in the evidence being combined, and normalization is required.It can be shown that if no renormalization is required in the combination of a group of belief functions by Dempster's rule, then there do exist consistent probabilities that simultaneously bound all the belief functions being combined as well as the belief function that results from the combination. We cannot count on this, however, when renormalization is required. Consequently, authors who favor a probability-bound interpretation of belief functions are uncomfortable with renormalization (see, e.g., Zadeh 1986).Probability bounds provide the basis for yet another mathematical theory of evidence, which I have called the “theory of lower probability” (Shafer 1981). In this theory, an analogy is drawn between actual evidence and knowledge of bounds on unknown true probabilities for the question of interest. This theory is not always useful, because unknown true probabilities exist only if a population and sampling scheme are well defined. An unknown true probability for the truth of Betty's statement, for example, exists only if a the population of true and false statements of witnesses like Betty is well-defined. In a problem where a reference population for the question of interest is well-defined, the theory of lower probability may be more useful than the theory of belief functions. But in other problems belief functions may be more useful.As Shafer and Tversky explain in our article in Chapter 2, both Bayesian and belief-function arguments involve analogies to sampling situations. But belief-function analogies are less complete than Bayesian analogies. They are useful when it is reasonable to evaluate certain evidence (e.g., Betty's reputation) using a sampling analogy, but this evidence will not support extending the analogy to all potentially relevant issues (what Betty would say were my good impression of her erroneous, how often limbs fall from that tree, etc.).There has been some confusion about the original relation between belief functions and probability bounds, because some of Dempster's early articles hinted at a probability-bound interpretation. But Dempster's “upper and lower probabilities” were not derived by bounding unknown true probabilities. My 1976 book, in which the term “belief function” was introduced, explicitly disavowed any probability-bound interpretation (p. ix). This disavowal was elaborated at length in Shafer (1981) and seconded by Dempster (1982).3. General Metaphors and Canonical ExamplesThe metaphor of the witness who may or may not be reliable can serve as a standard of comparison, or canonical example, for judging the strength of other evidence. We can assess given evidence by saying that it is comparable in strength to the evidence of a witness who has a certain chance of being reliable.A witness testifying to a specific proposition leads to a relatively simple belief function—one that gives a specified degree of belief to that proposition and its consequences, and zero degree of belief to all other propositions. Arbitrarily complex belief functions can be built up by combining such simple belief functions (Shafer 1976, p. 200), but in some cases we may want to produce complex belief functions more directly, in order to represent evidence that conveys a complex or mixed message but cannot be broken down into independent components. This requires more complex metaphors or canonical examples.Two distinct general metaphors have been suggested. Shafer (1981) suggests the metaphor of a randomly coded message. Pearl (1988) suggests the metaphor of random switches.Shafer's randomly coded messages are explained in detail in the article by Shafer and Tversky in Chapter 2. In this metaphor, we have probabilities for which of several codes was used to encode a message. We do not yet know what the message says, but we know it is true. We have this message in hand in its coded form, and we will try to decode it using each code, but the probabilities are judgments we make before this decoding. When we do decode using the different codes, we sometimes get nonsense, and we sometimes get a comprehensible statement. It seems sensible, in this situation, to condition our probabilities for the codes by eliminating the ones with which we get nonsense. The conditioned probability for each remaining code can then be associated with the statement we get by decoding using that code. These statements may be related in various ways; some may be inconsistent with each other, and some may be stronger than others. Thus we obtain the complexity of an arbitrary belief function.In this metaphor, the independence of two belief functions means that two different people independently choose codes with which to send two possibly different (though both true) messages. Our uncertainties about the codes in the two cases remain independent unless possible codes implycontradictory messages. If s1 is a possible code for the first person, and s2is a possible code for thesecond person, and the first message as decoded by s1contradicts the second message as decoded by s2, then it cannot be true that these were the two codes used. We eliminate such pairs of codes and renormalize the probabilities of the remaining possible pairs. The probability of each pair is then associated with the conjunction of the two implied messages. This is Dempster's rule.The metaphor can be presented in a way that forestalls the interpretation of belief-function degrees of belief in terms of bounds on probabilities. There is no probability model for the choice of the true message sent. The probabilities are only for the choice of codes. We might visualize these probabilities in terms of a repetition of the choice of codes, but since the true message can vary arbitrarily over this population of repetitions, the idea of this population does not lead to the idea of a true unknown probability for the true message or for the true answer to the question of interest. It leads only to an argument about what the true message says or implies—an argument whose strength can be represented in terms of the derived “degrees of belief.”Pearl's metaphor of random switches is explained in detail in his article in this chapter. In this metaphor, a switch oscillates randomly among propositions about the question of interest. Ourprobabilities are probabilities for the position of the switch, but when the switch points to a certain proposition, this indicates that the proposition is to be adopted as an assumption or axiom for further reasoning. A given proposition about the question of interest may be implied by several of these possible axioms, and its total degree of belief will be the total probability that an axiom implying it is adopted—the total probability, to speak more briefly, that it is proven.Pearl's metaphor seems well fitted for computer science, since it mixes the language of electrical engineering with that of symbolic logic. I find it difficult, however, to construe the metaphor in a way that completely avoids interpretation in terms of bounds on true probabilities. The random-code metaphor allows us to interpret the probabilities for a related question in terms of a population of repetitions completely unconnected with the question of interest. But since the switch positions in Pearl's metaphor are defined in terms of axioms about the question of interest, it seems to me that each repetition of the random selection of switch positions will generally constrain the true answer to the question of interest. Thus the probabilities do bear directly on the question of interest, and this leads to the objections to renormalization I discussed in the preceding section. In Pearl's view, however, his metaphor is compatible with renormalization (section 5.2 of his article in this section).4. Sorting Evide nce into Inde pe nde nt Ite msDempster's rule should be used to combine belief functions that represent independent items of evidence. But when are items of evidence independent? How can we tell? These are probably the questions asked most frequently about belief functions.The independence required by Dempster's rule is simply probabilistic independence, applied to the questions for which we have probabilities, rather than directly to the question of interest. In the metaphor of the randomly coded messages, this means that the codes are selected independently. In the more specialized metaphor of independent witnesses, it means that the witnesses (or at least their current properties as witnesses) are selected independently from well-defined populations.Whether two items of evidence are independent in a real problem is a subjective judgment, in the belief-function as in the Bayesian approach. There is no objective test.In practice, our task is to sort out the uncertainties in our evidence. When items of evidence are not subjectively independent, we can generally identify what uncertainties they have in common, thus arriving at a larger collection of items of evidence that are subjectively independent. Typically, this maneuver has a cost—it forces us to refine, or make more detailed, the frame over which our belief functions are defined.We can illustrate this by adapting an example from Pearl's “Bayes Decision Methods,” in Chapter 6. Suppose my neighbor Mr. Watson calls me at my office to say he has heard my burglar alarm. In order to assess this testimony in belief-function terms, I assess probabilities for the frameS= {Watson is reliable, Watson is not reliable}.1Here Watson being reliable means he is honest and he can tell whether it is my burglar alarm he is hearing. I can use these probabilities to get degrees of belief for the frameT = {My alarm sounded, My alarm did not sound}.Putting a probability of 90%, say, on Watson being reliable, I get a 90% degree of belief that my burglar alarm sounded, and a 0% degree of belief that my burglar alarm did not sound.I now call another neighbor, Mrs. Gibbons, who verifies that my alarm sounded. I can assess her testimony in the same way, by assessing probabilities for the frame= {Gibbons is reliable, Gibbons is not reliable}.S2Suppose I also put a probability of 95% on Gibbons being reliable, so that I again obtain a 95% degree of belief that my burglar alarm sounded, and a 0% degree of belief that it did not sound.Were I to combine these two belief functions by Dempster's rule, I would obtain an overall degree of belief of 99.5% that my burglar alarm sounded. This is inappropriate, however, for the two items of evidence involve a common uncertainty—whether there might have been some other noise similar to my burglar alarm.In order to deal with this problem, I must pull my skepticism about the possibility of a similar noise out of my assessment of Watson's and Gibbons' reliability, and identify my grounds for this skepticism as a separate item of evidence. So I now have three items of evidence—my evidence for Watson's honesty (I say honesty now instead of reliability, since I am not including here the judgment that there are no other potential noises in the neighborhood that Watson might confuse with my burglar alarm), my evidence for Gibbons' honesty, and my evidence that there are no potential noises in the neighborhood that sound like my burglar alarm.These three items of evidence are now independent, but their combination involves more than theframe T. In its place, we need the frame U = {u1,u2,u3}, whereu1= My alarm sounded,u2= There was a similar noise,u3= There was no noise.(Let us exclude, for simplicity of exposition, the possibility that there were two noises, my alarm and also a similar noise.) My first two items of evidence (my evidence for Watson's and Gibbons'honesty) both provide a high degree of belief in {u1,u2}, while the third item (my evidence against theexistence of other noise sources) provides a high degree of belief in {u1,u3}. Combining the three byDempster's rule produces a high degree of belief in {u1}.A Bayesian approach to this problem would be somewhat different, but it too would involve refining the frame T to U or something similar. In the Bayesian case, we would ask whether the events “Watson says he heard a burglar alarm” and “Gibbons says she heard a burglar alarm” are subjectively independent. They are not unconditionally independent, but they are independent conditional on a specification of what noise actually occurred. I can exploit this conditional independence in assessing my subjective probabilities, but in order to do so, I must bring the possibility of other noises into the frame.In the belief-function approach, one talks not about conditional independence of propositions, but rather about the overlapping and interaction of evidence. For further explanation and more examples, see Shafer (1976, Chapter 8), Shafer (1981), Shafer (1984), Shafer (1987) and Srivastava, Shenoy, and Shafer (1989).5. Fre que ncy ThinkingA common but dangerous temptation is to use Dempster's rule to combine opinions that are really fragments of information about a single probability distribution. This is usually inappropriate and can give misleading results.Suppose we are concerned with a bird Tweety. We want to know whether Tweety flies and whether Tweety is a penguin. We decide to make judgments about this by thinking of Tweety as randomly selected from a certain population of birds. We have guesses about the proportion of birds in this population that fly and the proportion that are penguins. Should these guesses be represented as belief functions over a set of statements about Tweety, and then combined by Dempster's rule? No. Both guesses bear on the particular bird only through their accuracy as guesses about the population. This means that they have in common the uncertainty involved in choosing Tweety at random from the population. Depending on how we obtained the guesses, they may also have other uncertainties in common.Like every problem of dependence, we can deal with this problem within the belief-function approach by sorting out the uncertainties and properly refining our frame. In this case, we must bring the possible values for the population frequencies into the frame. We can then formalize the connection between these frequencies and Tweety as one of our items of evidence. We must also identify our sources of evidence about the frequencies, sort our their uncertainties, and use them to assess belief functions about what these frequencies are.This brings us into the difficult realm of statistical inference. Belief functions provide only one of many approaches to statistical inference, and even the possible belief-function techniques are varied and complex (Dempster 1966, 1967a,b, 1968a,b, 1969; Chapter 11 of Shafer 1976; Shafer 1982a,b). No statistical approach is likely to be of much value unless considerable frequency information is available.An alternative is the Bayesian approach, in which (according to the constructive interpretation) we compare our hunches about Tweety to fairly precise knowledge of frequencies in a population. One of the messages of Pearl's article on probabilistic semantics in Chapter 9, with which I agree, is that many of the intuitions discussed in the literature on non-monotonic logic are best represented in this way, in spite of the authors' protestations that their intuitions are not probabilistic.In section 7 of his article in this chapter, Pearl gives a number of examples in which Dempster's rule gives inappropriate results. I believe the intuitions in most of these examples are based on frequency thinking, and I agree that Dempster's rule is not appropriate for such problems. If there is sufficient evidence on which to base detailed probability judgments in analogy to frequency knowledge, then the Bayesian approach will be far more useful. In some other cases, a lower。

相关文档
最新文档