Research of Face Location System Based on Human Vision Simulations

合集下载

香港理工大学的高光谱人脸数据库(PolyU-HSFD)_图像处理_科研数据集

香港理工大学的高光谱人脸数据库(PolyU-HSFD)_图像处理_科研数据集

香港理工大学的高光谱人脸数据库(PolyU-HSFD)(The Hong Kong Polytechnic University Hyperspectral Face Database (PolyU-HSFD))数据介绍:The Biometric Research Centre (UGC/CRC) at The Hong Kong Polytechnic University has developed a Hyperspectral Face database to advance research and to provide researchers working in the area of face recognition with an opportunity to compare the effectiveness of face recognition algorithms. The indoor hyperspectral face acquisition system was built which mainly consists of a CRI’s VariSpec LCTF and a Halogen Light (Illustrated in Fig. 1), and includes a hyperspectral dataset of 300 hyperspectral image cubes from 25 volunteers with age range from 21 to 33 (8 female and 17 male). For each individual, several sessions were collected with an average time space of 5 month. The minimal interval is 3 months and the maximum is 10 months. Each session consists of three hyperspectral cubes-- frontal, right and left views with neutral-expression. The spectral range is from 400nm to 720nm with a step length of 10nm, producing 33 bands in all. Fig. 2 shows an example of the hyperspectral face. Sinc关键词:香港理工大学,高光谱,人脸,UGC/CRC,识别,PolyU,Hyperspectral,Face,UGC/CRC,recognition,数据格式:IMAGE数据详细介绍:The Hong Kong Polytechnic University Hyperspectral Face Database(PolyU-HSFD)The Biometric Research Centre (UGC/CRC)at The Hong Kong Polytechnic University has developed a Hyperspectral Face database to advance research and to provide researchers working in the area of face recognition with an opportunity to compare the effectiveness of face recognition algorithms. The indoor hyperspectral face acquisition system was built which mainly consists of a CRI’s VariSpec LCTF and a Halogen Light (Illustrated in Fig. 1), and includes a hyperspectral dataset of 300 hyperspectral image cubes from 25 volunteers with age range from 21 to 33 (8 female and 17 male). For each individual, several sessions were collected with an average time space of 5 month. The minimal interval is 3 months and the maximum is 10 months. Each session consists of three hyperspectral cubes-- frontal, right and left views with neutral-expression. The spectral range is from 400nm to 720nm with a step length of 10nm, producing 33 bands in all. Fig. 2 shows an example of the hyperspectral face. Since the database was constructed over a long period of time, significant appearance variations of the subjects, e.g. changes of hair style and skin condition, are presented in the data. In data collection, positions of the camera, light and subject are fixed, which allows us to concentrate on the spectral characteristics for face recognition without masking from environmental changes.Fig. 1 The established hyperspectral face imaging systemFig. 2 Illustration of a set of 33 hyperspectral face bandsData description(A) Mat-fileEach Mat file is 3-D datacube with size: 220 (height) *180 (width) *33 (no. ofbands), cropped from the original images, which contains the radiance data from the original source .tif image.Three types of images were acquired: front, right, and left. The Mat-files are stored separately as:Front-Mat file: in the file of SampleImages_F, totally 151 datacubes.Right-Mat file: in the file of SampleImages_R, totally 125 datacubes.Left-Mat file: in the file of SampleImages_L, totally 124 datacubes.(B) Source DataSource data are documented according to the acquisition date. Totally four data sets (I to IV) were acquired at different time (10/17/2007, 11/21/2007-11/23/2007, 12/12/2007, and 3/21/2008). Some subjects may have several sessions at each time. Each session (e.g. Sample 17 from the 3/21/2008 data set) contains 3 sub-files: f-front image, r-right image and l-left image.For a detailed and completed data description, you are advised to read the DATA_TEXT and DATA_TABLE. Also, we manually located the eye coordinates from the front images (151 subjects) for registration and cropping. Here is the format:F_1_1.jpg61 93 118 89F_1_1.jpg: the name of the front face image61: the X-coordinate of the right eye (from the view of the sample face, R_x) 93: the Y-coordinate of the right eye (from the view of the sample face, R_y) 118: the X-coordinate of the left eye (from the view of the sample face, L_x) 89: the Y-coordinate of the left eye (from the view of the sample face, L_y) Related Publication:1. Wei Di, Lei Zhang, David Zhang, and Quan Pan “Studies on Hyperspectral Face Recognition in Visible Spectrum with Feature Band Selection” IEEE Trans. on System, Man and Cybernetics, Part A, vol. 40, issue 6, pp. 1354 –1361, Nov. 2010The Announcement of the CopyrightAll rights of the PolyU-HSFD are reserved. The database is only available for research and noncommercial purposes. Commercial distribution or any act related to commercial use of this database is strictly prohibited. A clear acknowledgement should be made for any public work based on the PolyU-HSFD. A citation to “PolyU-HSFD, .hk/~biometrics/hyper_face.htm” and our related works must be added in the references. A soft copy of any released or public documents that use the PolyU-HSFD must be forwarded to: cslzhang@.hkHere we would like to thank Chen Chao and Yang Hongfang’s kind help for the data acquisition.Downloading Steps:Download ZIP to your local disk. Then, fill in the application forms. Send the application form to cslzhang@.hk. The successful applicants will receive the passwords for unzipping the files downloaded.HSFace databaseApplication FormContact Information:Dr. Lei Zhang, Associate ProfessorBiometric Research CentreThe Hong Kong Polytechnic UniversityHung Hom, Kowloon, Hong KongE-mail: cslzhang@.hk数据预览:点此下载完整数据集。

头部跟踪器的虚拟声源定位系统

头部跟踪器的虚拟声源定位系统

第8卷第2期2019乖3月Vol. 8 No. 2Mar. 2019网络新媒体技术头部跟踪器的虚拟声源定位系统**本文于2017 -03 - 16收到,2017 -03 -23收到修改稿。

*本文受国家自然科学基金(编号:61501449 J1404367)和中国科学院先导专项项目(编号:XDA06040501 )资助。

黄劲文“杨飞然I 杨军-2「中国科学院声学研究所噪声与振动重点实验室北京100190 $中国科学院大学北京100049)摘要:报道了基于Windows 平台个人电脑和C+ +编程语言的虚拟声源定位系统近期工作进展。

系统利用头部跟踪器实时追 踪倾听者头部的移动与转动,动态模拟声波从声源到双耳的传递路径,使耳机重放能够产生逼真的声学环境。

提出一种基于 虚声源镜像法的房间反射声模拟方法,该方法通过模拟声源在房间中的高阶反射声,提高耳机重放的声场逼真度与饱满感。

通过主观听觉实验表明,使用基于头部跟踪器的虚拟声源定位系统进行定位,比静态虚拟声源定位系统的定位误差降低84. 5%o 随着其性能的升级与扩展,系统可作为虚拟现实应用与双耳听觉研究的一个功能强大且灵活的平台。

关键词:头相关传递函数,虚拟声源定位,头部跟踪器,动态因素,房间脉冲响应,虚拟现实,虚拟听觉Virtual Sound Source Location System with Head TrackingHUANG Jinwen 1*2, YANG Feiran 1 , YANG Jun 1>2(1 Key Laborary of Noise and Vibration Research , Institute of Acoustics , Chinese Academy of Sciences , Beijing , 100190, China ,2University of Chinese Academy of Sciences , Beijing, 100049, China)Abstract : This paper reports the recent works and progress on a Windows platform PC and C+ + language - based virtual sound source location system. By using head tracker to track the real 一 time orientation and location of listener * s head and dynamically simulating the acoustic propagation from sound source to two ears , the system is capable of recreating free - field virtual sources at various direc ­tions and distances as well as auditory perception in reflective environment via headphone presentation. A scheme for improving virtual sound source location system performance , which is based on virtual sound source image method to simulate reflected sound wave , is proposed. This method can simulate high order reflected sound wave , which including room space information and sound source loca ­tion information in headphone presentation. Improving the system performance of reality and satiation. A set of psychoacoustic experi ­ments validate the performance of the system. The functions of the sound source location system is being upgraded and extended and the system serves as a powerful and flexible platform for future virtual reality applications and binaural hearing researches.Keywords : : Head - related transfer function , Virtual sound source location , Head tracking , Dynamic factor , Room impulse response , Virtual reality , Virtual auditory0引言虚拟声源定位系统由人工控制虚拟声源位置,产生虚拟听觉环境,使倾听者获得犹如置身于自然声学 环境的听觉感知系统不仅是多媒体与虚拟现实的重要组成部分,也是双耳听觉研究的基础实验平 台⑺",在医疗⑷、科学研究⑸、通信、影音娱乐、声学辅助设计、虚拟训练设备等方面具有广阔应用前景。

通用电气中国研发中心

通用电气中国研发中心

GE China Technology Center通用电气中国研发中心通用电气公司(GE)是一家集科技、传媒、金融服务于一身的多元化公司,致力于为客户解决世界上最棘手的问题。

GE的产品和服务范围广泛,从飞机发动机、发电设备、水处理和安防技术,到医疗成像、商务和消费者金融、媒体,客户遍及全球100多个国家,拥有30多万员工。

杰夫·伊梅尔特先生是现任董事长及首席执行官。

GE是全球500强企业中始终保持领先的企业,GE是道.琼斯工业指数1896年设立以来,唯一一家至今仍在指数榜上的公司。

GE中国研究开发中心(CTC)是GE公司四个全球研发中心之一,是跨业务集团、跨研究领域的研发机构,为GE各业务集团提供基础科学研究、新产品开发、工程开发和采购服务。

中心坐落于上海张江高科技园区,占地面积47,000平方米,于2003年5月投入使用,是国内最大的独立外资研发机构之一,也是国内极少数具有基础科学研究能力的企业研发中心。

GE中国研发中心现有研发人员1200多人,60多个拥有世界一流设备的实验室,其研究力量主要集中在以下领域:·清洁能源,包括洁净煤、风电、太阳能发电等;·海水淡化与水处理;·材料科学,包括纳米材料、晶体、陶瓷,材料分析;·电力电子和实时控制;·安防技术;·先进制造技术;·影像技术;·化学技术,如导电高分子,电化学技术GE中国研发中心每年完成100多个研究项目,其中包括三分之一左右的基础科研项目。

截至2008年底,GE 中国研发中心共申请了320多项专利。

The GE China Technology Center, driving GE’s growth in China and globallyOne of four global research and development facilities at the hub of GE’s worldwide technology development efforts, the GE China Technology Center (CTC) is a diversified, multi-disciplinary organization conducting fundamental research, new product development, engineering service and sourcing service.Located at Zhangjiang High-Tech Park, the China Technology Center (CTC) has been operational since May 2003 with a total physical area of 47,000 square meters. CTC is one of the biggest foreign invested R&D centers and among the very few enterprise R&D centers which have fundamental research capabilities.With more than 1,200 researchers and engineers and 60+ labs, CTC teams are driven to bring technology breakthroughs and product innovations to life in the following key areas:• Energy, including Clean Coal, Wind Power, Solar Power, etc. • Water, including Seawater desalination and water treatment • Material, including nano-material, crystal, etc. • Electronic and electric, real time control • Security systems• Advanced manufacture • Imaging Technology • ChemicalGE China Technology Center accomplished more than 100 R&D projects every year, one third of which are fundamental research projects. Till the end of 2008, CTC has filed more than 320 patent applications.如何申请职位:选择1. · 请访问/careers/ · 根据职位编号申请职位; ·在线投递简历.选择2. 请将简历直接投递至 yuxiao.tang@ (请在邮件中注明申请得职位编码)更多信息请访问GE 网站 : or How to apply:Option 1: Please visit: /careers/Please search and apply the jobs by job code Please submit your CV on-lineOption 2: Please send your CV directly to yuxiao.tang@ (Please remark the job code in your mail)More information, please visit GE website: or All the Positions below will be Located in Shangha iPOSITION TITLE: Lead Engineer-Signal Processing (Algorithm)(Job Code: 1008431)Work Location: Shanghai Reqs: 1 Language: FluentEnglish3-5 yearsDegree: Master / PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:conceptual design and algorithm validation, contribute to intelligence-embedded electronic system for SDE future development.QUALIFICATIONS / REQUIREMENTS:PhD degree with major in Signal Processing, Automation Control or related areas;· 5 years above signal processing experience in medical/military/industry applications;· Strong analytical capability in system modeling and verification through simulation;· Strong experience in algorithm implement and optimization for embedded systems based on micro-processors/MCU/DSP;· In-depth knowledge in advanced signal processing techniques, e.g. adaptive/statistical/array/image signal processing;· Excellent programming skills: Matlab, C/C++ and assembly languages;· Knowledge in embedded OS, signal processing system will add advantages;. Demonstrated leadership and problem solving skills; Excellent language skill on English reading, speaking and writing.POSITION TITLE: Lead Engineer-Embedded System (Hardware)(Job Code: 1008436)Work Location: Shanghai Reqs: 1 Language: FluentEnglish3-5 yearsDegree: Master / PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:conceptual design, implement and validation of embedded systems based on micro-processors/ /MCU/DSP/FPGA. QUALIFICATIONS / REQUIREMENTS:. PhD degree with major in Embedded Systems, Automation Control, Signal Processing or related areas;· Must have 3 years above working experience in hardware design, at least 2 year hand-on experience in FPGA;· Solid knowledge in digital/analog electronics;· Strong experience in plan, design, implement, debug and optimizing embedded system based on micro-processors/MCU/DSP/FPGA;· Proficiency in FPGA development tool-chain and design flow;· Excellent programming skills: VHDL/Verilog, C/C++ and assembly languages;· Knowledge in embedded OS, signal processing system will add advantages;· Demonstrated leadership and problem solving skills; Excellent language skill on English reading, speaking and writing.POSITION TITLE: Lead Engineer(Job Code: 1104592)Work Location: Shanghai Reqs: 1 Language: FluentEnglish3-5 yearsDegree: Master / PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:"· Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas.· Interfaces with related technology areas/labs.· Be responsible for project execution and technical deliverables. Leverages broad technical experience to ensure success of projects.· Has an expanded network through participation in development activities.· Builds customer relationships and communicates with the customer on technology development activities. Influences customer technical direction.QUALIFICATIONS / REQUIREMENTS:"· PhD in chemical engineering/thermal engineering· Industrial R&D experience in Coal Conversion/IGCC/Polygeneration· Proven record of technical accomplishments in the related field.· Strong leadership traits; experience in leading technology team.· Strong interpersonal skills and excellent communication skills.· Ability to build and maintain strong customer relationships.· Excellent problem solving skills - ability to consider overall problem, identify opportunities and implement major changes.· Self-starter & self-motivator, independent thinker, proactive problem solver.· Motivated by quality, cost and speed.· High energy with passion for excellence (Demonstrated ability to set and achieve aggressive goals and targets; Embrace change and technology evolution as an opportunity).· Strong environmental, health and safety ethics.· Fluent oral and written communication in EnglishDesired Characteristics:"· > 6 years industrial R&D experiences in Coal Conversion/IGCC/Polygeneration.· Good technical reputation in the area of expertise.· Technology vision and big picture in the understanding of the industrial trends.POSITION TITLE: Principal Engineer– Organic Materials & Surface Chemistry(Job Code: 1061942)Work Location: Shanghai Reqs: 1 Language: FluentEnglish10 yearsDegree: PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:- Strong strategic focus & influence to drive organization technologies- Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas- Responsible for technical growth of the organization. Recognized for technical expertise and breath- Considered a technical resource for complex, multi-disciplinary issues, works across discipline boundaries to integrate experience to ensure success of projects- Ability to lead projects and initiatives with broad scope and high impact to the business- Be responsible for program execution and technical deliverables- Leverages broad technical experience to ensure success of projects. Leverage technical expertise and experience to provide direction to the team in technology development and transfer of technology- External presence, recognized technical expert in industry, strong connection to the businesses. Has an expanded network through participation in development activities- Viewed as a technical guide. Provide technical consultant and mentoring for junior scientists & engineers. Coaches and mentors others in technical reviews- Builds customer relationships and communicates with the customer on technology development activities. Influences customer technical direction- Applies the GE values and GE Growth Traits to personal leadership style, behavior and team activities- Embraces EHS & plays an active role in creating culture of safety.QUALIFICATIONS / REQUIREMENTS:- PhD in polymer, material science, organic chemistry, biochemistry & processing- Proven track record with demonstrating strategic technical leadership skills- At least 10 years Industrial R&D experience in chemical and material systems development for applications in energy storage and conversion, renewable energies, environmental technologies, etc- Proven record of technical accomplishments in the related field & recognized technical expert in the related area - Ability to make effective resource decisions, identify and remove project obstacles or barriers on behalf of the team- Strong interpersonal skills and excellent communication skills, ability to give clear, understandable instructions and coaching, explain complex problems in simple terms, foster cross-organizational communications- Ability to build and maintain strong customer relationships, anticipate and address customer needs, accelerate the pace of change to meet business objectives; analyze competitors and share insights and information with the group or team- Excellent problem solving skills - ability to consider overall problem, identify opportunities and implement major changes- Self-starter & self-motivator, independent thinker, proactive problem solver- Motivated by quality, cost and speed- High energy with passion for excellence (Demonstrated ability to set and achieve aggressive goals and targets; Embrace change and technology evolution as an opportunity)- Strong environmental, health and safety ethics & play an active role in creating culture of safety- Fluent oral and written communication in EnglishPOSITION TITLE: Principal Engineer- Inorganic Materials(Job Code: 1061944)Work Location: Shanghai Reqs: 1 Language: FluentEnglish10 yearsDegree: PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:- Strong strategic focus & influence to drive organization technologies- Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas- Responsible for technical growth of the organization. Recognized for technical expertise and breath- Considered a technical resource for complex, multi-disciplinary issues, works across discipline boundaries to integrate experience to ensure success of projects- Ability to lead projects and initiatives with broad scope and high impact to the business- Be responsible for program execution and technical deliverables- Leverages broad technical experience to ensure success of projects. Leverage technical expertise and experience to provide direction to the team in technology development and transfer of technology- External presence, recognized technical expert in industry, strong connection to the businesses. Has an expanded network through participation in development activities- Viewed as a technical guide. Provide technical consultant and mentoring for junior scientists & engineers. Coaches and mentors others in technical reviews- Builds customer relationships and communicates with the customer on technology development activities. Influences customer technical direction- Applies the GE values and GE Growth Traits to personal leadership style, behavior and team activities- Embraces EHS & plays an active role in creating culture of safetyQUALIFICATIONS / REQUIREMENTS:- PhD in Physics, Chemical Engineering, Materials Science or related fields; strong academic credentials; a solid history of technical accomplishments including publications and patents- Experience with Crystal growth or thin film technology- Proven track record with demonstrating strategic technical leadership skills- At least 10 years Industrial R&D experience in chemical and material systems development for applications in energy storage and conversion, renewable energies, environmental technologies, etc- Proven record of technical accomplishments in the related field & recognized technical expert in the related area - Ability to make effective resource decisions, identify and remove project obstacles or barriers on behalf of the team- Strong interpersonal skills and excellent communication skills, ability to give clear, understandable instructions and coaching, explain complex problems in simple terms, foster cross-organizational communications- Ability to build and maintain strong customer relationships, anticipate and address customer needs, accelerate the pace of change to meet business objectives; analyze competitors and share insights and information with the group or team- Excellent problem solving skills - ability to consider overall problem, identify opportunities and implement major changes- Self-starter & self-motivator, independent thinker, proactive problem solver- Motivated by quality, cost and speed- High energy with passion for excellence (Demonstrated ability to set and achieve aggressive goals and targets; Embrace change and technology evolution as an opportunity)- Strong environmental, health and safety ethics & play an active role in creating culture of safety- Fluent oral and written communication in EnglishPOSITION TITLE: Principal Engineer-Coal Conversion/IGCC/Polygeneration(Job Code: 1061947)Work Location: Shanghai Reqs: 1 Language: FluentEnglish10 yearsDegree: PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:- Strong strategic focus & influence to drive organization technologies- Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas- Responsible for technical growth of the organization. Recognized for technical expertise and breath. Make high impact to the business or is a recognized expert in Coal Conversion/IGCC/Polygeneration field- Considered a technical resource for complex, multi-disciplinary issues, works across discipline boundaries to integrate experience to ensure success of projects- Ability to lead projects and initiatives with broad scope and high impact to the business- Be responsible for program execution and technical deliverables- Leverages broad technical experience to ensure success of projects. Leverage technical expertise and experience to provide direction to the team in technology development and transfer of technology- External presence, recognized technical expert in industry, strong connection to the businesses. Has an expanded network through participation in development activities- Viewed as a technical guide. Provide technical consultant and mentoring for junior scientists & engineers. Coaches and mentors others in technical reviews- Builds customer relationships and communicates with the customer on technology development activities. Influences customer technical direction- Applies the GE values and GE Growth Traits to personal leadership style, behavior and team activities- Embraces EHS & plays an active role in creating culture of safetyQUALIFICATIONS / REQUIREMENTS:- PhD in Chemical Engineering/Thermal Engineering- Proven track record with demonstrating strategic technical leadership skills- At least 10 years Industrial R&D experience in systems development for applications in energy storage and conversion, renewable energies, environmental technologies, etc- Proven record of technical accomplishments in the related field & recognized technical expert in the related area - Ability to make effective resource decisions, identify and remove project obstacles or barriers on behalf of the team- Strong interpersonal skills and excellent communication skills, ability to give clear, understandable instructionsand coaching, explain complex problems in simple terms, foster cross-organizational communications- Ability to build and maintain strong customer relationships, anticipate and address customer needs, accelerate the pace of change to meet business objectives; analyze competitors and share insights and information with the group or team- Excellent problem solving skills - ability to consider overall problem, identify opportunities and implement major changes- Self-starter & self-motivator, independent thinker, proactive problem solver- Motivated by quality, cost and speed- High energy with passion for excellence (Demonstrated ability to set and achieve aggressive goals and targets; Embrace change and technology evolution as an opportunity)- Strong environmental, health and safety ethics & play an active role in creating culture of safety- Fluent oral and written communication in EnglishPOSITION TITLE: Principle Mechanical Engineer(Job Code: 1061954)Work Location: Shanghai Reqs: 1 Language: FluentEnglish10 yearsDegree: PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:"Provide leadership for a cross-functional team to design and develop digital manufacturing technologies.- Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas- Responsible for technical growth of the organization. Recognized for technical expertise and breath- Considered a technical resource for complex, multi-disciplinary issues, works across discipline boundaries to integrate experience to ensure success of projects- Ability to lead projects and initiatives with broad scope and high impact to the business- Be responsible for program execution and technical deliverables- Leverages broad technical experience to ensure success of projects. Leverage technical expertise and experience to provide direction to the team in technology development and transfer of technology- External presence, recognized technical expert in industry, strong connection to the businesses. Has an expanded network through participation in development activities- Viewed as a technical guide. Provide technical consultant and mentoring for junior scientists & engineers. Coaches and mentors others in technical reviews- Applies the GE values to personal leadership style, behavior and team activities- Embraces EHS & plays an active role in creating culture of safetyQUALIFICATIONS / REQUIREMENTS:"1. Deeply understand the manufacturing processes and procedure, and have solid shop floor working experience.2. Deeply understand and use the philosophy and tools of total quality, lean manufacturing, design for manufacturability and assembly, and serve as an engineering resource to others within the research center and also the customer site.3. Focus on optimizing CAD definition with NC Programming techniques, machine/fixture design, and machine dynamics.4. Specialize in system level manufacturing process including inspection, geometry analysis, process variable analysis, connectivity between digital definition/controls and the feedback loops, manufacturing IT systems connectivity to shop floor process.5. Strong capability of Modeling & simulation - analyze shop floor data; develop optimization algorithms, presentdecision assistance to shop floor management.6. Expertise in the use of modern design tools including 3D CAD, Pro-Engineer or SolidWorks, Factory flow simulation.7. Coating and welding process level knowledge is a plus.Desired Characteristics:Successful candidate will champion collaboration with industrial design, engineering and manufacturing to lead the manufacturing development process.Strong strategic focus & influence to drive organization technologiesPOSITION TITLE: Principal Engineer - Electrochemistry & Chemical Engineering(Job Code: 1061958)Work Location: Shanghai Reqs: 1 Language: FluentEnglish10 yearsDegree: PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:- Strong strategic focus & influence to drive organization technologies- Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas- Responsible for technical growth of the organization. Recognized for technical expertise and breath- Considered a technical resource for complex, multi-disciplinary issues, works across discipline boundaries to integrate experience to ensure success of projects- Ability to lead projects and initiatives with broad scope and high impact to the business- Be responsible for program execution and technical deliverables- Leverages broad technical experience to ensure success of projects. Leverage technical expertise and experience to provide direction to the team in technology development and transfer of technology- External presence, recognized technical expert in industry, strong connection to the businesses. Has an expanded network through participation in development activities- Viewed as a technical guide. Provide technical consultant and mentoring for junior scientists & engineers. Coaches and mentors others in technical reviews- Builds customer relationships and communicates with the customer on technology development activities. Influences customer technical direction- Applies the GE values and GE Growth Traits to personal leadership style, behavior and team activities- Embraces EHS & plays an active role in creating culture of safetyQUALIFICATIONS / REQUIREMENTS:- PhD in Chemical Engineering (process development, system design & integration), Electrochemistry, Materials science (composites, materials structure property relationship, etc.)- Proven track record with demonstrating strategic technical leadership skills- At least 10 years Industrial R&D experience in chemical and material systems development for applications in energy storage and conversion, renewable energies, environmental technologies, etc- Proven record of technical accomplishments in the related field & recognized technical expert in the related area - Ability to make effective resource decisions, identify and remove project obstacles or barriers on behalf of the team- Strong interpersonal skills and excellent communication skills, ability to give clear, understandable instructions and coaching, explain complex problems in simple terms, foster cross-organizational communications- Ability to build and maintain strong customer relationships, anticipate and address customer needs, accelerate the pace of change to meet business objectives; analyze competitors and share insights and information with the group or team- Excellent problem solving skills - ability to consider overall problem, identify opportunities and implement major changes- Self-starter & self-motivator, independent thinker, proactive problem solver- Motivated by quality, cost and speed- High energy with passion for excellence (Demonstrated ability to set and achieve aggressive goals and targets; Embrace change and technology evolution as an opportunity)- Strong environmental, health and safety ethics & play an active role in creating culture of safety- Fluent oral and written communication in EnglishPOSITION TITLE: Principal Engineer - Optical Instrumentation(Job Code: 1061961)Work Location: Shanghai Reqs: 1 Language: FluentEnglish10 yearsDegree: PhD WorkingExperience:JOB DESCRIPTION:ESSENTIAL FUNCTION / RESPONSIBILITIES:- Strong strategic focus & influence to drive organization technologies- Formulates, implements and executes on new programs or integrates major programs to meet key technical objectives. Champion new ideas- Responsible for technical growth of the organization. Recognized for technical expertise and breath- Considered a technical resource for complex, multi-disciplinary issues, works across discipline boundaries to integrate experience to ensure success of projects- Ability to lead projects and initiatives with broad scope and high impact to the business- Be responsible for program execution and technical deliverables- Leverages broad technical experience to ensure success of projects. Leverage technical expertise and experience to provide direction to the team in technology development and transfer of technology- External presence, recognized technical expert in industry, strong connection to the businesses. Has an expanded network through participation in development activities- Viewed as a technical guide. Provide technical consultant and mentoring for junior scientists & engineers. Coaches and mentors others in technical reviews- Builds customer relationships and communicates with the customer on technology development activities. Influences customer technical direction- Applies the GE values and GE Growth Traits to personal leadership style, behavior and team activities- Embraces EHS & plays an active role in creating culture of safety.QUALIFICATIONS / REQUIREMENTS:- PhD in Electrical/Mechanical/Optical Engineering or related fields- Proven track record with demonstrating strategic technical leadership skills- At least 10 years Industrial R&D experience in optical systems development for applications in industrial inspection, biomedical engineering - Proven record of technical accomplishments in the related field & recognized technical expert in the related area- Ability to make effective resource decisions, identify and remove project obstacles or barriers on behalf of the team- Strong interpersonal skills and excellent communication skills, ability to give clear, understandable instructions and coaching, explain complex problems in simple terms, foster cross-organizational communications- Ability to build and maintain strong customer relationships, anticipate and address customer needs, accelerate the pace of change to meet business objectives; analyze competitors and share insights and information with the group or team- Excellent problem solving skills - ability to consider overall problem, identify opportunities and implement major changes- Self-starter & self-motivator, independent thinker, proactive problem solver- Motivated by quality, cost and speed- High energy with passion for excellence (Demonstrated ability to set and achieve aggressive goals and targets; Embrace change and technology evolution as an opportunity)- Strong environmental, health and safety ethics & play an active role in creating culture of safety- Fluent oral and written communication in English。

基于面部深度空时特征的抑郁症识别算法

基于面部深度空时特征的抑郁症识别算法

文献引用格式:12 - 18.YU M,XU X Y,SHI S,et al.Depression Recognition Algorithm Based on Facial Deep Spatio - Temporal Features[J].中图分类号:基于面部深度空时特征的抑郁症识别算法摘要:提出基于残差注意力网络和金字塔扩大卷积长短时记忆(Convolutional Long Short-Term Memory网络提取人脸图像空时特征的抑郁症识别算法。

首先构建残差注意力网络提取人脸图像不同权值的空间特征,征,显示在两个数据集上,特征10%可见,关键词Abstract:the automatic diagnosis of depression from facial expressions, which extracted spatio-temporal features based on the residual attention network and pyramidal dilated convolutional LSTM network. Firstly, the residual attention network was constructed to extract the spatial features with different weight from facial expressions. Then based on the convLSTM network, a pyramid expansion strategy was added to extract the temporal features with different scales on the resulting spatial features. Finally the spatio-temporal features were input into the DNN network for the regression analysis of depression inventory score. Validation was performed on the AVECthe results were shown on both data sets, the Mae and RMSE values between the predicted and true depression degree of the proposed algorithm are better than those based on manual feature and manual feature + depth feature. In the AVEC2030年,抑郁症将成为全世界导致死亡和残疾的最大因素[2]。

人脸识别介绍_IntroFaceDetectRecognition

人脸识别介绍_IntroFaceDetectRecognition

Knowledge-based Methods: Summary
Pros:
Easy to come up with simple rules Based on the coded rules, facial features in an input image are extracted first, and face candidates are identified Work well for face localization in uncluttered background
Template-Based Methods: Summary
Pros:
Simple
Cons:
Templates needs to be initialized near the face images Difficult to enumerate templates for different poses (similar to knowledgebased methods)
Knowledge-Based Methods
Top Top-down approach: Represent a face using a set of human-coded rules Example:
The center part of face has uniform intensity values The difference between the average intensity values of the center part and the upper part is significant A face often appears with two eyes that are symmetric to each other, a nose and a mouth

人脸识别技术设计论文

人脸识别技术设计论文

人脸识别算法摘要人脸自动识别是模式识别领域的一项热门研究课题,有着十分广泛的应用前景。

本文对人脸位置矫正,人脸的特征提取和识别这些方面进行了研究,并提出了相应的实现算法。

人脸位置矫正作为人脸检测定位的一个环节,在计算机人脸识别中具有重要的意义。

本文第二章提出了一种基于单人脸灰度图像中眼睛定位的人脸位置矫正方法,它是针对人眼灰度变化特点、人眼几何形状特征及双眼的轴对称性而设计的。

实验结果表明,该方法对于双眼可见单人脸灰度图像能实现快速有效矫正,并能在矫正结果中精确给出双眼瞳孔位置。

本文第三章提出了一种基于神经网络的主元分析人脸图像识别方法。

该方法利用非线性主元分析神经网络对人脸图像提取人脸特征(矢量),并在BP神经网络上实现了对人脸图像的识别。

实验结果证明了该方法的有效性和稳定性。

关键词人脸位置矫正,人脸特征提取,人脸识别,神经网络,灰度图像,图像块纵向复杂度,主元分析法,1-iThe Design and Implementation of Algorithms for Human FaceRecognitionAbstractThe automatic recognition of human faces is a hot spot in the field of pattern recognition , which has a wide range of potential applications . As the results of our in-depth research ,two algorithms are proposed : one for face pose adjustment , the other for facial feature extraction and face identification .Face pose adjustment , as a loop of human face location, is very important in computer face recognition. Chapter 2 of this thesis presents a new approach to automatic face pose adjustment on gray-scale static images with a single face . In a first stage , the right positions of eyes are precisely detected according to several designed parameters which well characterize the complex changes of the gray parameter in and around eyes and the geometrical shape of eyes . During the second stage , based on the location and the symmetry feature of eyes , the inclination angle is calculated and the face position is redressed . The experimentation shows that the algorithm performs very well both in terms of rate and of efficiency . What’s more , due to the precise location of eyes , the apples of the eyes are detected .In chapter 3, a novel approach to human face image recognition based on principal component analysis and neural networks has been proposed . By using BP neural networks , human face images are successfully classified and recognized according to the output of BPNN whose input is the eigenvector extracted from the human face images via nonlinear principal1-iicomponent analysis of a single layer neural network . Simulation results demonstrate the effectiveness and stability of the approach .KeywordsFace Pose Adjustment, Facial Feature Extraction , Human Face Recognition , Neural Networks , Gray-scale Static Image , Vertical-complexity of Image Block, Principal Component Analysis1-iii致谢首先要感谢我的毕业设计导师曹文明教授,他是我在人脸识别领域研究的启蒙老师。

人脸识别外文文献

人脸识别外文文献

Method of Face Recognition Based on Red-BlackWavelet Transform and PCAYuqing He, Huan He, and Hongying YangDepartment of Opto-Electronic Engineering,Beijing Institute of Technology, Beijing, P.R. China, 10008120701170@。

cnAbstract。

With the development of the man—machine interface and the recogni—tion technology, face recognition has became one of the most important research aspects in the biological features recognition domain. Nowadays, PCA(Principal Components Analysis) has applied in recognition based on many face database and achieved good results. However, PCA has its limitations: the large volume of computing and the low distinction ability。

In view of these limitations, this paper puts forward a face recognition method based on red—black wavelet transform and PCA. The improved histogram equalization is used to realize image pre-processing in order to compensate the illumination. Then, appling the red—black wavelet sub—band which contains the information of the original image to extract the feature and do matching。

Python使用face_recognition人脸识别

Python使用face_recognition人脸识别

Python使⽤face_recognition⼈脸识别Python 使⽤ face_recognition ⼈脸识别官⽅说明:⼈脸识别 face_recognition 是世界上最简单的⼈脸识别库。

使⽤ dlib 最先进的⼈脸识别功能构建建⽴深度学习,该模型准确率在99.38%。

Python模块的使⽤ Python可以安装导⼊ face_recognition 模块轻松操作,对于简单的⼏⾏代码来讲,再简单不过了。

Python操作 face_recognition API ⽂档:⾃动查找图⽚中的所有⾯部import face_recognitionimage = face_recognition.load_image_file("my_picture.jpg")face_locations = face_recognition.face_locations(image)# face_locations is now an array listing the co-ordinates of each face!还可以选择更准确的给予深度学习的⼈脸检测模型import face_recognitionimage = face_recognition.load_image_file("my_picture.jpg")face_locations = face_recognition.face_locations(image, model="cnn")# face_locations is now an array listing the co-ordinates of each face!⾃动定位图像中⼈物的⾯部特征import face_recognitionimage = face_recognition.load_image_file("my_picture.jpg")face_landmarks_list = face_recognition.face_landmarks(image)# face_landmarks_list is now an array with the locations of each facial feature in each face.# face_landmarks_list[0]['left_eye'] would be the location and outline of the first person's left eye.识别图像中的⾯部并识别它们是谁import face_recognitionpicture_of_me = face_recognition.load_image_file("me.jpg")my_face_encoding = face_recognition.face_encodings(picture_of_me)[0]# my_face_encoding now contains a universal 'encoding' of my facial features that can be compared to any other picture of a face! unknown_picture = face_recognition.load_image_file("unknown.jpg")unknown_face_encoding = face_recognition.face_encodings(unknown_picture)[0]# Now we can see the two face encodings are of the same person with `compare_faces`!results = face_pare_faces([my_face_encoding], unknown_face_encoding)if results[0] == True:print("It's a picture of me!")else:print("It's not a picture of me!")face_recognition ⽤法要在项⽬中使⽤⾯部识别,⾸先导⼊⾯部识别库,没有则安装:import face_recognition基本思路是⾸先加載圖⽚:# 导⼊⼈脸识别库import face_recognition# 加载图⽚image = face_recognition.load_image_file("1.jpg")上⾯这⼀步会将图像加载到 numpy 数组中,如果已经有⼀个 numpy 数组图像则可以跳过此步骤。

基于wifi的室内定位系统毕业设计论文

基于wifi的室内定位系统毕业设计论文

本科毕业论文题目基于wifi的室内定位系统摘要本文设计及实现了一个基于WiFi 射频信号强度指纹匹配的移动终端定位系统,并设计实现了一种基于权重值选择的定位算法。

该算法为每个扫描到的AP 的RSSI 设定了选择区间,指纹库中落在此区间的所有位置点设平均权值,最后选取权重值最大者为待定位点的位置估计,如有相同权重值,则比较信号强度距离,取最小者,这种算法在一定程度上克服了RSSI 信号随机抖动对定位的影响,提高了定位的稳定性和精度。

经实验测试,此系统在 4 米范围内具有良好的定位效果。

可部署在展馆、校园、公园等公共场所,为客户提供定位导航服务。

定位算法运行于服务端,客户端为配备WiFi 模块的Android手机。

借助该定位系统,基于Android系统的移动终端可方便地查询自身位置,并获取各种基于位置服务。

关键词: 接收信号强度;无线室内定位;射频指纹;Android 操作系统AbstractThis paper designs and implements an indoor location system based on WiFi for mobile user with Android handset. A locating arithmetic based on Weight-Select is introduced to filter the random noise of RSSI. For each location in Radio Map, a weight is set if the RSSI of the AP scanned is in the interval preset. Then max-weighted location or the min-RSSI-distance among them will be selected as the estimated position. According to experiments, 4-metre locating precision is available. It can be used for locating and navigating in such scene as exhibition center, campus, park, and so on. Users equipped with Android handset could get its location and some intelligent services. It is also an open and extensible system. Some locating arithmetic also could be tested on this system.Key words:Received Signal Strength, Wireless Indoor Locating, Radio Map, Android Operating System第一章绪论 (6)1.1关于位置信息确定的意义及方法 (6)1.1.1位置信息确定的意义及方法 (6)1.1.2本文主要介绍的定位系统 (7)1.2本文的主要研究内容以及各章安排 (7)1.2.1主要内容 (7)1.2.2本文安排 (7)第二章目前主要定位方式及各种测量方法 (7)2.1 GPS定位系统介绍 (8)2.1.1GPS的发展 (8)2.1.2 GPS国内外动态 (10)2.2 wifi定位技术 (11)2.2.1 wifi的利用原理 (11)2.2.2定位需要两个先决条件 (12)2.3定位运用的各种测量方法 (12)2.3.1 通过传播时间测量方法 (13)2.3.2信号衰减测量方法 (13)2.3.3改进的TOA算法 (13)2.4本章总结 (14)第三章无线定位系统和物联定位系统的介绍 (14)3.1无线定位系统方案 (14)3.1.1系统方案 (14)3.1.2特点与指标 (16)3.2 LocateSYS物联定位系统 (17)3.2.1系统概述 (17)3.2.2工作原理 (18)3.2.3特点与指标 (18)3.2.4产品资料 (19)3.2.5应用领域 (21)3.3 本章总结 (21)第四章基于WiFi 的室内定位系统设计与实现 (21)4.1系统设计 (21)4.2系统的实现 (23)4.2.1客户端设计 (23)4.2.4. Activity 生命周期 (24)4.2.5.获取周边AP 信号强度 (25)4.3 程序流程 (26)4.4. 服务端软件设计 (27)4.4.1. Web 服务器 (27)4.4.2. 定位服务器 (28)4.5.客户端与服务端通信 (28)4.6. 2算法描述 (31)4.6. 3算法分析 (31)4. 7实验 (32)4.7. 1实验过程 (32)4.7.2. 实验结果 (33)4.8. 总结 (33)致谢 (34)参考文献 (35)第一章绪论1.1关于位置信息确定的意义及方法1.1.1位置信息确定的意义及方法位置信息在人们的日常生活中扮演着重要的作用。

wifi障报告常见英语句子

wifi障报告常见英语句子

wifi障报告常见英语句子1.Fault Restoration For Distribution Network Base On Dissipated Network;基于耗散网络的配电网网络故障恢复。

2.Research Of Network Fault Location System Based On Framework Of Network Topology;基于网络拓扑结构的网络故障定位系统的研究。

3.The Research Of Analog Circuit Diagnosis Approaches Based On Network Decomposition And Neural Networks;基于网络撕裂和神经网络故障诊断方法研究。

puter Networks, Data Link Layer Of Network Troubleshooting计算机网络数据链路层网络故障排除初探。

5.Research And Implement Of Network Fault Diagnosis In Local Area Network;局域网络故障诊断技术的研究与实现。

6.The Research And Implementation Of IPNetwork Fault Inspecting System;IP网络故障监测系统的研究与实现。

7.Research And Implementation Of Computer Network Fault Management System;计算机网络故障管理系统研究及实现。

8.Fault Diagnosis Model Based On Fault Tree And Bayesian Networks基于故障树和贝叶斯网络的故障诊断模型。

9.Fault Diagnosis With Fault Gradation Using Neural Network Group基于神经网络组与故障分级的故障诊断。

Presentation Preference Oral Presentation or Poster Presentation

Presentation Preference Oral Presentation or Poster Presentation

Paper Title3D Face Recognition based on Geodesic DistancesAuthorsShalini GuptaDepartment of Electrical and Computer EngineeringThe University of Texas at Austin1University Station C0800Austin,TX78712+1.512.471.8660+1.512.471.0616(fax)shalinig@Mia K.MarkeyDepartment of Biomedical EngineeringThe University of Texas at Austin1University Station C0800Austin,TX78712+1.512.471.8660+1.512.471.0616(fax)mia.markey@Jake AggarwalDepartment of Electrical and Computer EngineeringThe University of Texas at Austin1University Station C0803Austin,TX78712+1.512.471.1369+1.512.471.5532(fax)aggarwaljk@Alan C.BovikDepartment of Electrical and Computer EngineeringThe University of Texas at Austin1University Station C0803Austin,TX78712+1.512.471.5370+1.512.471.1225(fax)bovik@Presentation PreferenceOral Presentation or Poster PresentationPrincipal Author’s BiographyShalini Gupta received a BE degree in Electronics and Electrical Communication Engineering from Punjab Engineering College,India.She received a MS degree in Electrical and Computer Engi-neering from the University of Texas at Austin,where she is currently a PhD student.During her masters,she developed techniques for computer aided diagnosis of breast cancer.She is currently investigating techniques for3D human face recognition.KeywordsGeodesic distances,three-dimensional face recognition,range image,biometricsExtended AbstractProblem Statement:Automated human identification is required in applications such as access control,passenger screening,passport control,surveillance,criminal justice and human computer interaction.Face recognition is one of the most widely investigated biometric techniques for human identification. Face recognition systems require less user co-operation than systems based on other biometrics(e.g.fingerprints and iris).Although considerable progress has been made on face recognition systems based on two dimensional(2D)intensity images,they are inadequate for robust face recognition. Their performance is reported to decrease significantly with varying facial pose and illumination conditions[1].Three-dimensional face recognition systems are less sensitive to changes in ambient illumination conditions than2D systems[2].Three-dimensional face models can also be rigidly transformed to a canonical pose.Hence,considerable research attention is now being directed toward developing3D face recognition systems.Review of Previous Work:Techniques employed for3D face recognition include those based upon global appearance of face range images,surface matching,and local facial geometric features.Techniques based on global appearance of face range images are straight-forward extensions of statistical learning techniques that were successful to a degree with2D face images.They involve statistical learning of the3D face space through an ensemble of range images.A popular3D face recognition technique is based on principal component analysis(PCA)[3]and is often taken as the baseline for assessing the performance of other algorithms[4].While appearance based techniques have met with a degree of success,it is intuitively less obvious exactly what discriminatory information about faces they encode.Furthermore,since they employ information from large range image regions,their recog-nition performance is affected by changes in facial pose,expression,occlusions,and holes.Techniques based on surface matching use an iterative procedures to rigidly align two face surfaces as closely as possible[5].A metric quantifies the difference between the two face surfaces after alignment,and this is employed for recognition.The computational load of such techniques can be considerable,especially when searching large3D face databases.Their performance is also affected by changes in facial expression.For techniques based on local geometric facial features,characteristics of localized regions of the face surface,and their relationships to others,are quantified and employed as features.Some local geometric features that have been used previously for face recognition include surface curva-tures,Euclidean distances and angles betweenfiducial points on the face[6,7,8],point signatures [9],and shape variations of facial sub regions[10].Techniques based on local features require an additional step of localization and segmentation of specific regions of the face.A pragmatic issue affecting the success of these techniques is the choice of local regions andfiducial points.Ideally the choice of such regions should be based on an understanding of the variability of different parts of the face within and between individuals.Three dimensional face recognition techniques based on local feature have been shown to be robust to a degree to varying facial expression[9].Recently,methods for expression invariant3D face recognition have been proposed[11].They are based on the assumption that different facial expressions can be regarded as isometric deformations of the face surface.These deformations preserve intrinsic properties of the surface,one of which is the geodesic distance between a pair of points on the surface.Based on these ideas we present a preliminary study aimed at investigating the effectiveness of using geodesic distances between all pairs of25fiducial points on the face as features for face recognition.To the best of our knowledge,this is thefirst study of its kind.Another contribution of this study is that instead of choosing a random set of points on the face surface,we considered facial landmarks relevant to measuring anthropometric facial proportions employed widely in fa-cial plastic surgery and art[12].The performance of the proposed face recognition algorithm was compared against other established algorithms.Proposed Approach:Three dimensional face models for the study were acquired by an MU-2stereo imaging systemby3Q Technologies Ltd.(Atlanta,GA).The system simultaneously acquires both shape and tex-ture information.The data set contained1128head models of105subjects.It was partitioned intoa gallery set containing one image each of the105subjects with a neutral expression.The probeset contained another663images of the gallery subjects with a neutral or an arbitrary expression.The probe set had a variable number of images per subject(1-55).Models were rigidly aligned to frontal orientation and range images were constructed.Theywere medianfiltered and interpolated to remove holes.Twenty-fivefiducial points,as depicted inFigure1were manually located on each face.Three face recognition algorithms were implemented.Thefirst employed300geodesic distances(between all pairs offiducial points)as features for recog-nition.The fast marching algorithm for front propagation was employed to calculate the geodesicdistance between pairs of points[13].The second algorithm employed300Euclidean distancesbetween all pairs offiducial points as features.The normalized L1norm where each dimensionwas divided by its variance,was used as the metric for matching faces with both the Euclideandistance and geodesic distance features.The third3D face recognition algorithm implemented was based on PCA.For this algorithm,a subsection of each face range image of size354pixels,enclosing the main facial features wasemployed.The gallery and probe sets employed to test the performance of this algorithm were thesame as those used in thefirst and second algorithms.Additionally a separate set of360rangeimages of12subjects(30images per subjects),was used to train the PCA classifier.Face rangeimages were projected on to42eigen vectors accounting for99%of the variance in the data.Again,the L1norm was employed for matching faces in the42dimensional PCA sub space.Verification performance of all algorithms was evaluated using the receiver operating charac-teristic(ROC)methodology,from which the equal error rates(EER)were noted.Identificationperformance was evaluated by means of the cumulative match characteristic curves(CMC)andthe rank1recognition rates(RR)were observed.The performance of each technique for the entireprobe set,for neutral probes only and for expressive probes only were evaluated separately. Experimental Results:Table1presents the equal error rates for verification performance and the rank1recognitionrates for identification performance of the three face recognition algorithms.Figure2(a)presentsROC curves of the three systems for neutral expression probes only.Figure2(b)presents the CMCcurves for the three systems for neutral expression probes only.It is evident that the two algorithmsbased on Euclidean or geodesic distances between anthropometric facial landmarks(EER∼5%, RR∼89%)performed substantially better than the baseline PCA algorithm(EER=16.5%, RR=69.7%).The algorithms based on geodesic distance features performed on a par with the algorithm based on Euclidean distance features.Both were effective,to a degree,at recognizing3D faces.In this study the performance of the proposed algorithm based on geodesic distancesbetween anthropometric facial landmarks decreased when probes with arbitrary facial expressionswere matched against a gallery of neutral expression3D faces.This suggests that geodesic distancesbetween pairs of landmarks on a face may not be preserved when the facial expression changes.This was contradictory to Bronstein et al.’s assumption regarding facial expressions being isometricdeformations of facial surfaces[11].In conclusion,geodesic distances between anthropometric landmarks were observed to be ef-fective features for recognizing3D faces,however they were not more effective than Euclideandistances between the same landmarks.The3D face recognition algorithm based on geodesic dis-tance features was affected by changes in facial expression.In the future,we plan to investigatemethods for reducing the dimensionality of the proposed algorithm and to identify the more dis-criminatory geodesic distance features.Acknowledgments:The authors would like to gratefully acknowledge Advanced Digital Imaging Research,LLC(Houston,TX)for providing support in terms of funding and3D face data for the study. Figures and Tables:Figure1:Thefigures show the25anthropometric landmarks that were considered on a color and range image of a human face.(a)ROC(b)CMCFigure2:Thisfigure presents the2(a)verification performance in terms of an ROC curve;2(b) the cumulative match characteristic curves for the identification performance of the three face recognition algorithms with the neutral expression probes only.Method EER(%)Rank1RR(%)N-N N-E N-All N-N N-E N-AllGEODESIC2.78.55.693.181.489.9EUCLIDEAN2.26.74.192.978.188.8PCA18.113.416.570.268.369.7Table1:Verification and identification performance statistics for the face recognition systems based on PCA,Euclidean distances and geodesic distances.N-N represents performance of a system for the neutral probes only,N-E for the expressive probes only and N-All for all probes. References[1]P.J.Phillips,P.Grother,R.J.Micheals,D.M.Blackburn,E.Tabassi,and J.M.Bone.Frvt2002:Overview and summary.available at ,March2003.[2]E.P.Kukula,S.J.Elliott,R.Waupotitsch,and B.Pesenti.Effects of illumination changes onthe performance of geometrix facevision/spl reg/3d frs.In Security Technology,2004.38th Annual2004International Carnahan Conference on,pages331–337,2004.[3]K.I.Chang,K.W.Bowyer,and P.J.Flynn.An evaluation of multimodal2d+3d facebiometrics.Pattern Analysis and Machine Intelligence,IEEE Transactions on,27(4):619–624,2005.[4]P.J.Phillips,P.J.Flynn,T.Scruggs,K.W.Bowyer,and W.Worek.Preliminary face recog-nition grand challenge results.In Automatic Face and Gesture Recognition,2006.FGR2006.7th International Conference on,pages15–24,2006.[5]Xiaoguang Lu,A.K.Jain,and D.Colbry.Matching2.5d face scans to3d models.PatternAnalysis and Machine Intelligence,IEEE Transactions on,28(1):31–43,2006.[6]G.G.Gordon.Face recognition based on depth and curvature features.In Computer Vi-sion and Pattern Recognition,1992.Proceedings CVPR’92.,1992IEEE Computer Society Conference on,pages808–810,1992.[7]A.B.Moreno,A.Sanchez,J.Fco,V.Fco,and J.Diaz.Face recognition using3d surface-extracted descriptors.In Irish Machine Vision and Image Processing Conference(IMVIP 2003),Sepetember2003.[8]Y.Lee,H.Song,U.Yang,H.Shin,and K.Sohn.Local feature based3d face recognition.InAudio-and Video-based Biometric Person Authentication,2005International Conference on, LNCS,volume3546,pages909–918,2005.[9]Yingjie Wang,Chin-Seng Chua,and Yeong-Khing Ho.Facial feature detection and facerecognition from2d and3d images.Pattern Recognition Letters,23(10):1191–1202,2002. [10]Chenghua Xu,Yunhong Wang,Tieniu Tan,and Long Quan.Automatic3d face recognitioncombining global geometric features with local shape variation information.In Automatic Face and Gesture Recognition,2004.Proceedings.Sixth IEEE International Conference on, pages308–313,2004.[11]A.M.Bronstein,M.M.Bronstein,and R.Kimmel.Three-dimensional face recognition.International Journal of Computer Vision,64(1):5–30,2005.[12]L.Farkas.Anthropometric Facial Proportions in Medicine.Thomas Books,1987.[13]R.Kimmel and puting geodesic paths on manifolds.Proceedings of theNational Academy of Sciences,USA,95:84318435,1998.。

Accurate Passive Location Estimation Using TOA Measurements

Accurate Passive Location Estimation Using TOA Measurements

Accurate Passive Location Estimation Using TOA MeasurementsJunyang Shen,Andreas F.Molisch,Fellow,IEEE,and Jussi Salmi,Member,IEEEAbstract—Localization of objects is fast becoming a major aspect of wireless technologies,with applications in logistics, surveillance,and emergency response.Time-of-arrival(TOA) localization is ideally suited for high-precision localization of objects in particular in indoor environments,where GPS is not available.This paper considers the case where one transmitter and multiple,distributed,receivers are used to estimate the location of a passive(reflecting)object.It furthermore focuses on the situation when the transmitter and receivers can be synchronized,so that TOA(as opposed to time-difference-of-arrival(TDOA))information can be used.We propose a novel, Two-Step estimation(TSE)algorithm for the localization of the object.We then derive the Cramer-Rao Lower Bound(CRLB) for TOA and show that it is an order of magnitude lower than the CRLB of TDOA in typical setups.The TSE algorithm achieves the CRLB when the TOA measurements are subject to small Gaussian-distributed errors,which is verified by analytical and simulation results.Moreover,practical measurement results show that the estimation error variance of TSE can be33dB lower than that of TDOA based algorithms.Index Terms—TOA,TDOA,location estimation,CRLB.I.I NTRODUCTIONO BJECT location estimation has recently received inten-sive interests for a large variety of applications.For example,localization of people in smoke-filled buildings can be life-saving[1];positioning techniques also provide useful location information for search-and-rescue[2],logistics[3], and security applications such as localization of intruders[4].A variety of localization techniques have been proposed in the literature,which differ by the type of information and system parameters that are used.The three most important kinds utilize the received signal strength(RSS)[5],angle of arrival(AOA)[6],and signal propagation time[7],[8],[9], respectively.RSS algorithms use the received signal power for object positioning;their accuracies are limited by the fading of wireless signals[5].AOA algorithms require either directional antennas or receiver antenna arrays1.Signal-propagation-time based algorithms estimate the object location using the time it takes the signal to travel from the transmitter to the target and from there to the receivers.They achieve very accurate Manuscript received April15,2011;revised September28,2011and Jan-uary18,2012;accepted February12,2012.The associate editor coordinating the review of this paper and approving it for publication was X.Wang.J.Shen and A.F.Molisch are,and J.Salmi was with the Department of Electrical Engineering,Viterbi School of Engineering,University of Southern California(e-mail:{junyangs,molisch,salmi}@).J.Salmi is currently with Aalto University,SMARAD CoE,Espoo,Finland.This paper is partially supported by the Office of Naval Research(ONR) under grant10599363.Part of this work was presented in the IEEE Int.Conference on Ultrawide-band Communications2011.Digital Object Identifier10.1109/TWC.2012.040412.1106971Note that AOA does not provide better estimation accuracy than the signal propagation time based methods[10].estimation of object location if combined with high-precision timing measurement techniques[11],such as ultrawideband (UWB)signaling,which allows centimeter and even sub-millimeter accuracy,see[12],[13],and Section VII.Due to such merits,the UWB range determination is an ideal candidate for short-range object location systems and also forms the basis for the localization of sensor nodes in the IEEE802.15.4a standard[14].The algorithms based on signal propagation time can be fur-ther classified into Time of Arrival(TOA)and Time Difference of Arrival(TDOA).TOA algorithms employ the information of the absolute signal travel time from the transmitter to the target and thence to the receivers.The term“TOA”can be used in two different cases:1)there is no synchronization between transmitters and receivers and then clock bias between them exist;2)there is synchronization between transmitters and receivers and then clock bias between them does not exist. In this paper,we consider the second situation with the synchronization between the transmitter and receivers.Such synchronization can be done by cable connections between the devices,or sophisticated wireless synchronization algo-rithms[15].TDOA is employed if there is no synchronization between the transmitter and the receivers.In that case,only the receivers are synchronized.Receivers do not know the signal travel time and therefore employ the difference of signal travel times between the receivers.It is intuitive that TOA has better performance than the TDOA,since the TDOA loses information about the signal departure time[7].The TDOA/TOA positioning problems can furthermore be divided into“active”and“passive”object cases.“Active”means that the object itself is the transmitter,while“passive”means that it is not the transmitter nor receiver,but a separate (reflecting/scattering)object that just interacts with the signal stemming from a separate transmitter2.There are numerous papers on the TOA/TDOA location estimation for“active”objects.Regarding TDOA,the two-stage method[16]and the Approximate Maximum Likelihood Estimation[17]are shown to be able to achieve the Cramer-Rao Lower Bound(CRLB)of“active”TDOA[8].As we know,the CRLB sets the lower bound of the estimation error variance of any un-biased method.Two important TOA methods of“active”object positioning are the Least-Square Method[18]and the Approximate Maximum Likelihood Es-timation Method[17],both of which achieve the CRLB of “active”TOA.“Active”object estimation methods are used, e.g,for cellular handsets,WLAN,satellite positioning,and active RFID.2The definitions of“active”and“passive”here are different from those in radar literature.In radar literature,“passive radar”does not transmit signals and only detects transmission while“active radar”transmits signals toward targets.1536-1276/12$31.00c 2012IEEE“Passive”positioning is necessary in many practical situa-tions like crime-prevention surveillance,assets tracking,and medical patient monitoring,where the target to be localized is neither transmitter nor receiver,but a separate(reflect-ing/scattering)object.The TDOA positioning algorithms for “passive”objects are essentially the same as for“active”objects.For TOA,however,the synchronization creates a fundamental difference between“active”and“passive”cases. Regarding the“passive”object positioning,to the best of our knowledge,no TOA algorithms have been developed.This paper aims tofill this gap by proposing a TOA algorithm for passive object location estimation,which furthermore achieves the CRLB of“passive”TOA.The key contributions are:•A novel,two step estimation(TSE)method for the passive TOA based location estimation.It borrows an idea from the TDOA algorithm of[16].•CRLB for passive TOA based location estimation.When the TOA measurement error is Gaussian and small,we prove that the TSE can achieve the CRLB.Besides,it is also shown that the estimated target locations by TSE are Gaussian random variables whose covariance matrix is the inverse of the Fisher Information Matrix(FIM)related to the CRLB.We also show that in typical situations the CRLB of TOA is much lower than that of TDOA.•Experimental study of the performances of TSE.With one transmitter and three receivers equipped with UWB antennas,we perform100experimental measurements with an aluminium pole as the target.After extracting the signal travel time by high-resolution algorithms,the location of the target is evaluated by TSE.We show that the variance of estimated target location by TSE is much (33dB)lower than that by the TDOA method in[16]. The remainder of this paper is organized as follows.Section II presents the architecture of positioning system.Section III derives the TSE,followed by comparison between CRLB of TOA and TDOA algorithms in Section IV.Section V analyzes the performance of TSE.Section VI presents the simulations results.Section VII evaluates the performance of TSE based on UWB measurement.Finally Section VIII draws the conclusions.Notation:Throughout this paper,a variable with“hat”ˆ•denotes the measured/estimated values,and the“bar”¯•denotes the mean value.Bold letters denote vectors/matrices. E(•)is the expectation operator.If not particularly specified,“TOA”in this paper denotes the“TOA”for a passive object.II.A RCHITECTURE OF L OCALIZATION S YSTEMIn this section,wefirst discuss the challenges of localization systems,and present the focus of this paper.Then,the system model of individual localization is discussed.A.Challenges for target localizationFor easy understanding,we consider an intruder localization system using UWB signals.Note that the intruder detection can also be performed using other methods such as the Device-free Passive(DfP)approach[19]and Radio Frequency Identification(RFID)method[20].However,both the DfP and RFID methods are based on preliminary environmental measurement information like“Radio Map Construction”[19] and“fingerprints”[20].On the other hand,the TOA based approach considered in our framework does not require the preliminary efforts for obtaining environmental information. With this example,we show the challenges of target po-sitioning system:Multiple Source Separation,Indirect Path Detection and Individual Target Localization.The intruder detection system localizes,and then directs a camera to capture the photo of the targets(intruders).This localization system consists of one transmitter and several receivers.The transmitter transmits signals which are reflected by the targets,then,the receivers localize the targets based on the received signals.Multiple Source Separation:If there are more than one intruders,the system needs to localize each of them.With multiple targets,each receiver receives impulses from several objects.Only the information(such as TOA)extracted from impulses reflected by the same target should be combined for localization.Thus,the Multiple Source Separation is very important for target localization and several techniques have been proposed for this purpose.In[21],a pattern recognition scheme is used to perform the Multiple Source Separation. Video imaging and blind source separation techniques are employed for target separation in[22].Indirect Path Detection:The transmitted signals are not only reflected by the intruders,but also by surrounding objects,such as walls and tables.To reduce the adverse impact of non-target objects in the localization of target, the localization process consists of two steps.In the initial/first stage,the system measures and then stores the channel impulses without the intruders.These impulses are reflected by non-target objects,which is referred to as reflectors here.The radio signal paths existing without the target are called background paths.When the intruders are present,the system performs the second measurement. To obtain the impulses related to the intruders,the system subtracts the second measurement with thefirst one. The remaining impulses after the subtraction can be through one of the following paths:a)transmitter-intruders-receivers,b)transmitter-reflectors-intruders-receivers,c) transmitter-intruders-reflectors-receivers,d)transmitter-reflectors-intruders-reflectors-receivers3.Thefirst kind of paths are called direct paths and the rest are called indirect paths.In most situations,only direct paths can be used for localization.In the literature,there are several methods proposed for indirect path identification[23],[24]. Individual Target Localization:After the Multiple Source Separation and Indirect Path Detection,the positioning system knows the signal impulses through the direct paths for each target.Then,the system extracts the characteristics of direct paths such as TOA and AOA.Based on these characteristics, the targets arefinally localized.Most researches on Individual Target Localization assumes that Multiple Source Separation and Indirect Path Detection are perfectly performed such as [16],[25]and[26].Note that the three challenges sometimes 3Note that here we omit the impulses having two or more interactions with the intruder because of the resulted low signal-to-noise radio(SNR)by multiple reflections.Cable for synchronizationFig.1.Illustration of TOA based Location Estimation System Model.are jointly addressed,so that the target locations are estimated in one step such as the method presented in [27].In this paper,we focus on the Individual Target Local-ization,under the same framework of [16],[25]and [26],assuming that Multiple Source Separation and Indirect Path Detection are perfectly performed in prior.In addition,we only use the TOA information for localization,which achieves very high accuracy with ultra-wideband signals.The method to ex-tract TOA information using background channel cancelation is described in details in [28]and also Section VII.B.System Model of Individual LocalizationFor ease of exposition,we consider the passive object (target)location estimation problem in a two-dimensional plane as shown in Fig.1.There is a target whose location [x,y ]is to be estimated by a system with one transmitter and M receivers.Without loss of generality,let the location of the transmitter be [0,0],and the location of the i th receiver be [a i ,b i ],1≤i ≤M .The transmitter transmits an impulse;the receivers subsequently receive the signal copies reflected from the target and other objects.We adopt the assumption also made in [16],[17]that the target reflects the signal into all ing (wired)backbone connections be-tween the transmitter and receivers,or high-accuracy wireless synchronization algorithms,the transmitter and receivers are synchronized.The errors of cable synchronization are negli-gible compared with the TOA measurement errors.Thus,at the estimation center,signal travel times can be obtained by comparing the departure time at the transmitter and the arrival time at the receivers.Let the TOA from the transmitter via the target to the i th receiver be t i ,and r i =c 0t i ,where c 0is the speed of light,1≤i ≤M .Then,r i = x 2+y 2+(x −a i )2+(y −b i )2i =1,...M.(1)For future use we define r =[r 1,r 2,...,r M ].Assuming each measurement involves an error,we haver i −ˆri =e i ,1≤i ≤M,where r i is the true value,ˆr i is the measured value and e i is the measurement error.In our model,the indirect paths areignored and we assume e i to be zero mean.The estimation system tries to find the [ˆx ,ˆy ],that best fits the above equations in the sense of minimizing the error varianceΔ=E [(ˆx −x )2+(ˆy −y )2].(2)Assuming the e i are Gaussian-distributed variables with zeromean and variances σ2i ,the conditional probability functionof the observations ˆr are formulated as follows:p (ˆr |z )=Ni =11√2πσi ·exp −(ˆr i −( x 2+y 2+ (x −a i )2+(y −b i )2))22σ2i,(3)where z =[x,y ].III.TSE M ETHODIn this section,we present the two steps of TSE andsummarize them in Algorithm 1.In the first step of TSE,we assume x ,y , x 2+y 2are independent of each other,and obtain temporary results for the target location based on this assumption.In the second step,we remove the assumption and update the estimation results.A.Step 1of TSEIn the first step of TSE,we obtain an initial estimate of[x,y, x 2+y 2],which is performed in two stages:Stage A and Stage B.The basic idea here is to utilize the linear approximation [16][29]to simplify the problem,considering that TOA measurement errors are small with UWB signals.Let v =x 2+y 2,taking the squares of both sides of (1)leads to2a i x +2b i y −2r i v =a 2i +b 2i −r 2i .Since r i −ˆr i =e i ,it follows that−a 2i +b 2i −ˆr 2i 2+a i x +b i y −ˆr i v=e i (v −ˆr i )−e 2i 2=e i (v −ˆr i )−O (e 2i ).(4)where O (•)is the Big O Notation meaning that f (α)=O (g (α))if and only if there exits a positive real number M and a real number αsuch that|f (α)|≤M |g (α)|for all α>α0.If e i is small,we can omit the second or higher order terms O (e 2i )in Eqn (4).In the following of this paper,we do this,leaving the linear (first order)term.Since there are M such equations,we can express them in a matrix form as followsh −S θ=Be +O (e 2)≈Be ,(5)whereh=⎡⎢⎢⎢⎢⎣−a21+b21−ˆr212−a22+b22−ˆr222...−a2M+b2M−ˆr2M2⎤⎥⎥⎥⎥⎦,S=−⎡⎢⎢⎢⎣a1b1−ˆr1a2b2−ˆr2...a Mb M−ˆr M⎤⎥⎥⎥⎦,θ=[x,y,v]T,e=[e1,e2,...,e M]T,andB=v·I−diag([r1,r2,...,r M]),(6) where O(e2)=[O(e21),O(e22),...,O(e2M)]T and diag(a) denotes the diagonal matrix with elements of vector a on its diagonal.For notational convenience,we define the error vectorϕ=h−Sθ.(7) According to(5)and(7),the mean ofϕis zero,and its covariance matrix is given byΨ=E(ϕϕT)=E(Bee T B T)+E(O(e2)e T B T)+E(Be O(e2)T)+E(O(e2)O(e2)T)≈¯BQ¯B T(8)where Q=diag[σ21,σ22,...,σ2M].Because¯B depends on the true values r,which are not obtainable,we use B(derived from the measurementsˆr)in our calculations.From(5)and the definition ofϕ,it follows thatϕis a vector of Gaussian variables;thus,the probability density function (pdf)ofϕgivenθisp(ϕ|θ)≈1(2π)M2|Ψ|12exp(−12ϕTΨ−1ϕ)=1(2π)M2|Ψ|12exp(−12(h−Sθ)TΨ−1(h−Sθ)).Then,lnp(ϕ|θ)≈−12(h−Sθ)TΨ−1(h−Sθ)+ln|Ψ|−M2ln2π(9)We assume for the moment that x,y,v are independent of each other(this clearly non-fulfilled assumption will be relaxed in the second step of the algorithm).Then,according to(9),the optimumθthat maximizes p(ϕ|θ)is equivalent to the one minimizingΠ=(h−Sθ)TΨ−1(h−Sθ)+ln|Ψ|. IfΨis a constant,the optimumθto minimizeΠsatisfies dΠdθθ=0.Taking the derivative ofΠoverθ,we havedΠdθθ=−2S TΨ−1h+2S TΨ−1Sθ.Fig.2.Illustration of estimation ofθin step1of TSE.Thus,the optimumθsatisfiesˆθ=arg minθ{Π}=(S TΨ−1S)−1S TΨ−1h,(10)which provides[ˆx,ˆy].Note that(10)also provides the leastsquares solution for non-Gaussian errors.However,for our problem,Ψis a function ofθsince Bdepends on the(unknown)values[x,y].For this reason,themaximum-likelihood(ML)estimation method in(10)can notbe directly used.Tofind the optimumθ,we perform theestimation in two stages:Stage A and Stage B.In Stage A,themissing data(Ψ)is calculated given the estimate of parameters(θ).Note thatθprovides the values of[x,y]and thus thevalue of B,therefore,Ψcan be calculated usingθby(8).In the Stage B,the parameters(θ)are updated according to(10)to maximize the likelihood function(which is equivalentto minimizingΠ).These two stages are iterated until con-vergence.Simulations in Section V show that commonly oneiteration is enough for TSE to closely approach the CRLB,which indicates that the global optimum is reached.B.Step2of TSEIn the above calculations,ˆθcontains three componentsˆx,ˆy andˆv.They were previously assumed to be independent;however,ˆx andˆy are clearly not independent ofˆv.As amatter of fact,we wish to eliminateˆv;this will be achievedby treatingˆx,ˆy,andˆv as random variables,and,knowing thelinear mapping of their squared values,the problem can besolved using the LS solution.Letˆθ=⎡⎣ˆxˆyˆv⎤⎦=⎡⎣x+n1y+n2v+n3⎤⎦(11)where n i(i=1,2,3)are the estimation errors of thefirststep.Obviously,the estimator(10)is an unbiased one,and themean of n i is zero.Before proceeding,we need the following Lemma.Lemma 1:By omitting the second or higher order errors,the covariance of ˆθcan be approximated as cov (ˆθ)=E (nn T )≈(¯S T Ψ−1¯S )−1.(12)where n =[n 1,n 2,n 3]T ,and Ψand ¯S(the mean value of S )use the true/mean values of x ,y,and r i .Proof:Please refer to the Appendix.Note that since the true values of x ,y,and r i are not obtain-able,we use the estimated/measured values in the calculationof cov (ˆθ).Let us now construct a vector g as followsg =ˆΘ−G Υ,(13)where ˆΘ=[ˆx 2,ˆy 2,ˆv 2]T ,Υ=[x 2,y 2]T and G =⎡⎣100111⎤⎦.Note that here ˆΘis the square of estimation result ˆθfrom the first step containing the estimated values ˆx ,ˆy and ˆv .Υis the vector to be estimated.If ˆΘis obtained without error,g =0and the location of the target is perfectly obtained.However,the error inevitably exists and we need to estimate Υ.Recalling that v =x 2+y 2,substituting (11)into (13),and omitting the second-order terms n 21,n 22,n 23,it follows that,g =⎡⎣2xn 1+O (n 21)2yn 2+O (n 22)2vn 3+O (n 23)⎤⎦≈⎡⎣2xn 12yn 22vn 3⎤⎦.Besides,following similar procedure as that in computing(8),we haveΩ=E (gg T )≈4¯D cov (ˆθ)¯D ,(14)where ¯D =diag ([¯x ,¯y ,¯v ]).Since x ,y are not known,¯Dis calculated as ˆD using the estimated values ˆx ,ˆy from the firststep.The vector g can be approximated as a vector of Gaussian variables.Thus the maximum likelihood estimation of Υis theone minimizing (ˆΘ−G Υ)T Ω−1(ˆΘ−G Υ),expressed by ˆΥ=(G T Ω−1G )−1G T Ω−1ˆΘ.(15)The value of Ωis calculated according to (14)using the valuesof ˆx and ˆy in the first step.Finally,the estimation of target location z is obtained byˆz =[ˆx ,ˆy ]=[±ˆΥ1,± ˆΥ2],(16)where ˆΥi is the i th item of Υ,i =1,2.To choose the correct one among the four values in (16),we can test the square error as followsχ=M i =1( ˆx 2+ˆy 2+ (ˆx −a i )2+(ˆy −b i )−ˆr i )2.(17)The value of z that minimizes χis considered as the final estimate of the target location.In summary,the procedure of TSE is listed in Algorithm 1:Note that one should avoid placing the receivers on a line,since in this case (S T Ψ−1S )−1can become nearly singular,and solving (10)is not accurate.Algorithm 1TSE Location Estimation Method1.In the first step,use algorithm as shown in Fig.2to obtain ˆθ,2.In the second step,use the values of ˆx and ˆy from ˆθ,generate ˆΘand D ,and calculate Ω.Then,calculate the value of ˆΥby (15),3.Among the four candidate values of ˆz =[ˆx ,ˆy ]obtained by (16),choose the one minimizing (17)as the final estimate for target location.IV.C OMPARISON OF CRLB BETWEEN TDOA AND TOA In this section,we derive the CRLB of TOA based estima-tion algorithms and show that it is much lower (can be 30dB lower)than the CRLB of TDOA algorithms.The CRLB of “active”TOA localization has been studied in [30].The “passive”localization has been studied before under the model of multistatic radar [31],[32],[33].The difference between our model and the radar model is that in our model the localization error is a function of errors of TOA measurements,while in the radar model the localization error is a function of signal SNR and waveform.The CRLB is related to the 2×2Fisher Information Matrix (FIM)[34],J ,whose components J 11,J 12,J 21,J 22are defined in (18)–(20)as follows J 11=−E (∂2ln(p (ˆr |z ))∂x 2)=ΣM i =11σ2i (x −a i (x −a i )2+(y −b i )2+xx 2+y2)2,(18)J 12=J 21=−E (∂2ln(p (ˆr |z ))∂x∂y )=ΣM i =11σ2i (x −a i (x −a i )2+(y −b i )2+x x 2+y 2)×(y −b i (x −a i )2+(y −b i )2+yx 2+y 2),(19)J 22=−E (∂2ln(p (ˆr |z ))∂y 2)=ΣM i =11σ2i (y −b i (x −a i )2+(y −b i )2+yx 2+y2)2.(20)This can be expressed asJ =U T Q −1U ,(21)where Q is defined after Eqn.(8),and the entries of U in the first and second column are{U }i,1=x ¯r i −a ix 2+y 2(x −a i )2+(y −b i )2 x 2+y 2,(22)and{U }i,2=y ¯r i −b ix 2+y 2(x −a i )2+(y −b i )2 x 2+y 2,(23)with ¯r i =(x −a i )2+(y −b i )2+ x 2+y 2.The CRLB sets the lower bound for the variance of esti-mation error of TOA algorithms,which can be expressed as [34]E [(ˆx −x )2+(ˆy −y )2]≥ J −1 1,1+J −1 2,2=CRLB T OA ,(24)where ˆx and ˆy are the estimated values of x and y ,respec-tively,and J −1 i,j is the (i,j )th element of the inverse matrix of J in (21).For the TDOA estimation,its CRLB has been derived in [16].The difference of signal travel time between several receivers are considered:(x −a i )2+(y −b i )2−(x −a 1)2+(y −b 1)2=r i −r 1=l i ,2≤i ≤M.(25)Let l =[l 2,l 3,...,l M ]T ,and t be the observa-tions/measurements of l ,then,the conditional probability density function of t is p (t |z )=1(2π)(M −1)/2|Z |12×exp(−12(t −l )T Z −1(t −l )),where Z is the correlation matrix of t ,Z =E (tt T ).Then,the FIM is expressed as [16]ˇJ=ˇU T Z −1ˇU (26)where ˇUis a M −1×2matrix defined as ˇU i,1=x −a i (x −a i )2+(y −b i )2−x −a 1(x −a 1)2+(y −b 1)2,ˇUi,2=y −b i (x −a i )2+(y −b i )2−y −b 1(x −a 1)2+(y −b 1)2.The CRLB sets the lower bound for the variance of esti-mation error of TDOA algorithms,which can be expressed as [34]:E [(ˆx −x )2+(ˆy −y )2]≥ ˇJ −1 1,1+ ˇJ −1 2,2=CRLB T DOA .(27)Note that the correlation matrix Q for TOA is different from the correlation matrix Z for TDOA.Assume the variance of TOA measurement at i th (1≤i ≤M )receiver is σ2i ,it follows that:Q (i,j )=σ2i i =j,0i =j.and Z (i,j )= σ21+σ2i +1i =j,σ21i =j.As an example,we consider a scenario wherethere is a transmitter at [0,0],and four receivers at [−6,2],[6.2,1.4],[1.5,4],[2,2.3].The range of the targetlocations is 1≤x ≤10,1≤y ≤10.The ratio of CRLB of TOA over that of TDOA is plotted in Fig.3.Fig.3(a)shows the contour plot while Fig.3(b)shows the color-coded plot.It can be observed that the CRLB of TOA is always —in most cases significantly —lower than that of TDOA.xy(a )xy0.10.20.30.40.50.60.70.80.9Fig.3.CRLB ratio of passive TOA over passive TDOA estimation:(a)contour plot;(b)pcolor plot.V.P ERFORMANCE OF TSEIn this section,we first prove that the TSE can achieve the CRLB of TOA algorithms by showing that the estimation error variance of TSE is the same as the CRLB of TOA algorithms.In addition,we show that,for small TOA error regions,the estimated target location is approximately a Gaussian random variable whose covariance matrix is the inverse of the Fisher Information Matrix (FIM),which in turn is related to the CRLB.Similar to the reasoning in Lemma 1,we can obtain the variance of error in the estimation of Υas follows:cov (ˆΥ)≈(G T Ω−1G )−1.(28)Let ˆx =x +e x ,ˆy=y +e y ,and insert them into Υ,omitting the second order errors,we obtainˆΥ1−x 2=2xe x +O (e 2x )≈2xe x ˆΥ2−y 2=2ye y +O (e 2y)≈2ye y (29)Then,the variance of the final estimate of target location ˆzis cov (ˆz )=E (e x e ye x e y )≈14C −1E ( Υ1−x 2Υ2−y 2Υ1−x 2Υ2−y 2 )C −1=14C −1cov (ˆΥ)C −1,(30)where C = x 00y.Substituting (14),(28),(12)and (8)into (30),we can rewrite cov (ˆz )as cov (ˆz )≈(W T Q −1W )−1(31)where W =B −1¯SD−1GC .Since we are computing an error variance,B (19),¯S(5)and D (14)are calculated using the true (mean)value of x ,y and r i .Using (19)and (1),we can rewrite B =−diag ([d 1,d 2,...,d M ]),whered i=(x−a i)2+(y−b i)2.Then B−1¯SD−1is given by B−1¯SD−1=⎡⎢⎢⎢⎢⎢⎣a1xd1b1yd1−¯r1√x2+y2d1a2xd2b2yd2−¯r2√x2+y2d2.........a Mxd Mb Myd M−¯r M√x2+y2d M⎤⎥⎥⎥⎥⎥⎦.(32)Consequently,we obtain the entries of W as{W}i,1=x¯r i−a ix2+y2(x−a i)2+(y−b i)2x2+y2,(33){W}i,2=y¯r i −b ix2+y2(x−a i)2+(y−b i)2x2+y2.(34)where{W}i,j denotes the entry at the i th row and j th column.From this we can see that W=paring(21)and (31),it followscov(ˆz)≈J−1.(35) Then,E[(ˆx−x)2+(ˆy−y)2]≈J−11,1+J−12,2.Therefore,the variance of the estimation error is the same as the CRLB.In the following,wefirst employ an example to show that[ˆx,ˆy]obtained by TSE are Gaussian distributed with covariance matrix J−1,and then give the explanation for this phenomenon.Let the transmitter be at[0,0],target at[0.699, 4.874]and four receivers at[-1,1],[2,1],[-31.1]and[4 0].The signal travel distance variance at four receivers are [0.1000,0.1300,0.1200,0.0950]×10−4.The two dimensional probability density function(PDF)of[ˆx,ˆy]is shown in Fig.4 (a).To verify the Gaussianity of[ˆx,ˆy],the difference between the PDF of[ˆx,ˆy]and the PDF of Gaussian distribution with mean[¯x,¯y]and covariance J−1is plotted in Fig.4(b).The Gaussianity of[ˆx,ˆy]can be explained as follows.Eqn.(35)means that the covariance of thefinal estimation of target location is the FIM related to CRLB.We could further study the distribution of[e x,e y].The basic idea is that by omitting the second or high order and nonlinear errors,[e x,e y]can be written as linear function of e:1)According to(29),[e x,e y]are approximately lineartransformations ofˆΥ.2)(15)means thatˆΥis approximately a linear transfor-mation ofˆΘ.Here we could omit the nonlinear errors occurred in the estimate/calculation ofΩ.3)According to(11),ˆΘ≈¯θ2+2¯θn+n2,thus,omittingthe second order error,thus,ˆΘis approximately a linear transformation of n.4)(10)and(39)mean that n is approximately a lineartransformation of e.Here we could omit the nonlinear errors accrued in the estimate of S andΨ.Thus,we could approximately write[e x,e y]as a linear trans-formation of e,thus,[e x,e y]can be approximated as Gaussian variables.Fig.4.(a):PDF of[ˆx,ˆy]by TSE(b):difference between the PDF of[ˆx,ˆy] by TSE and PDF of Gaussian distribution with mean[¯x,¯y]and covariance J−1.Fig.5.Simulation results of TSE for thefirst configuration.VI.S IMULATION R ESULTSIn this section,wefirst compare the performance of TSE with that TDOA algorithm proposed in[16]and CRLBs.Then, we show the performance of TSE at high TOA measurement error scenario.For comparison,the performance of a Quasi-Newton iterative method[35]is shown.To verify our theoretical analysis,six different system con-figurations are simulated.The transmitter is at[0,0]for all six configurations,and the receiver locations and error variances are listed in Table I.Figures5,6and7show simulation results comparing the distance to the target(Configuration1vs. Configuration2),the receiver separation(Configuration3vs. Configuration4)and the number of receivers(Configuration5 vs.Configuration6),respectively4.In eachfigure,10000trails are simulated and the estimation variance of TSE estimate is compared with the CRLB of TDOA and TOA based localization schemes.For comparison,the simulation results of error variance of the TDOA method proposed in[16]are also drawn in eachfigure.It can be observed that1)The localization error of TSE can closely approach theCRLB of TOA based positioning algorithms.4During the simulations,only one iteration is used for the calculation of B(19).。

Face Recognition A Literature Review

Face Recognition A Literature Review

Abstract—The task of face recognition has been actively researched in recent years. This paper provides an up-to-date review of major human face recognition research. We first present an overview of face recognition and its applications. Then, a literature review of the most recent face recognition techniques is presented. Description and limitations of face databases which are used to test the performance of these face recognition algorithms are given. A brief summary of the face recognition vendor test (FRVT) 2002, a large scale evaluation of automatic face recognition technology, and its conclusions are also given. Finally, we give a summary of the research results.Keywords—Combined classifiers, face recognition, graph matching, neural networks.I.I NTRODUCTIONACE recognition is an important research problem spanning numerous fields and disciplines. This because face recognition, in additional to having numerous practical applications such as bankcard identification, access control, Mug shots searching, security monitoring, and surveillance system, is a fundamental human behaviour that is essential for effective communications and interactions among people.A formal method of classifying faces was first proposed in[1]. The author proposed collecting facial profiles as curves, finding their norm, and then classifying other profiles by their deviations from the norm. This classification is multi-modal, i.e. resulting in a vector of independent measures that could be compared with other vectors in a database.Progress has advanced to the point that face recognition systems are being demonstrated in real-world settings [2]. The rapid development of face recognition is due to a combination of factors: active development of algorithms, the availability of a large databases of facial images, and a method for evaluating the performance of face recognition algorithms.In the literatures, face recognition problem can be formulated as: given static (still) or video images of a scene, identify or verify one or more persons in the scene by comparing with faces stored in a database.When comparing person verification to face recognition, there are several aspects which differ. First, a client – an authorized user of a personal identification system – is Manuscript received February 22, 2005.A. S. Tolba is with the Information Systems Department, Mansoura University, Egypt, (e-mail: tolba1954@)).A. H. EL-Baz is with the Mathematics Department, Damietta Faculty of Science, New Damietta, Egypt, and doing PhD research on pattern recognition (phone: 0020-57-403980; Fax: 0020-57–403868; e-mail: ali_elbaz@).A. H. EL-Harby is with the Mathematics Department, Damietta Faculty of Science, New Damietta, Egypt, (e-mail: elharby@). assumed to be co-operative and makes an identity claim. Computationally this means that it is not necessary to consult the complete set of database images (denoted model images below) in order to verify a claim. An incoming image (referred to as a probe image) is thus compared to a small number of model images of the person whose identity is claimed and not, as in the recognition scenario, with every image (or some descriptor of an image) in a potentially large database. Second, an automatic authentication system must operate in near-real time to be acceptable to users. Finally, in recognition experiments, only images of people from the training database are presented to the system, whereas the case of an imposter (most likely a previously unseen person) is of utmost importance for authentication.Face recognition is a biometric approach that employs automated methods to verify or recognize the identity of a living person based on his/her physiological characteristics. In general, a biometric identification system makes use of either physiological characteristics (such as a fingerprint, iris pattern, or face) or behaviour patterns (such as hand-writing, voice, or key-stroke pattern) to identify a person. Because of human inherent protectiveness of his/her eyes, some people are reluctant to use eye identification systems. Face recognition has the benefit of being a passive, non intrusive system to verify personal identity in a “natural” and friendly way.In general, biometric devices can be explained with a three-step procedure (1) a sensor takes an observation. The type of sensor and its observation depend on the type of biometric devices used. This observation gives us a “Biometric Signature” of the individual. (2) a computer algorithm “normalizes” the biometric signature so that it is in the same format (size, resolution, view, etc.) as the signatures on the system’s database. The normalization of the biometric signature gives us a “Normalized Signature” of the individual.(3) a matcher compares the normalized signature with the set (or sub-set) of normalized signatures on the system's database and provides a “similarity score” that compares the individual's normalized signature with each signature in the database set (or sub-set). What is then done with the similarity scores depends on the biometric system’s application?Face recognition starts with the detection of face patterns in sometimes cluttered scenes, proceeds by normalizing the face images to account for geometrical and illumination changes, possibly using information about the location and appearance of facial landmarks, identifies the faces using appropriate classification algorithms, and post processes the results using model-based schemes and logistic feedback [3].The application of face recognition technique can be categorized into two main parts: law enforcement application and commercial application. Face recognition technology isFace Recognition: A Literature ReviewA. S. Tolba, A.H. El-Baz, and A.A. El-HarbyFprimarily used in law enforcement applications, especially Mug shot albums (static matching) and video surveillance (real-time matching by video image sequences). The commercial applications range from static matching of photographs on credit cards, ATM cards, passports, driver’s licenses, and photo ID to real-time matching with still images or video image sequences for access control. Each application presents different constraints in terms of processing.All face recognition algorithms consistent of two major parts: (1) face detection and normalization and (2) face identification. Algorithms that consist of both parts are referred to as fully automatic algorithms and those that consist of only the second part are called partially automatic algorithms. Partially automatic algorithms are given a facial image and the coordinates of the center of the eyes. Fully automatic algorithms are only given facial images. On the other hand, the development of face recognition over the past years allows an organization into three types of recognition algorithms, namely frontal, profile, and view-tolerant recognition, depending on the kind of images and the recognition algorithms. While frontal recognition certainly is the classical approach, view-tolerant algorithms usually perform recognition in a more sophisticated fashion by taking into consideration some of the underlying physics, geometry, and statistics. Profile schemes as stand-alone systems have a rather marginal significance for identification, (for more detail see [4]). However, they are very practical either for fast coarse pre-searches of large face database to reduce the computational load for a subsequent sophisticated algorithm, or as part of a hybrid recognition scheme. Such hybrid approaches have a special status among face recognition systems as they combine different recognition approaches in an either serial or parallel order to overcome the shortcoming of the individual components.Another way to categorize face recognition techniques is to consider whether they are based on models or exemplars. Models are used in [5] to compute the Quotient Image, and in [6] to derive their Active Appearance Model. These models capture class information (the class face), and provide strong constraints when dealing with appearance variation. At the other extreme, exemplars may also be used for recognition. The ARENA method in [7] simply stores all training and matches each one against the task image. As far we can tell, current methods that employ models do not use exemplars, and vice versa. This is because these two approaches are by no means mutually exclusive. Recently, [8] proposed a way of combining models and exemplars for face recognition. In which, models are used to synthesize additional training images, which can then be used as exemplars in the learning stage of a face recognition system.Focusing on the aspect of pose invariance, face recognition approaches may be divided into two categories: (i) global approach and (ii) component-based approach. In global approach, a single feature vector that represents the whole face image is used as input to a classifier. Several classifiers have been proposed in the literature e.g. minimum distance classification in the eigenspace [9,10], Fisher’s discriminant analysis [11], and neural networks [12]. Global techniques work well for classifying frontal views of faces. However, they are not robust against pose changes since global features are highly sensitive to translation and rotation of the face. To avoid this problem an alignment stage can be added before classifying the face. Aligning an input face image with a reference face image requires computing correspondence between the two face images. The correspondence is usually determined for a small number of prominent points in the face like the center of the eye, the nostrils, or the corners of the mouth. Based on these correspondences, the input face image can be warped to a reference face image.In [13], an affine transformation is computed to perform the warping. Active shape models are used in [14] to align input faces with model faces. A semi-automatic alignment step in combination with support vector machines classification was proposed in [15]. An alternative to the global approach is to classify local facial components. The main idea of component based recognition is to compensate for pose changes by allowing a flexible geometrical relation between the components in the classification stage.In [16], face recognition was performed by independently matching templates of three facial regions (eyes, nose and mouth). The configuration of the components during classification was unconstrained since the system did not include a geometrical model of the face. A similar approach with an additional alignment stage was proposed in [17]. In [18], a geometrical model of a face was implemented by a 2D elastic graph. The recognition was based on wavelet coefficients that were computed on the nodes of the elastic graph. In [19], a window was shifted over the face image and the DCT coefficients computed within the window were fed into a 2D Hidden Markov Model.Face recognition research still face challenge in some specific domains such as pose and illumination changes. Although numerous methods have been proposed to solve such problems and have demonstrated significant promise, the difficulties still remain. For these reasons, the matching performance in current automatic face recognition is relatively poor compared to that achieved in fingerprint and iris matching, yet it may be the only available measuring tool for an application. Error rates of 2-25% are typical. It is effective if combined with other biometric measurements.Current systems work very well whenever the test image to be recognized is captured under conditions similar to those of the training images. However, they are not robust enough if there is variation between test and training images [20]. Changes in incident illumination, head pose, facial expression, hairstyle (include facial hair), cosmetics (including eyewear) and age, all confound the best systems today.As a general rule, we may categorize approaches used to cope with variation in appearance into three kinds: invariant features, canonical forms, and variation- modeling. The first approach seeks to utilize features that are invariant to the changes being studied. For instance, the Quotient Image [5] is (by construction) invariant to illumination and may be used to recognize faces (assumed to be Lambertian) when lighting conditions change.The second approach attempts to “normalize” away the variation, either by clever image transformations or by synthesizing a new image (from the given test image) in some“canonical” or “prototypical” form. Recognition is then performed using this canonical form. Examples of this approach include [21,22]. In [21], for instance, the test image under arbitrary illumination is re-rendered under frontal illumination, and then compared against other frontally-illuminated prototypes.The third approach of variation-modeling is self explanatory: the idea is to learn, in some suitable subspace, the extent of the variation in that space. This usually leads to some parameterization of the subspace(s). Recognition is then performed by choosing the subspace closest to the test image, after the latter has been appropriately mapped. In effect, the recognition step recovers the variation (e.g. pose estimation) as well as the identity of the person. For examples of this technique, see [18, 23, 24 and 25].Despite the plethora of techniques, and the valiant effort of many researchers, face recognition remains a difficult, unsolved problem in general. While each of the above approaches works well for the specific variation being studied, performance degrades rapidly when other variations are present. For instance, a feature invariant to illumination works well as long as pose or facial expression remains constant, but fails to be invariant when pose or expression is changed. This is not a problem for some applications, such as controlling access to a secured room, since both the training and test images may be captured under similar conditions. However, for general, unconstrained recognition, none of these techniques are robust enough.Moreover, it is not clear that different techniques can be combined to overcome each other’s limitations. Some techniques, by their very nature, exclude others. For example, the Symmetric Shape-from-Shading method of [22] relies on the approximate symmetry of a frontal face. It is unclear how this may be combined with a technique that depends on side profiles, where the symmetry is absent.We can make two important observations after surveying the research literature: (1) there does not appear to be any feature, set of features, or subspace that is simultaneously invariant to all the variations that a face image may exhibit, (2) given more training images, almost any technique will perform better. These two factors are the major reasons why face recognition is not widely used in real-world applications. The fact is that for many applications, it is usual to require the ability to recognize faces under different variations, even when training images are severely limited.II.L ITERATURE R EVIEW OF F ACE R ECOGNITION T ECHNIQUES This section gives an overview on the major human face recognition techniques that apply mostly to frontal faces, advantages and disadvantages of each method are also given. The methods considered are eigenfaces (eigenfeatures), neural networks, dynamic link architecture, hidden Markov model, geometrical feature matching, and template matching. The approaches are analyzed in terms of the facial representations they used.A.EigenfacesEigenface is one of the most thoroughly investigated approaches to face recognition. It is also known as Karhunen- Loève expansion, eigenpicture, eigenvector, and principal component. References [26, 27] used principal component analysis to efficiently represent pictures of faces. They argued that any face images could be approximately reconstructed by a small collection of weights for each face and a standard face picture (eigenpicture). The weights describing each face are obtained by projecting the face image onto the eigenpicture. Reference [28] used eigenfaces, which was motivated by the technique of Kirby and Sirovich, for face detection and identification.In mathematical terms, eigenfaces are the principal components of the distribution of faces, or the eigenvectors of the covariance matrix of the set of face images. The eigenvectors are ordered to represent different amounts of the variation, respectively, among the faces. Each face can be represented exactly by a linear combination of the eigenfaces. It can also be approximated using only the “best” eigenvectors with the largest eigenvalues. The best M eigenfaces construct an M dimensional space, i.e., the “face space”. The authors reported 96 percent, 85 percent, and 64 percent correct classifications averaged over lighting, orientation, and size variations, respectively. Their database contained 2,500 images of 16 individuals.As the images include a large quantity of background area, the above results are influenced by background. The authors explained the robust performance of the system under different lighting conditions by significant correlation between images with changes in illumination. However, [29] showed that the correlation between images of the whole faces is not efficient for satisfactory recognition performance. Illumination normalization [27] is usually necessary for the eigenfaces approach.Reference [30] proposed a new method to compute the covariance matrix using three images each was taken in different lighting conditions to account for arbitrary illumination effects, if the object is Lambertian. Reference [31] extended their early work on eigenface to eigenfeatures corresponding to face components, such as eyes, nose, and mouth. They used a modular eigenspace which was composed of the above eigenfeatures (i.e., eigeneyes, eigennose, and eigenmouth). This method would be less sensitive to appearance changes than the standard eigenface method. The system achieved a recognition rate of 95 percent on the FERET database of 7,562 images of approximately 3,000 individuals. In summary, eigenface appears as a fast, simple, and practical method. However, in general, it does not provide invariance over changes in scale and lighting conditions. Recently, in [32] experiments with ear and face recognition, using the standard principal component analysis approach , showed that the recognition performance is essentially identical using ear images or face images and combining the two for multimodal recognition results in a statistically significant performance improvement. For example, the difference in the rank-one recognition rate for the day variation experiment using the 197-image training sets is90.9% for the multimodal biometric versus 71.6% for the ear and 70.5% for the face.There is substantial related work in multimodal biometrics. For example [33] used face and fingerprint in multimodal biometric identification, and [34] used face and voice. However, use of the face and ear in combination seems more relevant to surveillance applications.B.Neural NetworksThe attractiveness of using neural networks could be due to its non linearity in the network. Hence, the feature extraction step may be more efficient than the linear Karhunen-Loève methods. One of the first artificial neural networks (ANN) techniques used for face recognition is a single layer adaptive network called WISARD which contains a separate network for each stored individual [35]. The way in constructing a neural network structure is crucial for successful recognition. It is very much dependent on the intended application. For face detection, multilayer perceptron [36] and convolutional neural network [37] have been applied. For face verification, [38] is a multi-resolution pyramid structure. Reference [37] proposed a hybrid neural network which combines local image sampling, a self-organizing map (SOM) neural network, and a convolutional neural network. The SOM provides a quantization of the image samples into a topological space where inputs that are nearby in the original space are also nearby in the output space, thereby providing dimension reduction and invariance to minor changes in the image sample. The convolutional network extracts successively larger features in a hierarchical set of layers and provides partial invariance to translation, rotation, scale, and deformation. The authors reported 96.2% correct recognition on ORL database of 400 images of 40 individuals.The classification time is less than 0.5 second, but the training time is as long as 4 hours. Reference [39] used probabilistic decision-based neural network (PDBNN) which inherited the modular structure from its predecessor, a decision based neural network (DBNN) [40]. The PDBNN can be applied effectively to 1) face detector: which finds the location of a human face in a cluttered image, 2) eye localizer: which determines the positions of both eyes in order to generate meaningful feature vectors, and 3) face recognizer. PDNN does not have a fully connected network topology. Instead, it divides the network into K subnets. Each subset is dedicated to recognize one person in the database. PDNN uses the Guassian activation function for its neurons, and the output of each “face subnet” is the weighted summation of the neuron outputs. In other words, the face subnet estimates the likelihood density using the popular mixture-of-Guassian model. Compared to the AWGN scheme, mixture of Guassian provides a much more flexible and complex model for approximating the time likelihood densities in the face space. The learning scheme of the PDNN consists of two phases, in the first phase; each subnet is trained by its own face images. In the second phase, called the decision-based learning, the subnet parameters may be trained by some particular samples from other face classes. The decision-based learning scheme does not use all the training samples for the training. Only misclassified patterns are used. If the sample is misclassified to the wrong subnet, the rightful subnet will tune its parameters so that its decision-region can be moved closer to the misclassified sample.PDBNN-based biometric identification system has the merits of both neural networks and statistical approaches, and its distributed computing principle is relatively easy to implement on parallel computer. In [39], it was reported that PDBNN face recognizer had the capability of recognizing up to 200 people and could achieve up to 96% correct recognition rate in approximately 1 second. However, when the number of persons increases, the computing expense will become more demanding. In general, neural network approaches encounter problems when the number of classes (i.e., individuals) increases. Moreover, they are not suitable for a single model image recognition test because multiple model images per person are necessary in order for training the systems to “optimal” parameter setting.C.Graph MatchingGraph matching is another approach to face recognition. Reference [41] presented a dynamic link structure for distortion invariant object recognition which employed elastic graph matching to find the closest stored graph. Dynamic link architecture is an extension to classical artificial neural networks. Memorized objects are represented by sparse graphs, whose vertices are labeled with a multiresolution description in terms of a local power spectrum and whose edges are labeled with geometrical distance vectors. Object recognition can be formulated as elastic graph matching which is performed by stochastic optimization of a matching cost function. They reported good results on a database of 87 people and a small set of office items comprising different expressions with a rotation of 15 degrees.The matching process is computationally expensive, taking about 25 seconds to compare with 87 stored objects on a parallel machine with 23 transputers. Reference [42] extended the technique and matched human faces against a gallery of 112 neutral frontal view faces. Probe images were distorted due to rotation in depth and changing facial expression. Encouraging results on faces with large rotation angles were obtained. They reported recognition rates of 86.5% and 66.4% for the matching tests of 111 faces of 15 degree rotation and 110 faces of 30 degree rotation to a gallery of 112 neutral frontal views. In general, dynamic link architecture is superior to other face recognition techniques in terms of rotation invariance; however, the matching process is computationally expensive.D.Hidden Markov Models (HMMs)Stochastic modeling of nonstationary vector time series based on (HMM) has been very successful for speech applications. Reference [43] applied this method to human face recognition. Faces were intuitively divided into regions such as the eyes, nose, mouth, etc., which can be associated with the states of a hidden Markov model. Since HMMs require a one-dimensional observation sequence and images are two-dimensional, the images should be converted into either 1D temporal sequences or 1D spatial sequences.In [44], a spatial observation sequence was extracted from a face image by using a band sampling technique. Each face image was represented by a 1D vector series of pixel observation. Each observation vector is a block of L lines and there is an M lines overlap between successive observations. An unknown test image is first sampled to an observation sequence. Then, it is matched against every HMMs in the model face database (each HMM represents a different subject). The match with the highest likelihood is considered the best match and the relevant model reveals the identity of the test face.The recognition rate of HMM approach is 87% using ORL database consisting of 400 images of 40 individuals. A pseudo 2D HMM [44] was reported to achieve a 95% recognition rate in their preliminary experiments. Its classification time and training time were not given (believed to be very expensive). The choice of parameters had been based on subjective intuition.E.Geometrical Feature MatchingGeometrical feature matching techniques are based on the computation of a set of geometrical features from the picture of a face. The fact that face recognition is possible even at coarse resolution as low as 8x6 pixels [45] when the single facial features are hardly revealed in detail, implies that the overall geometrical configuration of the face features is sufficient for recognition. The overall configuration can be described by a vector representing the position and size of the main facial features, such as eyes and eyebrows, nose, mouth, and the shape of face outline.One of the pioneering works on automated face recognition by using geometrical features was done by [46] in 1973. Their system achieved a peak performance of 75% recognition rate on a database of 20 people using two images per person, one as the model and the other as the test image. References [47,48] showed that a face recognition program provided with features extracted manually could perform recognition apparently with satisfactory results. Reference [49] automatically extracted a set of geometrical features from the picture of a face, such as nose width and length, mouth position, and chin shape. There were 35 features extracted form a 35 dimensional vector. The recognition was then performed with a Bayes classifier. They reported a recognition rate of 90% on a database of 47 people.Reference [50] introduced a mixture-distance technique which achieved 95% recognition rate on a query database of 685 individuals. Each face was represented by 30 manually extracted distances. Reference [51] used Gabor wavelet decomposition to detect feature points for each face image which greatly reduced the storage requirement for the database. Typically, 35-45 feature points per face were generated. The matching process utilized the information presented in a topological graphic representation of the feature points. After compensating for different centroid location, two cost values, the topological cost, and similarity cost, were evaluated. The recognition accuracy in terms of the best match to the right person was 86% and 94% of the correct person's faces was in the top three candidate matches.In summary, geometrical feature matching based on precisely measured distances between features may be most useful for finding possible matches in a large database such as a Mug shot album. However, it will be dependent on the accuracy of the feature location algorithms. Current automated face feature location algorithms do not provide a high degree of accuracy and require considerable computational time.F.Template MatchingA simple version of template matching is that a test image represented as a two-dimensional array of intensity values is compared using a suitable metric, such as the Euclidean distance, with a single template representing the whole face. There are several other more sophisticated versions of template matching on face recognition. One can use more than one face template from different viewpoints to represent an individual's face.A face from a single viewpoint can also be represented by a set of multiple distinctive smaller templates [49,52]. The face image of gray levels may also be properly processed before matching [53]. In [49], Bruneli and Poggio automatically selected a set of four features templates, i.e., the eyes, nose, mouth, and the whole face, for all of the available faces. They compared the performance of their geometrical matching algorithm and template matching algorithm on the same database of faces which contains 188 images of 47 individuals. The template matching was superior in recognition (100 percent recognition rate) to geometrical matching (90 percent recognition rate) and was also simpler. Since the principal components (also known as eigenfaces or eigenfeatures) are linear combinations of the templates in the data basis, the technique cannot achieve better results than correlation [49], but it may be less computationally expensive. One drawback of template matching is its computational complexity. Another problem lies in the description of these templates. Since the recognition system has to be tolerant to certain discrepancies between the template and the test image, this tolerance might average out the differences that make individual faces unique.In general, template-based approaches compared to feature matching are a more logical approach. In summary, no existing technique is free from limitations. Further efforts are required to improve the performances of face recognition techniques, especially in the wide range of environments encountered in real world.G.3D Morphable ModelThe morphable face model is based on a vector space representation of faces [54] that is constructed such that any convex combination of shape and texture vectors of a set of examples describes a realistic human face.Fitting the 3D morphable model to images can be used in two ways for recognition across different viewing conditions: Paradigm 1. After fitting the model, recognition can be based on model coefficients, which represent intrinsic shape and texture of faces, and are independent of the imaging conditions: Paradigm 2. Three-dimension face reconstruction can also be employed to generate synthetic views from gallery probe images [55-58]. The synthetic views are then。

人脸识别的英文文献15篇

人脸识别的英文文献15篇

人脸识别的英文文献15篇英文回答:1. Title: A Survey on Face Recognition Algorithms.Abstract: Face recognition is a challenging task in computer vision due to variations in illumination, pose, expression, and occlusion. This survey provides a comprehensive overview of the state-of-the-art face recognition algorithms, including traditional methods like Eigenfaces and Fisherfaces, and deep learning-based methods such as Convolutional Neural Networks (CNNs).2. Title: Face Recognition using Deep Learning: A Literature Review.Abstract: Deep learning has revolutionized the field of face recognition, leading to significant improvements in accuracy and robustness. This literature review presents an in-depth analysis of various deep learning architecturesand techniques used for face recognition, highlighting their strengths and limitations.3. Title: Real-Time Face Recognition: A Comprehensive Review.Abstract: Real-time face recognition is essential for various applications such as surveillance, access control, and biometrics. This review surveys the recent advances in real-time face recognition algorithms, with a focus on computational efficiency, accuracy, and scalability.4. Title: Facial Expression Recognition: A Comprehensive Survey.Abstract: Facial expression recognition plays a significant role in human-computer interaction and emotion analysis. This survey presents a comprehensive overview of facial expression recognition techniques, including traditional approaches and deep learning-based methods.5. Title: Age Estimation from Facial Images: A Review.Abstract: Age estimation from facial images has applications in various fields, such as law enforcement, forensics, and healthcare. This review surveys the existing age estimation methods, including both supervised and unsupervised learning approaches.6. Title: Face Detection: A Literature Review.Abstract: Face detection is a fundamental task in computer vision, serving as a prerequisite for face recognition and other facial analysis applications. This review presents an overview of face detection techniques, from traditional methods to deep learning-based approaches.7. Title: Gender Classification from Facial Images: A Survey.Abstract: Gender classification from facial imagesis a widely studied problem with applications in gender-specific marketing, surveillance, and security. This surveyprovides an overview of gender classification methods, including both traditional and deep learning-based approaches.8. Title: Facial Keypoint Detection: A Comprehensive Review.Abstract: Facial keypoint detection is a crucialstep in face analysis, providing valuable information about facial structure. This review surveys facial keypoint detection methods, including traditional approaches anddeep learning-based algorithms.9. Title: Face Tracking: A Survey.Abstract: Face tracking is vital for real-time applications such as video surveillance and facial animation. This survey presents an overview of facetracking techniques, including both model-based andfeature-based approaches.10. Title: Facial Emotion Analysis: A Literature Review.Abstract: Facial emotion analysis has become increasingly important in various applications, including affective computing, human-computer interaction, and surveillance. This literature review provides a comprehensive overview of facial emotion analysis techniques, from traditional methods to deep learning-based approaches.11. Title: Deep Learning for Face Recognition: A Comprehensive Guide.Abstract: Deep learning has emerged as a powerful technique for face recognition, achieving state-of-the-art results. This guide provides a comprehensive overview of deep learning architectures and techniques used for face recognition, including Convolutional Neural Networks (CNNs) and Deep Residual Networks (ResNets).12. Title: Face Recognition with Transfer Learning: A Survey.Abstract: Transfer learning has become a popular technique for accelerating the training of deep learning models. This survey presents an overview of transferlearning approaches used for face recognition, highlighting their advantages and limitations.13. Title: Domain Adaptation for Face Recognition: A Comprehensive Review.Abstract: Domain adaptation is essential foradapting face recognition models to new domains withdifferent characteristics. This review surveys various domain adaptation techniques used for face recognition, including adversarial learning and self-supervised learning.14. Title: Privacy-Preserving Face Recognition: A Comprehensive Guide.Abstract: Privacy concerns have arisen with the widespread use of face recognition technology. This guide provides an overview of privacy-preserving face recognition techniques, including anonymization, encryption, anddifferential privacy.15. Title: The Ethical and Social Implications of Face Recognition Technology.Abstract: The use of face recognition technology has raised ethical and social concerns. This paper explores the potential risks and benefits of face recognition technology, and discusses the implications for society.中文回答:1. 题目,人脸识别算法综述。

基于YOLOv3的人脸关键点检测

基于YOLOv3的人脸关键点检测

^mmmm2021年第01期(总第217期)基于YOLOv3的人脸关键点检测屈金山,朱泽群,万秋波(三峡大学计算机与信息学院,湖北宜昌443002)摘要:深度学习中神经网络强大的特征提取能力,使非约束场景下的人脸检测不再困难,于是人脸关键点的检测逐渐成 为人脸检测的关注点,但目前为止较少算法具备对人脸关键点的检测能力。

Y O L O v3作为精度和速度均表现优异的算 法,同样不具备关键点检测的能力。

因此,文章提出基于Y O L O v3的人脸关键点检测算法,该算法对Y O L O v3改进,设 计关键点损失函数,实现对人脸关键点的定位,最终实现Y O L O v3在人脸检测中同时输出人脸包围框和人脸关键点。

实验表明,提出的方法在Y O L O v3上成功实现了对人脸矩形包围框和人脸关键点的同时输出。

关键词:人脸检测;深度学习;Y O L O v3;关键点检测;损失函数中图分类号:TP391 文献标识码:B文章编号=2096-9759(2021)01-0055-04F a ceLandmarkDetectionBasedOn Y O LOv3Qu Jinshan,Zu Zequn,Wan Qiubo(School of Computer and Information science,China Three Gorges University,Yichang Hubei 443002) Abstract: The powerful feature extraction ability of neural network in deep learning makes face detection in uncon-strained scenes no longer diffic ult,so the detection of face key points has gradually become the focus of face detection,but so fa r,few algorithms have the ability to detect face key points.As an algorithm with excellent accuracy and speed,yolov3 also does not have the ability of key point detection.Therefore,this paper proposes a face key point detection al-gorithm based on yolov3.The algorithm improves yolov3,designs the key point loss function,realizes the location of the face key points,and finally re­alizes the simultaneous output of face bounding box and face key points in yolov3 face de-tection.Experimental results show that the proposed method can successfully output the rectangular bounding box and key points of human face on yolov3.Key words: face detection;deep learning;Y O L O v3; face landmark;loss function〇引言人脸检测是机器视觉领域被深入研宄的经典问题,早期 人脸检测作为人脸识别的一部分,待检测的人脸通常为受到 约束的特征明显的人脸,具有清晰的五官特征以及较小的尺 度变化。

Application of Face Recognition Technology in Intelligent Vehicle Security System

Application of Face Recognition Technology in Intelligent Vehicle Security System

Application of Face Recognition Technology in IntelligentVehicle Security SystemAbstract. One being developed intelligent vehicle security system,need to estimate if anyone on a certain range ahead is authorized users then intelligently open the car door or not,in order to ensure work convenience and anti-theft security. This paper proposed a method using face recognition technology to predict the data of image sensor. The experimental results show that,the proposed algorithm is practical and reliable,and good outcome have been achieved in the application of instruction.Keywords:intelligent vehicle security system;face recognition;diagnostic faceIntroductionThe traditional mechanical car key is not only discommodious when a person is holding a bundle of goods,but also poor in anti-theft performance. As for the new-style keyless go system,its signal may be intercepted by wireless decoder of criminals.So,we develop an intelligent vehicle security systembased on face recognition technology successfully. If anyone on a certain range ahead is authorized users,it automatically opens the car door. Or else,it keeps locking. This intelligent vehicle security system enjoys a reputation of high security because it is photo-communication which is difficult to grab.We present an approach to the detection and identification of human faces and describe a working,near-real-time face recognition system which tracks a subject’s head and then recognizes the person by comparing characteristics of the face to those of known individuals. Our approach treats face recognition as a two-dimensional recognition problem. Taking advantage of the fact that faces are normally upright and thus may be described by a small set of 2-D characteristic views. Face images are projected onto a feature space (“face space”)that best encodes the variation among known face images. The face space is defined by the “diagnostic―faces”,which are the diagnostic vectors of the set of races. They do not necessarily correspond to isolated features such as eyes,ears and noses. Automatically learning and later recognizing new faces is practical within this framework.Recognition under reasonably varying conditions is achieved by training on a limited number of characteristic views(e g. a “straight on”view, a 40°view,and a profile view).Diagnostic faces for recognitionMuch of the previous work on automated face recognition has ignored the issue of just what aspects of the face stimulus are important for identification assuming that predefined measurements were relevant and sufficient. This suggested to us that all information theory approach of coding and decoding lace image may give insight into the information content of face images,emphasizing the significant local and global “features”. Such features may or may not be directly related to our intuitive notion of face features such as the eyes,noses,lips and hair. We want to extract the relevant information in a face image encode it as efficiently as possible and compare one face encoding with a database of models encoded similarly. A simply approach to extract the information contained in an image of a face is to somehow capture the variation in a collection of images,independent of any judge of features and use this information to encode and compare individual face images.In other words,we wish to find diagnostic vectors of the covariance matrix of the set of face images. These diagnosticvectors can be thought of a set of features which together characterize the variation between face images. Each image location contributes more or less to each diagnostic vectors. So we can display the diagnostic vector as a sort of ghostly face which we call a diagnostic face.Each face image in the training set can be represented exactly in terms of a linear combination of the diagnostic faces. The number of possible diagnostic faces is equal to the number of face images in the training set. However the faces can also be approximated used only the “best”diagnostic faces―those that have the largest diagnostic values,and which therefore account for the most variance within the set of face images. The primary reason for using fewer diagnostic faces is computational efficiency. The best M’ diagnostic faces span an M’ dimensional subspace―“face space”―of all possible images. As s inusoids of varying frequency and phase are the basis functions of a fourier decomposition (and are in fact diagnostic functions of linear systems). The diagnostic faces are the basis vectors of the diagnostic face decomposition.If a multitude of face images can be reconstructed by weighted sums of a small collection of characteristic images,then an efficient way to learn and recognize faces might be tobuild the characteristic features from known face images and to recognize particular faces by comparing the feature weights needed to (approximately)reconstruct them with the weights associated with the known individuals.The following steps summarize the recognition process:l. Initialization:Acquire the training set of face images and calculate the diagnostic faces which define the face space.2 When a new face image is encountered,calculate a set of weights based on the input image and the M diagnostic faces by projecting the input image onto each of the diagnostic faces.3. Determine whether the image is a face at all (whether known or unknown)by checking to see if the image is sufficiently close to “face space”.4. If it is a face,classify the weight pattern as either a known person or as unknown.Calculating diagnostic facesLet a face image I(x,y)be a two-dimensional N by N array of intensity values,or a vector of dimension N2.A typical image of size 256 by 256 describes a vector of dimension 65,536,or,equivalently,a point in 65,536-dimensional space. An ensemble of images,then,maps to a collection of points in this huge space.Images of faces,being similar in overall configuration will not be randomly distributed in this huge image space and thus can be described by a relatively low dimensional subspace. The main idea of us is to find the vectors which best account for the distribution of face images within the entire image space. These vectors define the subspace of face images,which we call “face space”. Each vector is of length N2,describes a N by N image and is a linear combination of the original face images. Because these vectors are the diagnostic vector of the covariance matrix corresponding to the original face images. and because they are face like in appearance ,we refer to them as diagnostic faces.With this analysis the calculation are greatly reduced from the order of the number of pixels in the images N2 to the order of the number of images in the training set M.In practice,the training set of images will be relatively small(),and the calculations become quite manageable. The associated diagnostic values allow us to rank the diagnostic vectors according to their usefulness in characterizing the variation among the images Normally the background is removed by cropping training images so that the diagnostic face have zero value outside of the face area.Using diagnostic faces to classify a face imageOnce the diagnostic faces are created,identification becomes a pattern recognition task. The diagnostic faces span an M'-dimensional subspace of the original N2 image space .TheM' significant diagnostic vectors of the L matrix are chosen as those with the largest associated diagnostic values. The number of diagnostic faces to be used is chosen heuristically based on the diagnostic values.A new face image L is transformed into its diagnostic face component (projected into “face space”)by a simple operation,ωk=ukT(L-φ)for .This describes a set of point-by-point image multiplications and summations.The weights from a vector ?T=[ω1,ω2,ω3…ωM']that describes the contribution of each diagnostic face in representing the input face image,treating the diagnostic faces as a basis set for face images. The vector is used to find which of a number of pre-defined face classes,if any,best describes the face. The simplest method for determining which face class provides the best description of an input face image is to find the face class k that minimizes the Euclidian distance εk=||?-?k||,where ?k is a vector describing the kth face class. A face is classified as belonging to class k when the minimum εkbecome some chosen threshold θε.Otherwise the face is classified as “unknown”. Using diagnostic faces to detect facesWe can also use knowledge of the face space to detect and locate faces in single images. This allows us to recognize the presence of faces apart from the task of identifying them.Creating the vector of weights for an image is equivalent to projecting the image onto the low dimensional face space. The distance εbetween the image and its projection onto the face space is simply the distance between mean-adjusted input image φ=L-φand ,its projection onto the face space.In this paper,we propose a face recognition method which is applied in intelligent vehicle security system. Performance in the appliance shows that the method is of high accuracy and trustworthy. The deficiency is that it can only enable face recognition of single person,which points out the direction for the improvement of our future work.References[1] Tong Lin. Research on Face Recognition and Tracking Algorithm based on Video [J]. Computer and Modernization. 2013;02(15);84-92.[2] Yuan Li,Chen Qinghu. Multi-modal Face Recognitionbased on A Few Feature Points [J]. Computer Engineering and Applications,2013;02(01);17-25.[3] Zhu Wenzhong. Comparative Analysis of Face Recognition Algorithm based on Data Set [J]. Science Technology and Engineering. 20123;01(08);22-27.[4] Zhang Yousai,Yang Shu. Multi-pose face recognition algorithm based on local weighted average virtual samples [J]. Journal of Jiangsu University of Science and Technology,2013;02(15);65-72.[5] Gao Xiaojing,Pan Xin. Face Recognition based on GLOH Operator and Local Feature [J]. Computer Applications and Software. 2013;05 (15):37-41.[6] Han Juan. The Research of Face Recognition Algorithm based on Graph Matching [J]. Computer Knowledge and Technology.2013;01(05):75-82.[7] Cui Qi,Du Haishun. Research on Face Recognition Method based on Local Matching [J]. Information and Computer. 2013;03(15):80-83.[8] Tong Xiaonian,Wen Weiyu. Face Recognition Method Using MapReduce Model for Training Support Vector Machines [J]. Journal of South-Central University For Nationalities. 2013;03(13):17-21.。

生成对抗网络人脸生成及其检测技术研究

生成对抗网络人脸生成及其检测技术研究

1 研究背景近年来,AIGC(AI Generated Content)技术已经成为人工智能技术新的增长点。

2023年,AIGC开启了人机共生的时代,特别是ChatGPT的成功,使得社会对AIGC的应用前景充满了期待。

但AIGC在使用中也存在着诸多的风险隐患,主要表现在:利用AI生成不存在的虚假内容或篡改现有的真实内容已经达到了以假乱真的效果,这降低了人们对于虚假信息的判断力。

例如,2020年,MIT利用深度伪造技术制作并发布了一段美国总统宣布登月计划失败的视频,视频中语音和面部表情都高度还原了尼克松的真实特征,成功地实现了对非专业人士的欺骗。

人有判断力,但AI没有,AI 生成的内容完全取决于使用者对它的引导。

如果使用者在使用这项技术的过程中做了恶意诱导,那么AI所生成的如暴力、极端仇恨等这样有风险的内容会给我们带来很大的隐患。

因此,对相关生成技术及其检测技术的研究成为信息安全领域新的研究内容。

本文以A I G C在图片生成方面的生成技术为目标,分析现有的以生成对抗网络(G e n e r a t i v e Adversarial Network,GAN)为技术基础的人脸生成技术。

在理解GAN的基本原理的同时,致力于对现有的人像生成技术体系和主要技术方法进行阐述。

对于当前人脸伪造检测的主流技术进行综述,并根据实验的结果分析检测技术存在的问题和研究的方向。

2 GAN的基本原理GAN由Goodfellow等人[1]于2014年首次提出。

生成对抗网络人脸生成及其检测技术研究吴春生,佟 晖,范晓明(北京警察学院,北京 102202)摘要:随着AIGC的突破性进展,内容生成技术成为社会关注的热点。

文章重点分析基于GAN的人脸生成技术及其检测方法。

首先介绍GAN的原理和基本架构,然后阐述GAN在人脸生成方面的技术模式。

重点对基于GAN在人脸语义生成方面的技术框架进行了综述,包括人脸语义生成发展、人脸语义生成的GAN实现。

基于Python的人脸识别技术研究

基于Python的人脸识别技术研究

信I ■与足1B China Computer & Communication 人工智饨与识别就术2021年第2期基于Python 的人脸识别技术研究魏天琪(南京工程学院国际教育学院,江苏南京211167)摘 要:由于基因的多样性,人脸在现今社会已成为判别身份的重要标准之一,对社会安全等领域具有重要的意义. 本文对基于Python 的人脸识别技术进行研究,使用Python 中内置的OpenCV 库函数、Keras 的Tensorflow 版建立卷积 神经模型并训练得到人脸模型,多次使用模型进行识别,证明模型一般有效。

关键词:人脸识别;Python; OpenCV; Tensorflow中图分类号:TP391.41 文献标识码:A 文章编号:1003-9767 (2021) 02-162-03Research on Face Recognition Technology Based on PythonWEI Tianqi(School of International Education, Nanjing Institute of Technology, Nanjing Jiangsu 211167, China)Abstract : Due to the diversity of genes, in today *s society, human faces have become one of the important criteria used to identify identity, which is of great significance to social security and other fields. This article researches and analyzes the face recognition technology of Python. Using the built-in OpenCV library functions in Python and the Tensorflow version of Keras, a convolutional neural model is established and trained to obtain a face model. The model is used for recognition many times, and the accuracy can be Improve it to prove that the model is generally effective.Keywords : face recognition; Python; OpenCV; Tensorflow1人脸识别的原理对Python 而言,人脸识别有多种方式,本实验主要采用 两种方式完成人脸识别:一是利用Python 自带的级联分类器 进行人脸识别,可以识别人脸特征,但不能识别身份;二是利用卷积神经网络建立人脸模型进行识别,既可以识别人脸 特征又可以识别身份。

彩色序列图像中的人脸快速定位研究

彩色序列图像中的人脸快速定位研究

2. 2 肤色分割 待检测的彩色图像经运动区域检测后 , 大大缩 小了可疑人脸区域的搜索范围 。接下来对图像进行 肤色分割 ,其目的是希望在图像中进行是否是肤色 的判断 ,并把肤色与非肤色用二值图像表现出来 。 许多 研 究 者对 人 脸 肤色 分 布 的 深入 研 究 表 明 :人脸在外观的差异实际是由亮度引起的 , 在某 些色度空间中 , 各种人脸肤色分布是一致的 , 而且集 中在一个很小的区域 。因此 , 采用一定的色度空间 可以有 效地 将 人 脸 与周 围 环 境 区分 开 。常 用 的 RG B 表示方法不适合于皮肤模型 , 在 RG B 空间 , 三 基色 R G B 不仅代表颜色 , 还表示了亮度 。为利用肤 色在色度空间的聚类性 , 需要把颜色表达式中的色 度信息与亮度信息分开 ,将 RG B 空间转换为色度与 亮度分开的空间可以达到这个目的 。 本文选择了 YCbCr 色度空间 。Y CbCr 与常见的 RG B 三原色的色度空间有如下的转换关系 :
Abstra ct : Face location is an important research in pattern recognition and image processing. This paper pro2 posed a method of fast face location i n color sequent images. Fi rstly , frame difference is used to ascertain t he motion area preliminaryly in the color image. Then , the ski n segmentation algorithm is used to remove the non skin color parts in the motion area . Finally ,the candidate face regions are examined based on the knowledge of human face and located. The test results indicate that the system not only coul d be real - time ,but also has a good robustness. Key wor ds : face location; skin segmentation ; motion detection
  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

Research of Face Location System Based on Human Vision SimulationsMo Huayi, Li Weijun, Lai Jiangliang, Dai LeiLab of Artificial Neural Networks, Institute of Semiconductors, CAS, Beijing 100083, Chinamohuayi@AbstractIn this paper we present a robust face location system based on human vision simulations to automatically locate faces in color static images. Our method is divided into four stages. In the first stage we use a gauss low-pass filter to remove the fine information of images, which is useless in the initial stage of human vision. During the second and the third stages, our technique approximately detects the image regions, which may contain faces. During the fourth stage, the existence of faces in the selected regions is verified. Having combined the advantages of Bottom-Up Feature Based Methods and Appearance-Based Methods, our algorithm performs well in various images, including those with highly complex backgrounds.Key words: face location, human vision, low-pass filter, image segmentation, region selecting and merging1. IntroductionAutomatic face location (also named face detection) is a very important task, which constitutes the "first step” of a large area of applications: face recognition, face retrieval by similarity, face tracking, surveillance, etc. Many face-location approaches have been proposed in the past 30 years, and according to the face information used, they can be roughly divided into tow categories [1, 2]: 1) Bottom-Up Feature Based Methods which include facial feature (such as eyes, nose, mouth, face contour) methods[3-9], skin color methods[10], active contour model methods[11], and template matching methods [3, 12]; 2) Appearance-Based Methods which include linear subspace methods [13], neural network methods[14], SVM methods[15], SNoW methods[16], and adaBoost methods[17]. Of the two categories, the bottom-up feature based methods have the advantages of vividness, easily being implemented and designed in various circumstances, but also have the disadvantages of being easily affected by noise and complex backgrounds of the images, meanwhile, Appearance-Based Methods generally have the advantages of strong robustness, great accuracy, but also have the disadvantages of high complexity and great difficulty in generalization and collecting training simples.Considering the advantages and disadvantages of these two kinds of methods, we design a novel face location system based on human vision simulations, which not only have the advantages of vividness, easily being implemented and designed in various circumstances of Bottom-up F eature Based Methods, but also have the advantages of strong robustness and great accuracy of Appearance-Based Methods mentioned.2. Human vision process of locating facesNowadays the human vision is still the best approach to locate faces in images, whose process consists of the following steps:Firstly, the human vision system perceives a scene image by means of the peripheral vision to get a low-resolution image[18, 19]. Secondly, before the eyes fixating on a certain region of the image, the human vision system segments the low-resolution image into several discrete regions[20]. Thirdly, the eyes focus on a certain region of interest and get the high-resolution image of the region (each represents an object) with the fovea vision, and in most cases, a face region can be focused on easily[21, 22]. Finally, the vision system identifies the high-resolution image, usually in accordance with the internal features of the region[20]. 3. Experiment steps based on human vision simulationsOur face location system is to simulate the human vision program and divided into four stages as follows.2008 International Conference on Intelligent Computation Technology and Automation3.1. Low-pass filteringThe input image is a colorful one with one or more faces in it. In our experiment, we use gauss low pass filter to process the initial image. The transfer function of Gauss low pass filter is:22(,)/2(,)Du v H u v e VHere ı is the standard deviation, let ı=D 0(Ceiling cutoff frequency), we can get the following expression:220(,)/2(,)D u v D H u v eWhile0(,)D u v D, the filter value drops from its maximum 1 to 0.607. In our experiment, we set01min(,)6D height widthFor in a human vision system, the sharpness of the fovea vision is about 6 multiples of that of the peripheral vision. Here height and width represents those of the image.After filtering the original image, we get a blurred image (fig. 2(b), and fig. 2(a) is the original image) as follows.3.2 Image segmentationHuman vision can segment the perceived image into several discrete regions. The way that the human vision perceives a scene image is similar to the way that artificial satellites perceive a terrain region imageon the earth from the sky. In order to get the segmented regions, we adopt a watershed algorithm based on immersion simulations[23] to segment the blurred image.(a) (b)F ig. 2. The original image (a) and the filtered image (b).To enhance the effect of the watershed transition, we use the gradient amplitude image of the filtered image. Here the sobel operator is employed to detect edges:121000121 §·¨¸¨¸¨¸©¹101202101 §·¨¸ ¨¸¨¸ ©¹G x G yF g Because the filtered image is a colorful one and has three color components (each is a 2-D image), we can get the gradient amplitude images of each component and combine them to form the final gradient amplitude image.g Here g r , g b, g g are the gradient amplitudes of the R,G,B components.To avoid the excess segment regions during the following watershed segmentation, we can make some changes in the gradient amplitude image:or a M×N gradient amplitude image g, let g(x,y)=0, while 0111(,),(,)M Ni j g x y c g i j MNd ¦¦ or1x c or 1y c or 1y N c ! .In our experiment, 00.8c ,1max(0.02min(,),2)c M N .After the watershed segmentation, the filtered image is segmented into several discrete regions (fig.3 (a)).In many cases (Fig. 3(a) for example), a face region will be segmented into several regions, and we need tomerge them into one. The desirable target is to merge the adjacent regions with high similarity. Here we employ the gray level average difference and texture difference to measure the similarity.(a) (b)F ig. 3. The segmenting result (a) and the merging result (b).As to the two adjacent regions R i and R j , each has three components of R, G, B, the gray level average difference of Rand R is defined as:M iR , M iG , M iB represent the gray level averages of the three components of R i , and M jR ,M jG ,M jB represent the gray level averages of the three components of R j . The texture difference is defined as:2(,)d i j iR iG iB V V V ǃǃare the standard deviations of the R,G, B components of R i , etc.The similarity between R and R is defined as:(,)d i j If d (i,j)<40, R i and R j are merged into one, and fig.3 (b) is the merging result.3.3. Face candidate region selectionAfter the image segmentation stage, the filtered image is segmented into several regions, and a face is contained by one of them. Since our target is to find out the regions containing faces, we can use the transcendental face information such as the complexion, size and shape of the face.Here we create a complexion model to detect the face region candidates. The color space we have chose n in our experiments is YCbCr space, since in YCbCr space the clustering result of the complexion points is good.YCbCr format can be transformed from RGB format in this way:0.299000.587000.1140000.168740.331260.500001280.500000.418690.08131128Y R Cb G Cr B ªºªºªºªº«»«»«»«» «»«»«»«»«»«»«»«» ¬¼¬¼¬¼¬¼The distribution of our complexion simple pixels (34800 pixels from 24 images) in YCbCr space and CbCr space is:(a) (b)F ig. 4. The distribution of our complexion simple pixels in Ycbcr space (a) and CbCr space (b).We use a 2-D gauss distribution to describe fig.4 (b)10000000110011(,)exp(()cov ())2(,),(,),11,,348001cov ()(),(,)T NN i i i i NTi i i i i i Cb Cr Cb Cr Cb Cr Cb Cb Cr Cr N N N Cb Cr N P P P P P P P P P P P ­½U °°°°°°°°®¾ °°°°°° °°¯¿¦¦ F rom the sample points we get ȝ0 =(111.7751.,148.2287)ˈcov=¸¸¹·¨¨©§137.30449.50659.506587.6061Select a proper threshold Į (in our experiment Į=0.2), while ȡ(Cb,Cr)>Į, we believe that the pixel is a complexion one. A face region candidate must have more than 50% complexion pixels.In the other hand, a face region candidate also must have enough pixels, for it is hard to verify a candidate with few pixels. In our experiment, the minimum of the pixel number of a face region candidate is 1/600 of that of the input image.At last, we attain the height and width of the minimum rectangle containing the a segmented region, if the region is a face region, the ratio of height to width will distribute in [a,b], in our experiment, a=0.8, b=1.6.The selecting result of fig.3 (b) is showed in fig5. (b), and fig. 5 (a) is the selecting result showed in origin input image.3.4 Verifying the face region candidates in the original image.1(,)d i jWhile human vision system perceives a certain region of the images with the fovea vision, the fine, high-resolution information of the region is attained and identified. We adopt an eyes-location method mentioned in [24](our previous work) and a template-matching method mentioned in[10] to verify the corresponding face region candidates (while verifying a region candidate, ignore other regions) in the original color image respectively , the final detection result is show in fig.5 (c).(a) (b) (c)Fig. 5. The selected regions showed in original image (a) and filtered image (b), the verifying result (c).4. Experiment ResultsWe have evaluated our algorithm on several face databases, including ceruleansky 1.0 face database of our lab, family collections, photos downloaded from websites with the hardware of a personal computer with a P4 1.8G CPU, 512M memories and software of windows XP OS and matlab 7.0.Each image is colorful and contains one or more complete, upright, moderate size faces in the normal light condition. In the image set with simple backgrounds (93 images with 108 faces, for example fig. 4(c), fig.5 (a)), we got True Positive Rate (TPR) of 98% and none False Detections (FD). In the image set with moderately complex backgrounds (84 images with 101 faces, fig. 5(b) for example), we got 94% TPR and 4 FD. In the image set with highly complex backgrounds (77 images with 95 faces, fig. 5(c), fig. 5(d) for example), we got TPR of 90% and 12 FD.5. DiscussionThis work proposes a four-stage approach, which is a “coarse to fine” course similar to that of human vision, to locate faces in color static images. Our approach combines the advantages of Bottom-Up F eature Based Methods and Appearance-Based Methods, and works well in various images even those with highly complex backgrounds.Therefore, we believe that the human vision, which works very well on image processing, will provide useful cues to the computer vision, and computer vision researchers should pay more attention to the researches of the rules of the human vision.(a) (b)(c) (d)Fig. 6. The experiment results. (a) simple background image set; (b) moderately complex background image set; (c), (d) highly complex background image set.6. AcknowledgementThis paper is supported by 863 project, 2006AA01Z12.7. Reference[1] Huang fuzhen and S. jianbo, Face Detection (book inChinese). Shanghai: Shanghai Jiao Tong Universitys press, 2006.[2] Y. Ming-Hsuan, D. J. Kriegman, and N. Ahuja,"Detecting faces in images: a survey," Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 24, pp. 34-58, 2002.[3] A. Lanitis, C. J. Taylor, and T. F. Cootes, "Automaticface identification system using flexible appearance models," Image and Vision Computing, vol. 13, pp. 393-401, 1995.[4] D. W. Hansen and R. I. Hammoud, "An improvedlikelihood model for eye tracking," Computer Visionand Image Understanding, vol. 106, pp. 220-230, 2007.[5] I. Craw, H. Elis, and J. R. Lishman, "Automaticextraction of face-features," Pattern Recogn. Lett., vol.5, pp. 183-187, 1987.[6] A. L. Yuille, P. W. Hallinan, and D. S. Cohen, "Featureextraction from faces using deformable templates,"International Journal of Computer Vision, vol. 8, pp.99-111, 1992.[7] D. Maio and D. Maltoni, "Real-time face location ongray-scale static images," Pattern Recognition, vol. 33,pp. 1525-1539, 2000.[8] L. C. De Silva, K. Aizawa, and M. Hatori, "Detectionand tracking of facial features," in VisualCommunications and Image P rocessing '95, Taipei,Taiwan, 1995, pp. 1161-1172.[9] V. Govindaraju, S. N. Srihari, and D. B. Sher, "Acomputational model for face location," in ComputerVision, 1990. P roceedings, Third International Conference on, 1990, pp. 718-721.[10] R. L. Hsu, M. Abdel-Mottaleb, and A. K. Jain, "F acedetection in color images," P attern Analysis andMachine Intelligence, IEEE Transactions on vol. 24,pp. 696-706, 2002.[11] S. R. Gunn and M. S. Nixon, "Global and local activecontours for head boundary extraction," InternationalJournal of Computer Vision, vol. 30, pp. 43-54, 1998. [12] H.-S. Kim, W.-S. Kang, J.-I. Shin, and S.-H. Park,"F ace detection using template matching and ellipsefitting," IEICE Transactions on Information andSystems, vol. 383-D, pp. 2008-2011, 2000.[13] M. S. Bartlett, M. S. Bartlett, J. R. Movellan, and T. J.Sejnowski, "F ace recognition by independent component analysis," Neural Networks, IEEE Transactions on, vol. 13, pp. 1450-1464, 2002.[14] S. Kah-Kay and P. Tomaso, "Example-Based Learningfor View-Based Human Face Detection," IEEE Trans.Pattern Anal. Mach. Intell., vol. 20, pp. 39-51, 1998. [15] E. Osuna, R. Freund, and F. Girosit, "Training supportvector machines: an application to face detection," inComputer Vision and P attern Recognition P roceedings., 1997 IEEE Computer SocietyConference on, 1997, pp. 130-136.[16] M. Nilsson, J. Nordberg, and I. Claesson, "F aceDetection using Local SMQT F eatures and Split upSnow Classifier," in Acoustics, Speech and SignalProcessing, 2007. IEEE International Conference on,2007, pp. II-589-II-592.[17] Z. ZhenQiu, Z. Long, S. Z. Li, and A. H. Z. HongJiangZhang, "Real-time multi-view face detection," inAutomatic Face and Gesture Recognition, 2002.Proceedings. Fifth IEEE International Conference on,2002, pp. 142-147.[18] K. Rayner, "Eye movements in reading andinformation processing: 20 years of research,"Psychological Bulletin, vol. 124, pp. 372-422, 1998. [19] J. M. Henderson and A. Hollingworth, "High-levelscene perception," Annual Review of P sychology, vol.50, pp. 243-271, 1999.[20] Jing qicheng, Jiao shulan, and J. guiping, HumanVision (book in Chinese). Beijing: Science Press, 1987. [21] S. K. Mannan, K. H. Ruddock, and D. S. Wooding,"The relationship between the locations of spatialfeatures and those of fixations made during visualexamination of briefly presented images " SpatialVision, vol. 10, pp. 165-188, 1996.[22] P. G. Schyns and A. Oliva, "F ROM BLOBS TOBOUNDARY EDGES:. Evidence for Time- andSpatial-Scale-Dependent Scene Recognition,"Psychological Science, vol. 5, pp. 195-200, 1994.[23] L. Vincent and P. Soille, "Watersheds in digital spaces:An efficient algorithm based on immersionsimulations," IEEE Transactions on P attern Analysisand Machine Intelligence, vol. 13, pp. 583-598, 1991. [24] Z.-H. Wang, H.-Y. Mo, H.-X. Lu, and A. W. S.-J.Wang Shou-Jue, "A method of biomimetic patternrecognition for face recognition," in Neural Networks,2003. P roceedings of the International JointConference on, 2003, pp. 2216-2221 vol.3.。

相关文档
最新文档