数据结构外文翻译外文文献英文文献
MVC框架中英文对照外文翻译文献

MVC框架中英⽂对照外⽂翻译⽂献中英⽂对照外⽂翻译⽂献(⽂档含英⽂原⽂和中⽂翻译)译⽂:Web 2.0下的Spring MVC框架摘要 - 当要建⽴丰富⽤户体验的WEB应⽤时,有⼤量的WED应⽤框架可以使⽤,却很少有该选择哪⼀种的指导。
WEB 2.0应⽤允许个体管理他们⾃⼰的在线⽹页,并能与其他在线⽤户和服务器共享。
这样分享需要访问控制器来实现。
然⽽,现有的访问控制器解决⽅案不是令⼈很满意。
因为在开放且由⽤户主导的WEB环境下,它满⾜不了⽤户的功能需求。
MVC框架是在所有的WEB开发框架中最受欢迎的。
模型-视图-控制器(MVC)是⼀种软件架构,如今被认为是⼀种体系结构在软件⼯程模式中使⽤。
该模式从⽤户界⾯(输⼊和演⽰)分离出了“领域逻辑”(基于⽤户的应⽤逻辑),它允许独⽴地开发,测试和维护每个分离的部分。
模型-视图-控制器(MVC)模型创建的应⽤分离为不同的层次应⽤,同时在每两者之间建⽴松散的耦合。
关键字 - Spring MVC, 结构, XStudio, SOA, 控制器I.绪论如何确切地定义⼀个⽹站为“WEB 2.0”的呢?关于这有着许多不同见解,使它很难精确地下⼀个确切的定论。
但当我们将所有的WEB开发框架过⼀遍之后它就会变得清晰了。
各种基于WEB开发的架构如下:●Ntier架构(Ntier Architecture)在软件⼯程中,多层架构(常被称为n-tier架构)是⼀种表⽰层,应⽤处理层和数据管理层在逻辑上分开处理的客户端-服务器架构。
例如,⼀个应⽤在⽤户与数据库之间使⽤中间件提供数据请求服务就⽤到了多层体系结构。
最为⼴泛应⽤的多层体系结构是三层架构。
N-tier 应⽤架构为开发者提供了⽤来创建了⼀个灵活且可复⽤的模型。
通过打破应⽤层次,开发者只需修改或添加⼀个特定的层,⽽不是要去重写⼀遍整个应⽤。
它需要有⼀个表⽰层,⼀个业务层或者数据访问层和⼀个数据层。
层(layer)和层(tier)之间的概念常常是可以互换的。
数据结构 外文翻译 外文文献 英文文献

外文翻译原文Computer programming data structure is an important theoretical basis for the design, it is not only the core curriculum of computer disciplines, and has become a popular elective course other Polytechnic professional, so studied this course well and studied computer are closely related.一、the concept of data structureComputer data structure is the foundation of science and technology professional classes, is the essential core curriculum. All computer system software and application software to use various types of data structures. Therefore, if we want to make better use of computers to solve practical problems, only to several computer programming languages are difficult to cope with the many complex issues. To the effective use of computers, give full play to computer performance, but also must learn and master relevant knowledge of data structure. A solid foundation of "data structure"for learning other computer professional courses, such as operating systems, translation theory, database management systems, software engineering, artificial intelligence, etc. are very useful.二、why should learn from data structure?In the early development of computers, the use of computer designed primarilyto deal with terms. When we use the computer to solve a specific problem, the following general needs through several steps : the first is a specific problem of appropriate abstract mathematical models, and then design or choose a mathematical model of the algorithm,the final procedures for debugging, testing, until they have the ultimate answer.Since then the object is INTEGER, REAL, BOOLEAN, the procedures of the main designers of energy is focused on programming skills, without attention to the data structure. With the expansion of computer applications and development of software and hardware, the issue of non-terms increasing importance. According to statistics, Now dealing with the issue of non-occupancy of more than 90% of the machine time. Such issues involve more complex data structure, the relationshipsbetween data elements generally can not be described by mathematical formula. Therefore, the key to solving such problems is no longer mathematical analysis and calculations, but to devise appropriate data structure, can effectively address the problem.Description of the terms of such non-mathematical model is not a mathematical equation, but such as tables, trees, such as map data structure. Therefore, it can be said that data structure courses primarily designed to study the issue of non-value calculation procedures as a computer operations and the relationship between objects and their operating disciplines.The purpose of the study is to understand the structure of data for computer processing of the identity object to the practical problems involved in dealing with that subject at the computer out and deal with them. At the same time, through training algorithms to improve the thinking ability of students through procedures designed to promote student skills integrated applications and professional qualities.三、the concepts and terminologySystematic study of knowledge in the data structure before some of the basic concepts and terminology to give a precise meaning.Data (Data) is the information carrier, it could be computer identification, storage and processing. It is the computer processing of raw materials, a variety of data processing applications. Computer science, computer processing is the so-called data objects, which can be numerical data can be non - numerical data. Numerical data are integer, the actual number or plural, mainly for engineering computing, scientific computing and commercial processing; Non - numerical data, including characters, text, graphics, images, voice and so on.Data elements (Data Element) is the basic unit of data. In different conditions, data elements can be called elements, nodes, the peak, recording. For example, students information retrieval system table information, a record high, 8 Queen's issue of a state tree, teaching programming issues such as a peak, known as a data element. Sometimes, a data from a number of data elements (Data Item), for example, the student information management system students each data element table is a studentrecord. It includes students of the school, name, sex, nationality, date of birth, performance data items. These data items can be divided into two types : one called early such as student gender, origin, etc., these data were no longer divided in data processing, the smallest units; Another called portfolio, the performance of students who, it can be divided into mathematics, physics, chemistry and other smaller items. Normally, in addressing the question of the practical application of each student is recorded as a basic unit for a visit and treatment.Data objects (Data Object) or data element type (Data Element Class) is the nature of the data elements with the same pool. In a specific issue, the data elements have the same nature (not necessarily equal value elements), belonging to the same data objects (data element type), the data element is an example of such data elements. For example, traffic information systems in the transportation network, is a culmination of all the data elements category, peak a and B each represent an urban middle is the data elements of the two types of examples of the value of their data elements a and B respectively.Data structure (Data Structure) refers to the mutual relationship that exists between one or more data elements together. In any case, between data elements will not be isolated in between them exist in one way or another, such as the relationship between the data element structure. According to the data elements of the relationship between different characteristics, usually have the following four basic categories of the structure :1 assembly structures. In the assembly structure, the relationship between data elements is "belonging to the same pool." Assembly elements relations is a very loose structure.2 linear structures. The structure of the data elements exist between one-to-one relationship.3 tree structure. The structure of the data elements exist between hierarchical relationship.4graphics structure. The structure of the data elements of the relationship that existed between Duoduiduo, graphics structure also known as network structure.C++Builder programming experience一、Database programmingAnd the use of Delphi, Borland C++Builder BDE (Borland Database Engine) database interface, in particular its use BDE Administrator unified management database alias, the database operation has nothing to do with the location of the database documents, thus enabling database development easier operation. But in a database application procedures at the same time we have to "release" BDE, the database for some simple procedures may BDE than our own design procedures big, but as the use of BDE InstallShield, add database alias is likely allocation failure. Therefore, we can use the following methods : still in the design stage procedure using BDE alias management database for debugging, but in procedures substantially (as in the main Chuangti OnCreate event processing function) to Table components DatabaseName attributes, such as the use of similar phrases as follows :Table1->DatabaseName = ExtractFilePath (Application->ExeName); OrTable1->DatabaseName= ExtractFilePath (Application->ExeName+ "DB");Thus, no impact on the debugging phase, will be issued if the application procedures Table1 document on the use of databases or their current catalogue "DB" virus, database procedures can be normal operation. You can even be a database to catalogue the documents in the form of character string Register (installed in the installation process), then the procedure in the acquisition of substantially from the catalogue of payrolls, Fuzhi DatabaseName attribute to be. Anyway, you do not need to install relatively large BDE forced users.二、the Registry visitAs in the design process we often required 9x/NT Windows Registry information visit, such as retrieval of information procedures, preservation of information. Register write a subroutine to visit necessary. When the Register to visit, the library will be directly available without always some duplication operation. The following can be used to access cosmetic Licheng, the character string type Jianzhi, and the retrieval of failure to return default value Default.#include < Registry.hpp >int ReadIntFromReg(HKEY Root, AnsiString Key, AnsiString KeyName, int Default) {int KeyValue;TRegistry *Registry = new TRegistry();Registry->RootKey = Root;Registry->OpenKey(Key, false);try {KeyValue = Registry->ReadInteger(KeyName);}catch(...) {KeyValue = Default;}delete Registry;return KeyValue;}void SaveIntToReg(HKEY Root, AnsiString Key, AnsiString KeyName, int KeyValue) {TRegistry *Registry = new TRegistry();Registry->RootKey = Root;Registry->OpenKey(Key, true);Registry->WriteInteger(KeyName, KeyValue);delete Registry;}char *ReadStringFromReg(HKEY Root, AnsiString Key, AnsiString KeyName, char *Default) {AnsiString KeyValue;TRegistry *Registry = new TRegistry();Registry->RootKey = Root;Registry->OpenKey(Key, false);try {KeyValue = Registry->ReadString(KeyName);}catch(...) {KeyValue = (AnsiString)Default;}delete Registry;return KeyValue.c_str();}void SaveStringToReg(HKEY Root, AnsiString Key,AnsiString KeyName, char *KeyValue) {TRegistry *Registry = new TRegistry();Registry->RootKey = Root;Registry->OpenKey(Key, true);Registry->WriteString(KeyName, (AnsiString)KeyValue);delete Registry;}We may use the following access methods (to Windows wallpaper documents) : AnsiString WallPaperFileName =ReadStringFromReg(HKEY_CURRENT_USER,"\\Control Panel\\Desktop", "Wallpaper", "");三、show / hide icons task columnStandard Windows applications generally operating in the mission mandate column on the chart shows, users can directly use the mouse clicking column logo for the mission task cut over, but some applications do not use task column signs, such as the typical Office tools, There are also procedures that can be shown or hidden customization tasks column icon, such as Winamp. We can do the procedure, as long as access Windows SetWindowLong function can drive, as follows : // hidden task column chart :SetWindowLong (Application->Handle.GWL_EXSTYLE, WS_EX_TOOLWINDOW);// task column shows signs :SetWindowLong (Application->Handle.GWL_EXSTYLE, WS_EX_APPWINDOW);四、the establishment of a simple "on" windowA complete Windows applications typically contain a "on the" window to show version information. We customized a dialog box as usual "on the" window of the "on" free customized window, indicates that more information, even including super links. If only show simple version information,Windows ShellAbout function shelf items have sufficient, following this line of code can be "on" Duihuakuang and is Windows standard "on the" Duihuakuang and procedures may show signs such as the use of resources and systems.ShellAbout (Handle, ( "on" +Application->Title+ "#"). C_str ()( "\n"+Application->Title+ "V1.0\n\n" + "夏登城版权所有!"). C_str ()Application->Icon->Handle);五、the two methods to choice catalogueIn our applications, allowing users to choose the regular catalogue, such as software manufacturers, users choose catalogue. This involves catalogue option, we may use the following methods for users to choose one of the catalogue : 1, use SHBrowseForFolder and SHGetPathFromIDList function; Company affirms its function as follows :WINSHELLAPI LPITEMIDLIST WINAPISHBrowseForFolder(LPBROWSEINFO lpbi); WINSHELLAPI BOOL WINAPI SHGetPathFromIDList(LPCITEMIDLIST pidl, LPSTR pszPath); LPBROWSEINFO 和LPITEMIDLIST structure refer Win32 files. This method of selecting catalogues available Windows desktop all available inventory, including networks of other computers sharing catalogue neighbors, but not the new catalogue. Li Cheng allows users to choose the following directory, the directory of choice Licheng return at all trails character string.#include < shlobj.h >char *GetDir(char *DisplayName, HWND Owner) {char dir[MAX_PATH] = "";BROWSEINFO *bi = new BROWSEINFO;bi->hwndOwner = Owner;bi->pidlRoot = NULL;bi->pszDisplayName = NULL;bi->lpszTitle = DisplayName;bi->ulFlags = BIF_RETURNONLYFSDIRS;bi->lpfn = NULL;bi->lParam = NULL;bi->iImage = 0;ITEMIDLIST *il = SHBrowseForFolder(bi);if(il!=NULL) {SHGetPathFromIDList(il, dir);}delete bi;return dir;}We can use the following list to be chosen from :AnsiString at Dir = (AnsiString) GetDir ( "Please select catalogue :" Handle);2, the use of SelectDirectory function. C++Builder the function SelectDirectory achievable catalogue of options, which showed that similar "open" / "preserve" Duihuakuang, but its advantage is to use / non-use keyboard input catalogue members, and allow the creation of new directories. Its original definition as follows : Extern package bool __fastcall SelectDirectory (AnsiString &Directory, TSelectDirOpts Options, 103-116 HelpCtx);Licheng SelectDir allow you to choose the following directory :#include < FileCtrl.hpp >AnsiString SelectDir(AnsiString Dir) {if(SelectDirectory(Dir, TSelectDirOpts()<< sdAllowCreate << sdPerformCreate << sdPrompt,0))return Dir;elsereturn "";}for the following redeployed to the users choice catalogue :AnsiString SelectedDir = SelectDir ( "C:\\My Documents");外文翻译译文数据结构是计算机程序设计的重要理论设计基础,它不仅是计算机学科的核心课程,而且已成为其他理工专业的热门选修课,所以学好这门课程是与学好计算机专业是息息相关的。
数据结构英语作文加翻译

数据结构英语作文加翻译Title: The Importance of Data Structures in Computer Science。
Data structures play a crucial role in the field of computer science. They are fundamental concepts that enable efficient storage, retrieval, and manipulation of data in computer programs. In this essay, we will explore the significance of data structures, their types, and their applications in various domains.Firstly, let us delve into the importance of data structures. In computer science, data is the foundation of every software application. However, raw data alone is not sufficient; it needs to be organized in a structured manner to be processed efficiently. Here comes the role of data structures. They provide a way to organize and store datain such a way that it can be easily accessed and manipulated. By choosing appropriate data structures, programmers can optimize the performance of theiralgorithms, leading to faster execution times and more efficient resource utilization.There are several types of data structures, each with its unique characteristics and use cases. One of the most basic data structures is the array, which stores elements of the same type in contiguous memory locations. Arrays are widely used due to their simplicity and constant-time access to elements. Another commonly used data structure is the linked list, which consists of nodes where each node contains a data field and a reference (or pointer) to the next node in the sequence. Linked lists are efficient for insertion and deletion operations but may have slower access times compared to arrays.Apart from arrays and linked lists, there are more complex data structures such as stacks, queues, trees, and graphs. Stacks follow the Last-In-First-Out (LIFO)principle and are often used in algorithms involving function calls, expression evaluation, and backtracking. Queues, on the other hand, adhere to the First-In-First-Out (FIFO) principle and are commonly used in scenarios liketask scheduling, job processing, and breadth-first search algorithms. Trees are hierarchical data structures consisting of nodes connected by edges, with a root node at the top and leaf nodes at the bottom. They are utilized in applications like hierarchical data storage, binary search trees, and decision trees. Graphs are collections of nodes (vertices) and edges connecting these nodes, and they find applications in various fields such as social networks, routing algorithms, and network flow optimization.Now, let's discuss the applications of data structures across different domains. In software development, data structures are extensively used in designing databases, implementing algorithms, and building user interfaces. For example, databases rely on data structures like B-trees and hash tables for efficient storage and retrieval of information. In algorithm design, efficient data structures are crucial for optimizing time and space complexity. Many popular algorithms such as sorting, searching, and graph traversal algorithms heavily rely on data structures for their implementation. Moreover, in user interface development, data structures like trees and graphs are usedto represent the hierarchical structure of UI components and their relationships.In addition to software development, data structures find applications in fields like artificial intelligence, bioinformatics, and computational biology. In artificial intelligence, data structures are used to represent knowledge, make decisions, and solve complex problems. For instance, knowledge graphs are used to represent relationships between entities in a knowledge base, while decision trees are employed in decision-making processes. In bioinformatics and computational biology, data structures are used to store and analyze biological data such as DNA sequences, protein structures, and metabolic pathways. Efficient data structures and algorithms are essential for tasks like sequence alignment, genome assembly, and protein folding prediction.In conclusion, data structures are the building blocks of computer science. They enable efficient storage, retrieval, and manipulation of data in computer programs, leading to faster execution times and more efficientresource utilization. With various types of data structures available and their applications spanning across different domains, it is evident that a solid understanding of data structures is essential for every computer scientist and software developer. By mastering data structures and their applications, programmers can write more efficient and scalable software solutions, thereby advancing the field of computer science as a whole.(翻译)。
数据库中英文对照外文翻译文献

中英文对照外文翻译Database Management SystemsA database (sometimes spelled data base) is also called an electronic database , referring to any collection of data, or information, that is specially organized for rapid search and retrieval by a computer. Databases are structured to facilitate the storage, retrieval , modification, and deletion of data in conjunction with various data-processing operations .Databases can be stored on magnetic disk or tape, optical disk, or some other secondary storage device.A database consists of a file or a set of files. The information in these files may be broken down into records, each of which consists of one or more fields. Fields are the basic units of data storage , and each field typically contains information pertaining to one aspect or attribute of the entity described by the database . Using keywords and various sorting commands, users can rapidly search , rearrange, group, and select the fields in many records to retrieve or create reports on particular aggregate of data.Complex data relationships and linkages may be found in all but the simplest databases .The system software package that handles the difficult tasks associated with creating ,accessing, and maintaining database records is called a database management system(DBMS).The programs in a DBMS package establish an interface between the database itself and the users of the database.. (These users may be applications programmers, managers and others with information needs, and various OS programs.)A DBMS can organize, process, and present selected data elements form the database. This capability enables decision makers to search, probe, and query database contents in order to extract answers to nonrecurring and unplanned questions that aren’t available in regular reports. These questions might initially be vague and/or poorly defined ,but people can “browse” through the database until they have the needed information. In short, the DBMS will “manage” the stored data items and assemble the needed items from the common database in response to the queries of those who aren’t programmers.A database management system (DBMS) is composed of three major parts:(1)a storage subsystemthat stores and retrieves data in files;(2) a modeling and manipulation subsystem that provides the means with which to organize the data and to add , delete, maintain, and update the data;(3)and an interface between the DBMS and its users. Several major trends are emerging that enhance the value and usefulness of database management systems;Managers: who require more up-to-data information to make effective decisionCustomers: who demand increasingly sophisticated information services and more current information about the status of their orders, invoices, and accounts.Users: who find that they can develop custom applications with database systems in a fraction of the time it takes to use traditional programming languages.Organizations : that discover information has a strategic value; they utilize their database systems to gain an edge over their competitors.The Database ModelA data model describes a way to structure and manipulate the data in a database. The structural part of the model specifies how data should be represented(such as tree, tables, and so on ).The manipulative part of the model specifies the operation with which to add, delete, display, maintain, print, search, select, sort and update the data.Hierarchical ModelThe first database management systems used a hierarchical model-that is-they arranged records into a tree structure. Some records are root records and all others have unique parent records. The structure of the tree is designed to reflect the order in which the data will be used that is ,the record at the root of a tree will be accessed first, then records one level below the root ,and so on.The hierarchical model was developed because hierarchical relationships are commonly found in business applications. As you have known, an organization char often describes a hierarchical relationship: top management is at the highest level, middle management at lower levels, and operational employees at the lowest levels. Note that within a strict hierarchy, each level of management may have many employees or levels of employees beneath it, but each employee has only one manager. Hierarchical data are characterized by this one-to-many relationship among data.In the hierarchical approach, each relationship must be explicitly defined when the database is created. Each record in a hierarchical database can contain only one key field and only one relationship is allowed between any two fields. This can create a problem because data do not always conform to such a strict hierarchy.Relational ModelA major breakthrough in database research occurred in 1970 when E. F. Codd proposed a fundamentally different approach to database management called relational model ,which uses a table asits data structure.The relational database is the most widely used database structure. Data is organized into related tables. Each table is made up of rows called and columns called fields. Each record contains fields of data about some specific item. For example, in a table containing information on employees, a record would contain fields of data such as a person’s last name ,first name ,and street address.Structured query language(SQL)is a query language for manipulating data in a relational database .It is nonprocedural or declarative, in which the user need only specify an English-like description that specifies the operation and the described record or combination of records. A query optimizer translates the description into a procedure to perform the database manipulation.Network ModelThe network model creates relationships among data through a linked-list structure in which subordinate records can be linked to more than one parent record. This approach combines records with links, which are called pointers. The pointers are addresses that indicate the location of a record. With the network approach, a subordinate record can be linked to a key record and at the same time itself be a key record linked to other sets of subordinate records. The network mode historically has had a performance advantage over other database models. Today , such performance characteristics are only important in high-volume ,high-speed transaction processing such as automatic teller machine networks or airline reservation system.Both hierarchical and network databases are application specific. If a new application is developed ,maintaining the consistency of databases in different applications can be very difficult. For example, suppose a new pension application is developed .The data are the same, but a new database must be created.Object ModelThe newest approach to database management uses an object model , in which records are represented by entities called objects that can both store data and provide methods or procedures to perform specific tasks.The query language used for the object model is the same object-oriented programming language used to develop the database application .This can create problems because there is no simple , uniform query language such as SQL . The object model is relatively new, and only a few examples of object-oriented database exist. It has attracted attention because developers who choose an object-oriented programming language want a database based on an object-oriented model. Distributed DatabaseSimilarly , a distributed database is one in which different parts of the database reside on physically separated computers . One goal of distributed databases is the access of informationwithout regard to where the data might be stored. Keeping in mind that once the users and their data are separated , the communication and networking concepts come into play .Distributed databases require software that resides partially in the larger computer. This software bridges the gap between personal and large computers and resolves the problems of incompatible data formats. Ideally, it would make the mainframe databases appear to be large libraries of information, with most of the processing accomplished on the personal computer.A drawback to some distributed systems is that they are often based on what is called a mainframe-entire model , in which the larger host computer is seen as the master and the terminal or personal computer is seen as a slave. There are some advantages to this approach . With databases under centralized control , many of the problems of data integrity that we mentioned earlier are solved . But today’s personal computers, departmental computers, and distributed processing require computers and their applications to communicate with each other on a more equal or peer-to-peer basis. In a database, the client/server model provides the framework for distributing databases.One way to take advantage of many connected computers running database applications is to distribute the application into cooperating parts that are independent of one anther. A client is an end user or computer program that requests resources across a network. A server is a computer running software that fulfills those requests across a network . When the resources are data in a database ,the client/server model provides the framework for distributing database.A file serve is software that provides access to files across a network. A dedicated file server is a single computer dedicated to being a file server. This is useful ,for example ,if the files are large and require fast access .In such cases, a minicomputer or mainframe would be used as a file server. A distributed file server spreads the files around on individual computers instead of placing them on one dedicated computer.Advantages of the latter server include the ability to store and retrieve files on other computers and the elimination of duplicate files on each computer. A major disadvantage , however, is that individual read/write requests are being moved across the network and problems can arise when updating files. Suppose a user requests a record from a file and changes it while another user requests the same record and changes it too. The solution to this problems called record locking, which means that the first request makes others requests wait until the first request is satisfied . Other users may be able to read the record, but they will not be able to change it .A database server is software that services requests to a database across a network. For example, suppose a user types in a query for data on his or her personal computer . If the application is designed with the client/server model in mind ,the query language part on the personal computer simple sends the query across the network to the database server and requests to be notified when the data are found.Examples of distributed database systems can be found in the engineering world. Sun’s Network Filing System(NFS),for example, is used in computer-aided engineering applications to distribute data among the hard disks in a network of Sun workstation.Distributing databases is an evolutionary step because it is logical that data should exist at the location where they are being used . Departmental computers within a large corporation ,for example, should have data reside locally , yet those data should be accessible by authorized corporate management when they want to consolidate departmental data . DBMS software will protect the security and integrity of the database , and the distributed database will appear to its users as no different from the non-distributed database .In this information age, the data server has become the heart of a company. This one piece of software controls the rhythm of most organizations and is used to pump information lifeblood through the arteries of the network. Because of the critical nature of this application, the data server is also the one of the most popular targets for hackers. If a hacker owns this application, he can cause the company's "heart" to suffer a fatal arrest.Ironically, although most users are now aware of hackers, they still do not realize how susceptible their database servers are to hack attacks. Thus, this article presents a description of the primary methods of attacking database servers (also known as SQL servers) and shows you how to protect yourself from these attacks.You should note this information is not new. Many technical white papers go into great detail about how to perform SQL attacks, and numerous vulnerabilities have been posted to security lists that describe exactly how certain database applications can be exploited. This article was written for the curious non-SQL experts who do not care to know the details, and as a review to those who do use SQL regularly.What Is a SQL Server?A database application is a program that provides clients with access to data. There are many variations of this type of application, ranging from the expensive enterprise-level Microsoft SQL Server to the free and open source mySQL. Regardless of the flavor, most database server applications have several things in common.First, database applications use the same general programming language known as SQL, or Structured Query Language. This language, also known as a fourth-level language due to its simplistic syntax, is at the core of how a client communicates its requests to the server. Using SQL in its simplest form, a programmer can select, add, update, and delete information in a database. However, SQL can also be used to create and design entire databases, perform various functions on the returned information, and even execute other programs.To illustrate how SQL can be used, the following is an example of a simple standard SQL query and a more powerful SQL query:Simple: "Select * from dbFurniture.tblChair"This returns all information in the table tblChair from the database dbFurniture.Complex: "EXEC master..xp_cmdshell 'dir c:\'"This short SQL command returns to the client the list of files and folders under the c:\ directory of the SQL server. Note that this example uses an extended stored procedure that is exclusive to MS SQL Server.The second function that database server applications share is that they all require some form of authenticated connection between client and host. Although the SQL language is fairly easy to use, at least in its basic form, any client that wants to perform queries must first provide some form of credentials that will authorize the client; the client also must define the format of the request and response.This connection is defined by several attributes, depending on the relative location of the client and what operating systems are in use. We could spend a whole article discussing various technologies such as DSN connections, DSN-less connections, RDO, ADO, and more, but these subjects are outside the scope of this article. If you want to learn more about them, a little Google'ing will provide you with more than enough information. However, the following is a list of the more common items included in a connection request.Database sourceRequest typeDatabaseUser IDPasswordBefore any connection can be made, the client must define what type of database server it is connecting to. This is handled by a software component that provides the client with the instructions needed to create the request in the correct format. In addition to the type of database, the request type can be used to further define how the client's request will be handled by the server. Next comes the database name and finally the authentication information.All the connection information is important, but by far the weakest link is the authentication information—or lack thereof. In a properly managed server, each database has its own users with specifically designated permissions that control what type of activity they can perform. For example, a user account would be set up as read only for applications that need to only access information. Another account should be used for inserts or updates, and maybe even a third account would be used for deletes.This type of account control ensures that any compromised account is limited in functionality. Unfortunately, many database programs are set up with null or easy passwords, which leads to successful hack attacks.译文数据库管理系统介绍数据库(database,有时拼作data base)又称为电子数据库,是专门组织起来的一组数据或信息,其目的是为了便于计算机快速查询及检索。
数据库系统英文文献

Database Systems1. Fundamental Concepts of DatabaseDatabase and database technology are having a major impact on the growing use of computers. It is fair to say that database will play a critical role in almost all areas where computers are used, including business, engineering, medicine, law, education, and library science, to name a few. The word "database" is in such common use that we must begin by defining what a database is. Our initial definition is quit general.A database is a collection of related data. By data, we mean known facts that can be recorded and that have implicit meaning. For example, consider the names, telephone numbers, and addresses of all the people you know. Y ou may have recorded this data in an indexed address book, or you may have stored it on a diskette using a personal computer and software such as DBASE III or Lotus 1-2-3. This is a collection of related data with an implic it meaning and hence is a database.The above definition of database is quite general; for example, we may consider the collection of words that make up thispage of text to be related data and hence a database. However, the common use of the term database is usually more restricted.A database has the following implicit properties:.A database is a logically coherent collection of data with some inherent meaning. A random assortment of data cannot bereferred to as a database..A database is designed, built, and populated with data for a specific purpose. It has an intended group of users and somepreconceived applications in which these users are interested..A database represents some aspect of the real world, sometimes called the mini world. Changes to the mini world are reflected in the database.In other words, a database has some source from which data are derived, some degree of interaction with events in the real world, and an audience that is actively interested in the contents of the database.A database can be of any size and of varying complexity. For example, the list of names and addresses referred to earlier may have only a couple of hundred records in it, each with asimple structure. On the other hand, the card catalog of a large library may contain half a million cards stored under different categories-by primary author’s last name, by subject, by book title, and the like-with each category organized in alphabetic order. A database of even greater size and complexity may be that maintained by the Internal Revenue Service to keep track of the tax forms filed by taxpayers of the United States. If we assume that there are 100million taxpayers and each taxpayer files an average of five forms with approximately 200 characters of information per form, we would get a database of 100*(106)*200*5 characters(bytes) of information. Assuming the IRS keeps the past three returns for each taxpayer in addition to the current return, we would get a database of 4*(1011) bytes. This huge amount of information must somehow be organized and managed so that users can search for, retrieve, and update the data as needed.A database may be generated and maintained manually or by machine. Of course, in this we are mainly interested in computerized database. The library card catalog is an example of a database that may be manually created and maintained. A computerized database may be created and maintained either by a group of application programs written specifically for that task or by a database management system.A data base management system (DBMS) is a collection of programs that enables users to create and maintain a database. The DBMS is hence a general-purpose software system that facilitates the processes of defining, constructing, and manipulating databases for various applications. Defining a database involves specifying the types of data to be stored in the database, along with a detailed description of each type of data. Constructing the database is the process of storing the data itself on some storage medium that is controlled by the DBMS. Manipulating a database includes such functions as querying the database to retrieve specific data, updating the database to reflect changes in the mini world, and generating reports from the data.Note that it is not necessary to use general-purpose DBMS software for implementing a computerized database. We could write our own set of programs to create and maintain the database, in effect creating our own special-purpose DBMS software. In either case-whether we use a general-purpose DBMS or not-we usually have a considerable amount of software to manipulate the database in addition to the database itself. The database and software are together called a database system.2. Data ModelsOne of the fundamental characteristics of the database approach is that it provides some level of data abstraction by hiding details of data storage that are not needed by most database users. A data model is the main tool for providing this abstraction. A data is a set of concepts that can beused to describe the structure of a database. By structure of a database, we mean the data types, relationships, and constraints that should hold on the data. Most data models also include a set of operations for specifying retrievals and updates on the database.Categories of Data ModelsMany data models have been proposed. We can categorize data models based on the types of concepts they provide to describe the database structure. High-level or conceptual data models provide concepts that are close to the way many users perceive data, whereas low-level or physical data models provide concepts that describe the details of how data is stored in the computer. Concepts provided by low-level data models are generally meant for computer specialists, not for typical end users. Between these two extremes is a class of implementation data models, which provide concepts that may be understood by end users but that are not too far removed from the way data is organized within the computer. Implementation data models hide some details of data storage but can be implemented on a computer system in a direct way.High-level data models use concepts such as entities, attributes, and relationships. An entity is an object that is represented in the database. An attribute is a property that describes some aspect of an object. Relationships among objects are easily represented in high-level data models, which are sometimes called object-based models because they mainly describe objects and their interrelationships.Implementation data models are the ones used most frequently in current commerc ial DBMSs and include the three most widely used data models-relational, network, and hierarchical. They represent data using record structures and hence are sometimes called record-based data modes.Physical data models describe how data is stored in the computer by representing information such as record formats, record orderings, and access paths. An access path is a structure that makes the search for particular database records much faster.3. Classification of Database Management SystemsThe main criterion used to classify DBMSs is the data model on which the DBMS is based. The data models used most often in current commercial DBMSs are the relational, network, and hierarchical models. Some recent DBMSs are based on conceptual or object-oriented models. We will categorize DBMSs as relational, hierarchical, and others.Another criterion used to classify DBMSs is the number of users supported by the DBMS. Single-user systems support only one user at a time and are mostly used with personal computer. Multiuser systems include the majority of DBMSs and support many users concurrently.A third criterion is the number of sites over which the database is distributed. Most DBMSs are centralized, meaning that their data is stored at a single computer site. A centralized DBMS can support multiple users, but the DBMS and database themselves reside totally at a single computer site. A distributed DBMS (DDBMS) can have the actual database and DBMS software distributed over many sites connected by a computer network. Homogeneous DDBMSs use the same DBMS software at multiple sites. A recent trend is to develop software to access several autonomous preexisting database stored under heterogeneous DBMSs. This leads to a federated DBMS (or multidatabase system),, where the participating DBMSs are loosely coupled and have a degree of local autonomy.We can also classify a DBMS on the basis of the types of access paty options available for storing files. One well-known family of DBMSs is based on inverted file structures. Finally, a DBMS can be general purpose of special purpose. When performance is a prime consideration, a special-purpose DBMS can be designed and built for a specific application and cannot be used for other applications, Many airline reservations and telephone directory systems are special-purpose DBMSs.Let us briefly discuss the main criterion for classifying DBMSs: the data mode. The relational data model represents a database as a collection of tables, which look like files. Mos t relational databases have high-level query languages and support a limited form of user views.The network model represents data as record types and also represents a limited type of 1:N relationship, called a set type. The network model, also known as the CODASYL DBTG model, has an associated record-at-a-time language that must be embedded in a host programming language.The hierarchical model represents data as hierarchical tree structures. Each hierarchy represents a number of related records. There is no standard language for the hierarchical model, although most hierarchical DBMSs have record-at-a-time languages.4. Client-Server ArchitectureMany varieties of modern software use a client-server architecture, in which requests by one process (the client) are sent to another process (the server) for execution. Database systems are no exception. In the simplest client/server architecture, the entire DBMS is a server, except for the query interfaces that interact with the user and send queries or other commands across to the server. For example, relational systems generally use the SQL language for representing requests from the client to the server. The database server then sends the answer, in the form of a table or relation, back to the client. The relationship between client and server can get more work in theclient, since the server will e a bottleneck if there are many simultaneous database users.。
计算机专业毕业设计外文翻译---GIS软件和数据结构 (1)

外文文献GIS software and data structures(bilingual)1、Arc/InfoArc/Info is a long-lived full functions GIS package that has been ported to the microcomputer, the workstation, and the mainframe. Arc/Info is used to automate, manipulate, analyze, and display geographic data and incorporates hundreds of sophisticated tools for map automation, data conversion, database management, map overlay and spatial analysis, interactive display and query, graphic editing, and address geocoding. The software includes a relational database interface for integration with commercial database management systems and macro language for developing customized applications called AML (ARC Macro Language). Arc/Info is a generic nonapplication specific approach to geographic information systems, allowing the software to address virtually any geographic application. The software runs both on higher and microcomputers and is available on several Unix workstations and for Windows NT.2、ArcViewArcView is available for Windows, Macintosh, and a variety of Unix platforms. It is a desktop system for storing, modifying, querying, analyzing, and displaying information about geographic space. Support for spatial and tabular queries, “hot links” to other desktop applications and data types, business graphic functions such as charting, bar and pie charts, and map symbolization, desing, and layout capabilities are supported. Gocoding and address matching are also possible. ArcView is also a product of ESRI, who makes Arc/Info. There is compatibility between the two systems, with ArcView being more oriented toward map display than database management. When the ArcView version II software was introduced, the original ArcView software was placed into the public domain and is available over the Internet.3、Atlas GISAtlasGIS is available for both DOS and as version 3.0 for Windows. The original vendor has recently sold the software to Clartas, which in turn was purchased in 1996 by ESRI. This GIS lets you display, edit, and analyze information from a database or spreadsheet on a map, and can turn statistics and tabular data into graphics for decision-making and presentation purposes. Atlas GIS has a database management system with spreadsheet-style presentation, map editing, drawing tools, and reporting. Atlas GIS for Windows includes features that give SQL data access and street-level geocoding of addresses. Interfacing with the Oracle DBMS is also possible.The graphics allow full-featured geographical analysis and integrated database connectivity. In recent comparative reviews, three GIS and computer magazines awarded AtlasGIS for Windows the highest rating in value and ease of use. The GUI builds on Windows to allow “point and click” access, a button bar, and a page layout system that displays the map automatically. Several map “style sheets” come with the program. Legends, titles, scale bars and other elements update automatically. A CD ROM of U.S. addresses allows nationwide geocoding by address matching and supports mapping. A map layer management system let’s you click to change colors, settings, styles, visibility, and so on. Interface via Windows OLE, allowing cut and paste to other Windows applications, is possible.4、GRASSThe U.S. Army Construction Engineering Research Laboratories has developed publicdomain software called the Geographical Resources Analysis Support System (GRASS). Grass is raster based, was the first Unix GIS software, and has been considerably enhanced by the addition of user contributions, for example in hydrologic modeling. GRASS is available free over the Internet. Many users run GRASS on PC’s under the Linux version of UNIX. Since 1985, CERL has released upgrades and enhancements to GRASS and provided technical user support. CERL terminated GRASS related work in the spring of 1996. Under formal agreements, CERL now works with commercial vendors both to support GRASS and to integrate its capabilities into commercial system. Existing information on the GRASS WorldWide Web sites will be maintained for some time as background.5、IDRISIThe IDRISI software system has been developed, distributed, and supported on a notfor-profit basis by the IDRISI Project, Clark University Graduate School of Geography. To date, there are over 15,000 registered users of IDRISI software in over 130 countries, making it the most widely used raster GIS in the world. IDRISI is designed to be easy to use, yet provide professional-level GIS, image processing and spatial statistics analytical capability on both DOS- and Windows-based personal computers. It is intended to be affordable to all levels of users and to run on the most basic of common computer platforms. Expensive graphics cards or peripheral devices are not required to make use of the analytical power of the system. The system is designed with an open architecture so that researchers can integrate their own modules.IDRISI for Windows, first released in 1995, added a graphical user interface, flexible cartographic composition facilities, and integrated database management system to the analytical toolkit. Special routines for change and time-series analysis, spatial decision support, and uncertainty analysis and incorporation are included. The software comes with a set of tutorial exercises and data that guide the new user through the concepts of GIS and image processing while also introducing the features of IDRISI. The tutorial exercises are appropriate for use in self-training or in classroom settings.6、MapInfoMapInfo was one of the first GIS programs to do desktop mapping. The vendor is MapInfo Corporation of Troy, New York. The software is well distributed and has many user groups and a broad variety of applications worldwide. The software runs under DOS, Windows, Macintosh, and on various Unix platforms. While MapInfo’s GIS retrieval and analysis functions are fewer than those of full-blown GIS packages, MapInfo includes a link to the Basic programming language via a language called MapBasic. This development environment permits the creation of customized “mapplications”, extending MapInfo’s built-in functionality and allowing use of acommon graphical interface.MapInfo also supplies information products spanning geographic, economic, political, cultural, and industry application-specific content, each derived from leading worldwide sources to work the software. MapInfo also has an extensive training program, with classes at introductory and advanced levels for MapInfo and MapBasic.7、Microstation MGEMGE is widely distributed layer-based GIS with a tradition in computer assisted design by the Intergraph Corporation of Huntsville, Alabama. The software runs on workstations, PCs and under the Windows NT system. An extensive set of add on modules allow users to configure GIS capability around their specific needs. The layered implementation allows efficient storage structures for the geometry and linkages to relational database records. Geographic element is represented in the GIS as features. Features are grouped into the same categories as the maps on which they appear.For the attribute data, MGE incorporates use of the relational interface system, which facilitates client-server network communication to the relational DBMS so that multiple workstations communicate with the database server simultaneously. MGE contains tools for building and maintaining topologically clean data without the processing and storage overhead of building and maintaining topology. MGE supports the open geodata interoperability specification and the spatial data transfer standard.中文译文GIS软件和数据结构Arc/Info是一个产生早、功能齐全的地理信息系统软件包,它已经被安装到微型计算机、工作站以及电算机的主机中。
大数据外文翻译文献

大数据外文翻译文献(文档含中英文对照即英文原文和中文翻译)原文:What is Data Mining?Many people treat data mining as a synonym for another popularly used term, “Knowledge Discovery in Databases”, or KDD. Alternatively, others view data mining as simply an essential step in the process of knowledge discovery in databases. Knowledge discovery consists of an iterative sequence of the following steps:· data cleaning: to remove noise or irrelevant data,· data integration: where multiple data sources may be combined,·data selection : where data relevant to the analysis task are retrieved from the database,·data transformation : where data are transformed or consolidated into forms appropriate for mining by performing summary or aggregation operations, for instance,·data mining: an essential process where intelligent methods are applied in order to extract data patterns,·pattern evaluation: to identify the truly interesting patterns representing knowledge based on some interestingness measures, and ·knowledge presentation: where visualization and knowledge representation techniques are used to present the mined knowledge to the user .The data mining step may interact with the user or a knowledge base. The interesting patterns are presented to the user, and may be stored as new knowledge in the knowledge base. Note that according to this view, data mining is only one step in the entire process, albeit an essential one since it uncovers hidden patterns for evaluation.We agree that data mining is a knowledge discovery process. However, in industry, in media, and in the database research milieu, the term “data mining” is becoming more popular than the longer term of “knowledge discovery in databases”. Therefore, in this book, we choose to use the term “data mining”. We adop t a broad view of data mining functionality: data mining is the process of discovering interestingknowledge from large amounts of data stored either in databases, data warehouses, or other information repositories.Based on this view, the architecture of a typical data mining system may have the following major components:1. Database, data warehouse, or other information repository. This is one or a set of databases, data warehouses, spread sheets, or other kinds of information repositories. Data cleaning and data integration techniques may be performed on the data.2. Database or data warehouse server. The database or data warehouse server is responsible for fetching the relevant data, based on the user’s data mining request.3. Knowledge base. This is the domain knowledge that is used to guide the search, or evaluate the interestingness of resulting patterns. Such knowledge can include concept hierarchies, used to organize attributes or attribute values into different levels of abstraction. Knowledge such as user beliefs, which can be used to assess a pattern’s interestingness based on its unexpectedness, may also be included. Other examples of domain knowledge are additional interestingness constraints or thresholds, and metadata (e.g., describing data from multiple heterogeneous sources).4. Data mining engine. This is essential to the data mining system and ideally consists of a set of functional modules for tasks such ascharacterization, association analysis, classification, evolution and deviation analysis.5. Pattern evaluation module. This component typically employs interestingness measures and interacts with the data mining modules so as to focus the search towards interesting patterns. It may access interestingness thresholds stored in the knowledge base. Alternatively, the pattern evaluation module may be integrated with the mining module, depending on the implementation of the data mining method used. For efficient data mining, it is highly recommended to push the evaluation of pattern interestingness as deep as possible into the mining process so as to confine the search to only the interesting patterns.6. Graphical user interface. This module communicates between users and the data mining system, allowing the user to interact with the system by specifying a data mining query or task, providing information to help focus the search, and performing exploratory data mining based on the intermediate data mining results. In addition, this component allows the user to browse database and data warehouse schemas or data structures, evaluate mined patterns, and visualize the patterns in different forms.From a data warehouse perspective, data mining can be viewed as an advanced stage of on-1ine analytical processing (OLAP). However, data mining goes far beyond the narrow scope of summarization-styleanalytical processing of data warehouse systems by incorporating more advanced techniques for data understanding.While there may be many “data mining systems” on the market, not all of them can perform true data mining. A data analysis system that does not handle large amounts of data can at most be categorized as a machine learning system, a statistical data analysis tool, or an experimental system prototype. A system that can only perform data or information retrieval, including finding aggregate values, or that performs deductive query answering in large databases should be more appropriately categorized as either a database system, an information retrieval system, or a deductive database system.Data mining involves an integration of techniques from mult1ple disciplines such as database technology, statistics, machine learning, high performance computing, pattern recognition, neural networks, data visualization, information retrieval, image and signal processing, and spatial data analysis. We adopt a database perspective in our presentation of data mining in this book. That is, emphasis is placed on efficient and scalable data mining techniques for large databases. By performing data mining, interesting knowledge, regularities, or high-level information can be extracted from databases and viewed or browsed from different angles. The discovered knowledge can be applied to decision making, process control, information management, query processing, and so on. Therefore,data mining is considered as one of the most important frontiers in database systems and one of the most promising, new database applications in the information industry.A classification of data mining systemsData mining is an interdisciplinary field, the confluence of a set of disciplines, including database systems, statistics, machine learning, visualization, and information science. Moreover, depending on the data mining approach used, techniques from other disciplines may be applied, such as neural networks, fuzzy and or rough set theory, knowledge representation, inductive logic programming, or high performance computing. Depending on the kinds of data to be mined or on the given data mining application, the data mining system may also integrate techniques from spatial data analysis, Information retrieval, pattern recognition, image analysis, signal processing, computer graphics, Web technology, economics, or psychology.Because of the diversity of disciplines contributing to data mining, data mining research is expected to generate a large variety of data mining systems. Therefore, it is necessary to provide a clear classification of data mining systems. Such a classification may help potential users distinguish data mining systems and identify those that best match their needs. Data mining systems can be categorized according to various criteria, as follows.1) Classification according to the kinds of databases mined.A data mining system can be classified according to the kinds of databases mined. Database systems themselves can be classified according to different criteria (such as data models, or the types of data or applications involved), each of which may require its own data mining technique. Data mining systems can therefore be classified accordingly.For instance, if classifying according to data models, we may have a relational, transactional, object-oriented, object-relational, or data warehouse mining system. If classifying according to the special types of data handled, we may have a spatial, time -series, text, or multimedia data mining system , or a World-Wide Web mining system . Other system types include heterogeneous data mining systems, and legacy data mining systems.2) Classification according to the kinds of knowledge mined.Data mining systems can be categorized according to the kinds of knowledge they mine, i.e., based on data mining functionalities, such as characterization, discrimination, association, classification, clustering, trend and evolution analysis, deviation analysis , similarity analysis, etc.A comprehensive data mining system usually provides multiple and/or integrated data mining functionalities.Moreover, data mining systems can also be distinguished based on the granularity or levels of abstraction of the knowledge mined, includinggeneralized knowledge(at a high level of abstraction), primitive-level knowledge(at a raw data level), or knowledge at multiple levels (considering several levels of abstraction). An advanced data mining system should facilitate the discovery of knowledge at multiple levels of abstraction.3) Classification according to the kinds of techniques utilized.Data mining systems can also be categorized according to the underlying data mining techniques employed. These techniques can be described according to the degree of user interaction involved (e.g., autonomous systems, interactive exploratory systems, query-driven systems), or the methods of data analysis employed(e.g., database-oriented or data warehouse-oriented techniques, machine learning, statistics, visualization, pattern recognition, neural networks, and so on ) .A sophisticated data mining system will often adopt multiple data mining techniques or work out an effective, integrated technique which combines the merits of a few individual approaches.什么是数据挖掘?许多人把数据挖掘视为另一个常用的术语—数据库中的知识发现或KDD的同义词。
数据分析外文文献+翻译

数据分析外文文献+翻译文献1:《数据分析在企业决策中的应用》该文献探讨了数据分析在企业决策中的重要性和应用。
研究发现,通过数据分析可以获取准确的商业情报,帮助企业更好地理解市场趋势和消费者需求。
通过对大量数据的分析,企业可以发现隐藏的模式和关联,从而制定出更具竞争力的产品和服务策略。
数据分析还可以提供决策支持,帮助企业在不确定的环境下做出明智的决策。
因此,数据分析已成为现代企业成功的关键要素之一。
文献2:《机器研究在数据分析中的应用》该文献探讨了机器研究在数据分析中的应用。
研究发现,机器研究可以帮助企业更高效地分析大量的数据,并从中发现有价值的信息。
机器研究算法可以自动研究和改进,从而帮助企业发现数据中的模式和趋势。
通过机器研究的应用,企业可以更准确地预测市场需求、优化业务流程,并制定更具策略性的决策。
因此,机器研究在数据分析中的应用正逐渐受到企业的关注和采用。
文献3:《数据可视化在数据分析中的应用》该文献探讨了数据可视化在数据分析中的重要性和应用。
研究发现,通过数据可视化可以更直观地呈现复杂的数据关系和趋势。
可视化可以帮助企业更好地理解数据,发现数据中的模式和规律。
数据可视化还可以帮助企业进行数据交互和决策共享,提升决策的效率和准确性。
因此,数据可视化在数据分析中扮演着非常重要的角色。
翻译文献1标题: The Application of Data Analysis in Business Decision-making The Application of Data Analysis in Business Decision-making文献2标题: The Application of Machine Learning in Data Analysis The Application of Machine Learning in Data Analysis文献3标题: The Application of Data Visualization in Data Analysis The Application of Data Visualization in Data Analysis翻译摘要:本文献研究了数据分析在企业决策中的应用,以及机器研究和数据可视化在数据分析中的作用。
计算机 数学 外文翻译 外文文献 英文文献 模糊决策森林

模糊决策森林Cezary Z. JanikowDepartment of Math and Computer ScienceUniversity of Missouri – St. Louiscjanikow@摘要:过去,我们已经提出了模糊决策树,现在作为一种扩展就称本文中的方法为模糊决策森林。
森林背后的想法不仅是要代表多个树,而且还要代表在每棵树的各级进行的测试选择。
这样产生的树其实是一个三维树。
森林允许在决策树的一些或所有的结点进行测试的多种选择。
然而,有多个测试选择的主要优点是在测试数据的特征是不可靠或丢失的情况下,有选择测试决策。
在本文中,我们概述了模糊决策森林背后的想法,并且用特征值缺失的数据进行了大量的实验,证明了这种方法的增强能力。
一引言当今时代,面对海量的的数据,开发能够处理和挖掘数据的计算机程序显得尤为重要。
对于分类任务,决策树被证明是最成功的方法之一[1] [6] [7]。
用决策树的形式以及推理步骤的来获取知识,一直以准确性,效率和可理解性为人称道。
决策树方法原本是为符号域和一个简单的决策过程提出的[6],它有着许多方法论的进步性,如能产生二叉树处理连续数据[1],新的推理过程,例如,计算决策的概率[7],最后纳入模糊集和不确定性推理推论法说明噪音和不确定的状况[2] [8]。
决策树是由两个要素组成:一个自上而下的划分递归过程,生成决策树,然后从得到的树推出规则。
该过程开始于训练集,根据可用的变量和域通过特征的组合来表达,并划分为若干类。
划分过程一次选择一个测试,通常是一个特征,然后根据测试特征将数据分成几个子集。
选定的测试是为了最大限度地提高一些目标,如将不同类的样例分离[7]。
一旦样例被完美的分类或者达到一些其他目标,递归过程就停止[7]。
随后的推理规则使用树来分配新的测试数据,到达一些相同的类。
模糊集与逻辑被提出用来处理语言和数据有关的不确定性[9]。
同不确定性推理相结合,模糊表达提供了更大的稳定性和鲁棒性。
关系数据库的结构外文翻译外文文献英文文献

关系数据库的结构关系模型是任何关系数据库管理系统(RDBMS)的基础。
一个关系模型有三个重要组成部分:对象或关系的集合,作用于对象或关系上的操作,以及数据完整性规则。
换句话说,关系数据库有一个存储数据的地方,一种创建和检索数据的方法,以及一种确保数据的逻辑一致性的方法。
一个关系数据库使用关系或二维表来存储支持某个事物所需的信息。
让我们了解一下一个传统的关系数据库系统的基本组件并学习关系数据库的设计。
一旦你对于行、列、表和关联是什么有了深刻理解,你就能够充分发挥关系数据库的强大功能。
表,行和列在关系数据库中,一个表是一个用于保存相关信息的二维结构。
一个数据库由一个或者多个相关联的表组成。
表中的一行是一种事物的集合或实例,比如一个员工或发票上的一项。
表中的一列包含了一类信息;而且行列相交点上的数据、字段,就是能够用数据库查询语言检索到的最小片信息。
举个例子来说,一个员工信息表可能有一列,其列名为“LAST_NAME”,列中就包含所有员工的名字。
数据是通过对行、列进行过滤而从表中检索出来的。
主码、数据类型和外码本篇文章均以假设的斯科特·史密斯的工厂为例,他是数据库的建立者和企业的主办人。
他刚开办了一个饰品公司并目想要使用关系数据库的几项基本功能来管理人力资源部门。
关系:用来保存相关信息的一个二维结构,也就是表。
行:在一个数据库表中的一组单数据或多数据元素,用于描述一个人、地方或事物。
列:列是数据库表的组件,它包含所有行中同名和同类型的所有数据。
你会在下面章节学到如何设计数据库,现在让我们假设数据库大部分己经设计完成并且有一些表需要被执行。
斯科特创建了EMP表来保存基本的员工信息,就像这样:你可能注意到佣金列和管理人列中有一些单元格中没有值;它们是空值。
一个关系数据库能够规定列中的一个单元格是否为空。
如此,可以明确那些非销售部的员工佣金单元为空。
同样也明确了公司董事长的管理人单元为空,因为这个员工不需要向任何人汇报工作。
大数据文献综述英文版

欢迎阅读The development and tendency of Big DataTang Xia(Guilin University of electronic technology, electronic engineering?and?automation, Guilin)Abstract: "Big Data"?is the most popular IT word after the "Internet of things"?and "Cloud computing". From the source, development, status quo and tendency of big data, we can understand every aspect of it. Big data is one of the most important technologies around the world and every country has their own way to develop the technology.Key words : big data; IT; technology“Big Data ” so that this products other Google federal improve the ability to extract information and viewpoint of big data [1]. Thus, it can speed science and engineering discovery up, and it is a major move to push some research institutions making innovations. The federal government put big data development into a strategy place, which has a big impact on every country. At present, many big European institutions is still at the primary stage to use big data and seriously lack technology about big data. Most improvements and technology of big data are come from America. Therefore, there are kind of challenges of Europe to keep in step with the development of big data. But, in the financial service industry especially investment banking in London is one of the earliest industries in Europe. The experiment and technology of big data is as good as the giant institution of America. And, the investment of big data has been maintained promising efforts. January 2013, British government announced 1.89 million pound will be invested in big data and calculation of energy savingtechnology in earth observation and health care[3].Japanese government timely takes the challenge of big data strategy. July 2013, Japan’s communications ministry proposed a synthesize strategy called “Energy ICT of Japan” which focused on big data application. June 2013, the abe cabinet formally announced the new IT strategy----“The announcement of creating the most advanced IT country”. This announcement comprehensively expounded that Japanese new IT national strategy is with the core of developing opening public data and big data in 2013 to 2020[4].Big data has also drawn attention of China government.《Guiding?opinions of the State Council on promoting?the healthy and orderly development?of the?Internet of things》promote to quicken the core technology including sensor network、intelligent terminal、big data processing、intelligent analysis and service integration. December 2012, the national development and reform commission add data analysisto data, data.projects4.3 Development direction of big dataThe storage technology of big data is relational database at primary. But due to the canonical design, friendly query language, efficient ability dealing with online affair, Big data dominate the market a long term. However, its strict design pattern, it ensures consistency to give up function, its poor expansibility these problems are exposed in big data analysis. Then, NoSQL data storage model and Bigtable propsed by Google start to be in fashion[5].Big data analysis technology which uses MapReduce technological frame proposed by Google is used to deal with large scale concurrent batch transaction. Using file system to store unstructured data is not lost function but also win the expansilility. Later, there are big data analysis platform like HA VEn proposed by HP and Fusion Insight proposed by Huawei . Beyond doubt, this situation will be continued, new technology and measures will come out such as next generation data warehouse, Hadoop distribute and soon[6].ConclusionThis paper we analysis the development and tendency of big data. Based on this, we know that the big data is still at a primary stage, there are too many problems need to deal with. But the commercial value and market value of big data are the direction of development to information age.[1] Li Chunwei, Development report of China’s E-Commerce enterprises, Beijing , 2013, pp.268-270[2] Li Fen, Zhu Zhixiang, Liu Shenghui, The development status and the problems of large data, Journal of Xi’th ACM。
data structures and algorithm analysi英文原版 pdf (2)

data structures and algorithm analysi英文原版 pdfTitle: Data Structures and Algorithm Analysis: A Comprehensive ReviewIntroduction:Data structures and algorithm analysis are fundamental concepts in computer science. They form the backbone of efficient and optimized software development. This article aims to provide a comprehensive review of the book "Data Structures and Algorithm Analysis" in its English original version PDF format. The review will cover the key points, structure, and significance of the book.I. Overview of the Book:1.1 Importance of Data Structures:- Discuss the significance of data structures in organizing and manipulating data efficiently.- Explain how data structures enhance the performance and scalability of software applications.1.2 Algorithm Analysis:- Describe the role of algorithm analysis in evaluating the efficiency and performance of algorithms.- Highlight the importance of selecting appropriate algorithms for different problem-solving scenarios.1.3 Book Structure:- Outline the organization of the book, including chapters, sections, and topics covered.- Emphasize the logical progression of concepts, starting from basic data structures to advanced algorithm analysis.II. Data Structures:2.1 Arrays and Linked Lists:- Explain the characteristics, advantages, and disadvantages of arrays and linked lists.- Discuss the implementation details, operations, and time complexities of these data structures.2.2 Stacks and Queues:- Define stacks and queues and their applications in various scenarios.- Elaborate on the implementation, operations, and time complexities of stacks and queues.2.3 Trees and Graphs:- Introduce the concepts of trees and graphs and their real-world applications.- Discuss different types of trees (binary, AVL, B-trees) and graphs (directed, undirected, weighted).III. Algorithm Analysis:3.1 Asymptotic Notation:- Explain the significance of asymptotic notation in analyzing the efficiency of algorithms.- Discuss the Big-O, Omega, and Theta notations and their usage in algorithm analysis.3.2 Sorting and Searching Algorithms:- Describe various sorting algorithms such as bubble sort, insertion sort, merge sort, and quicksort.- Discuss searching algorithms like linear search, binary search, and hash-based searching.3.3 Dynamic Programming and Greedy Algorithms:- Define dynamic programming and greedy algorithms and their applications.- Provide examples of problems that can be solved using these approaches.IV. Advanced Topics:4.1 Hashing and Hash Tables:- Explain the concept of hashing and its applications in efficient data retrieval.- Discuss hash functions, collision handling, and the implementation of hash tables.4.2 Graph Algorithms:- Explore advanced graph algorithms such as Dijkstra's algorithm, breadth-first search, and depth-first search.- Discuss their applications in solving complex problems like shortest path finding and network analysis.4.3 Advanced Data Structures:- Introduce advanced data structures like heaps, priority queues, and self-balancing binary search trees.- Explain their advantages, implementation details, and usage in various scenarios.V. Summary:5.1 Key Takeaways:- Summarize the main points covered in the book, emphasizing the importance of data structures and algorithm analysis.- Highlight the significance of selecting appropriate data structures and algorithms for efficient software development.5.2 Practical Applications:- Discuss real-world scenarios where the concepts from the book can be applied.- Illustrate how understanding data structures and algorithm analysis can lead to optimized software solutions.5.3 Conclusion:- Conclude the review by emphasizing the relevance and usefulness of the book "Data Structures and Algorithm Analysis."- Encourage readers to explore the book further for a deeper understanding of the subject.In conclusion, "Data Structures and Algorithm Analysis" is a comprehensive guide that covers essential concepts in data structures and algorithm analysis. The book's structure, detailed explanations, and practical examples make it a valuable resource for computer science students, software developers, and anyone interested in optimizing their software solutions. Understanding these fundamental concepts is crucial for building efficient and scalable software applications.。
计算机 数据库 外文文献翻译 中英文

科技外文文献Microsoft Future "Soul" - SQL Server 2005 Exploration SecretAuthor : CHEN Bao-linSQL Server development "Brief History"At the beginning of this before, let us look at Microsoft SQL Server development "Brief History."1988 : SQL Server from Microsoft and Sybase common development, running on OS / 2 platform.1993-09-14 : SQL Server 4.2, a desktop database system contains less functional. Integration with Windows and to provide easy-to-use user interface.1994 : Microsoft and Sybase database in cooperation in the development of suspension.1995 : SQL Server 6.0, code-named "SQL95" Microsoft rewriting most of the core system. Provide a low-cost small business application database program.1996-04-16 : SQL Server 6.5, This version brings significant performance improvement and providing a wide variety of useful functions.1998-11-16 : SQL Server 7.0, code-named "Sphinx." Completely rewritten core database engine, providing small and medium business applications database program, contains the initial Web support. SQL Server starting from this version has been widely used.2000-08-07 : the birth of SQL Server 2000, code-named "Shiloh." Microsoft to produce the product has been defined as enterprise-class database system, which includes three components (DB, OLAP, English Query). Rich front-end tools, improved development tools, and XML support, the promotion of this version of the promotion and application. And contains the following several versions.Enterprise Edition : through the deployment of cluster TB-class support services giant databases and thousands of concurrent users online.Standard Edition : to support SMEs.Personal version : support desktop applications.Developer : staff development for enterprises and Windows CE build enterprise applications.Window CE Version : can be applied to any Windows CE mobile devices.2003-04-24 : SQL Server 2000, 64-bit version. Codenamed "Liberty" has been and Unix / Linux Oracle compete.2005-11-07 : SQL Server 2005, codenamed "Yukon" Microsoft SQL Server products to the latest version. Microsoft commented that the status of this product took five years of major changes, a landmark product. Microsoft SQL Server 4.2 to 2005. Microsoft since the early 1990s to enter the database market, SQL Server 2005 until the launch, behaved like an enterprise database from the market to lead the followers of the restructuring, sword was sharpened for 10 years, through many a storm, Microsoft already enterprises database management perspective extends to a broader and deeper realm, the paper attempts to explore the history, Aggregate Microsoft SQL Server formative history.1987 Sysbase developed Unix systems running SQL Server version. In 1988, Microsoft invited the then momentum in the database fields are busy Sysbase. joint development of SQL server. "Sima heart erased", Microsoft tried to enter the database market moves obviously, and, database market is bound to whip up some wind action. Sure enough, after 10 years of market access database for the intense period of the Warring States. 1993-04-12, Microsoft SQL Server version 4.2. And before the introduction of Windows NT echoed that Microsoft officially entered the enterprise applications market. And the SQL Server database and the enterprise is the most important. Although SQL Server 4.2 while still just a desktop version, but there has been considerablepotential. 1994, Microsoft and Sybase formal suspension of the database development cooperation This meaningfully.From 1995 to 2000, Microsoft has adopted 6.0, 6.5,7.0, 2000 Version 4. From the perspective view, SQL Server 2000 version has been able to provide the following services.Online Services (On-line services) : "On-Line" refers to real-time online users use data services.Online transaction processing OLTP (On-Line Transaction Processing) : OLTP operation by the order-processing services transactions, or transactions follow completion or undoes all the principles. It also did not include the type of services. This is a sector that is the most universal and most widely forms of service. Analysis of online services OLAP (On-Line Analytical Processing) : OLAP is a kind of multidimensional data display (such as data warehousing, data mart, data cube), usually to do data mining. As OLTP used to operate and SQL data definition, OLAP is used and MDX (MultiDimensional Expressions) visit and definitions of data. From the technical structure of SQL Server 2000, as follows.Data structure•physical structure of data structure.•logical framework : how to define Tables, ro ws, columns, and other data objectsData Processing• data processing storage engine : it is responsible for dealing with how the data retention.• engine : it is responsible for how the data for the visit and relations.• SQL Server Agent : it is respo nsible for task scheduling and events management.Data manipulation• DB APIs : ADO (ActiveX Data Objects).OLE DB (linking and embedding data objects).DB-Library for C + +.ODBC (Open Data Internet).ESQL (Embedded SQL.)• URLs (uniform resource locat or address).• English inquiries (English Query).SQL Server Enterprise Manager.Tools : Inquiry analyzers, DTS (Data Transformation Services), Backup and restore and replication, metadata services, storage expansion process, SQL tracking, can be used for performance tuning.Experiences from users, SQL Server 2000 version of a number of new characteristics, such as XML support, many examples of support, data warehouse and business intelligence to enhance performance and scalability will improve, operating guide, and the inquiries, DTS, Transact SQL enhancements.From the license price, Microsoft SQL Server 2000, the price and total cost of ownership (TCO) only to the Oracle or D B2 2 / 1 to 1 / 3.In summary, Microsoft high-performance low-cost access to the product concept on the market success SQL Server 2000 database can meet the OLTP and OLAP application deployment, and better performance, and prices relative Oracle, DB2 and other databases low. Meanwhile, SQL Server 2000 Enterprise Edition also includes the standard version and other versions to meet different levels of user demand, These factors prompted the SQL Server 2000 was a significant part of the SME market share Microsoft has the opportunity to enter the mainstream database vendors ranks.At the same time, we should realize that SQL Server 2000 and Oracle launched late in the G 10 high-end enterprise-level functions in surviving deficient, so bridging the gap to catch up on the historic mission to the code-named "Yukon," the new version.Killer code-named "Yukon"From the 1989 release of Microsoft SQL Server 1.0 is now a full 15 years. In that 15 years of SQL Server fromscratch, from small to large, experiencing a once legendary. It has not only eroded with IBM, Oracle database market share, and the next generation of SQL Server has begun to gradually become the next Windows operating system core. China and the Bill Gates mouth • The constant repetition of "seamless calculation" is the core of Yukon, The code-named "Yukon," the next generation of our database will be brought into what kind of world? Internet "soft" pillarIn today's era of the network, data searching,data storage, classification of data, etc. All this has become the Internet network constitutes the "soft" pillars, and the database system is the pillar of the most critical. If there is no database support, we would never be able to Google or Baidu in the search for the information they need. can not use the convenient electronic mailbox, but that Network World because it is a large database consisting of.According to IDC's latest data show that the global database software market seems to be stirring Tension 2003 total revenue reached 13.6 billion U.S. dollars, compared with 2002's 12.6 billion U.S. dollars have increased. Oracle, IBM and Microsoft now controls 75% market share. Oracle last year for a market share of 39.8%, 31.3% for IBM, Microsoft to 12.1%.What is the database? In the University's computer textbooks, the database is being interpreted in this way : The database is the computer application system in a specialized data resource management system. There are many forms of data, such as text, digital, symbols, graphics, images and voices, and so on. All computer data system to deal with the subject. People familiar approach of a document is produced, will soon compile a program processing documents, will be covered by the procedural requirements of data organized into data files, documentation of procedures to call. Data files and program files maintain a certain relationship. Computer Application in the rapid development of the situation, by means of such a document will highlight deficiencies. For example, it allows poor definitive data, facilitate transplantation, in different documents stored information much duplication and waste of storage space, Update inconvenience. Database system will solve this problem. Database systems from the application of specific procedures, but based on the data management, All data will be stored in a database, scientific organizations, and by means of the database management system, using it as an intermediary, with a variety of applications or application interface to make it easy access to the data in the database.This note describes is indeed very detailed, but you may not always seem dizziness, In fact, a simple database that is after a group of computer collation of data stored in one or more documents, and the management of the database software called on the database management system. A general database system (104217) can be divided into the database (Database ) and Data Management System (Database Management System, DBMS) in two parts, all of these constitute the Internet is a "soft" pillars all.Microsoft's SQL Server database software, as many of the upgrade from 6.5 to the 7.0 version, gradually become mainstream database software, and SQL Server 2000 also proved that the Windows operating system can bear the same high-end data application, as the mainstream business application of database management software. It broke the rule by the large Unix database software myth and the next generation of SQL Server 2005 there will be what kind of change?Live Yukon core secretsMicrosoft in the next version of SQL Server (codenamed "Yukon") at the planning stage , considered more of the future development of the database, and SQL Server programming capabilities. Microsoft's internal development staff had long been aware that the future must introduce a more unified programming model but for a different data model to provide more flexibility. The unified programming model means that the ordinary data access and operation tasks can be carried out through various channels. For example, you can choose to use XML or Framework, or Transact-S QL (T-SQL) code, and so on.Such planning will result is a new database programming platform, which in many ways a natural extension. First, host. NET Framework common language runtime (CLR) to the function of the process of expansion of database programming and managed code area. Secondly,. NET framework provides a host integration from within SQL Server powerful object database functions. XML is the in-depth support functions through the XML data typeto achieve, and It has a data type of relationship between all the functions. In addition, also added a pair of XML Query (XQuery) and XML structure definition language (XSD) standard server support. Finally, SQL Server Yukon includes T-SQL language to enhance the important function.XML in SQL Server Yukon's history really began with SQL Server 2000. SQL Server 2000 with the introduction of the XML format to relational data. large load and segmentation XML documents and databases will be open targets for XML-based Web services, and other functions, However Yukon provide a more senior XML Query function, After perfecting the Y ukon will be full play all of the advantages of XML. XML Why so critical? In fact, from the initial XML an alternative HTML said the technical development of a line format, now be seen as a storage format. XML lasting memory has drawn widespread attention, the Internet has also been a lot of XML data type applications. XML itself can be an across any platform data format, It started as a file format for use, as XML in the enterprise has been widely recognized, Users began to use XML to solve thorny business problems, such as data integration. This makes as a data storage format XML development today, Because XML can be displayed on any platform to produce the same results, XML has become a mainstream database storage format. This built-in the Yukon comprehensive XML support will trigger a new database technology revolution.These new programming models and enhanced common language to create a series of programmable, They complement and expand the current relational database model. This architecture has the ultimate aim is to build more scalable, more reliable, more robust applications, and to enhance the development of efficiency. These models Another result is a service called SQL Agent new application framework -- for Asynchronous sources delivering the Distributed Application Framework.Yukon joining century gambleConstantly talking before we say a string of technology advantages, then you may very curious, Why should we introduce this appears to be a high-end database application software technologies? Perhaps we should kick the answer.The richest on Earth doing computer predictions for the future, he believes, in the next world, every one ordinary computer will have a large enough super hard disks, At that time the hard disk is no longer simply an 80 GB is likely to be 80 TB, Although it is only a change GB TB, but that means hard disk capacity of a full upgrade of 1000 times. And the existing Windows disk data storage NTFS format, simply unable to cope with such a large capacity hard disk data search. Said an image of the example, if the 100 TB of disk space on your computer, At that time, or you use Windows XP, You collate debris disk of the time required is likely to be for two days and two nights, if you want to find a particular document, You will have waited for several hours. That feeling is like to return to 286 times.In order to solve this thorny problem, the next generation Windows operating system Longhorn decided with the previous non-Windows diametrically with the programming model. The core is Avalon (development code). Avalon is the new Windows GUI library. New Longhorn into the Indigo (Web services) and WinFS (file system) of the new function. Including Avalon, these three new function called hell. Longhorn is the founder of a new "local" API. Although now is to the Win32 API compatibility and grow, However, to use the new Longhorn functions, under normal circumstances the use of hell. Max belongs to the present. NET Framework in the city. Present. NET Framework used in the category, which has hell, DLL support for the procedural mechanisms and the operation. NET basically the same.. NET Framework in SQL Server Yukon Availability when major version upgrade ( Major VersionUp), the specific date is the end of 2004. In the Yukon. NET Framework to run. In the storage process (Stored Procedures) use. NET Framework The class library. Yukon operations. NET Framework version 2.0. Supplementary to the present. NET Framework 1.1 is no relevant category of multimedia. WinFS use Yukon engines. In other words, Longhorn, the file system will use database engine.This time you understand, the next generation Windows operating system, the whole document data management will be introduced SQL Server configuration management, when Our computer data querycapabilities, data integration capability will be greatly enhanced. This of course, that the rich keep saying that the "seamless calculation" is a critical step on Microsoft, Let database software and operating systems integration projects century is undoubtedly a gamble, which, if successful, Microsoft will gradually become the dominant database, but if it fails, The almost even harden the next generation Windows listing of the normal schedule.Microsoft has provided some tools for SQL server and client applications on the network between the transmission of data increases secret. However, the Microsoft product manager said Kirsten Ward, plans to release next year a new SQL Server database will be stored in the data encryption, Hacker attacks increase defense capabilities.Microsoft earlier this year "SQL Server 2005" release time postponed until the first half of next year. The database software will enhance the launch of Microsoft database computing power and better with Oracle and IBM compete. Microsoft will also introduce a unified storage concept, locating and retrieving data more convenient. Oracle in Windows and Unix database market has been in a leading position. However, the recently adopted this year, Microsoft SQL Server to increase more advanced functions have also made remarkable progress.In addition, Microsoft will also provide a service called "Best Practices Analyzer Tool" (best practice analyzer tool) software. Database administrators can use the software using Microsoft editor of the Guide database software debugging. This applies to software tools for Microsoft database software current version "SQL Server 2000" and to provide a database administrator in various fields Operations Guide, For example, how to improve performance and how to conduct more effective data backup and so on.Ward said that the software tool also includes an "Upgrade Advisor" procedure. This procedure can scan database programs and warned "SQL Server 2000" users to make the necessary amendments changed so that the procedures compatible with the upcoming launch of the "SQL Server 2005."(Source : China Computer Education)中文译文微软未来的“灵魂”—SQL Server 2005探密作者:陈宝林SQL Server的发展“简史”在开始本文之前,先让我们来看一下微软SQL Server的发展“简史”。
数据库外文参考文献及翻译.

数据库外文参考文献及翻译数据库外文参考文献及翻译数据库管理系统——实施数据完整性一个数据库,只有用户对它特别有信心的时候。
这就是为什么服务器必须实施数据完整性规则和商业政策的原因。
执行SQL Server的数据完整性的数据库本身,保证了复杂的业务政策得以遵循,以及强制性数据元素之间的关系得到遵守。
因为SQL Server的客户机/服务器体系结构允许你使用各种不同的前端应用程序去操纵和从服务器上呈现同样的数据,这把一切必要的完整性约束,安全权限,业务规则编码成每个应用,是非常繁琐的。
如果企业的所有政策都在前端应用程序中被编码,那么各种应用程序都将随着每一次业务的政策的改变而改变。
即使您试图把业务规则编码为每个客户端应用程序,其应用程序失常的危险性也将依然存在。
大多数应用程序都是不能完全信任的,只有当服务器可以作为最后仲裁者,并且服务器不能为一个很差的书面或恶意程序去破坏其完整性而提供一个后门。
SQL Server使用了先进的数据完整性功能,如存储过程,声明引用完整性(DRI),数据类型,限制,规则,默认和触发器来执行数据的完整性。
所有这些功能在数据库里都有各自的用途;通过这些完整性功能的结合,可以实现您的数据库的灵活性和易于管理,而且还安全。
声明数据完整性声明数据完整原文请找腾讯3249114六,维-论'文.网 定义一个表时指定构成的主键的列。
这就是所谓的主键约束。
SQL Server使用主键约束以保证所有值的唯一性在指定的列从未侵犯。
通过确保这个表有一个主键来实现这个表的实体完整性。
有时,在一个表中一个以上的列(或列的组合)可以唯一标志一行,例如,雇员表可能有员工编号( emp_id )列和社会安全号码( soc_sec_num )列,两者的值都被认为是唯一的。
这种列经常被称为替代键或候选键。
这些项也必须是唯一的。
虽然一个表只能有一个主键,但是它可以有多个候选键。
SQL Server的支持多个候选键概念进入唯一性约束。
关于大数据的学术英文文献

关于大数据的学术英文文献Big Data: Challenges and Opportunities in the Digital Age.Introduction.In the contemporary digital era, the advent of big data has revolutionized various aspects of human society. Big data refers to vast and complex datasets generated at an unprecedented rate from diverse sources, including social media platforms, sensor networks, and scientific research. While big data holds immense potential for transformative insights, it also poses significant challenges and opportunities that require thoughtful consideration. This article aims to elucidate the key challenges and opportunities associated with big data, providing a comprehensive overview of its impact and future implications.Challenges of Big Data.1. Data Volume and Variety: Big data datasets are characterized by their enormous size and heterogeneity. Dealing with such immense volumes and diverse types of data requires specialized infrastructure, computational capabilities, and data management techniques.2. Data Velocity: The continuous influx of data from various sources necessitates real-time analysis and decision-making. The rapid pace at which data is generated poses challenges for data processing, storage, andefficient access.3. Data Veracity: The credibility and accuracy of big data can be a concern due to the potential for noise, biases, and inconsistencies in data sources. Ensuring data quality and reliability is crucial for meaningful analysis and decision-making.4. Data Privacy and Security: The vast amounts of data collected and processed raise concerns about privacy and security. Sensitive data must be protected fromunauthorized access, misuse, or breaches. Balancing data utility with privacy considerations is a key challenge.5. Skills Gap: The analysis and interpretation of big data require specialized skills and expertise in data science, statistics, and machine learning. There is a growing need for skilled professionals who can effectively harness big data for valuable insights.Opportunities of Big Data.1. Improved Decision-Making: Big data analytics enables organizations to make informed decisions based on comprehensive data-driven insights. Data analysis can reveal patterns, trends, and correlations that would be difficult to identify manually.2. Personalized Experiences: Big data allows companies to tailor products, services, and marketing strategies to individual customer needs. By understanding customer preferences and behaviors through data analysis, businesses can provide personalized experiences that enhancesatisfaction and loyalty.3. Scientific Discovery and Innovation: Big data enables advancements in various scientific fields,including medicine, genomics, and climate modeling. The vast datasets facilitate the identification of complex relationships, patterns, and anomalies that can lead to breakthroughs and new discoveries.4. Economic Growth and Productivity: Big data-driven insights can improve operational efficiency, optimize supply chains, and create new economic opportunities. By leveraging data to streamline processes, reduce costs, and identify growth areas, businesses can enhance their competitiveness and contribute to economic development.5. Societal Benefits: Big data has the potential to address societal challenges such as crime prevention, disease control, and disaster management. Data analysis can empower governments and organizations to make evidence-based decisions that benefit society.Conclusion.Big data presents both challenges and opportunities in the digital age. The challenges of data volume, velocity, veracity, privacy, and skills gap must be addressed to harness the full potential of big data. However, the opportunities for improved decision-making, personalized experiences, scientific discoveries, economic growth, and societal benefits are significant. By investing in infrastructure, developing expertise, and establishing robust data governance frameworks, organizations and individuals can effectively navigate the challenges and realize the transformative power of big data. As thedigital landscape continues to evolve, big data will undoubtedly play an increasingly important role in shaping the future of human society and technological advancement.。
计算机外文文献英文文献外文翻译信息系统开发和数据库开发

Information System Development and Database Development In many organizations, database development from the beginning of enterprise data modeling, data modeling enterprises determine the scope of the database and the general content. This step usually occurs in an organization's information system planning process, it aims to help organizations create an overall data description or explanation, and not the design of a specific database. A specific database for one or more information systems provide data and the corporate data model (which may involve a number of databases) described by the organization maintaining the scope of the data. Data modeling in the enterprise, you review of the current system, the need to support analysis of the nature of the business areas, the need for further description of the abstract data, and planning one or more database development project. Figure 1 shows Pine Valley furniture company's enterprise data model of a part.Information System ArchitectureAs shown in figure 1, senior data model is only general information system architecture (ISA) or a part of an organization's information system blueprint. In the information system planning, you can build an enterprise data model as a whole information system architecture part. According to Zachman (1987), Sowa and Zachman (1992) views of an information system architecture consists of the following six key components: Data (Figure 1 shows, but there are other methods that).Manipulation of data processing (of a data flow diagram can be used, with the object model methods, or other symbols that).Networks, which organizations and in organizations with its main transmission of data between business partners (it can connect through the network topology map and to demonstrate).People who deal with the implementation of data and information and is the source and receiver (in the process model for the data shows that the sender and the receiver).Implementation of the events and time points (they can use state transition diagram and other means.)The reasons for the incident and data processing rules (often in the form of text display, but there are also a number of charts for the planning tools such as decision tables).Information EngineeringInformation systems planners in accordance with the specific information system planning methods developed information system architecture. Information engineering is a popular and formal methods. Information engineering is a data-oriented creation and maintenance of the information system. Information engineering is because the data-oriented, so when you begin to understand how the database is defined by the logo and when information engineering a concise explanation is very helpful. Information Engineering follow top-down planning approach, in whichspecific information systems from a wide range of information needs in the understanding derived from (for example, we need about customers, products, suppliers, sales and processing of the data center), rather than merging many detailed information requested ( orders such as a screen or in accordance with the importation of geographical sales summary report). Top-down planning will enable developers to plan more comprehensive information system, consider system components provide an integrated approach to enhance the information system and the relationship between the business objectives of the understanding, deepen their understanding of information systems throughout the organization in understanding the impact.Information Engineering includes four steps: planning, analysis, design and implementation. The planning stage of project information generated information system architecture, including enterprise data model.Information System PlanningInformation systems planning objective is to enable IT organizations and the business strategy closely integrated, such integration for the information systems and technology to make the most of the investment interest is very important. As the table as a description, information engineering approach the planning stage include three steps, we in the follow-up of three sections they discussed.1. Critical factors determining the planningPlanning is the key factor that organizational objectives, critical success factors and problem areas. These factors determine the purpose of the establishment of planning and environment planning and information systems linked to strategic business planning. Table 2 shows the Pine Valley furniture company's key planning a number of possible factors, these factors contribute to the information systems manager for the new information systems and databases clubs top priority to deal with the demand. For example, given the imprecise sales forecasts this problem areas, information systems managers in the organization may be stored in the database additional historical sales data, new market research data and new product test data.2. The planning organizations set targetsOrganizations planning targets defined scope of business, and business scope will limit the subsequent analysis and information systems may change places. Five key planning targets as follows:●organizational units in the various sectors.●organizations location of the place of business operations.●functions of the business support organizations handling mission of the relevant group. Unlike business organizations function modules, in fact a function can be assigned to various organizations modules (for example, product development function is the production and sale of thecommon responsibility of the Ministry).●types of entities managed by the organization on the people, places and things of the major types of data.●Information System data set processing software applications and support procedures.3. To set up a business modelA comprehensive business model including the functions of each enterprise functional decomposition model, the enterprise data model and the various planning matrix. Functional decomposition is the function of the organization for a more detailed decomposition process, the functional decomposition is to simplify the analysis of the issue, distracted and identify components and the use of the classical approach. Pine Valley furniture company in order to function in the functional decomposition example in figure 2 below. In dealing with business functions and support functions of the full set, multiple databases, is essential to a specific database therefore likely only to support functions (as shown in Figure 2) provide a subset of support. In order to reduce data redundancy and to make data more meaningful, has a complete, high-level business view is very helpful.The use of specific enterprise data model to describe the symbol. Apart from the graphical description of this type of entity, a complete enterprise data model should also include a description of each entity type description of business operations and a summary of that business rules.Business rules determine the validity of the data.An enterprise data model includes not only the types of entities, including the link between the data entities, as well as various other objects planning links. Showed that the linkage between planning targets a common form of matrix. Because of planning matrix need not be explicit modeling database can be clearly described business needs, planning matrix is an important function. Regular planning matrix derived from the operational rules, it will help social development activities that top priority will be sorting and development activities under the top-down view through an enterprise-wide approach for the development of these activities. There are many types of planning matrix is available, their commonalities are:●locations - features show business function in which the implementation of operational locations.●unit - functions which showed that business function or business unit responsible for implementation.●Information System - data entities to explain how each information system interact with each data entity (for example, whether or not each system in each entity have the data to create, retrieve, update and delete).●support functions - data in each functional entities in the data set for the acquisition, use, update and delete.●Information System - target indication for each information systemto support business objectives.Figure 3 illustrate a possible functions - data entities matrix. Such a matrix can be used for a variety of purposes, including the following three objectives:1) identify gaps in the data entities to indicate the types of entities not use any function or functions which do not use any entity.2) found that the loss of each functional entities involved in the inspection staff through the matrix to identify any possible loss of the entity.3) The distinction between development activities if the priority to the top of a system development function for a high-priority (probably because it important organizational objectives related), then this area used by entities in the development of the database has a high priority.Hoffer, George and Valacich (2002) are the works of the matrix on how to use the planning and completion of the Information Engineering The planning system more complete description.2 database development processBased on information engineering information systems planning database is a source of development projects. These new database development projects is usually in order to meet the strategic needs of organizations, such as improving customer support, improve product and inventory management, or a more accurate sales forecast. However, manymore database development project is the bottom-up approach emerging, such as information system user needs specific information to complete their work, thus beginning a project request, and as other information systems experts found that organizations need to improve data management and begin new projects. Bottom-up even in the circumstances, to set up an enterprise data model is also necessary to understand the existing database can provide the necessary data, otherwise, the new database, data entities and attributes can be added to the current data resources to the organization.Both the strategic needs or operational information needs of each database development projects normally concentrated in a database. Some projects only concentrated in the database definition, design and implementation of a database, as a follow-up to the basis of the development of information systems. However, in most cases, the database and associated information processing function as a complete information systems development project was part of the development.System Development Life CycleGuide management information system development projects is the traditional process of system development life cycle (SDLC). System development life cycle is an organization of the database designers and programmers information system composed of the Panel of Experts detailed description, development, maintenance and replacement of theentire information system steps. This process is because Waterfall than for every step into the adjacent the next step, that is, the information system is a specification developed by a piece of land, every piece of the output is under an input. However shown in the figure, these steps are not purely linear, each of the steps overlap in time (and thus can manage parallel steps), but when the need to reconsider previous decisions, but also to roll back some steps ahead. (And therefore water can be put back in the waterfall!) Figure 4 on the system development life cycle and the purpose of each stage of the product can be delivered concise notes. The system development life cycle including each stage and database development-related activities, therefore, the question of database management systems throughout the entire development process. In Figure 5 we repeat of the system development life cycle stage of the seven, and outlines the common database at each stage of development activities. Please note that the systems development life cycle stages and database development steps一一对应exists between the relationship between the concept of modeling data in both systems development life cycle stages between.Enterprise ModelingDatabase development process from the enterprise modeling (system development life cycle stage of the project feasibility studies, and to choose a part), Organizations set the scope and general database content.Enterprise modeling in information systems planning and other activities, these activities determine which part of information systems need to change and strengthen the entire organization and outlines the scope of data. In this step, check the current database and information systems, development of the project as the main areas of the nature of the business, with a very general description of each term in the development of information systems when needed data. Each item only when it achieved the expected goals of organizations can be when the next step.Conceptual Data ModelingOne has already begun on the Information System project, the concept of data modeling phase of the information systems needs of all the data. It is divided into two stages. First, it began the project in the planning stage and the establishment of a plan similar to Figure 1. At the same time outlining the establishment of other documents to the existing database without considering the circumstances specific development projects in the scope of the required data. This category only includes high-level data (entities), and main contact. Then in the system development life-cycle analysis stage must have a management information system set the entire organization Details of the data model definition of all data attributes, listing all data types that all data inter-entity business linkages, defining description of the full data integrity rules. In the analysis phase, but also the concept of inspection data model (also called the concept behind the model) and thegoal of information systems used to explain other aspects of the model of consistency categories, such as processing steps, rules and data processing time of timing. However, even if the concept is such detailed data model is only preliminary, because follow-up information system life cycle activities in the design of services, statements, display and inquiries may find that missing element or mistakes. Therefore, the concept of data often said that modeling is a top-down manner, its areas of operation from the general understanding of the driver, rather than the specific information processing activities by the driver.3. Logical Database DesignLogical database design from two perspectives database development. First, the concept of data model transform into relational database theory based on the criteria that means - between. Then, as the design of information systems, every computer procedures (including procedures for the input and output format), database support services, statements, and inquiries revealed that a detailed examination. In this so-called Bottom-up analysis, accurate verification of the need to maintain the database and the data in each affairs, statements and so on the needs of those in the nature of the data.For each separate statements, services, and so on the analysis must take into account a specific, limited but complete database view. When statements, services, and other analysis might be necessary to change theconcept of data model. Especially in large-scale projects, the different analytical systems development staff and the team can work independently in different procedures or in a centralized, the details of their work until all the logic design stage may be displayed. In these circumstances, logic database design stage must be the original concept of data model and user view these independent or merged into a comprehensive design. In logic design information systems also identify additional information processing needs of these new demands at this time must be integrated into the logic of earlier identified in the database design.Logical database design is based on the final step for the formation of good data specifications and determine the rules, the combination, the data after consultation specifications or converted into basic atomic element. Most of today's database, these rules from the relational database theory and the process known as standardization. This step is the result of management of these data have not cited any database management system for a complete description of the database map. Logical database design completed, we began to identify in detail the logic of the computer program and maintenance, the report contents of the database for inquiries.4. Physical database design and definitionPhysical database design and definition phase decisions computer memory (usually disk) database in the organization, definition of According to the library management system for physical structure, the proceduresoutlined processing services, produce the desired management information and decision support statements. The objective of this stage is to design an effective and safe management of all data-processing database, the physical database design to closely integrate the information systems of other physical aspects of the design, including procedures, computer hardware, operating systems and data communications networks.5. Database ImplementationThe database prepared by the realization stage, testing and installation procedures for handling databases. Designers can use the standard programming language (such as COBOL, C or Visual Basic), the dedicated database processing languages (such as SQL), or the process of the non-exclusive language programming in order to produce a statement of the fixed format, the result will be displayed, and may also include charts. In achieving stage, but also the completion of all the database files, training users for information systems (database) user setup program. The final step is to use existing sources of information (documents legacy applications and databases and now needs new data) loading data. Loading data is often the first step in data from existing files and databases to an intermediate format (such as binary or text files) and then to turn intermediate loading data to a new database. Finally, running databases and related applications for the actual user maintenance and retrieval of data. In operation, the regular backup database and the database when damaged or affected resumedatabase.6. Database maintenanceDuring the database in the progressive development of database maintenance. In this step, in order to meet changing business conditions, in order to correct the erroneous database design, database applications or processing speed increase, delete or change the structure of the database. When a procedure or failure of the computer database affect or damage the database may also be reconstruction. This step usually is the longest in the database development process step, as it continued to databases and related applications throughout the life cycle, the development of each database can be seen as a brief database development process and data modeling concepts arise, logical and physical database design and database to achieve dealing with the changes.Information System developed by other meansSystem Development Life Cycle minor changes in law or its variant of the often used to guide information systems and database development. Information System is a life-cycle methodology, it is highly structured approach, which includes many checks and balances to ensure that every step of produce accurate results, and new or alternative information system and it must communications or data definitions consistent existing system needs consistency. System development life cycle because of the regular need to have a working system for a long time been criticized because onlywork in the system until the end of the whole process generated. More and more organizations now use rapid application development method, it is a includes analysis, design and implementation of steps to repeat the rapid iterative process until convergence to users the system so far. Rapid Application Development Act required the database has been in existence, and enhance system is mainly to the application of data retrieval application, but not to those who generate and modify database applications.The most widely used method of rapid application development is one of the prototype. The prototype system is a method of iterative development process, analysts and users through close co-operation, continuing to revise the system will eventually convert all the needs of a working system. Figure 6 shows prototype of the process. In this diagram we contains notes, briefly describes each stage of the prototype of the database development activities. Normally, when information systems problems were identified, tried only a rough concept of data modeling. In the development of the initial prototype, the design of the user wants to display and statements, and that any new database needs and define a term prototype database. This is usually a new database, copy the part of the existing system, but might also added some new content. When the need for new content, these elements are usually from external data sources, such as market research data, the general economic indicators or industry standards.When a prototype of a new version to repeat the achievement andmaintenance of database activities. Usually only a minimum level of security and integrity control, because at this time the focus is as soon as possible to produce a prototype version can be used. But document management project also deferred to the final, only be used in the delivery of user training. Finally, once constructed an acceptable prototype, developers, and users will be the final decision of whether to prototype delivery and the use of the database. If the system (including database) efficiency is very low, then the system and database will be re-programming and re-organization in order to achieve the desired performance.Along with visual programming tools (such as Visual Basic, Java, Visual C + + and fourth generation language) increasingly popular use of visual programming tools can easily change the user interface with the system, the prototype is becoming the choice of system development methodology. Customers using the prototype method statements and show changes to the content and layout is quite easy. In the process, the new database needs were identified, so it is the development of the use of the existing database should be amended. There is even the possibility of a need for a new database system prototype method, in such circumstances, when the system demand in the iterative process of development in the ever-changing needs access to sample data, the construction or reconstruction of the database prototype.3 database development of the three-tier architecture modelIn this article on the front of the database development process mentioned in the interpretation of a system development project on the establishment of the several different, but related database view or model:●conceptual model (in the analysis stage of the establishment).●external model or user view (in the analysis phase and the establishment of logical design phase).●physical model or internal model (in the physical design phase of the establishment).Figure 7 describes the database view that the relationship between the three, it is important to remember that they are the same organizations database view or model. In other words, each organization has a database of the physical model, a concept model and one or more users view. Therefore, the three-tier architecture model using the same data set observe the different ways definition database.Concept models on the full database structure, has nothing to do with the technical specifications. Conceptual model definition do not involve the entire database data stored in the computer how the secondary memory. Usually, the conceptual model by entities - links (E-R) map or object modeling symbols such a graphical format to describe, we have this type of concept model called the data model. In addition, the conceptual model specification as a metadata stored in the database or data dictionary.Physical models including conceptual model of how data stored incomputer memory in the two specifications. Analysts and the database design is as important to the physical database (physical mode) definition, it provides information on the distribution and management of data storage and access of the physical memory space of two full database technology specifications.Database development and database technology database is among the three models divided into basis. Database development projects may have a role to only deal with these three views of a related work. For example, a beginner may be designed for one or more procedures external model, and an experienced developer will design the physical model or conceptual model. Database design issues at different levels are quite different.4 three-tier structure of the database positioning systemObviously, all the good things in the database are, and the "three"!When designing a database, you have to choose where to store data. This option in the physical database design stage. Database is divided into individual databases, the Working Group database, departmental databases, corporate databases and the Internet database. Individuals often by the end-user database design and development of their own, just by database experts to give training and advice to help, it only contains individual end-users interested in the data. Sometimes, personal database from the database or enterprise Working Group extracted from the database, such circumstances database prepared by some experts from the regular routineto create local database. Sector Working Group database and the database is often the end-user, business experts and the central database system experts development. The collaborative work of these officers is necessary because in the design of the database to be shared by a large number of issues weigh: processing speed, ease of use, data definition differences and other similar problems. Due to corporate databases and the Internet database broad impact, large-scale, it is normally concentrated in the database development team has received professional training to develop a database of experts.1. Customers layerA desktop or notebook also known as that layer, which specialized management user interface and system localization data in this layer can be implemented on the Web scripting tasks.2. Server / Web serverHTTP protocol handling, scripting tasks, the implementation of computing and provide data access, the layer known as processing services layer.3. Enterprise Server (Minicomputer or mainframe) layerThe implementation of complex computing and inter-organizational management from multiple data sources of data integration, also known as data services layer.In an organization, hierarchical database and information system architecture for distributed computing and the client / server architecture ofthe concept of correlation. Client / server architecture based on a LAN environment, including servers (referred to as database server or database engine) database software implementation from the client workstation database orders, each customer applications focus on their user interface functions. In fact, the whole concept of the database (as well as the application of these databases to handle routine) as a distributed database or the separate but related physical database distribution in the local PC workstation, server intermediate (working group or sector) and one center server (departments or enterprises ). Simply said that the use of client / server architecture for:●it can handle multiple processors on the same application at the same time, improve application response time and data processing speed.●It can use each computer platform of the best data processing (such as PC Minicom Advanced user interface with the mainframe and computing speed).●can mix various client technology (Intel or Motorola processor assembly of personal computers, computer networks, information kiosks, etc.) and public data sharing. In addition, you can change the technology at any layer and other layers only a small influence on the system module.●able to handle close to the data source to be addressed to improve response time and reduce network traffic.●accept it to allow and encourage open systems standards.。
数据库英文参考文献(最新推荐120个)

由于我国经济的高速发展,计算机科学技术在当前各个科技领域中迅速发展,成为了应用最广泛的技术之一.其中数据库又是计算机科学技术中发展最快,应用最广泛的重要分支之一.它已成为计算机信息系统和计算机应用系统的重要技术基础和支柱。
下面是数据库英文参考文献的分享,希望对你有所帮助。
数据库英文参考文献一:[1]Nú?ez Matías,Weht Ruben,Nú?ez Regueiro Manuel. Searching for electronically two dimensional metals in high-throughput ab initio databases[J]. Computational Materials Science,2020,182.[2]Izabela Karsznia,Marta Przychodzeń,Karolina Sielicka. Methodology of the automatic generalization of buildings, road networks, forests and surface waters: a case study based on the Topographic Objects Database in Poland[J]. Geocarto International,2020,35(7).[3]Alankrit Chaturvedi. Secure Cloud Migration Challenges and Solutions[J]. Journal of Research in Science and Engineering,2020,2(4).[4]Ivana Nin?evi? Pa?ali?,Maja ?uku?i?,Mario Jadri?. Smart city research advances in Southeast Europe[J]. International Journal of Information Management,2020.[5]Jongseong Kim,Unil Yun,Eunchul Yoon,Jerry Chun-Wei Lin,Philippe Fournier-Viger. One scan based high average-utility pattern mining in static and dynamic databases[J]. Future Generation Computer Systems,2020.[6]Jo?o Peixoto Martins,António Andrade-Campos,Sandrine Thuillier. Calibration of Johnson-Cook Model Using Heterogeneous Thermo-Mechanical Tests[J]. Procedia Manufacturing,2020,47.[7]Anna Soriani,Roberto Gemignani,Matteo Strano. A Metamodel for the Management of Large Databases: Toward Industry 4.0 in Metal Forming[J]. Procedia Manufacturing,2020,47.[8]Ayman Elbadawi,Karim Mahmoud,Islam Y. Elgendy,Mohammed Elzeneini,Michael Megaly,Gbolahan Ogunbayo,Mohamed A. Omer,Michelle Albert,Samir Kapadia,Hani Jneid. Racial disparities in the utilization and outcomes of transcatheter mitral valve repair: Insights from a national database[J]. Cardiovascular Revascularization Medicine,2020.[9]Maurizio Boccia,Antonio Sforza,Claudio Sterle. Simple Pattern Minimality Problems: Integer Linear Programming Formulations and Covering-Based Heuristic Solving Approaches[J]. INFORMS Journal on Computing,2020.[10]. Inc.; Patent Issued for Systems And User Interfaces For Dynamic Access Of Multiple Remote Databases And Synchronization Of Data Based On User Rules (USPTO 10,628,448)[J]. Computer Technology Journal,2020.[11]. Bank of America Corporation; Patent Issued for System For Electronic Data Verification, Storage, And Transfer (USPTO 10,628,058)[J]. Computer Technology Journal,2020.[12]. Information Technology - Database Management; Data from Technical University Munich (TU Munich) Advance Knowledge in Database Management (Make the most out of your SIMD investments: counter control flow divergence in compiled query pipelines)[J]. Computer Technology Journal,2020.[13]. Information Technology - Database Management; Studies from Pontifical Catholic University Update Current Data on Database Management (General dynamic Yannakakis: conjunctive queries with theta joins under updates)[J]. Computer Technology Journal,2020.[14]Kimothi Dhananjay,Biyani Pravesh,Hogan James M,Soni Akshay,Kelly Wayne. Learning supervised embeddings for large scale sequence comparisons.[J]. PloS one,2020,15(3).[15]. Information Technology; Studies from University of California San Diego (UCSD) Reveal New Findings on Information Technology (A Physics-constrained Data-driven Approach Based On Locally Convex Reconstruction for Noisy Database)[J]. Information Technology Newsweekly,2020.[16]. Information Technology; Researchers from National Institute of Information and Communications Technology Describe Findings in Information Technology (Efficient Discovery of Weighted Frequent Neighborhood Itemsets in Very Large Spatiotemporal Databases)[J]. Information Technology Newsweekly,2020.[17]. Information Technology; Investigators at Gdansk University of Technology Report Findings in Information Technology (A Framework for Accelerated Optimization of Antennas Using Design Database and Initial Parameter Set Estimation)[J]. Information Technology Newsweekly,2020.[18]. Information Technology; Study Results from Palacky University Update Understanding of Information Technology (Evaluation of Replication Mechanisms on Selected Database Systems)[J]. Information Technology Newsweekly,2020.[19]Runfola Daniel,Anderson Austin,Baier Heather,Crittenden Matt,Dowker Elizabeth,Fuhrig Sydney,Goodman Seth,Grimsley Grace,Layko Rachel,MelvilleGraham,Mulder Maddy,Oberman Rachel,Panganiban Joshua,Peck Andrew,Seitz Leigh,Shea Sylvia,Slevin Hannah,Youngerman Rebecca,Hobbs Lauren. geoBoundaries: A global database of political administrative boundaries.[J]. PloS one,2020,15(4).[20]Dupré Damien,Krumhuber Eva G,Küster Dennis,McKeown Gary J. A performance comparison of eight commercially available automatic classifiers for facial affect recognition.[J]. PloS one,2020,15(4).[21]Partha Pratim Banik,Rappy Saha,Ki-Doo Kim. An Automatic Nucleus Segmentation and CNN Model based Classification Method of White Blood Cell[J]. Expert Systems With Applications,2020,149.[22]Hang Dong,Wei Wang,Frans Coenen,Kaizhu Huang. Knowledge base enrichment by relation learning from social tagging data[J]. Information Sciences,2020,526.[23]Xiaodong Zhao,Dechang Pi,Junfu Chen. Novel trajectory privacy-preserving method based on clustering using differential privacy[J]. Expert Systems With Applications,2020,149.[24]. Information Technology; Researchers at Beijing University of Posts and Telecommunications Have Reported New Data on Information Technology (Mining top-k sequential patterns in transaction database graphs)[J]. Internet Weekly News,2020.[25]Sunil Kumar Sharma. An empirical model (EM: CCO) for clustering, convergence and center optimization in distributive databases[J]. Journal of Ambient Intelligence and Humanized Computing,2020(prepublish).[26]Naryzhny Stanislav,Klopov Nikolay,Ronzhina Natalia,Zorina Elena,Zgoda Victor,Kleyst Olga,Belyakova Natalia,Legina Olga. A database for inventory of proteoform profiles: "2DE-pattern".[J]. Electrophoresis,2020.[27]Noel Varela,Jesus Silva,Fredy Marin Gonzalez,Pablo Palencia,Hugo Hernandez Palma,Omar Bonerge Pineda. Method for the Recovery of Images in Databases of Rice Grains from Visual Content[J]. Procedia Computer Science,2020,170.[28]Ahmad Rabanimotlagh,Prabhu Janakaraj,Pu Wang. Optimal Crowd-Augmented Spectrum Mapping via an Iterative Bayesian Decision Framework[J]. Ad Hoc Networks,2020.[29]Ismail Boucherit,Mohamed Ould Zmirli,Hamza Hentabli,Bakhtiar Affendi Rosdi. Finger vein identification using deeply-fused Convolutional Neural Network[J]. Journal of King Saud University - Computer and Information Sciences,2020.[30]Sachin P. Patel,S.H. Upadhyay. Euclidean Distance based Feature Ranking andSubset Selection for Bearing Fault Diagnosis[J]. Expert Systems With Applications,2020.[31]Julia Fomina,Denis Safikanov,Alexey Artamonov,Evgeniy Tretyakov. Parametric and semantic analytical search indexes in hieroglyphic languages[J]. Procedia Computer Science,2020,169.[32]Selvine G. Mathias,Sebastian Schmied,Daniel Grossmann. An Investigation on Database Connections in OPC UA Applications[J]. Procedia Computer Science,2020,170.[33]Abdourrahmane Mahamane Atto,Alexandre Benoit,Patrick Lambert. Timed-image based deep learning for action recognition in video sequences[J]. Pattern Recognition,2020.[34]Yonis Gulzar,Ali A. Alwan,Abedallah Zaid Abualkishik,Abid Mehmood. A Model for Computing Skyline Data Items in Cloud Incomplete Databases[J]. Procedia Computer Science,2020,170.[35]Xiaohan Yang,Fan Li,Hantao Liu. Deep feature importance awareness based no-reference image quality prediction[J]. Neurocomputing,2020.[36]Dilana Hazer-Rau,Sascha Meudt,Andreas Daucher,Jennifer Spohrs,Holger Hoffmann,Friedhelm Schwenker,Harald C. Traue. The uulmMAC Database—A Multimodal Affective Corpus for Affective Computing in Human-Computer Interaction[J]. Sensors,2020,20(8).[37]Tomá? Pohanka,Vilém Pechanec. Evaluation of Replication Mechanisms on Selected Database Systems[J]. ISPRS International Journal of Geo-Information,2020,9(4).[38]Verheggen Kenneth,Raeder Helge,Berven Frode S,Martens Lennart,Barsnes Harald,Vaudel Marc. Anatomy and evolution of database search engines-a central component of mass spectrometry based proteomic workflows.[J]. Mass spectrometry reviews,2020,39(3).[39]Moscona Leon,Casta?eda Pablo,Masrouha Karim. Citation analysis of the highest-cited articles on developmental dysplasia of the hip.[J]. Journal of pediatric orthopedics. Part B,2020,29(3).[40]Nasseh Daniel,Schneiderbauer Sophie,Lange Michael,Schweizer Diana,Heinemann Volker,Belka Claus,Cadenovic Ranko,Buysse Laurence,Erickson Nicole,Mueller Michael,Kortuem Karsten,Niyazi Maximilian,Marschner Sebastian,Fey Theres. Optimizing the Analytical Value of Oncology-Related Data Based on an In-Memory Analysis Layer: Development and Assessment of the Munich OnlineComprehensive Cancer Analysis Platform.[J]. Journal of medical Internet research,2020,22(4).数据库英文参考文献二:[41]Meiling Chai,Changgeng Li,Hui Huang. A New Indoor Positioning Algorithm of Cellular and Wi-Fi Networks[J]. Journal of Navigation,2020,73(3).[42]Mandy Watson. How to undertake a literature search: a step-by-step guide[J]. British Journal of Nursing,2020,29(7).[43]. Patent Application; "Memorial Facility With Memorabilia, Meeting Room, Secure Memorial Database, And Data Needed For An Interactive Computer Conversation With The Deceased" in Patent Application Approval Process (USPTO 20200089455)[J]. Computer Technology Journal,2020.[44]. Information Technology; Data on Information Technology Detailed by Researchers at Complutense University Madrid (Hr-sql: Extending Sql With Hypothetical Reasoning and Improved Recursion for Current Database Systems)[J]. Computer Technology Journal,2020.[45]. Science - Metabolomics; Study Data from Wake Forest University School of Medicine Update Knowledge of Metabolomics (Software tools, databases and resources in metabolomics: updates from 2018 to 2019)[J]. Computer Technology Journal,2020.[46]. Sigma Computing Inc.; Researchers Submit Patent Application, "GeneratingA Database Query To Dynamically Aggregate Rows Of A Data Set", for Approval (USPTO 20200089796)[J]. Computer Technology Journal,2020.[47]. Machine Learning; Findings on Machine Learning Reported by Investigators at Tongji University (Comparing Machine Learning Algorithms In Predicting Thermal Sensation Using Ashrae Comfort Database Ii)[J]. Computer Technology Journal,2020.[48]. Sigma Computing Inc.; "Generating A Database Query Using A Dimensional Hierarchy Within A Graphical User Interface" in Patent Application Approval Process (USPTO 20200089794)[J]. Computer Technology Journal,2020.[49]Qizhi He,Jiun-Shyan Chen. A physics-constrained data-driven approach based on locally convex reconstruction for noisy database[J]. Computer Methods in Applied Mechanics and Engineering,2020,363.[50]José A. Delgado-Osuna,Carlos García-Martínez,JoséGómez-Barbadillo,Sebastián Ventura. Heuristics for interesting class association rule mining a colorectal cancer database[J]. Information Processing andManagement,2020,57(3).[51]Edival Lima,Thales Vieira,Evandro de Barros Costa. Evaluating deep models for absenteeism prediction of public security agents[J]. Applied Soft Computing Journal,2020,91.[52]S. Fareri,G. Fantoni,F. Chiarello,E. Coli,A. Binda. Estimating Industry 4.0 impact on job profiles and skills using text mining[J]. Computers in Industry,2020,118.[53]Estrela Carlos,Pécora Jesus Djalma,Dami?o Sousa-Neto Manoel. The Contribution of the Brazilian Dental Journal to the Brazilian Scientific Research over 30 Years.[J]. Brazilian dental journal,2020,31(1).[54]van den Oever L B,Vonder M,van Assen M,van Ooijen P M A,de Bock G H,Xie X Q,Vliegenthart R. Application of artificial intelligence in cardiac CT: From basics to clinical practice.[J]. European journal of radiology,2020,128.[55]Li Liu,Deborah Silver,Karen Bemis. Visualizing events in time-varying scientific data[J]. Journal of Visualization,2020,23(2–3).[56]. Information Technology - Database Management; Data on Database Management Discussed by Researchers at Arizona State University (Architecture of a Distributed Storage That Combines File System, Memory and Computation In a Single Layer)[J]. Information Technology Newsweekly,2020.[57]. Information Technology - Database Management; New Findings from Guangzhou Medical University Update Understanding of Database Management (GREG-studying transcriptional regulation using integrative graph databases)[J]. Information Technology Newsweekly,2020.[58]. Technology - Laser Research; Reports from Nicolaus Copernicus University in Torun Add New Data to Findings in Laser Research (Nonlinear optical study of Schiff bases using Z-scan technique)[J]. Journal of Technology,2020.[59]Loeffler Caitlin,Karlsberg Aaron,Martin Lana S,Eskin Eleazar,Koslicki David,Mangul Serghei. Improving the usability and comprehensiveness of microbial databases.[J]. BMC biology,2020,18(1).[60]Caitlin Loeffler,Aaron Karlsberg,Lana S. Martin,Eleazar Eskin,David Koslicki,Serghei Mangul. Improving the usability and comprehensiveness of microbial databases[J]. BMC Biology,2020,18(1).[61]Dean H. Barrett,Aderemi Haruna. Artificial intelligence and machine learningfor targeted energy storage solutions[J]. Current Opinion in Electrochemistry,2020,21.[62]Chenghao Sun. Research on investment decision-making model from the perspective of “Internet of Things + Big data”[J]. Future Generation Computer Systems,2020,107.[63]Sa?a Adamovi?,Vladislav Mi?kovic,Nemanja Ma?ek,Milan Milosavljevi?,Marko ?arac,Muzafer Sara?evi?,Milan Gnjatovi?. An efficient novel approach for iris recognition based on stylometric features and machine learning techniques[J]. Future Generation Computer Systems,2020,107.[64]Olivier Pivert,Etienne Scholly,Grégory Smits,Virginie Thion. Fuzzy quality-Aware queries to graph databases[J]. Information Sciences,2020,521.[65]Javier Fernando Botía Valderrama,Diego José Luis Botía Valderrama. Two cluster validity indices for the LAMDA clustering method[J]. Applied Soft Computing Journal,2020,89.[66]Amer N. Kadri,Marie Bernardo,Steven W. Werns,Amr E. Abbas. TAVR VS. SAVR IN PATIENTS WITH CANCER AND AORTIC STENOSIS: A NATIONWIDE READMISSION DATABASE REGISTRY STUDY[J]. Journal of the American College of Cardiology,2020,75(11).[67]. Information Technology; Findings from P. Sjolund and Co-Authors Update Knowledge of Information Technology (Whole-genome sequencing of human remains to enable genealogy DNA database searches - A case report)[J]. Information Technology Newsweekly,2020.[68]. Information Technology; New Findings from P. Yan and Co-Researchers in the Area of Information Technology Described (BrainEXP: a database featuring with spatiotemporal expression variations and co-expression organizations in human brains)[J]. Information Technology Newsweekly,2020.[69]. IDERA; IDERA Database Tools Expand Support for Cloud-Hosted Databases[J]. Information Technology Newsweekly,2020.[70]Adrienne Warner,David A. Hurley,Jonathan Wheeler,Todd Quinn. Proactive chat in research databases: Inviting new and different questions[J]. The Journal of Academic Librarianship,2020,46(2).[71]Chidentree Treesatayapun. Discrete-time adaptive controller based on IF-THEN rules database for novel architecture of ABB IRB-1400[J]. Journal of the Franklin Institute,2020.[72]Tian Fang,Tan Han,Cheng Zhang,Ya Juan Yao. Research and Construction of the Online Pesticide Information Center and Discovery Platform Based on Web Crawler[J]. Procedia Computer Science,2020,166.[73]Dinusha Vatsalan,Peter Christen,Erhard Rahm. Incremental clustering techniques for multi-party Privacy-Preserving Record Linkage[J]. Data & Knowledge Engineering,2020.[74]Ying Xin Liu,Xi Yuan Li. Design and Implementation of a Business Platform System Based on Java[J]. Procedia Computer Science,2020,166.[75]Akhilesh Kumar Bajpai,Sravanthi Davuluri,Kriti Tiwary,Sithalechumi Narayanan,Sailaja Oguru,Kavyashree Basavaraju,Deena Dayalan,Kavitha Thirumurugan,Kshitish K. Acharya. Systematic comparison of the protein-protein interaction databases from a user's perspective[J]. Journal of Biomedical Informatics,2020,103.[76]P. Raveendra,V. Siva Reddy,G.V. Subbaiah. Vision based weed recognition using LabVIEW environment for agricultural applications[J]. Materials Today: Proceedings,2020,23(Pt 3).[77]Christine Rosati,Emily Bakinowski. Preparing for the Implementation of an Agnis Enabled Data Reporting System and Comprehensive Research Level Data Repository for All Cellular Therapy Patients[J]. Biology of Blood and Marrow Transplantation,2020,26(3).[78]Zeiser Felipe André,da Costa Cristiano André,Zonta Tiago,Marques Nuno M C,Roehe Adriana Vial,Moreno Marcelo,da Rosa Righi Rodrigo. Segmentation of Masses on Mammograms Using Data Augmentation and Deep Learning.[J]. Journal of digital imaging,2020.[79]Dhaked Devendra K,Guasch Laura,Nicklaus Marc C. Tautomer Database: A Comprehensive Resource for Tautomerism Analyses.[J]. Journal of chemical information and modeling,2020,60(3).[80]Pian Cong,Zhang Guangle,Gao Libin,Fan Xiaodan,Li Fei. miR+Pathway: the integration and visualization of miRNA and KEGG pathways.[J]. Briefings in bioinformatics,2020,21(2).数据库英文参考文献三:[81]Marcello W. M. Ribeiro,Alexandre A. B. Lima,Daniel Oliveira. OLAP parallel query processing in clouds with C‐ParGRES[J]. Concurrency and Computation: Practice and Experience,2020,32(7).[82]Li Gao,Peng Lin,Peng Chen,Rui‐Zhi Gao,Hong Yang,Yun He,Jia‐Bo Chen,Yi ‐Ge Luo,Qiong‐Qian Xu,Song‐Wu Liang,Jin‐Han Gu,Zhi‐Guang Huang,Yi‐Wu Dang,Gang Chen. A novel risk signature that combines 10 long noncoding RNAs to predict neuroblastoma prognosis[J]. Journal of Cellular Physiology,2020,235(4).[83]Julia Krzykalla,Axel Benner,Annette Kopp‐Schneider. Exploratory identification of predictive biomarkers in randomized trials with normal endpoints[J]. Statistics in Medicine,2020,39(7).[84]Jianye Ching,Kok-Kwang Phoon. Measuring Similarity between Site-Specific Data and Records from Other Sites[J]. ASCE-ASME Journal of Risk and Uncertainty in Engineering Systems, Part A: Civil Engineering,2020,6(2).[85]Anne Kelly Knowles,Justus Hillebrand,Paul B. Jaskot,Anika Walke. Integrative, Interdisciplinary Database Design for the Spatial Humanities: the Case of the Holocaust Ghettos Project[J]. International Journal of Humanities and Arts Computing,2020,14(1-2).[86]Sheng-Feng Sung,Pei-Ju Lee,Cheng-Yang Hsieh,Wan-Lun Zheng. Medication Use and the Risk of Newly Diagnosed Diabetes in Patients with Epilepsy: A Data Mining Application on a Healthcare Database[J]. Journal of Organizational and End User Computing (JOEUC),2020,32(2).[87]Rashkovits Rami,Lavy Ilana. Students' Difficulties in Identifying the Use of Ternary Relationships in Data Modeling[J]. International Journal of Information and Communication Technology Education (IJICTE,2020,16(2).[88]Yusuf Akhtar,Dipti Prasad Mukherjee. Context-based ensemble classification for the detection of architectural distortion in a digitised mammogram[J]. IET Image Processing,2020,14(4).[89]Gurpreet Kaur,Sukhwinder Singh,Renu Vig. Medical fusion framework using discrete fractional wavelets and non-subsampled directional filter banks[J]. IET Image Processing,2020,14(4).[90]Qian Liu,Bo Jiang,Jia-lei Zhang,Peng Gao,Zhi-jian Xia. Semi-supervised uncorrelated dictionary learning for colour face recognition[J]. IET Computer Vision,2020,14(3).[91]Yipo Huang,Leida Li,Yu Zhou,Bo Hu. No-reference quality assessment for live broadcasting videos in temporal and spatial domains[J]. IET Image Processing,2020,14(4).[92]Panetta Karen,Wan Qianwen,Agaian Sos,Rajeev Srijith,Kamath Shreyas,Rajendran Rahul,Rao Shishir Paramathma,Kaszowska Aleksandra,Taylor Holly A,Samani Arash,Yuan Xin. A Comprehensive Database for Benchmarking Imaging Systems.[J]. IEEE transactions on pattern analysis and machine intelligence,2020,42(3).[93]Rahnev Dobromir,Desender Kobe,Lee Alan L F,Adler William T,Aguilar-Lleyda David,Akdo?an Ba?ak,Arbuzova Polina,Atlas Lauren Y,Balc? Fuat,Bang Ji Won,Bègue Indrit,Birney Damian P,Brady Timothy F,Calder-Travis Joshua,Chetverikov Andrey,Clark Torin K,Davranche Karen,Denison Rachel N,Dildine Troy C,Double Kit S,Duyan Yaln A,Faivre Nathan,Fallow Kaitlyn,Filevich Elisa,Gajdos Thibault,Gallagher Regan M,de Gardelle Vincent,Gherman Sabina,Haddara Nadia,Hainguerlot Marine,Hsu Tzu-Yu,Hu Xiao,Iturrate I?aki,Jaquiery Matt,Kantner Justin,Koculak Marcin,Konishi Mahiko,Ko? Christina,Kvam Peter D,Kwok Sze Chai,Lebreton Ma?l,Lempert Karolina M,Ming Lo Chien,Luo Liang,Maniscalco Brian,Martin Antonio,Massoni Sébastien,Matthews Julian,Mazancieux Audrey,Merfeld Daniel M,O'Hora Denis,Palser Eleanor R,Paulewicz Borys?aw,Pereira Michael,Peters Caroline,Philiastides Marios G,Pfuhl Gerit,Prieto Fernanda,Rausch Manuel,Recht Samuel,Reyes Gabriel,Rouault Marion,Sackur Jér?me,Sadeghi Saeedeh,Samaha Jason,Seow Tricia X F,Shekhar Medha,Sherman Maxine T,Siedlecka Marta,Skóra Zuzanna,Song Chen,Soto David,Sun Sai,van Boxtel Jeroen J A,Wang Shuo,Weidemann Christoph T,Weindel Gabriel,WierzchońMicha?,Xu Xinming,Ye Qun,Yeon Jiwon,Zou Futing,Zylberberg Ariel. The Confidence Database.[J]. Nature human behaviour,2020,4(3).[94]Taipalus Toni. The Effects of Database Complexity on SQL Query Formulation[J]. Journal of Systems and Software,2020(prepublish).[95]. Information Technology; Investigators from Deakin University Target Information Technology (Conjunctive query pattern structures: A relational database model for Formal Concept Analysis)[J]. Computer Technology Journal,2020.[96]. Machine Learning; Findings from Rensselaer Polytechnic Institute Broaden Understanding of Machine Learning (Self Healing Databases for Predictive Risk Analytics In Safety-critical Systems)[J]. Computer Technology Journal,2020.[97]. Science - Library Science; Investigators from Cumhuriyet University Release New Data on Library Science (Scholarly databases under scrutiny)[J]. Computer Technology Journal,2020.[98]. Information Technology; Investigators from Faculty of Computer Science and Engineering Release New Data on Information Technology (FGSA for optimal quality of service based transaction in real-time database systems under different workload condition)[J]. Computer Technology Journal,2020.[99]Muhammad Aqib Javed,M.A. Naveed,Azam Hussain,S. Hussain. Integrated data acquisition, storage and retrieval for glass spherical tokamak (GLAST)[J]. Fusion Engineering and Design,2020,152.[100]Vinay M.S.,Jayant R. Haritsa. Operator implementation of Result Set Dependent KWS scoring functions[J]. Information Systems,2020,89.[101]. Capital One Services LLC; Patent Issued for Computer-Based Systems Configured For Managing Authentication Challenge Questions In A Database And Methods Of Use (USPTO 10,572,653)[J]. Journal of Robotics & Machine Learning,2020.[102]Ikawa Fusao,Michihata Nobuaki. In Reply to Letter to the Editor Regarding "Treatment Risk for Elderly Patients with Unruptured Cerebral Aneurysm from a Nationwide Database in Japan".[J]. World neurosurgery,2020,135.[103]Chen Wei,You Chao. Letter to the Editor Regarding "Treatment Risk for Elderly Patients with Unruptured Cerebral Aneurysm from a Nationwide Database in Japan".[J]. World neurosurgery,2020,135.[104]Zhitao Xiao,Lei Pei,Lei Geng,Ying Sun,Fang Zhang,Jun Wu. Surface Parameter Measurement of Braided Composite Preform Based on Faster R-CNN[J]. Fibers and Polymers,2020,21(3).[105]Xiaoyu Cui,Ruifan Cai,Xiangjun Tang,Zhigang Deng,Xiaogang Jin. Sketch‐based shape‐constrained fireworks simulation in head‐mounted virtual reality[J]. Computer Animation and Virtual Worlds,2020,31(2).[106]Klaus B?hm,Tibor Kubjatko,Daniel Paula,Hans-Georg Schweiger. New developments on EDR (Event Data Recorder) for automated vehicles[J]. Open Engineering,2020,10(1).[107]Ming Li,Ruizhi Chen,Xuan Liao,Bingxuan Guo,Weilong Zhang,Ge Guo. A Precise Indoor Visual Positioning Approach Using a Built Image Feature Database and Single User Image from Smartphone Cameras[J]. Remote Sensing,2020,12(5).[108]Matthew Grewe,Phillip Sexton,David Dellenbach. Use Risk‐Based Asset Prioritization to Develop Accurate Capital Budgets[J]. Opflow,2020,46(3).[109]Jose R. Salvador,D. Mu?oz de la Pe?a,D.R. Ramirez,T. Alamo. Predictive control of a water distribution system based on process historian data[J]. Optimal Control Applications and Methods,2020,41(2).[110]Esmaeil Nourani,Vahideh Reshadat. Association extraction from biomedicalliterature based on representation and transfer learning[J]. Journal of Theoretical Biology,2020,488.[111]Ikram Saima,Ahmad Jamshaid,Durdagi Serdar. Screening of FDA approved drugs for finding potential inhibitors against Granzyme B as a potent drug-repurposing target.[J]. Journal of molecular graphics & modelling,2020,95.[112]Keiron O’Shea,Biswapriya B. Misra. Software tools, databases and resources in metabolomics: updates from 2018 to 2019[J]. Metabolomics,2020,16(D1).[113]. Information Technology; Researchers from Virginia Polytechnic Institute and State University (Virginia Tech) Describe Findings in Information Technology (A database for global soil health assessment)[J]. Energy & Ecology,2020.[114]Moosa Johra Muhammad,Guan Shenheng,Moran Michael F,Ma Bin. Repeat-Preserving Decoy Database for False Discovery Rate Estimation in Peptide Identification.[J]. Journal of proteome research,2020,19(3).[115]Huttunen Janne M J,K?rkk?inen Leo,Honkala Mikko,Lindholm Harri. Deep learning for prediction of cardiac indices from photoplethysmographic waveform: A virtual database approach.[J]. International journal for numerical methods in biomedical engineering,2020,36(3).[116]Kunxia Wang,Guoxin Su,Li Liu,Shu Wang. Wavelet packet analysis for speaker-independent emotion recognition[J]. Neurocomputing,2020.[117]Fusao Ikawa,Nobuaki Michihata. In Reply to Letter to the Editor Regarding “Treatment Risk for Elderly Patients with Unruptured Cerebral Aneurysm from a Nationwide Database in Japan”[J]. World Neurosurgery,2020,135.[118]Wei Chen,Chao You. Letter to the Editor Regarding “Treatment Risk for Elderly Patients with Unruptured Cerebral Aneurysm from a Nationwide Database in Japan”[J]. World Neurosurgery,2020,135.[119]Lindsey A. Parsons,Jonathan A. Jenks,Andrew J. Gregory. Accuracy Assessment of National Land Cover Database Shrubland Products on the Sagebrush Steppe Fringe[J]. Rangeland Ecology & Management,2020,73(2).[120]Jing Hua,Yilu Xu,Jianjun Tang,Jizhong Liu,Jihao Zhang. ECG heartbeat classification in compressive domain for wearable devices[J]. Journal of Systems Architecture,2020,104.以上就是关于数据库英文参考文献的全部内容,希望看完后对你有所启发。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
外文翻译原文Computer programming data structure is an important theoretical basis for the design, it is not only the core curriculum of computer disciplines, and has become a popular elective course other Polytechnic professional, so studied this course well and studied computer are closely related.一、the concept of data structureComputer data structure is the foundation of science and technology professional classes, is the essential core curriculum. All computer system software and application software to use various types of data structures. Therefore, if we want to make better use of computers to solve practical problems, only to several computer programming languages are difficult to cope with the many complex issues. To the effective use of computers, give full play to computer performance, but also must learn and master relevant knowledge of data structure. A solid foundation of "data structure" for learning other computer professional courses, such as operating systems, translation theory, database management systems, software engineering, artificial intelligence, etc. are very useful.二、why should learn from data structure?In the early development of computers, the use of computer designed primarily to deal with terms. When we use the computer to solve a specific problem, the following general needs through several steps : the first is a specific problem of appropriate abstract mathematical models, and then design or choose a mathematical model of the algorithm,the final procedures for debugging, testing, until they have the ultimate answer.Since then the object is INTEGER, REAL, BOOLEAN, the procedures of the main designers of energy is focused on programming skills, without attention to the data structure. With the expansion of computer applications and development of software and hardware, the issue of non-terms increasing importance. According to statistics, Now dealing with the issue of non-occupancy of more than 90% of the machine time. Such issues involve more complex data structure, the relationships between data elements generally can not be described by mathematical formula. Therefore, the key to solvingsuch problems is no longer mathematical analysis and calculations, but to devise appropriate data structure, can effectively address the problem.Description of the terms of such non-mathematical model is not a mathematical equation, but such as tables, trees, such as map data structure. Therefore, it can be said that data structure courses primarily designed to study the issue of non-value calculation procedures as a computer operations and the relationship between objects and their operating disciplines.The purpose of the study is to understand the structure of data for computer processing of the identity object to the practical problems involved in dealing with that subject at the computer out and deal with them. At the same time, through training algorithms to improve the thinking ability of students through procedures designed to promote student skills integrated applications and professional qualities.三、the concepts and terminologySystematic study of knowledge in the data structure before some of the basic concepts and terminology to give a precise meaning.Data (Data) is the information carrier, it could be computer identification, storage and processing. It is the computer processing of raw materials, a variety of data processing applications. Computer science, computer processing is the so-called data objects, which can be numerical data can be non - numerical data. Numerical data are integer, the actual number or plural, mainly for engineering computing, scientific computing and commercial processing; Non - numerical data, including characters, text, graphics, images, voice and so on.Data elements (Data Element) is the basic unit of data. In different conditions, data elements can be called elements, nodes, the peak, recording. For example, students information retrieval system table information, a record high, 8 Queen's issue of a state tree, teaching programming issues such as a peak, known as a data element. Sometimes, a data from a number of data elements (Data Item), for example, the student information management system students each data element table is a student record. It includes students of the school, name, sex, nationality, date of birth, performance data items. These data items can be divided into two types : one called early such as studentgender, origin, etc., these data were no longer divided in data processing, the smallest units; Another called portfolio, the performance of students who, it can be divided into mathematics, physics, chemistry and other smaller items. Normally, in addressing the question of the practical application of each student is recorded as a basic unit for a visit and treatment.Data objects (Data Object) or data element type (Data Element Class) is the nature of the data elements with the same pool. In a specific issue, the data elements have the same nature (not necessarily equal value elements), belonging to the same data objects (data element type), the data element is an example of such data elements. For example, traffic information systems in the transportation network, is a culmination of all the data elements category, peak a and B each represent an urban middle is the data elements of the two types of examples of the value of their data elements a and B respectively.Data structure (Data Structure) refers to the mutual relationship that exists between one or more data elements together. In any case, between data elements will not be isolated in between them exist in one way or another, such as the relationship between the data element structure. According to the data elements of the relationship between different characteristics, usually have the following four basic categories of the structure :1 assembly structures. In the assembly structure, the relationship between data elements is "belonging to the same pool." Assembly elements relations is a very loose structure.2 linear structures. The structure of the data elements exist between one-to-one relationship.3 tree structure. The structure of the data elements exist between hierarchical relationship.4graphics structure. The structure of the data elements of the relationship that existed between Duoduiduo, graphics structure also known as network structure.C++Builder programming experience一、Database programmingAnd the use of Delphi, Borland C++Builder BDE (Borland Database Engine) database interface, in particular its use BDE Administrator unified management databasealias, the database operation has nothing to do with the location of the database documents, thus enabling database development easier operation. But in a database application procedures at the same time we have to "release" BDE, the database for some simple procedures may BDE than our own design procedures big, but as the use of BDE InstallShield, add database alias is likely allocation failure. Therefore, we can use the following methods : still in the design stage procedure using BDE alias management database for debugging, but in procedures substantially (as in the main Chuangti OnCreate event processing function) to Table components DatabaseName attributes, such as the use of similar phrases as follows : Table1->DatabaseName = ExtractFilePath (Application->ExeName); Or Table1->DatabaseName = ExtractFilePath(Application->ExeName+ "DB");Thus, no impact on the debugging phase, will be issued if the application procedures Table1 document on the use of databases or their current catalogue "DB" virus, database procedures can be normal operation. You can even be a database to catalogue the documents in the form of character string Register (installed in the installation process), then the procedure in the acquisition of substantially from the catalogue of payrolls, Fuzhi DatabaseName attribute to be. Anyway, you do not need to install relatively large BDE forced users.二、the Registry visitAs in the design process we often required 9x/NT Windows Registry information visit, such as retrieval of information procedures, preservation of information. Register write a subroutine to visit necessary. When the Register to visit, the library will be directly available without always some duplication operation. The following can be used to access cosmetic Licheng, the character string type Jianzhi, and the retrieval of failure to return default value Default.#include < Registry.hpp >int ReadIntFromReg(HKEY Root, AnsiString Key, AnsiString KeyName, int Default) { int KeyValue;TRegistry *Registry = new TRegistry(); Registry->RootKey = Root;Registry->OpenKey(Key, false);try {KeyValue = Registry->ReadInteger(KeyName);}catch(...) {KeyValue = Default;}delete Registry; return KeyValue;}void SaveIntToReg(HKEY Root, AnsiString Key, AnsiString KeyName, int KeyValue) {TRegistry *Registry = new TRegistry(); Registry->RootKey = Root;Registry->OpenKey(Key, true);Registry->WriteInteger(KeyName, KeyValue); delete Registry;}char *ReadStringFromReg(HKEY Root, AnsiString Key, AnsiString KeyName, char *Default) {AnsiString KeyValue;TRegistry *Registry = new TRegistry(); Registry->RootKey = Root;Registry->OpenKey(Key, false);try {KeyValue = Registry->ReadString(KeyName);}catch(...) {KeyValue = (AnsiString)Default;}delete Registry;return KeyValue.c_str();}void SaveStringToReg(HKEY Root, AnsiString Key,AnsiString KeyName, char *KeyValue) {TRegistry *Registry = new TRegistry();Registry->RootKey = Root;Registry->OpenKey(Key, true);Registry->WriteString(KeyName, (AnsiString)KeyValue);delete Registry;}We may use the following access methods (to Windows wallpaper documents) : AnsiString WallPaperFileName = ReadStringFromReg(HKEY_CURRENT_USER, "\\Control Panel\\Desktop", "Wallpaper", "");三、show / hide icons task columnStandard Windows applications generally operating in the mission mandate column on the chart shows, users can directly use the mouse clicking column logo for the mission task cut over, but some applications do not use task column signs, such as the typical Office tools, There are also procedures that can be shown or hidden customization tasks column icon, such as Winamp. We can do the procedure, as long as access Windows SetWindowLong function can drive, as follows :// hidden task column chart :SetWindowLong (Application->Handle.GWL_EXSTYLE, WS_EX_TOOLWINDOW);// task column shows signs :SetWindowLong (Application->Handle. GWL_EXSTYLE, WS_EX_APPWINDOW);四、the establishment of a simple "on" windowA complete Windows applications typically contain a "on the" window to show version information. We customized a dialog box as usual "on the" window of the "on" free customized window, indicates that more information, even including super links. If only show simple version information,Windows ShellAbout function shelf items have sufficient, following this line of code can be "on" Duihuakuang and is Windows standard "on the" Duihuakuang and procedures may show signs such as the use of resources and systems.ShellAbout (Handle, ( "on" +Application->Title+ "#"). C_str () ( "\n"+Application->Title+ "V1.0\n\n" + " 夏登城版权所有!"). C_str ()Application->Icon->Handle);五、the two methods to choice catalogueIn our applications, allowing users to choose the regular catalogue, such as software manufacturers, users choose catalogue. This involves catalogue option, we may use the following methods for users to choose one of the catalogue :1, use SHBrowseForFolder and SHGetPathFromIDList function; Company affirms its function as follows :WINSHELLAPI LPITEMIDLIST WINAPI SHBrowseForFolder(LPBROWSEINFO lpbi); WINSHELLAPI BOOL WINAPI SHGetPathFromIDList(LPCITEMIDLIST pidl, LPSTR pszPath); LPBROWSEINFO 和LPITEMIDLIST structure refer Win32 files. This method of selecting catalogues available Windows desktop all available inventory, including networks of other computers sharing catalogue neighbors, but not the new catalogue. Li Cheng allows users to choose the following directory, the directory of choice Licheng return at all trails character string.#include < shlobj.h >char *GetDir(char *DisplayName, HWND Owner) {char dir[MAX_PATH] = "";BROWSEINFO *bi = new BROWSEINFO;bi->hwndOwner = Owner;bi->pidlRoot = NULL;bi->pszDisplayName = NULL;bi->lpszTitle = DisplayName;bi->ulFlags = BIF_RETURNONLYFSDIRS;bi->lpfn = NULL;bi->lParam = NULL;bi->iImage = 0;ITEMIDLIST *il = SHBrowseForFolder(bi);if(il!=NULL) {SHGetPathFromIDList(il, dir);}delete bi;return dir;}We can use the following list to be chosen from :AnsiString at Dir = (AnsiString) GetDir ( "Please select catalogue :" Handle);2, the use of SelectDirectory function. C++Builder the function SelectDirectory achievable catalogue of options, which showed that similar "open" / "preserve" Duihuakuang, but its advantage is to use / non-use keyboard input catalogue members, and allow the creation of new directories. Its original definition as follows : Extern package bool __fastcall SelectDirectory ( AnsiString &Directory, TSelectDirOpts Options, 103-116 HelpCtx);Licheng SelectDir allow you to choose the following directory :#include < FileCtrl.hpp >AnsiString SelectDir(AnsiString Dir) {if(SelectDirectory(Dir, TSelectDirOpts()<< sdAllowCreate << sdPerformCreate << sdPrompt,0))return Dir; elsereturn "";}for the following redeployed to the users choice catalogue :AnsiString SelectedDir = SelectDir ( "C:\\My Documents");外文翻译译文数据结构是计算机程序设计的重要理论设计基础,它不仅是计算机学科的核心课程,而且已成为其他理工专业的热门选修课,所以学好这门课程是与学好计算机专业是息息相关的。