一种Web数据源索引方法(IJITCS-V6-N9-7)
基于Web的搜索引擎算法的研究
定义 二: 息素 更新规律及 公式: 信
m
检索服 务, 是快速 查找检索信 息的一种 网络 工具 。 蚁群算法 是由意大利 学者 M. oio等人在 2 Dr g O世 纪 9 O 年代 初首 先提 出来 的, 它是 继模 拟退 火算 法 、 传 算法 、 遗 禁
3 .用二叉树 算法存储信息 网络 中信 息是海量 的, 那么 怎么存储是个值得 我们思 索
的 问题 。 以树状 结构组织存储 , 有助于 用户提高查询率 , 先把
相 同或相似 内容的资源组织 起来 , 用户只要查到树 根就可 以
用 遍历算法遍 历整个树 , 进而查询 到整个信息 。这种 方法在 很 大程度 上提 高了查询 效率 , 省 了用 户 的时间 , 节 使用 户有 个 更好的交互体验 。
其中 =
,C 为经过路径 ( , 所 需的花费 。 【 p ij ) o和
两 个参数 ,分别 用来 控制信 息素和 路径长 度 的相 对重 要程 度 。 ik ct 是第 k个蚂蚁下 一步可 以选择 的城市 的集合 。 i  ̄ y cy t=
{,, ,…n1-r gn , 实际蚁 群算 法不 同 , no ̄ 0 12 -} e og - n 。与 r gn , e
放 入 缓 冲 区
Se : tp3
Whl(evr ] F l i S re[  ̄ u1 e s )
下
{
F rk lk - k + o(= ;< m;- 1 卜 Fn Netev r id xSre0;/根据公 式找下一个服务器 / 现 , 入数据 中 输 的 F AG字符 L /表示空树 , 回二叉树 的根指针 / 返 sa f %c , c 1 /读入一个字 符数据 cn( ” & h ;/ ” i(h— F A 1 fc L G
信息检索中的Web搜索引擎算法探索
信息检索中的Web搜索引擎算法探索随着互联网的迅速发展,人们在日常生活中越来越依赖搜索引擎来获取所需的信息。
对于搜索引擎来说,它的算法是其核心之一,直接决定了搜索结果的质量和相关性。
因此,对于信息检索中的Web搜索引擎算法进行探索就显得非常重要。
搜索引擎算法由数学、信息学和计算机科学的原理和方法组成。
这些算法旨在将用户的查询与索引中的网页进行匹配,并以某种方式将最相关的结果展示给用户。
本文将探讨几种常见的搜索引擎算法,以及它们如何工作。
一种常见的搜索引擎算法是基于关键词的搜索算法。
该算法通过识别用户查询中的关键词,并将其与网页的关键词匹配进行比较。
它使用一系列规则和权重来确定网页的相关性,然后按照相关性的程度对搜索结果进行排序。
这种算法的优势在于简单有效,但它可能无法解决查询含义模糊或多义性的问题。
为了解决查询含义模糊或多义性的问题,出现了基于语义的搜索算法。
这种算法使用自然语言处理和语义分析技术,以理解用户查询的语义,而不仅仅是关键词。
它通过理解用户查询的意图,将其与网页的内容进行匹配,并根据相关性对搜索结果进行排序。
基于语义的搜索算法可以更好地满足用户的需求,但它要求对自然语言的理解和语义分析有更高的要求。
另一种常见的搜索引擎算法是PageRank算法。
这个算法由Google创始人之一拉里·佩奇提出,并成为了Google搜索引擎的核心算法之一。
PageRank算法通过分析网页之间的链接关系,以及链接的权重和质量,来评估网页的重要性和相关性。
它认为一个被其他重要网页链接的网页也会具有一定的重要性。
因此,PageRank算法将重要的网页排在搜索结果的前面。
这个算法的优势在于较好地解决了垃圾网页和链接 farm 的问题,增加了搜索结果的质量和可靠性。
除了PageRank算法,还有一种被广泛应用的算法是TF-IDF算法。
这个算法用于评估文档中的单词的重要性和相关性。
TF-IDF的全称是词频-逆文档频率,它通过计算一个词在文档中的频率和在整个文档集合中的频率,来确定该词在文档中的重要性。
一种大数据索引方法及系统的制作方法
一种大数据索引方法及系统的制作方法专利名称:一种大数据索引方法及系统的制作方法技术领域:本发明涉及大数据背景下数据库索引技术领域,尤其涉及一种大数据索引方法及系统。
背景技术:在信息化的过程中,个人和企业的数据量都在迅速增长。
由于社交网络,电子商务和物联网技术的兴起,各种移动终端,传感器和传统设备时刻在产生着各种非结构化数据。
2011年麦肯锡公司全球研究所发表研究报告指出,目前数据已渗透到各行业和业务职能领域,逐渐成为人类社会中重要的生产要素。
截止2011年底,全球的数据总量达到1.9ZB(IZB=I X IO12GB),到2015年将达到8ZB,到2020年将达到大约35ZB。
面对数量巨大并且增长迅猛的数据,高效的管理和分析数据已经成为当前信息管理最关注的问题之一。
针对这一问题,人们提出了 “大数据”的概念。
与传统的海量数据不同,大数据除了有巨大的数据规模外,还有着复杂的数据类型和数据关联度。
IBM将大数据的特点概括为:海量化(Volume),快速化(Velocity)以及多样化(Variety)。
因此,传统的方法和技术已不适用于大数据的管理和分析。
支持时间连续性的大数据存储是一个非常复杂的问题。
为实现对数据的实时分析,必须实现数据的高效插入,并对查询请求做出实时响应。
为达到上述目的,需要对持续到来的数据流建立动态索引。
在经过实时分析后,暂存的数据需要被迁移到数据仓库进行进一步的数据分析。
因此,所建立的索引还需要支持高效率的批量删除操作。
LSM-Tree是一种支持数据库写优化的高效索引框架,在当前主流的NoSQL数据库系统中得到了广泛应用。
为实现上述目的,本发明提出了一种大数据索引方法及系统。
发明内容(一)技术问题大数据在经过实时分析后,暂存的数据需要被迁移到数据仓库进行进一步的数据分析,而根据现有技术建立的索引不支持高效率的批量删除操作,降低了数据的插入以及查询效率,本发明提供一种大数据索引方法及系统以实现数据的批量删除操作,提高大数据的插入及查询效率。
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
I.J. Information Technology and Computer Science, 2014, 09, 52-58Published Online August 2014 in MECS (/)DOI: 10.5815/ijitcs.2014.09.07An Approach for Indexing Web Data SourcesSaidi ImeneLITIO laboratory, University of Oran, BP 1524, El-M'Naouer, 31000, Oran, AlgeriaEmail: saidi.imene@univ-oran.dzNait Bahloul SafiaLITIO laboratory, University of Oran, BP 1524, El-M'Naouer, 31000, Oran, AlgeriaEmail: nait-bahloul.safia@univ-oran.dzAbstract— Web information sources such as forums, blogs, and news articles are becoming increasingly large and diverse. Even if advances in technology are helping to improve techniques for dealing with the large amounts of the generated data, such data sources are heterogeneous in structure (semi structured or unstructured sources) and nature (texts or images). Implementation of software solutions is then necessary to prepare data and access these sources in a homogenous way. In this paper we present an approach for indexing heterogeneous data sources. Our objective is to offer techniques for efficient indexing of web sources by storing only the necessary information. We propose automatic indexing for semi structured or unstructured sources (e.g., xml files, html files) and annotation for other sources (e.g., images, videos that exist within a page). We present our algorithms of indexing and propose the use of MapReduce model to build a scalable inverted index. Experiments on a real-world corpus show that our approach achieves a good performance.Index Terms—Information Retrieval, Indexing Techniques, Data Mining, MapreduceI. I NTRODUCTIONOne of the main difficulties encountered by the users of the web is heterogeneity of information sources and their diversity. Heterogeneity of sources may come from the format or the structure (structured sources: relational databases, semi structured sources: XML documents, or unstructured: texts). One of the objectives of the future web, which is called semantic web, is to provide mechanisms to access heterogeneous data sources in a standardized way. The systems known as mediation systems are then very useful, in the presence of heterogeneous data, because they give the impression of using homogeneous system.The main objective of this paper is to prepare data to be used in a global architecture for querying heterogeneous data sources of the web. We then index data sources syntactically and semantically and enable the exploitation of the generated indexes in order to use them in a global system.This paper is organized as follows: In the next section, we present some related works to our approach. Section 3, is reserved for the description of our approach in order to show the different steps for the creation of indexes. In section 4, we present our algorithms and in the section 5, we present experiments. A conclusion is presented to synthesize this work and a set of prospects are proposed in the end of this paper.II. R ELATED W ORKSThe scientific community has long been interested in indexing as a critical step of pretreatment of data and an important process in information retrieval systems. The implementation of software solutions which improve performance of the systems has always been necessary. The new models such as MapReduce handle quantities of data that are becoming increasingly large. To improve execution time and data processing, several implementations have been proposed, among these solutions, we can find Hadoop which is a free, Java-based programming framework that supports the processing of large data sets in a distributed computing environment [1]. Some research has been implemented to evaluate performance of the MapReduce model [2] [3].Several studies have been published about indexing strategies and MapReduce jobs. A good indexing strategy will allow fast access to desired data and a short time for the indexing process. Among the existing works on the problem of indexing, some approaches [4] [5] [6] analyze the query workload and decide which attributes to index. Another research direction [7] is the use of distributed key-value storage system as a generic infrastructure to accommodate structured and semi-structured data.Lin et al. [8] propose to build a full-text inverted index at the level of the blocks to optimize regular-expression queries over text fields.In our approach, we propose to index the concepts extracted from a semantic indexing phase after a filtering phase (syntactic phase). The creation of inverted index by storing only the concepts and the use of MapReduce model makes the index scalable.III. O UR A PPROACHIn this section, we present our approach of indexing heterogeneous data sources, i.e the steps of the construction of indexes, including how we adapt the classical techniques of indexing.First of all, we present our approach and remind some notions of indexing.Depending on the type of the source, two treatments are possible (Fig. 1):Fig. 1. Overall Scheme of our ApproachFig. 2. Syntactic indexingAutomatic indexing: concerns the semi structured or unstructured sources, i.e: textual documents, ex: text files.Annotation:(tagging) concerns annotations and using extracted concepts for the creation of inverted index. Annotation is for non-textual sources, e.g.video, image.A. Automatic indexingIn this type of indexation, we had been inspired by the work of Baziz et al [9] [10] [11].The construction of indexes in the automatic indexing is done in two main phases: the syntactic indexation and semantic indexation (Fig. 1).1. Syntactic IndexationThe objective of the syntactic indexation is to extract the terms from the documents and store them with their number of occurrences in a physical index. This type of indexing has been presented in several works [12] [13]. a-Extraction of wordsIn this first phase of construction of indexes, we extract the significant terms of the document (not a stop word). b-Calculation of the occurrence of the word:For the significant words, the frequency of occurrences must be calculated. Every time a word is encountered in the document, we increment its frequency (Term Frequency).The result of this phase is the file "Index_Syn": Syntactic Index related to each document. It is constituted as follows:For each document, you can find all the words which belong to the document with their term frequencies.An example of this type of indexing is given in the following Figure (Fig. 2).Each document of the corpus is treated and significant terms are extracted with their TF (Fig. 2).2. Semantic IndexingThis phase accepts as input information from the syntactic indexation.Semantic indexing consists of the following steps. In Index_Syn (created from the syntactic indexation) and for each document, we do:a-Detection of conceptsA term is said concept if it belongs to at least one entry in the ontology. WordNet [21] is used to do the matching. The detection of concepts is done by taking the words one by one and projecting them on the ontology to detect the concepts. Concepts are identified from WordNet and marked in a document.For example: "president" is a concept because it matches an entry in the ontology which is of the type “person”. This step is the projection of the document on the ontology. b-Expansion of concepts using the ontologyFor each detected concept, a specific treatment is done. It is to detect links between the different concepts and to link them together using WordNet.Concepts are then extended by their synonyms and hyponyms that exist in the document.A list of synonyms and meaning will be awarded to each concept. Creation of links and relations between concepts forms some kind of a network for each concept. This step is the projection of the ontology on the document.c-Construction of the semantic networkIn the previous two steps, we used ontology to represent the content of the documents in the form of:- Concepts.- Relations between concepts.We had as results and for each concept a number of terms which are associated with. We call this combination the semantic network of the term. Definition:The set of terms Sn = {t1… t p} forms a network with a term t of a document doc, means that: - Sn is formed by the union of the elements e i / e i are terms in relation with the term t, where the relation is = {synonymy, derivation, same familly, ... }.- AND t (e i) Sn, t (e i) doc.d-Weighting of conceptsThere are several approaches [19] [20] to weight the significant terms of a document or a query. Many of them are based on the factors T f and idf which allow the consideration of the local and global weights of a term. The measure tf*idf allows to approximate the representativeness of a term in a document.In our approach, we will use a variant of this measure, which is the Cf*idf.-Calculation of CFThe CF is not the frequency of the term but that of the concept. To be calculated, we must add the tf of the initial term to all tf of the terms that have been associated in the same document in the previous phase, i.e the tf of the terms of the semantic network, as shown in (1)..network)(semantictf+(t)tf=Cf (1)Such as t is the current term and tf (semantic network) are all terms of the semantic network.-Calculation of idfIn this work, we distinguished the frequency of occurrence of a term in a document (term frequency, tf) which was already calculated in the Syntactic index “index_syn” and the frequency of occurrence for the same term in the entire collection (inverse document frequency, idf ) , as shown in (2).Idf = (log ((total documents) / (number of documents with the term)).(2)Document_ID => {(term1, tf1)… (Term n, tf n)} / n: Number of Significant words of the document.Idf = (log ((total documents) / (number of documents with the term))1.e-Updating of index_SynIndex_Syn is updated by taking into account the concepts with an entry in the ontology (the detected concepts).The terms which belong to their network will be taken as the concepts with the same weight (which is the sum)In this phase, we enrich the index_Syn with the terms of the created network and we remove the terms that are not concepts.The result will be, a semantic index (Index_Sem) which is composed as follows:i, j: Weighting of the concept (Cf.idf already calculated). k: 1...m / m: the size of semantic network of the concept.B. AnnotationFor non-textual data sources, we propose the use of semantic annotation tools. Several works present tools for image annotation [15] [16] and video annotation [17] [18]. This part of the work will be presented in a future work. Concepts are related to sources and are then used to create the inverted index.Creation of the inverted indexThis step is necessary to allow exploitation of the indexes.The creation of inverted index is to correspond to all concepts, the documents that contain them with their weights.IV. A LGORITHMSIn this section, we propose efficient algorithms of the different phases of our approach. We also introduce the use of MapReduce model to create inverted index. Syntactic indexing phase (removes stop words and calculates Tf) is summarized by the following algorithm (Algorithm 1):The previous algorithm (Algorithm 1) extracts the terms, calculate the Tf (term frequency) of each term andgenerate a syntactic index of the documents.Semantic indexing phase is summarized by the following algorithm (Algorithm 2):1 /wiki/Tf-idfThe previous algorithm (Algorithm 2) extracts the concepts that will be considered in the semantic index (by projecting syntactic indexes of the documents on an ontology of the considered domain), and extends them by creating the semantic network and weighting the concepts. This algorithm generates semantic indexes of the documents.Algorithm 1: Syntactic indexingRequire: documents: corpusOutput: Index_Syn: syntactic indexFor each Doc_in (documents) do/* for all documents of the corpus, current is Doc */;Create (Doc.index_Syn);/* create index_Syn of each document */;For each line_in (Doc) faire/*for all the lines of document, current isline*/;For each word_in (line);/*for all the words of the line*/;If stop_word (word) = False then/* If the word is not a stop word */;If Doc.Index_Syn.Contains (word) = False then/* Index_Syn does not contain the word */;Doc.Index_Syn.Add (word, 1);/*add the word to index, tf =1 term frequency*/;ELSE/*Index_Syn contain the word: increment tf*/;Doc.Index_Syn.IncrementTf (word.TF);END If;END If;END For;ENDFor;ENDFor;END{{(concept_1k;i), (concept1_k+1;i)}, …, {(concept_n k;j), (concept_n k+1;j), …}}Concept 1 --> <Documents and weighting> ...Concept n --> <Documents and weighting> Algorithm 2: Semantic IndexingRequire: Index_Syn: indexOutput: Index Sem: index semanticFor each index_Syn do/*for all indexes Index_Syn of documents*/;For each term in index_Syn do/*for all terms of Index_Syn*/;If term.isConcept () then/*Term corresponds to an input ofontology*/;/*Projection on ontology*/Expansion _with_ontology ();/* detect links between concepts and extendthem*/;Construction_semantic_network ();/*concept and the terms associated with*/;Ponderation_Concept ();/*concept is weighted according to Cf.idf */;Else /*Term is ignored*/;END If;Update_Index_Syn ();/*update Index_Syn -> only concepts are considered*/; END For;END For;ENDThe following algorithm (Algorithm 3) creates an inverted index from semantic indexes of the documents:The previous algorithm (Algorithm 3) creates from all the semantic indexes of the documents, an inverted index (Index_Sem_Inv) that matches all the concepts with their documents and weightings. MapReduce AlgorithmMap-Reduce framework [22] is a popular way to deal with a large scale of data. In this paper, we focus on indexing which is a crucial step made offline. The use MapReduce framework grantee the performance over big data sets.To create an inverted index, we use the pseudo code given in [14] and we adapt it to our approach.The creation of the inverted index (Index_Sem_Inv) of our approach is as follows:The Map function calculates all the cf (Concept frequency) and emits the concept with a Posting List containing the identifier of the document and cf of theconcept in this document. The Reducer groups and ranks the positing lists of each concept, i.e, all documents that contain the concept and the corresponding cf. Reducer emits the concept and its Postindg list.V. E XPERIMENTSIn order to analyze our approach and study its performance, we considered the execution time and the size of the created indexes. We will present in this section some results and compare our approach of indexing using MapReduce to a classical one that uses Algorithm 3 of the previous section (Section 4) without MapReduce i.e without parallelization of execution; we name it “Approach 2”.We have created a corpus consisting of 20000 image-text pairs, retrieved from the Yahoo! News website 2. Experiment 1: Average execution timeIn this series of experiments, we measured the average execution time of the creation of indexes. We have launched the two approaches (with and without MapReduce) and calculate the time.Fig. 3. Execution time by varying the size of corpusFig. 3 shows the results of execution time of the two approaches by varying the size of corpus (we used various sizes of corpus from 500 files per corpus to 20000 files). The execution time (time for the creation of index) given by our approach is 282 seconds for example when the corpus size is 3500 when the Approach 2 is 564 seconds. In all cases, the required time for the creation of index of our approach is better than approach 2 (see Fig. 3). We can deduce from these results that our approach gives better results, which is estimated 49 % faster (the average considering the results of Fig. 3).Experiment 2: Index sizeThis second experiment is to study size of created index by varying the corpus size.Fig. 4 summarizes the results obtained by this experiment. We vary the size of corpus from 500 files to 20000 files, as shown in Fig. 4.21: Class Mapper2: procedure Map (docid id, doc d) 3: H <– new AssociativeArray;4: for all Concept C in Index_Sem do 5: cf{C} <– cf{C} + 1; 6: for all Concept C in H do7: Emit (Concept C, posting <id, cf{C}>)1: Class Reducer2: procedure Reduce (concept c, postings [< id1, cf1> ...]) 3: P <– new List;4: for all posting <a, f> in postings [<id1, cf1> ...] do 5: Append (P, <a, f>); 6: Sort (P)7: Emit (Concept C, postings P)Algorithm 3: Creation of inverted index Require : Index_Sem: semantic indexesOutput : Index_Sem_Inv: Inverted semantic indexFor each Index_Sem do/*for all indexes Index_Sem of documents*/; For each Concept in index_Sem do/* for all concepts of all the semantic indexes of the documents*/;If Concept.existsIn (Index_Sem Inv ()) = True Then/*If the Concept exists in Index_Sem_Inv */;Index_Sem_Inv [Concept].Add [Index_Sem.Do cument, weighting];/*add document of current Index_Sem to Index_Sem_Inv with the weighting of the concept*/Else /*Concept does not exist in the inverted index*/; Index_Sem_Inv [end].Add [Concept, Index_Sem.Document, weighting];/*add the concept with the document and weighting at the end of the inverted index*/;END If ; END For; END For; ENDFig. 4. Index size by varying the size of corpusWe note that our approach reduces the size of the created index. For the examples shown in Fig. 4, the size of index is 24Mo for our approach when the corpus size is 3500 and the index size of approach 2 is 48.5Mo when the corpus size is 3500, we have an average gain of 48% of space in our approach (the average considering the results of Fig. 4), we deduce that the proposed approach reduce the necessary space for index because of the performance of the MapReduce model and the storage of the necessary information (concepts only).VI. C ONCLUSIONIndexing plays an important role in information retrieval. There are several types of indexing but in all cases the principle is to convert data sources to computerized data. The objective of this article is to propose techniques for automatic semantic indexing (indexing linked to ontology) on heterogeneous sources. In this paper, we proposed an approach for indexing semi or unstructured sources. Annotation is proposed for non-textual sources and automatic indexing for textual sources by considering syntactic and semantic indexing. The goal of the integration of techniques for diverse sources is to manage the heterogeneity, and consider all sources of information of the web. To create our inverted index, we used MapReduce Framework to allow the scalability.Experiments show that our approach has a good response time and the size of the created index is improved because of the storing of necessary information and the use of MapReduce.Our approach is only a first phase in a global work of a research about new techniques for querying web data sources.To a continuation of our work, we will integrate our approach in a global system to exploit our indexes and test relevance of responses that will be returned to users queries.References[1]/definition/Hadoop. [2]J. Dean and S. Ghemawat. Mapreduce: Simplified dataProcessing on large clusters. OSDI ’04, pages 137–150,2008.[3]M.Isard, M.Budiu, Y.Yu, A.Birrell, and D.Fetterly. Dryad:distributed data-parallel programs from sequential building blocks. In Euro Sys ’07: Proceedings of the 2ndACM SIGOPS/EuroSys European Conference onComputer Systems 2007, pages 59–72. ACM2007.[4]S. Agrawal et al. Database Tuning Advisor for MicrosoftSQLServer2005. VLDB2004. Pages 1110–1121. 2004. [5]N. Bruno and S. Chaudhuri. Physical Design Refinement:The Merge-Reduce Approach.ACM TODS, 32(4), 2007. [6]S.Chaudhuri and V. R. Narasayya. Self-Tuning DatabaseSystems: A Decade of Progress. In VLDB, pages 3–14,2007.[7] F. Chang, J. Dean, S. Ghemawat, W. C. Hsieh, D. A.Wallach,M.Burrows,T.Chandra A.Fikes, and R.E. Gruber,“Bigtable: A distributed storage system for structured data”ACM Trans. Comput. Syst. vol. 26, no. 2, pp. 1–26. 2008. [8]J. Lin, D. Ryaboy, and K. Weil.Full-text indexing foroptimizing selection operations in large-scale dataanalytics. In MapReduce, pages 59–66, 2011.[9]Mustapha Baziz, Mohand Boughanem, Salam Traboulsi: Aconcept-based approach for indexing documents in IR.INFORSID 2005: 489-504.[10]Mustapha Baziz, Mohand Boughanem, and NathalieAussenac-Gilles: Conceptual Indexing Based on DocumentContent Representation. CoLIS 2005: 171 - 186.[11]Mustapha Baziz, Mohand Boughanem, Gabriella Pasi,Henri Prade: An Information Retrieval Driven by Ontology: from Query to Document Expansion. RIAO2007.[12]Gerard Salton: Syntactic Approaches to Automatic BookIndexing. ACL 1988: 204-210.[13]Christina Lioma, Iadh Ounis: Light Syntactically-BasedIndex Pruning for Information Retrieval. ECIR2007:88-100.[14]Lin, J. et C. Dyer (2010).Data-Intensive Text Processingwith MapReduce.Morgan & Claypool Publishers.[15]C., Schenk, S., Scherp, A.: Kat: the k-space annotation tool.In: Poster Session, Int. Conf. on Semantic and DigitalMedia Technologies (SAMT). Germany. (2008).[16]Russell, B., Torralba, A., Murphy, K., Freeman, W.:Labelme: A database and web-based tool for imageannotation. International Journal of Computer Vision 77(2008) 157-173.[17]Kipp, M.: Anvil - a generic annotation tool for multimodaldialogue. In: in Proc. 7th European Conf. on SpeechCommunication and Technology (Eurospeech), Aalborg,Denmark. (2001).[18]Schallauer, P., Ober, S., Neuschmied, H.: Efficientsemantic video annotation by object and shot re detection.In: Posters and Demos Session, 2nd InternationalConference on Semantic and Digital Media Technologies(SAMT), Koblenz, Germany. (2008).[19]Kang, B.-Y., & Lee S.-J. "Document indexing: a Concept-based approach to term weight estimation." InformationProcessing and Management: an International Journal41(5): 1065 – 1080. (2005).[20]Nick Craswell, Stephen E. Robertson, Hugo Zaragoza,Michael J. Taylor: Relevance weighting for queryindependent evidence. SIGIR 2005: 416-423 (2005). [21]Miller F., WordNet: A lexical database. Communication ofthe ACM, 38(11): 39-41, (1995).[22]G. Wang, Evaluating Mapreduce system performance: ASimulation approach. Ph.D. Thesis. Virginia PolytechnicInstitute and State University (2012)Authors’ ProfilesSaidi Imene:PhD candidate in computer science, at Litio Laboratory, University of Oran Es Senia, Algeria. She received 5-year engineering degree in computer science in 2010 and a Master Degree in 2011. Her research interests are data mining, information retrieval and web technology.Nait Bahloul Safia:Associate professor in department of computer science, University of Oran Es Senia, Algeria. Since 2011, she has been leading a team on the topic of data engineering and Web Technology. Her research covers advanced aspects of databases, Web technology and unsupervised classification.How to cite this paper: Saidi Imene, Nait Bahloul Safia,"An Approach for Indexing Web Data Sources", International Journal of Information Technology and Computer Science(IJITCS), vol.6, no.9, pp.52-58, 2014. DOI: 10.5815/ijitcs.2014.09.07。