Automated Path Ascend Forum Crawling

DOI : 10.17577/IJERTV2IS3641

Download Full-Text PDF Cite this Publication

Text Only Version

Automated Path Ascend Forum Crawling

Ms. Joycy Joy,

PG Scholar Department of CSE, Saveetha Engineering College,Thandalam,


Ms. Manju. A,

Assistant Professor, Department of CSE , Saveetha Engineering College,Thandalam, Chennai-602105

AbstractFoCUS (Forum Crawler Under Supervision), is a supervised web-scale forum crawler. The goal of FoCUS is to crawl relevant forum content from the web with minimal overhead. FoCUS is an automation engine that will dynamically crawl the relevant content in a forum . Forum threads contain information content that is the target of forum crawler. Cleanup of data and moving the contents to the appropriate web pages is the major scope of the project. The content of forum may be the queries asked by the users. After crawling the content, FoCUS will dynamically move the queries in the related forum, which will deal the particular query. Then FoCUS cleanup the unrelated query from the particular forum, and that free space is allocated to new queries posted by user. FoCUS take six path from entry page to thread page. It helps the frequent thread updation in forum. FoCUS makes use the technique called differential content extraction, which helps to maintain a record for already crawled data. In each time FoCUS will not crawl the forum data from the beginning, it will maintain a record of already crawled data and manipulates only the newly posted queries.

Keywords EIT Path, Forum Crawling, ITF Regex, URL Type.


    Internet forums [12] (also called web forums) are important services where users can request and exchange information with others. It helps to know users opinion about a product and understand what are their expectations. To harvest knowledge from forums, their content must be downloaded first. A web forum crawler which can collect the forum data automatically according to scheduled time such as once in a week. The collected data will be stored in the database. The data can be used for data mining or social network analysis.

    In the existing system iRobot forum crawler is used, which crawl the forum content. It does not deal with the frequent thread updation in forum.iRobots tree like traversal didnt allow more than one path from starting page node to same ending page node. So it takes only one path (first path that is entry-board-thread) from entry to thread page. Here sampling strategy and in formativeness estimation is not robust. Existing system doesnt follow the differential content

    extraction. That is it doesnt maintain a record of previously stored data. When new queries are posted by user, the crawler can start the crawling process from the beginning in every time. So it become a time consuming process.The main drawbacks of the existing system are:

    • No clear segregation of page identification is carried out

    • It takes only one path from entry to thread page.

    • It doesnt make use of differential content extraction technique.

    FoCUS tried to create an automation engine which will take care of traversing the contents dynamically. Moving towards the hyperlinks related to the forum and cleanup the related links. Integrating the missed out data pages in future were considered as the core proposed approaches included in the system. In our proposed system, we are utilizing the features of differential content extraction instead of an inefficient entire system scanning. This option will enhance the performance of the system very much. The option of differential content is done with the help of page indexes and number of links options or link value. In addition, amend and building the knowledge database enable the system a very efficient one in a longer vision. Scanning the entire web pages through Key match cum KnuthMorrisPratt algorithm is used. The proposed system maintain a record of already crawled data.The six paths from entry to thread page are given below:

    1. entry board thread

    2. entry list-of-board board thread

    3. entry list-of-board & thread thread

    4. entry list-of-board & thread board thread

    5. entry list-of-board list-of-board & thread thread 6.entry list-of-board list-of-board & threadboard


      The main advantages of Focus are given below :

      • Automation web crawling is done with this application.

      • FOCUS takes six paths from entry to thread page.

      • Differential content extraction is used.

    The major contributions of this paper are as follows:

    1. We create an automatic engine which will crawl the forum pages automatically..

    2. Focus make use of the technique called differential content extraction which helps to maintain the record of already crawled data.So the effectiveness becomes increased.

    3. Cleanup of data and moving the contents to the appropriate web pages is the major scope of FoCUS.

    4. After remove the unrelated links ,FoCUS allocate that space to the newly posted queries.


    Vidal Caj.R, Yang. J.M, Lai.W, Wang.Y, and Zhang.L[2]

    ,iRobot has an intelligence to understand the content and the structure of a forum site, and then decide how to choose traversal paths among different kinds of pages. Furthermore, it also achieve the following advantages: (1) significantly decreases the duplicate and invalid pages;(2) saves substantial network bandwidth and storage as it only fetches informative pages from a forum site;(3) It provides a great help for further indexing and data mining;(4) Effectiveness: it intelligently skip most invalid and duplicate pages, while keep informative and unique ones;(5) Efficiency: iRobot only need a few pages to rebuild the sitemap. It is also have some disadvantages such as; It follow a tree like traversal, so it didnt allow more than one path from starting page to ending page. It doesnt deal how to design a repository for forum archiving.

    Wang.Y, Yang.J.-M, Lai.W, Cai.R, Zhang.L, and Ma.W.-Y[5] , Exploring Traversal Strategy is a traversal strategy consists of the identification of the skeleton links and the detection of the page-flipping links. Furthermore, it achieve the following advantages: (1) The skeleton links instruct the crawler to only crawl valuable pages and meanwhile avoid duplicate and uninformative ones;(2) page- flipping links tell the crawler how to completely download a long discussion thread which is usually shown in multiple pages in Web forums. It has some demerits such as it doesnt deal with how to optimize the crawling schedule to incrementally update the archived forum content. And also it doesnt deal how to parse the crawled forum pages to separate replies in each post thread.

    Brin.S and Page.L[1] ,Web Search Engine it is a large-scale search engine which makes heavy use of the structure present in hypertext, for example is Google. Google is designed to crawl and index the web efficiently and produce much more satisfying search results than existing systems. It

    answers tens of millions of queries every day. This paper provides an in-depth description of large-scale web search engine. It makes heavy use of the structure present in hypertext.But it doesnt deal how to effectively deal with uncontrolled hypertext collections, where anyone can publish anything they want. The technical challenges involved with using the additional information present in hypertext to produce better search results.

    Guo.Y, Li. K, and Zhang.K[3] , Board Forum Crawling is a web crawlng method for web forum. This method exploits the organized characteristics of the Web forum sites and simulates human behavior of visiting Web Forums. Board Forum Crawling can crawl most meaningful information of a Web forum site efficiently and simply. Experiments have shown BFC is an efficient and economical method and has been used in a real project. Limiting to the space, the details of the method, such as link clustering based on URL is the main demerit of this paper.


      1. System Overview

        Fig. 1. The overall architecture of FoCUS

        Fig. 1 shows the overall architecture of FoCUS. The user come with a query, first point to the forum page.Given any page of a forum, FoCUS first finds its entry URL using the Entry URL Discovery module. Then, it uses the Index/Thread URL Detection module to detect index URLs and thread URLs on the entry page; the detected index URLs and thread URLs are saved to the URL training sets. It detect the keyword. Pre Built Page classifier makes the record of already crawled data.The ITF Regexes Learning module compare the

        new keywords with the keywords, that stored in the database.If any mismatch occurs it means that the data is irrelevant and Focus clean that data and the remaining data stored on the system.

      2. FoCUS Modules

        1. Main Forum

          This module, act as an integral portal of querying according to the basic doubts in the technology. Logged user can be able to post their queries by selecting the options of technology. An usual application with an authentication page with user creation can be done in the application.

        2. Forum Thread

          An individual thread will be created by the users based on the queries raised by the users. The threads will be segregated based on the technology of the project. Once the user wants to check an individual thread, open the website and move the appropriate technology and click the relative words to check it out.

        3. Authentication

          In this module, an authentication page is created which will enable the user to login into the system. The option of creating the registered user is provided to the system.

        4. Forum crawling

          The users were permitted to crawl the web pages automatically, once the user provide the necessary options of website details. An automatic recognition mechanism with an underlying Top down keyword based search algorithm is implemented to identify the exact URLs and the navigation of the web pages will happen automatically. The next page in the forum is moved and the thread in each individual category is scanned.

          Top down key word based search algorithm

          • Keyword search algorithm is an algorithm for finding an item with specified properties among a collection of items.

          • The items may be stored individually as records in a database. A Keyword search looks for words anywhere in the record.

          • The key words are searched top to down in the database.

        5. Keyword integrate engine

          Once the crawling engine has entered into the individual thread, the keywords were scanned through KMP Algorithm and the keywords were identified. The keywords were compared with the existing datasets. Once the keywords are matched. Immediately, the system will automatically move the web pages to the appropriate technology.

          KnuthMorrisPratt algorithm

          • KnuthMorrisPratt string searching algorithm (or KMP algorithm) searches for occurrences of a "word" W within a main "text string" S by employing the observation that when a mismatch occurs.

          • The word itself embodies sufficient information to determine where the next match could begin, thus bypassing re-examination of previously matched characters.

          • We proceed by comparing successive characters of W to "parallel" characters of S, moving from one to the next if they match. However, in the fourth step, we get S[3] is a space and W[3] = 'D', a mismatch.

          • Rather than beginning to search again at S[1], we note that no 'A' occurs between positions 0 and 3 in S except at 0; hence, having checked all those characters previously.

          • We know there is no chance of finding the beginning of a match if we check them again. Therefore we move on to the next character, setting m = 4 and i = 0.

        6. Forum manual cleanup

    Cleanup of data and moving the contents to the appropriate web pages is the major scope of the project. In this module, the unwanted data will be cleaned up and the forum data will be moved to the web pages according to the technology. The data will be cleaned up and the forums will be moved to the technology in turn, the display of the particular thread should be moved to the appropriate forum. In addition, the knowledge database will get accumulated with lots of knowledgeable information which will be used in future cases of getting more streamlined data processing.


FoCUS, a supervised forum crawler was implemented. Focus automatically crawl the forum data and it clean up the unwanted data. FoCUS made use of differential content extraction, which helps to make record of previously crawled data,so it reduce the crawling time for each new crawl. After

cleaning the unwanted data, FoCUS allocated that space to new queries posted by the user.

In future Focus will separate the sparms.Also in future an intimation mail wil send by the user after deleting his irrelevant queries.


  1. S. Brin and L. Page. The Anatomy of a Large-Scale Hypertextual Web Search Engine. Computer Networks and ISDN Systems, vol. 30, nos. 1-7, pp. 107-117, 1998.

  2. R. Cai, J.-M. Yang, W. Lai, Y. Wang, and L. Zhang. iRobot: An Intelligent Crawler for Web Forums. Proc. 17th Intl Conf. World Wide Web, pp. 447-456, 2008.

  3. Y. Guo, K. Li, K. Zhang, and G. Zhang. Board Forum Crawling: a Web Crawling Method for Web Forum. Proc. 2006 IEEE/WIC/ACM Intl Conf. Web Intelligence, pp. 475-478, 2006.

  4. C. Gao, L. Wang, C.-Y. Lin, and Y.-I. Song. Finding Question-Answer Pairs from Online Forums. Proc. 31st Ann. Intl ACM SIGIR Conf. Research and Development in Information Retrieval, pp. 467-474, 2008.

  5. Wang.Y, Yang.J.-M, Lai.W, Cai.R, Zhang.L, and Ma.W.-Y. ,Exploring Traversal Strategy for Web Forum Crawling. Proc. 31st Ann. Intl ACM SIGIR Conf. Research and Development in Information Retrieval, pp. 459- 466,2008.

  6. A. Dasgupta, R. Kumar, and A. Sasturkar. De-duping URLs via rewrite rules. Proc. 14th ACM SIGKDD Intl Conf. Knowledge Discovery

    and Data Mining, pp. 186-194, 2008.

  7. M. Henzinger. Finding near-duplicate Web pages: a large-scale evaluation of algorithms. Proc. 29th Ann. Intl ACM SIGIR Conf. Research

    and Development in Information Retrieval, pp. 284-291, 2006.

  8. H. S. Koppula, K.P. Leela, A. Agarwal, K.P. Chitrapura, S. Garg and

    A. Sasturkar. Learning URL Patterns for Webpage De-duplication. Proc. Third ACM Conf. Web Search and Data Mining, pp. 381-390, 2010.

  9. K. Li, X.Q. Cheng, Y. Guo, and K. Zhang. Crawling Dynamic Web

    Pages in WWW Forums. Computer Engineering, vol. 33, no. 6, pp. 80- 82,2007.

  10. G. S. Manku, A. Jain, and A. D. Sarma. Detecting near-duplicates for Web crawling. Proc. 16th Intl Conf. World Wide Web, pp. 141-150, 2007.

  11. M. L. A. Vidal, A. S. Silva, E. S. Moura, and J. M. B. Cavalcanti. Structure- driven Crawler Generation by Example. Proc. 29th Ann. Intl

    ACM SIGIR Conf. Research and Development in Information Retrieval, pp. 292-299, 2006.

  12. Internet Forum.

Leave a Reply