10-11-2014, 09:41 AM
Abstracts: Our Project Developing Hosting Layer Approach Traffic Analysis for Websites will run into the backend and when visitors visit website they will monitor each and every activity of the users depends on the various criteria specified by the users. This traffic analysis of web mining is to check when to go for the maintenance stage and down the website for the maintenance. And to check from which geo location (country) visitors are coming from so we can set the contents of the site accordingly. This project presents the procedures used for network traffic analysis, which provide a clear overview of the structure of traffic and enable the efficient detection of potential problems and irregularities. One of the most actual problems on web crawling (the most expensive task of any search engine, in terms of time and bandwidth consumption) is the detection of useless segments of Internet. In some cases such segments are purposely created to deceive the crawling engine while, in others, they simply do not contain any useful information. The typical approach to the problem consists in using a human-compiled blacklist of sites to avoid (e.g., advertising sites and web counters), but, due to the strongly dynamical nature of Internet, keeping them manually up-to-date is quite unfeasible. In this work we present a web usage statistics-based solution to the problem, aimed at automatically and, therefore, dynamically building blacklists of sites that the users of a monitored web-community. Our method performs a linear time complexity analysis on the traffic information which yields an abstraction of the linked web which can be incrementally up-dated, therefore allowing a streaming computation.