Dist-RIA Crawler: A Distributed Crawler for Rich Internet Applications
ieeexplore.ieee.org › document
To reduce the time to crawl a RIA, this paper presents a new distributed algorithm to crawl a RIA in parallel with multiple computers, called Dist-RIA Crawler.
Abstract—Crawling web applications is important for index- ing, accessibility and security assessment. Crawling traditional web applications is an old ...
A statistical approach for efficient crawling of rich internet applications. ICWE'12: Proceedings of the 12th international conference on Web Engineering.
In case of the so-called Rich Internet Application (RIA), a web crawler has to cover both all URLs and all application states, and this task can be mapped to ...
To crawl a RIA, the crawler has to reach every application state and execute every application event. On a large RIA, this operation takes a long time.
Oct 28, 2013 · Dist-RIA Crawler: A Distributed Crawler for. Rich Internet Applications. Software Security Research Group (SSRG). University of Ottawa. In ...
To reduce the time to crawl a RIA, this paper presents a new distributed algorithm to crawl a RIA in parallel with multiple computers, called Dist-RIA Crawler.
An adaptive load-balancing algorithm is introduced that enables the crawler to learn about the speed of the nodes and adapt to changes, thus better utilize ...
PDist-RIA doesn't have the same limitations in terms scalability while matching the performance of GDist-Ria, and is proposed as a fully peer-to-peer and ...
Web crawlers visit internet applications, collect data, and learn about new web pages from visited pages. Web crawlers have a long and interesting history.