A web crawler (also known in other terms like ants, automatic indexers, bots, web spiders, web robots or web scutters) is an automated program, or script, that methodically scans or "crawls" through web pages to create an index of the data it is set to look for. This process is called Web crawling or spidering. Read the article
here.
No comments:
Post a Comment
Note: only a member of this blog may post a comment.