If you were to create a robot (spider) to crawl the web, which of the following actions you should not perform?
1) Checking available crawled data from other robots, so you may not need to implement your own robot.
2) Announcing your intentions and using HTTP user-agent to identify your robot.
3) Keeping your crawler's raw data and sharing the results pubLiCly.
4) Implementing a fast-crawling robot to make sure the process finishes within a reasonable amount of time.