If you were to create a robot (spider) to crawl the web, which of the following actions should you be considering to do?
1) Trying to be anonymous during the crawl to avoid privacy issues.
2) Ignoring sites' crawling policies, so you don't skip any content.
3) Keeping your crawler's raw data, and sharing the results pubLiCly.
4) Running your crawler while out on vacation, because that may take a few days.