The basic function we need...
1. Crawl our own existing website mimicking a human searching our site
2. find and click on all external URL's leading to 3rd party websites
again from our existing website mimicking human behavior
3. every 3rd party URL that it finds and clicks on it will grab a
different IP address so that when the 3rd party site looks into their referrer logs this looks like multiple visitors from all over the united states.
4. Allow us to control the frequency and how many clicks we send to 3rd
5. Provide detailed reporting showing URL's from our site clicked on,
how many times and the ip addressed used.
6. We should be able to enter and perform the above function on any website we own simple by entering in the url we want to crawl then clicking on all the URLs on site entered delivery clicks to all 3rd party links on site all coming from DIFFERENT IP ADDRESSES.
Please Dont Bid Outrageous and Bid Accordingly !