We need a script/application that we can leave running on a PC that will crawl the web, preferably with editable limits (e.g. just .com domains, or just pages within a specific domain) and compile data.
It needs to log the page name, URL, and results for about 3 criteria....
1) Does the site use frames (y/n)
2) Does the page have a Doctype declaration (y/n)
3) Does the page code validate (might be too difficult, let me know your thoughts)
Please state how you will achieve this, how the logged data can be accessed (ideally a spreadsheet), and post examples if you have done similar things.
Hello, thanks for your bid invitation and think our price and time span are the best one as per your budget and project. Pls. See PMB. Regards.... Microsectec.
Dear Friend, The project is within our skills and abilities. We are a company with more than 10 years of experience in the IT-sphere. We offer flexible pricing and fast turnaround. If you are interested, please, contact us via PMB and provide more info on the project. Looking forward to working with you, Viaden Inc
To scan for frames or doc type is a sinch , however if you want to validate the code it would be difficult without a syntax checker.
If you want to attach a syntax checker that is already a program that you own or a component it will be just aseasy to do. Creating a syntax checker would be pointless. For further consultation please send a message.
Respected Sir . I have already done many Web Spidering projects. So I can assure you that it would be of good quality. Could you please check the Private Mesage board for my privately posted message and contact me for more discussions