• Status: Solved
  • Priority: Medium
  • Security: Public
  • Views: 239
  • Last Modified:

Protecting Web Site Content

Any suggestions on how to protect a web site from scrapers and other site capturing tools. I have reviewed a few products and methods, but wanted to get other thoughts.
0
tarrigo
Asked:
tarrigo
1 Solution
 
killbradCommented:
Nothing is really gonna work.  At best you will find a program that can rewrite your html into something that looks absolutely terrible, but the site will still look fine to 'scrapers'.  
0
 
mcse2007Commented:
If you referring to encrypting the traffic when the data travels from the public network to the web server, SSL certificate will do this for you. Also to protect your server from scrapers block all ports in the firewall and only port 80 for http and port 443 for ssl that you need to open, if you are hosting your server.
0
 
Hube02Commented:
check out this site: http://browsers.garykeith.com/

We have plans to put something similar in place. It requires adding a script to the beginning of every page (more likely an included file) that will check the data file from the above site against the browser that is requesting the page and not not send any content to the browsers listed as blocked in that file.

However, we are planning on doing this in PHP, and not ASP. There are several other tools available there.
0
Free Tool: Port Scanner

Check which ports are open to the outside world. Helps make sure that your firewall rules are working as intended.

One of a set of tools we are providing to everyone as a way of saying thank you for being a part of the community.

 
killbradCommented:
mcse2007:  Blocking other ports is not gonna stop someone from leeching his website with spider software.  
Hube:  It's a nice idea, but not realistic.  Many webleech programs give you the specific option to spoof the headers sent.  Also, trying to access the page you referenced above in Linux gives me this:

ACCESS DENIED
You do not appear to be using this form in accordance with my Terms of Use.
Continued abuse will eventually result in you losing access to this server!
It's also possible you are using security software that modifies the HTTP_REFERER header.
If I can't confirm the referrer is valid then you can't have access to my forms.

hmmm  :-/
0
 
Hube02Commented:
killbrad:  In your case I would say that you have a firewall or are going through a proxy server that blocks or modifies the HTTP_REFERER header. (that isn't my site) but it is a security measure that I also implement on forms on sites I build. All forms submit to themselves and if the form gets a post request from somewhere other than itself it will not process the request. This effectively stops 90% of spam bots submitting forms. But it also blocks those that have firewall or proxy servers set up like yours.
0
 
killbradCommented:
nope, it just won't let me use elinks to access it.. tried it from several places.
0
 
Computer101Commented:
Forced accept.

Computer101
EE Admin
0
Question has a verified solution.

Are you are experiencing a similar issue? Get a personalized answer when you ask a related question.

Have a better answer? Share it in a comment.

Join & Write a Comment

Featured Post

Get your problem seen by more experts

Be seen. Boost your question’s priority for more expert views and faster solutions

Tackle projects and never again get stuck behind a technical roadblock.
Join Now