We have a tracking system on our website that's begun to become a bit cumbersome on performance. Essentially - when someone hits a certain type of page - there's a sequence that:
1. Checks the IP address of the person hitting the page
2: Checks to see if that IP address has hit the page before
3: If the IP address is new/unique - logs it into the database for that page
4: If the IP address is not unique - adds to a counter for that IP address
The system works - but we're now at about 750k records and growing...and it's slowing down (obviously).
I know enough about SQL to make that process work - but need some guidance on how to optimize it. How exactly do you deal with large recordsets like that without a degradation in performance? Any suggestions/solutions that have worked for you?