We have been having issues with Chrome's malware blocking screen for about 6 months periodically. We manage roughly 110 websites for our clients and each one has a URL that we use when routing to external websites as a means of tracking the clicks. The problem is that if one of those links goes to a malware site, then the entire route ends up blocked as malware and now all outgoing URLs are blocked. In order to correct this problem, back in October we instituted a system where ALL outgoing URLs would be checked server-side against the Google Safe Browsing Go package. So in the browser the a tag would be written to /plugins/crm/count/?encodedString (to prevent against redirect attacks), then on the server it decodes the string, checks safe browsing and if it's safe redirects the user to the destination site.
It was working pretty well for the past month or two as we had no sites affected by the system. Every now and then a redirected site would get flagged and we would remove it from our database, but none of them caused cascading failure to block the whole /plugins/crm/count/ route. Now, we've had 2 sites with cascading blocking in the past week. The odd thing about both cases is that our server-side analytics tell us that we redirected exactly two, yes, two real users to the malware site in a period of time while it may be infected. This means we were flagged for merely having a URL which 301s to another site even if that 301 went through the safe browsing system. Unfortunately Google has been very quiet on how this whole system works, which leaves site operators like myself very frustrated. We want to comply with the rules, we will institute the necessary systems, but can you at least provide some guidance as to what we can do to prevent getting blocked?
So here are my questions:
1. It appears to me based on the evidence I have (because Google refuses to provide documentation on this subject!) that Google blocks sites that link to a site with malware, even if that site isn't actively sending traffic through those links and even if those links use the Google Safe Browsing to prevent redirects if they did become infected in the future. If this is true, then honestly, what is the point of the Safe Browsing API in the first place? The pattern we utilize is the exact same pattern used by Yelp and TripAdvisor for linking to external sites. Yet, they don't see to get their entire pathway blocked. Why?
2. The timing of how the whole system works is clouded in mystery. One of our sites showed up in the Google Search Console as being infected on 2/3/2017. If it shows up in Google Search Console at that time, does it mean that around that same time it was flagged as malware. According to our analytics only two users went through to the URL in the 5 days before 2/3/2017. Does that mean that sending through 2 clicks is enough to flag an entire pathway. It possible that it was flagged as malware months ago, and had been malware for quite some time, and only within the past week did Google escalate it up to the search console. If so, that would indeed be frustrating. If the search console is the method of notifying admins of the issue, then it would make sense to notify admins as soon as the problem is known so we can correct it before our urls are blocked as malware. Is there any documentation for when something is flagged about how long until it shows up in Search Console, how long before it gets blocked and linking sites.
3. Why is that sometimes our system only has issues with individual URLs being blocked /plugins/crm/count/?specificUrl, while others get /plugins/crm/count/ blocked in it's entirety. The times when the entire pathway are blocked do not seem to have either A) a high volume of traffic or B) a high number of individual infected URLs. In all of the cases of group blocking we've had only 1 corrupt URL out of thousands. In the ones where we have had a single URL blocked it was the same pattern. What causes one failure to cascade upstream to a whole pathway while others don't?
4. Is there a best practice we can utilize to solve this problem in a different way which will inoculate us from this issue. What we are doing is nothing different than
https://www.yelp.com/biz/sak%C3%A9-brisbane-2?osq=Restaurants when you look at their biz_redir pathway if you mouse over the restaurants website URL.
Example Client:
Any help is appreciated.