Need Help? Talk to Our Experts
Wednesday, May 26, 2021
As a website owner, you might provide a few channels where your users can interact, such as forums, guestbooks, social media platforms, file uploaders, hosting services, or internal search services. These services allow users to create an account to post content, upload a file, or search on your site. Unfortunately, spammers often take advantage of these types of services to generate hundreds of spam pages that add little or no value to the web. Under the principles set out in Google’s Webmaster Guidelines, this may result in Google taking manual actions against the affected pages. Here are some examples:
Spammy content like this can be harmful to your site and users in several ways:
In this blog post, we will provide some tips to prevent spammers from abusing your site.
When users create an account on your site, consider using Google’s CAPTCHAs service or similar verification tools (for example: Securimage or Jcaptcha) to only allow human submissions and prevent automated scripts from generating accounts and content on your site’s public platforms.
Requiring new users to validate a real email address when they sign up for a new account can also prevent many spam bots from automatically creating accounts. Additionally, you can set up filters to block email addresses that are suspicious or originating from email services that you don’t trust.
Consider enabling comment and profile creation moderation features that require users to have a certain reputation before links can be posted. If possible, change your settings so that you don’t allow anonymous posting, and make posts from new users require approval before they are publicly visible.
Register and verify your website ownership in Search Console. To see if there are any issues detected by Google, check the Security Issues report and Manual actions report. You can also check the Messages panel to learn more information.
In addition, it is good to occasionally check your site for unexpected or spammy content by using the site: operator in Google Search, together with commercial or adult keywords that are unrelated to your site’s topic. For example, search for [site:your-domain-name viagra] or [site:your-domain-name watch online] to detect the irrelevant content on your site, especially for:
site:
site:your-domain-name viagra
site:your-domain-name watch online
Monitor your web server log files for sudden traffic spikes, especially for newly created pages. For example, look for any URLs with keywords in URL patterns that are completely irrelevant to your website. To identify potential high traffic problematic URLs, use the Pages report in Google Analytics.
Block obviously inappropriate content from being published to your platform with a list of spammy terms (for example: streaming or download, adult, gambling, pharma related terms). Built-in features or plugins can delete or mark these content as spam for you.
Another great tool for this is Google Alerts. Set up a [site:your-domain-name spammy-keywords] alert using commercial or adult keywords that you wouldn’t expect to see on your site. Google Alerts is also a great tool for detecting hacked pages.
site:your-domain-name spammy-keywords
Monitor your web server log for user sign-ups and identify typical spam patterns, such as:
If your site allows users to create pages like profile pages, forum threads, or websites, you can deter spam abuse by preventing Google Search from showing or following new or untrusted content.
For example, you can use the noindex meta standard to block access to untrusted pages. Like this:
noindex
<html> <head> <meta name="googlebot" content="noindex"> </head> </html>
Or you can use the robots.txt standard to temporarily block the pages. For example:
Disallow: /guestbook/
You can also mark user-generated content (UGC) links, such as comments and forum posts, as UGC by using rel="ugc" or rel="nofollow". This helps you explain your relationship with the linked page to Google and request that Google not follow that link.
rel="ugc"
rel="nofollow"
With automated scripts or software, spammers can generate a large number of spammy pages on your site in a short time. Some of this content may be hosted in fragmented file paths or directories, which prevent site owners from effectively detecting and cleaning up spam. Some examples are like:
example.com/best-online-pharma-buy-red-viagra-online example.com/free-watch-online-2021-full-movie
It is also recommended to consolidate your user-generated content into a concentrated file path or directory for easier maintenance and spam detection. For example, the following file path would be recommended:
example.com/user-generated-content-dir-name/example01.html example.com/user-generated-content-dir-name/example02.html
Take the time to keep your software up-to-date and pay special attention to important security updates. Spammers may take advantage of security issues in older versions of blogs, bulletin boards, and other content management systems.
In addition, some comprehensive anti-spam systems like Akismet have plugins for many blogs and forum systems that are easy to install and do most of the spam fighting work for you. Additionally, there are trusted and well-known security plugins available for some platforms, which help secure the website, and may be able to catch abuse early.
Depending on your site’s situation, please check out our documentation for more detailed information:
You can also visit our Search Central Help Community if you need any help.
Posted by the Search Quality Team
[ad_2] Source link
Digital Strategy Consultants (DSC) © 2019 - 2024 All Rights Reserved|About Us|Privacy Policy
Refund Policy|Terms & Condition|Blog|Sitemap