Basic Steps To Avoid Red Flag On Your Scampage

Anoxymen

Member
Registered User
Jun 18, 2018
31
0
6
38
Stop Google from indexing your scampage
Step One Using robots.txt
What is a robots.txt file?
A robots.txt file tells search engine crawlers which pages or files the crawler can or can't request from your site. This is used mainly to avoid overloading your site with requests; it is not a mechanism for keeping a web page out of Google. To keep a web page out of Google, you should use noindex directives (Step two)
Note: If you’re using WordPress, you might see a robots.txt file when you go to yoursite.com/robots.txt, but you won’t be able to find it in your files.

This is because WordPress creates a virtual robots.txt file if there’s no robots.txt in the root directory.

If this happens to you, you’ll need to create a new robots.txt file.


Creating a robots.txt file

You can create a new robots.txt
Add the below code in the robot.txt file

User-agent: *
Disallow: /

this will block all search bots from indexing.
STEP TWO: noindex directives
Using the robots meta tag. Using the code below

<meta name="robots" content="noindex" />

Place the robots meta tag in the <head>section of a given page
Find the source code of the page you want to change, and make sure you’re in between the <head> tags.
image17-7.png

Then paste this line:

<meta name=”robots” content=”nofollow”>

So it should look like this:
image19-8.png

Make sure you’re not putting this line between any other tags––just the <head> tags.


for more info or help use the comments box or Inbox
 
Man, I love U!

This is awesome. Never knew that something like this is possible
 

About us

  • Our community has been around for many years and pride ourselves on offering unbiased, critical discussion among people of all different backgrounds. We are working every day to make sure our community is one of the best.

Quick Navigation

User Menu