Asked By
Scribner
10 points
N/A
Posted on - 11/15/2011
Hi Techyv expertise,
I need to know how I can remove my site from appearing in a certain Google search.
I have already tried contacting the webmaster but it didn’t work.
I hope you can help me.
Thanks
Scribner
Delete site from appearing on Google search
Hi Scribner,
This is the proper way how to prevent your file appearing to Google, kindly add a robot.txt to root server that block the image. Example, i want Google to exclude the picture.jpg image that appears on my site at www.mysite.com/images/picture, add the following to your robot.txt file:
User-agent: Googlebot-Image
Disallow: /images/picture.jpg
On the next time Google crawls your site, we'll see this directive and drop your image from our search results.
This is the proper way to remove all the images on your site from our index, place the following robots.txt file in your server root:
User-agent: Googlebot-Image
Disallow: /
In Addition to this, Google has introduced increased flexibility to the robots.txt file standard through the use (*) asterisk. By Disallowing patterns may include "*" to match any sequence of characters, and patterns may end in "$" dollar to indicate the end of a name. in order to delete all files of a specific file type (for example, to include .jpg but not .gif images), you'd use the following robots.txt entry:
User-agent: Googlebot-Image
Disallow: /*.gif$
The specifying Googlebot-Image as the User-agent, images will be excluded from Google Image Search. And It will also prevent cropping of the image for display within Mobile Image Search, image will be completely removed from Google's Image index. You would like to exclude the images from all Google searches (including Google web search and Google Images), specify User-agent Googlebot.
Thanks,
Mylene Yap
Delete site from appearing on Google search
I don’t think that’s possible. Once your site appears on Google, it is there for good. Your website starts appearing on Google search when you share a link from your website on another website or your site was searched on Google.
You can control how the pages of your website are crawled using the “robots.txt” file which is placed in the root of your website. When a spider finds a link, it strips the path component from the URL down to the first slash and puts “/robots.txt” in its place.
For example, in “https://www.techyv.com/pages/index.html”, the spider or crawler will remove the “/pages/index.html” and replaces it with “/robots.txt” which will end up with “https://www.techyv.com/robots.txt”. So, to make the file do its job, you need to place it in the right location to make the resulting URL work. This is also the same location where you put the default or main “index.html” page of your website.
The “robots.txt” file is just an ordinary text file you can create with Notepad and can contain something like this.
User-agent: *
Disallow: /cgi-bin/
Disallow: /tmp/
Disallow: /~joe/
Remember: a separate “Disallow” line is required for every URL you want to exclude. Also, you cannot have a blank line in the file because it is used to delimit multiple records. The “*” in the “User-agent” field is a special value which denotes “any robot”. Here are some examples. To exclude or not allow all robots from accessing your entire website, you can have:
User-agent: *
Disallow: /
On the other hand, to allow all robots to access your entire website and the total opposite of the previous instruction, you can have:
User-agent: *
Disallow:
Unfortunately, this doesn’t prevent or using “robots.txt” will not prevent “bad” spiders from crawling your website. The robots.txt is a de-facto standard and not owned by any standards body. Also, it is no longer being actively developed.