![]() |
#1 |
Registered User
Join Date: Sep 2012
Posts: 39
|
how to use robot.txt?
__________________
comprehensive internet marketing and seo service www.cimass.com |
![]() |
![]() |
![]() |
#2 |
Registered User
Join Date: Aug 2012
Posts: 120
|
robots.txt is a set of instructions to let the search engines know where and what to look at on your site.
you out it in the top-level directory of your web server. read a couple articles before doing it though because there is much more detail then this |
![]() |
![]() |
![]() |
#3 |
Registered User
|
If you don't have a robot.txt the search engines just crawl every page of your website. There's a lot of articles on how to use the file.
__________________
? ? XenLayer - Paravirtualization Professionals since 2008 - [ICQ ME] ? ? Reseller Hosting | OpenVZ VPS | XEN VPS | Dedicated Servers |
![]() |
![]() |
![]() |
#4 |
Registered User
Join Date: Nov 2012
Posts: 71
|
Robot.txt is a text file which have set of instructions for crawlers ot crawl the website or a specific page. If you have it then it must be placed in root directory.....and for more help you can read lots Articles related to this...thanks
|
![]() |
![]() |
![]() |
#5 |
Registered User
Join Date: Nov 2012
Location: India
Posts: 103
![]() |
The robots.txt is a simple text file in your web site that informs search engine bots how to crawl and index website or web pages. It is great when search engines frequently visit your site and index your content but often there are cases when indexing parts of your online content is not what you want.
__________________
Best Web Design Services - Corporate Web Design, Creative Design, Custom Web Design, Website Re-Design Quality SEO Services - - SEO, Pay Per Click (PPC), SEM, Link Popularity Building, SEO Copywriting, SMO, SMM |
![]() |
![]() |
![]() |
#6 |
Registered User
Join Date: Feb 2012
Posts: 225
|
Robots.txt is a file that is used to exclude content from the crawling process of search engine spiders / bots. Robots.txt is also called the Robots Exclusion Protocol.
|
![]() |
![]() |
![]() |
#7 |
Registered User
Join Date: Dec 2012
Posts: 657
|
type this in
User-agent: * Disallow: this is an example... open notepad, and just save a robots, then upload to your root. |
![]() |
![]() |
![]() |
#8 |
Registered User
Join Date: Mar 2012
Location: india
Posts: 218
![]() |
Robots.txt is a file that is used to exclude content from the crawling process of search engine spiders / bots. Robots.txt.
|
![]() |
![]() |
![]() |
#9 | |
Registered User
Join Date: Nov 2010
Location: us
Posts: 584
|
Quote:
By adding this code which pages search engines will not visit? |
|
![]() |
![]() |
![]() |
#10 |
Registered User
Join Date: Oct 2012
Posts: 149
|
robots.txt is a set of instructions to tell the search engines where and what to look at on your site.
|
![]() |
![]() |
![]() |
#11 |
Registered User
Join Date: Nov 2012
Posts: 71
|
robots.txt is a set of instructions to let the search engines know where and what to look for on your site.
|
![]() |
![]() |
![]() |
#12 |
Registered User
Join Date: Jul 2012
Posts: 102
|
Web site owners use the /robots.txt file to give instructions about their site to web robots; this is called The Robots Exclusion Protocol.
It works likes this: a robot wants to vists a Web site URL, say http://www.example.com/welcome.html. Before it does so, it firsts checks for http://www.example.com/robots.txt, and finds: User-agent: * Disallow: / The "User-agent: *" means this section applies to all robots. The "Disallow: /" tells the robot that it should not visit any pages on the site. There are two important considerations when using /robots.txt: robots can ignore your /robots.txt. Especially malware robots that scan the web for security vulnerabilities, and email address harvesters used by spammers will pay no attention. the /robots.txt file is a publicly available file. Anyone can see what sections of your server you don't want robots to use. |
![]() |
![]() |
![]() |
Currently Active Users Viewing This Thread: 1 (0 members and 1 guests) | |
|
|