Robots.txt

Hi there I just added a robots.txt on my site because I don't want to read some files from google, the problem is I have a 100 or so these index.html files on my site, they are in files separate album that I use for my carpet samples and they do not need to be indexed, the problem is even with the file are pages containing the index.html file when opened and they need to be indexed so I want to block the index files, but not the rest.

This is a link to one of the pages that contain the files: http://www.qualitycarpets.net/samples-carpet-berber/agadir-berber-carpet.php

If anyone can tell me how to write the text to do this, I will be very greatfull.

Thank you, Jeff Lane

If this should not be...

Tags: Dreamweaver

Similar Questions

  • I need the config of the robots.txt file

    OK I am running hmail Server im trying to webmail, part I need to config the robots.txt file, how can I do. also I can get the newspaper but after I log in I get 404 page not found what can I do

    http://www.hMailServer.com/Forum/

    Have you tried to ask in their Forums at the link above.

    http://www.hMailServer.com/index.php?page=support

    And they also have technical support Options.

    See you soon.

    Mick Murphy - Microsoft partner

  • Allowing resources Robot.txt

    I launched my site and it works great.  However, when I test the usability in googles webmaster tools they say that there are some resources that are refusing the analysis via robot.txt.  How can I change the settings to allow easier parsing through robots of google and Bing? I have presented a plan sitemap, but still need to resolve the deadlock.

    Thanks in advance.

    James

    Hello

    There is no robots.txt file in your root folder.

    http://www.heritagejewelryandloan.com/robots.txt

    You can manually create an and download to your host and come back in the Google Search (webmaster) console.

    You can take the little help from here the Web Robots Pages the works of the robots.txt file.

    Let me know if you have any question.

  • I have a paid site (it wasn't a demo for a few months) but robots.txt has this "# this file is automatically generated that your site is being tested." To remove it and to allow search engine indexing, please move your site to a paid plan. »

    I have a paid site (it wasn't a demo for a few months) but robots.txt has this "# this file is automatically generated that your site is being tested." To remove it and to allow search engine indexing, please update your site to a paid plan. ». How can I fix it?

    You should be able to create and upload a new file robtos.txt. Create a txt file and call robots. Put the content in the robots.txt file:

    User-agent: *.

    Allow: /.

    Site map: http://www.your-domian-goes-here.com/sitemap.xml

    Once created, download the robots.txt file in the root of your site.

  • I am looking for a SIMPLE way to add my robots.txt file

    I read up on this subject. But I wasn't able to do it again.

    I found this:

    1. re: file robots.txt for sites of the Muse

    Vinayak_GuptaEmployee Hosts

    You can follow the guidelines of Google to create a robots.txt file and place it in the root of your remote site.

    https://support.Google.com/webmasters/answer/156449?hl=en

    Thank you

    Vinayak

    ----------------

    Once you create a robots.txt like this:

    user-agent: ia_archive

    Disallow: /.

    (1) where do you put 'head labels? Do you need them?

    I would insert it in muse that looks like this

    < head >

    user-agent: ia_archive

    Disallow: /.

    < / head >

    or just put it anywhere inside the 'head' tag

    (2) put the robots.txt file in a folder?

    I've heard this, but it just doesn't seem right.

    (3) OK you have the

    -Properties Page

    -Metadata

    -HTML < head >

    Can I copy and paste my right robot.txt info in there? I don't think I can and make it work. According to the info I found (that I posted above), the robots.txt 'file' you 'place at the root of your remote site.

    (4) where is the 'root of my remote sit?

    How can I find that?

    I read other people having problems with this.

    I thank very you much for any help.

    Tim

    I need Terry White to make a video on it LOL

    Maybe I'll ask.

    I thought about it.

    However, with the help of Godaddy, the file was not placed between theand, so I'm still a little nervous.

    It is recommended to:

    ///////////////////////////////////

    1.  re: file robots.txt for sites of the Muse

      

       Vinayak_Gupta   , April 19, 2014 01:54 (in response to chuckstefen)

    You can follow the guidelines of Google to create a robots.txt file and place it in the root of your remote site.

    https://support.Google.com/webmasters/answer/156449?hl=en

    Thank you

    Vinayak

    /////////////////////////////////////

    Place the robots.txt file to the root "of your remote site.

    and that (Godaddy) is not between the

    I checked the robot file that I created here

    New syntax Robots.txt Checker: a validator for robots.txt files

    and other than me not to capitalize

    the 'u' in the user-agent, it seems to work. When my site is analyzed, she does not miss a robots.txt file

    user-agent: ia_archive

    Disallow: /.

    Problem solved, unless I find an easy way to place the robots.txt file placed between the head tags (and better).

    I'll keep my ears open, but don't worry too much on this subject.

    Step 1) write the code of robots that you want to

    Step 2) save the file as a txt

    Step 3) contact your Web hosting provider / upload in the root of the Web site in a single file

    Step 4) check with a robot's checker that I listed above

    What was shake me:

    -where to put

    -the difference between files and folders, it seemed I would be to load a file for some reason any.

    -I was expecting something like the list of the news LOL

  • robots.txt and sitemap.XML

    Hi all

    are there tools in QC to create and deploy the sitemap.xml and the robots.txt file?

    Thanks in advance,

    Michael

    Hi Michael,

    Unfortunately, there is no automated for that yet in QC tool.

    If you want to build something like this yourself, the best is to create a page somewhere in the structure of your site that is not displayed in the navigation (you would nominate this page "site map" or "robots" - without the extension). You must also create a component of page customized for each of these two types of special pages, where you control the output is exactly what you need (ie. no HTML code in the output)-for a first version, simply hard-code the content of each in JSPS, later, you can then make more sophisticated (eg. giving reading a page property or the iteration on the structure of your page). Using light of CRXDE, you obviously have to change ownership of Sling: resourceType of your page sitemap or robots to point to the new components of the appropriate page.

    To make these pages available on the root path, you can use the/etc/map parameters. Such a configuration is as follows in the case of the robots.txt file (CRXDE light to create the node and it's properties):

    card/etc/www-robots.txt

    JCR:primaryType = "Sling: Mapping ' (ie the type when you create a new node)

    Web: internalRedirect = ' / content/SITE/PATH/robots.html '.

    scarf: match = "http/www.SITE.com/robots.txt.

    BTW, this is the documentation of mapping: http://sling.apache.org/site/mappings-for-resource-resolution.html

    Hope that helps!

    Gabriel

  • Hi I need to hide a pdf file hosted on a site of glasses, I would normally add &lt; name meta = "robots" content = "nofollow" / &gt; tag meta on a html page, can I add this to the pdf format? I can't seem to find where to add this code, or is there a bett

    Hi I need to hide a pdf file hosted on a site of glasses, I would normally add

    tag Meta on a html page, can I add this to the pdf format? I can't seem to find where to add this code, or is there a better way?

    You cannot add these metadata in PDF format. You can use the robots.txt file.

  • No Robots

    No Robots.txt file found in the folder root of your site.

    It was one of the error massages I got back from the web ceo online. SEO analysis to date, can someone me a Robots txt file what do I need and if I do, do I need one for each page. Can someone help please.

    Google "robots.txt".

    for example http://en.wikipedia.org/wiki/Robots_exclusion_standard

  • robots - please help

    Hi all

    I created a site submitted to search engines .and have a few questions that I have created a robot.txt file.

    is there a list of bad bots that I should include in my robots.txt file and where I can find this list?

    I can understand outside the URL on my robot.txt file - such as a directory included our site on their but we never submitted to this site, and although they have keywords, our site is not actually on their directory - I also tried to do a search on their site which is powered by google, but when you click on search - it is said that it s banned by google as violation of googles terms that me rendering still more worried.

    Also, how can I password protect my files - is possible for the files on the server to be hacked? Should I password protect the root folder at all?

    I would really appreciate the help as I can't find a clear answer anywhere.

    Concerning
    Lorna

    Hey Joe,

    Point, I know the thing on the email address - that the customer did not get his e-mail address on the Web site that she has an alias - and even if we removed the catch all the script on the hosting server - Yes your right she gets a lot of spam emails.

    But at least I know now that I don't do something wrong or not left out something and that I'm on the right track so that a relief.

    Thanks a lot for your help. Very appreciated
    Good day
    Lorna

  • How to select only the URLS in the same page

    I'm used to be able to select only the URL with right click of the entire page. After reinstalling Firefox, I can't use this feature more. Please let me know what kind of add-on, I need to have. Thank you

    I looked in the Wayback Machine for an older version of this page. He said:

    Multi links allows you to open, copy or favorite of several links at the same time rather than having to do them all individually.

    To open, links copy or bookmark, you simply right-click and do to drag a rectangle around the links to the desire to act. When you release the right mouse button, you will open/copy/bookmark these links.

    Note: The Wayback Machine does not cache downloads real extension due to the limitations of robots.txt on the site of modules.

    One of these current extensions could be partial replacement:

    I have not tried any of them myself.

  • Publicly Accessible Podcast, but not published or available?

    We produce an internal podcast we want to be accessible to the public, but we do not want everyone to be able to find it easily.  Of course, as soon as the link is made public, it is essentially out there, but we do not want people to be able to search for "xxx" and it fine in the iTunes store.

    Put the XML file on our site works, but not as well as it would if in the iTunes directory.

    Is possible to submit a PodCast to the iTunes store/repository/directory but wouldn't be indexed/search/listed?

    Thank you

    PittCaleb

    It is not possible - once a podcast in the store it is available and there is no way to prevent this.

    If you place the flow file on your server and use a robots.txt file to prevent the Ad search engines while the people who you give the URL will be able to access. (If you want to know how to do this, please ask - it's pretty simple).

    If you use http to the beginning of the URL, then access it open without the visitor, RSS reader: you can place the file in a web page (also blocked from search engines) with a note explaining how to subscribe to it in iTunes. If you start the URL with itpc instead of http, then Mac users who have access to the URL (again, better made starting from a link on a web page) will be automatically subscribed in iTunes. However, Windows users who have not installed iTunes get an error telling them to install it.

    By the way, if you want to protect the podcast (and each episode) is possible in iTunes (although the store does not).

  • That means Firefox by "broken certificate"?

    I have a site that gives security WARNING 'impossible to obtain the status of identification for a given site' and then do not grant an exception. But only on some machines. It's OK in both Linux running 10.0.4 machines that it fails, in another one with 10.0.4. In fact, with most of the machines I tried, it's OK, and I can give a marked exception as "unsaved."

    I have traced the error in the source code, but lost the trail to gBroken in exceptionDialog.js
    If I look at the certificate with OpenSSL, it parses OK as X 509 version 1.

    I narrowed down it somewhat.
    If I create a new profile, there is no problem.
    I had trouble to access any page on the server, for example https://example.com/robots.txt, https://example.com/foobar
    I found an entry in the cache for https://example.com/ which is in reality a 301 moved
    response, and that includes a long security-info: string in base 64.
    If I clear the cache, the problem goes away - I get the challenge and it can grant a waiver. If I restore the profile and the cache from a saved copy, the problem reappears. I don't am not sure of what represents the State stored in the cache, because I play too much with the certificate on the server and had not documented when I had changed.

    If I go to https://example.com/robots.txt and use the refill, or shift-reload (which, as I recall would normally get a new copy independently cached copies), it does not help.

    When the problem occurs, I can't retrieve the certificate using Firefox. There is no intermediate certificate, it is self-signed.

    I had this problem on 3 machines. On one, I cleared the cache without saving it first, that fixed the problem. A third, running SeaMonkey, I presume still has the problem, and I assume that it can be fixed by disabling the cache.

    (later)
    The question appears if a HTTPS redirection is cached, and then passes the server (self-signed) certificate
    Clear the entrance of the cache solves the problem.
    I filed https://bugzilla.mozilla.org/show_bug.cgi?id=767611

  • Mobile and desktop links in Google

    I created a website with Adobe Muse. When in Google on your desktop looking for people in our office, their website bios desktop and mobile versions are available. How can I make sure the mobile links are not available on the desktop?

    You must create the robots.txt file by using Notepad. Add to that the guidelines and upload it to your hosting, in the root directory of your site

  • iFrames and blocks direct URL access to certain pages

    I've been designing a site in Muse using Widgets (iFrames) of the Composition. Because the site that I create is quite large with updates and frequent changes, I won't be publish and download the entire site, whenever I have make a change or add content (like muse seems to force me to do, even when I change a single image). So as a work, within the iFrame, I insert HTML that links to a separate "mini site" Muse in another folder in my folder root (in a manner similar to adding a blog or Twitter feed into a Widget). This way I can make changes and don't have to publish and download small pieces of the larger site. These mini-sites conducted small, contents are incomplete in regard to corporate image and layout of Web site and I don't want people to access it directly, but I DON'T want the search engines to access information in their breast.

    So how can I block a public direct URL access to the "mini-sites" without blocking the site parent to access and display them in the iFrame? I want the user to be redirected to the site parent if a picks up more search engine "mini site" content, rather than being directed to the page itself.

    I know that I won't be able to this in the Muse, using Dreamweaver or any other editor is fine. I'm not fluent HTML, PHP, CSS, or any other language, but I can muddle my way through it, if I have the direction.

    I hesitate now to present a link to the test site I've created (even if I could do it on request), so I hope I have explained myself well enough.

    Thanks for any help.

    This has nothing to do with the Muse or elsewhere in the HTML. You would have to put in place a whole bunch of rules server side to transfer users and extract content from specific referents, but in the end, there are a lot of mumbo jumbo for nothing. Search engines can pretend to be browsers browsers can pretend to be stupid to caterpillars and even obscured links can be followed in any way. You should just press F12 and cross browser debugging console. I'm afraid that it is something that you really can do it properly, using a dynamic system where you can use the ID session PHP, cookies, or personalised channels encoded in your URL. In your scenario current all you can do is to use .htaccess and robots.txt files to block search engines to dig in your records, but they still don't sign up under your main domain name and not necessarily pass. On the other hand, since the search engine still has the URL of the folder, little sleuths like me could pull off, stick it in a separate window and then apply the view of the folder for the site when possible or browse your files based on the URL in the iFrame code or their names. What you want is fundamentally mutually exclusive and goes against the work of sites HTML static how.

    Mylenium

  • Prevent crawlers search of the help output engines

    I'm trying to understand how we can prevent our webhelp HTML parsing by the search engines like Google. I found these instructions by digging through some of the debates on this forum: Stop engine search bots indexing your folders private by 'robots.txt'. | Internet marketing Blog

    However, it seems that we could add code in the project itself in order to stop search engines to explore using. We tried to add this code in our master page from the master page is applied to all subjects, but the code does not stay after the output has been generated:

    < name meta = "robots" content = "NOINDEX, NOFOLLOW" / >

    Does anyone know how we can prevent search engines to explore our help?

    The masterpage heading will not work for this. Personally, I would also do a find and replace in the output. It's the fastest way.

    Remember that search engines index only not your from site of meta tags is a courtesy, it does not completely block bots. Only the good guys such as Google will hear it. Not even a robots.txt file will block crawlers. (See for example: learn more about robots.txt - files help search on Console) If you don't want really not any unauthorized access, you can force authentication on your server.

Maybe you are looking for