robots.txt and sitemap.XML

Hi all

are there tools in QC to create and deploy the sitemap.xml and the robots.txt file?

Thanks in advance,

Michael

Hi Michael,

Unfortunately, there is no automated for that yet in QC tool.

If you want to build something like this yourself, the best is to create a page somewhere in the structure of your site that is not displayed in the navigation (you would nominate this page "site map" or "robots" - without the extension). You must also create a component of page customized for each of these two types of special pages, where you control the output is exactly what you need (ie. no HTML code in the output)-for a first version, simply hard-code the content of each in JSPS, later, you can then make more sophisticated (eg. giving reading a page property or the iteration on the structure of your page). Using light of CRXDE, you obviously have to change ownership of Sling: resourceType of your page sitemap or robots to point to the new components of the appropriate page.

To make these pages available on the root path, you can use the/etc/map parameters. Such a configuration is as follows in the case of the robots.txt file (CRXDE light to create the node and it's properties):

card/etc/www-robots.txt

JCR:primaryType = "Sling: Mapping ' (ie the type when you create a new node)

Web: internalRedirect = ' / content/SITE/PATH/robots.html '.

scarf: match = "http/www.SITE.com/robots.txt.

BTW, this is the documentation of mapping: http://sling.apache.org/site/mappings-for-resource-resolution.html

Hope that helps!

Gabriel

Tags: Adobe

Similar Questions

  • I have a paid site (it wasn't a demo for a few months) but robots.txt has this "# this file is automatically generated that your site is being tested." To remove it and to allow search engine indexing, please move your site to a paid plan. »

    I have a paid site (it wasn't a demo for a few months) but robots.txt has this "# this file is automatically generated that your site is being tested." To remove it and to allow search engine indexing, please update your site to a paid plan. ». How can I fix it?

    You should be able to create and upload a new file robtos.txt. Create a txt file and call robots. Put the content in the robots.txt file:

    User-agent: *.

    Allow: /.

    Site map: http://www.your-domian-goes-here.com/sitemap.xml

    Once created, download the robots.txt file in the root of your site.

  • Allowing resources Robot.txt

    I launched my site and it works great.  However, when I test the usability in googles webmaster tools they say that there are some resources that are refusing the analysis via robot.txt.  How can I change the settings to allow easier parsing through robots of google and Bing? I have presented a plan sitemap, but still need to resolve the deadlock.

    Thanks in advance.

    James

    Hello

    There is no robots.txt file in your root folder.

    http://www.heritagejewelryandloan.com/robots.txt

    You can manually create an and download to your host and come back in the Google Search (webmaster) console.

    You can take the little help from here the Web Robots Pages the works of the robots.txt file.

    Let me know if you have any question.

  • I need the config of the robots.txt file

    OK I am running hmail Server im trying to webmail, part I need to config the robots.txt file, how can I do. also I can get the newspaper but after I log in I get 404 page not found what can I do

    http://www.hMailServer.com/Forum/

    Have you tried to ask in their Forums at the link above.

    http://www.hMailServer.com/index.php?page=support

    And they also have technical support Options.

    See you soon.

    Mick Murphy - Microsoft partner

  • Presentation of sitemap.xml for Google

    Hello

    I just submitted my Sitemap to Google search Console. I did this by typing in my field, followed by sitemap.xml

    He says now that I have submitted 10 web pages:

    Screen Shot 2016-01-28 at 4.51.45 PM.png

    What does that mean?!?! Is this correct?

    Thanks in advance for the clarification.

    Your site has variations of office, Tablet and phone for 4 pages. By my calculations it is 12 HTML pages. Difficult to know exactly what Google account, but maybe your homepage for 3 devices is considered as a single page, but all the rest in the form of single pages?

  • Problem with Sitemap.XML do not regenerate

    Hello

    I recently built a Web site and I am trying to download my sitemap to google webmaster tools.

    Initially, my Sitemap was showing the wrong pages so I reuploaded the site with all the files.

    I disabled the Sitemap using utilities and I then activated the sitemap. He said it will take a day to the catylist company regenerating the sitemap.

    He was not close to 48 hours and I don't have a sitemap for my site yet-

    http://edenharlow.com/sitemap.XML

    I get the following error

    XML Parsing Error: not well-formed

    Location: http://www.edenharlow.com/sitemap.XML

    Line number 1, column 1: / * not * /.

    ^

    Does anyone have ideas, what the question is here?

    Kind regards

    Callum

    Hi Callum

    I checked the ticket you submitted BC, the ticket is up now and you will get an update soon.

    There seems to be some problem with sitemap.

    Thank you

    Sanjit

  • Robots.txt

    Hi there I just added a robots.txt on my site because I don't want to read some files from google, the problem is I have a 100 or so these index.html files on my site, they are in files separate album that I use for my carpet samples and they do not need to be indexed, the problem is even with the file are pages containing the index.html file when opened and they need to be indexed so I want to block the index files, but not the rest.

    This is a link to one of the pages that contain the files: http://www.qualitycarpets.net/samples-carpet-berber/agadir-berber-carpet.php

    If anyone can tell me how to write the text to do this, I will be very greatfull.

    Thank you, Jeff Lane

    If this should not be...

  • How can integrate phone pages to sitemap.xml

    I created a web page with Adobe Muse with plataform desk and telephone.  The problem I have is that the extract of phone did not appear in sitemap.xml

    I need to know if I did something wrong or this could really happen it's new software.

    Is it possible to contact you by phone?

    Everyone speeks spanih?

    The main problem is that Google don't bed not the phone as the sitemap.xml plataform does not a difference between the office and phone plataform as both have the same name in the HTML code.

    I'm not supposed to change any information that the pages come as default.

    Help!  from Chile, South America

    Stefano

    Sorry for the delay in response,

    Could you please share your with us .muse file to study? Please send it to [email protected]. If your file is larger than 20 MB, you can use something like Adobe SendNow SendThisFile or Dropbox. Don't forget to mention the link to this forum thread in your email so that we can identify the file.

  • Is there a way to create a sitemap xml in Dreamweaver 2014?

    I was wondering if there is a way to create a sitemap xml Dreamweaver 2014 inside.

    No automagic for her button.  You could do this manually in code view, but there are generators online and desktop applications that are better suited to the task.

    Create your Google Sitemap Online - XML Sitemaps Generator

    Xenu Link Sleuth can produce maps of site too.

    Find broken links on your site with Link Sleuth (TM) of Xenu

    Nancy O.

  • I am looking for a SIMPLE way to add my robots.txt file

    I read up on this subject. But I wasn't able to do it again.

    I found this:

    1. re: file robots.txt for sites of the Muse

    Vinayak_GuptaEmployee Hosts

    You can follow the guidelines of Google to create a robots.txt file and place it in the root of your remote site.

    https://support.Google.com/webmasters/answer/156449?hl=en

    Thank you

    Vinayak

    ----------------

    Once you create a robots.txt like this:

    user-agent: ia_archive

    Disallow: /.

    (1) where do you put 'head labels? Do you need them?

    I would insert it in muse that looks like this

    < head >

    user-agent: ia_archive

    Disallow: /.

    < / head >

    or just put it anywhere inside the 'head' tag

    (2) put the robots.txt file in a folder?

    I've heard this, but it just doesn't seem right.

    (3) OK you have the

    -Properties Page

    -Metadata

    -HTML < head >

    Can I copy and paste my right robot.txt info in there? I don't think I can and make it work. According to the info I found (that I posted above), the robots.txt 'file' you 'place at the root of your remote site.

    (4) where is the 'root of my remote sit?

    How can I find that?

    I read other people having problems with this.

    I thank very you much for any help.

    Tim

    I need Terry White to make a video on it LOL

    Maybe I'll ask.

    I thought about it.

    However, with the help of Godaddy, the file was not placed between theand, so I'm still a little nervous.

    It is recommended to:

    ///////////////////////////////////

    1.  re: file robots.txt for sites of the Muse

      

       Vinayak_Gupta   , April 19, 2014 01:54 (in response to chuckstefen)

    You can follow the guidelines of Google to create a robots.txt file and place it in the root of your remote site.

    https://support.Google.com/webmasters/answer/156449?hl=en

    Thank you

    Vinayak

    /////////////////////////////////////

    Place the robots.txt file to the root "of your remote site.

    and that (Godaddy) is not between the

    I checked the robot file that I created here

    New syntax Robots.txt Checker: a validator for robots.txt files

    and other than me not to capitalize

    the 'u' in the user-agent, it seems to work. When my site is analyzed, she does not miss a robots.txt file

    user-agent: ia_archive

    Disallow: /.

    Problem solved, unless I find an easy way to place the robots.txt file placed between the head tags (and better).

    I'll keep my ears open, but don't worry too much on this subject.

    Step 1) write the code of robots that you want to

    Step 2) save the file as a txt

    Step 3) contact your Web hosting provider / upload in the root of the Web site in a single file

    Step 4) check with a robot's checker that I listed above

    What was shake me:

    -where to put

    -the difference between files and folders, it seemed I would be to load a file for some reason any.

    -I was expecting something like the list of the news LOL

  • Submit files adobe muse site sitemap.xml for google?

    Can someone help me to submit my site for google.

    I realize adobe muse automatically creates a sitemap.xml file, but I don't know how to present this to Google?

    Is it possible to download the sitemap.xml file in the folder root of the sites of the muse.

    Thank you.

    google-sitemap-error.jpg

    Thus, the URL Sitemap of your site is: http://pixelmania.com.au/sitemap.xml and it will continue to be only. And as I mentioned earlier, this is the link that you can use when sending to the Webmasters. However, the plan of the site that this link is currently displayed is incorrect. So, I am offering measures to regeneate the good the site map for your site.

    1. go to this URL and connection: http://pixelmania.com.au/AdminConsole/Account/Login#! /Utilities/SEO.aspx

    Note that this page is not accessible directly from the administration console, and it's the only way to access it from the V3 administration console, so you can keep this useful link.

    2. once there, use the option that says 'Disable Sitemap'. Once disabled, check with the site map link that it is disabled and you no longer see a sitemap when access you the link.

    3. after the confirmation of this, activate the sitemap again of the same section of your administrative console and then wait for 24 to 36 hours, after which, when you click on the site map link, it will show you the new, correct sitemap.

    4. Once you see the new site map, proceed to upload on webmasters.

    I hope this helps.

    See you soon

    Parikshit

  • Rename sitemap.xml

    I have several subdirectories on my site.

    I use MUSE to redo each subdirectory. MUSE creates a file called sitemap.xml for each subdirectory

    Google will get confused if there are multiple sitemap files even if they are in the subdirectories differenent.

    How can I rename the sitemap.xml for each subdirectory in MUSE

    Thank you

    K

    So if I get this straight, you're creating several sites of Muse (with the same layout and design) and exporting to their individual directories to create a unique site.

    For now, there's an option to disable the creation of sitemap.xml when exporting the site in HTML or by using the built-in FTP download dialog. The only way around this would be to export the site in HTML format and manually manage sitemap.xml files before loading your third party hosting site using a FTP client.

    Guidelines for Sitemaps Google - https://support.google.com/webmasters/answer/183668?hl=en#1.

    If you have more than one Sitemap, you can list them in a Sitemap index file and then submit the Sitemap index file to Google. You don't need to submit each file individually.

    And here's how you create an index file Sitemap to make reference to several Sitemaps - https://support.google.com/webmasters/answer/71453.

    Thank you

    Vinayak

  • Sitemap.XML not updated

    Hi, on the site www.arcadiaceramiche.it every time I change Business Catalyst site, I can't export to this topic the sitemap updated; otherwise if I export the HTML code on the local drive, I extract the correct sitemap.

    What why?

    When you use FTP Upload or export HTML in Muse, Muse will generate a sitemap.xml file. Which edition to Business Catalyst the sitemap.xml file is created by Business Catalyst and it can take up to 24 hours for Business Catalyst update the sitemap.xml file.

    A similar thread here that can be useful - http://forums.adobe.com/message/5048293#5048293

    Thank you

    Vinayak

  • 2 movieClips and 1 xml

    Dear Sir

    I have 2 movieClip (sublink + TXT)

    My xml file:

    < name of the submenu "Company profile" = >
    < name of links = 'Our strategy' > COMMTECH strategy aims to provide the Egyptian market with the technology solutions State-of-The-Art through the most worthy of trust and reputable suppliers and technology providers in the world.
    Our role is to provide information and new communications innovative technology solutions to improve and develop small and medium-sized businesses IT infrastructure network capacity.
    We complete our strategy through high quality, reliable, effective and innovative high solutions for our clients, while providing full technical support customers to control and manage their IT resources thanks to integrate and standardize several key components to improve the SME IT business processes. < / links >
    < name of links = "Our Vision" > our Vision < / links >
    < name of links = 'Our Mission' > our Mission < / links >
    < name of links = "Our commitment" > our commitment < / links >
    < / submenu >

    I would like to know how to create a link to an xml file to Flash when I press on one of the links of the thesis to give me a text in the same movieclip "TXT".

    I'd appreciate your help...!

    Kind regards

    MELAS

    This function should be on the chronology of _root.all.sublink.subTXT.

  • Why my Sitemap XML plan has detected two versions of my homepage URL

    I am new Web Designer and just finished building my first customer main site using Dreamweaver CS4.

    I used www.xml-sitemaps.com to automatically build my XML sitemap and also to submit my site for Google Analytics.

    In doing so, it seems that plans Sitemap and Google Analytics see two versions of the original URL:

    www.euro-peanflooring.co.UK/

    www.euro-peanflooring.co.uk/index.htm

    Here is the code that generated sitemaps for me:

    < url >
    < loc > http://www.euro-peanflooring.co.UK/ < / loc >
    < lastmod > 2009-12-18T 11: + 00:00 48:07 < / lastmod >
    < changefreq > monthly < / changefreq >
    < priority > 1,00 < / priority >
    < / url >
    < url >
    < loc > http://www.euro-peanflooring.co.uk/index.htm < / loc >
    < lastmod > 2009-12-18T 11: + 00:00 48:07 < / lastmod >
    < changefreq > monthly < / changefreq >
    < priority > 0.80 < / priority >
    < / url >

    My question is I did nothing wrong to the above tools to recognize the two versions of the original URL?

    I checked the sitemaps xml to reputable sites and they don't have two versions of the original URL.

    Maybe, I have a problem with the structure of my root directory?

    Thank you in advance for help. I don't understand what is happening.

    I have no idea why it doesn't work. This method is used by ALL the sites I design.

    Check out my site: www.theindesignguy.com and go to the portfolio page where you will find links to some of my sites. Check out the code on those, and you'll see that they all use the / instead of index.xxx.

    You can post a link to the site that does not work?

    Bob

Maybe you are looking for

  • Impossible to delete the mail.

    I recently got e-mail from someone I know watch USA... but it cannot be deleted. The message of the computer says... the 11416.emix file could not be opened because there is no file.  I can read the mail, so I don't need to open it again; I now just

  • How to recover the music Apple folder in iTunes?

    I just nuked my Mac and copied my username on an external drive first folder. Music bought and torn is in place but I can't the files downloaded music Apple back on the computer. It seems that it is something, but the number of songs does not change.

  • How to trace the signal acquired against the clock all the time fixed by using Graph XY?

    Hello My goal is to trace the signal acquired against the clock all the fixed time dynamically. In General, it is like trace the waveform of the signal against the clock in a graph, but don't keep the recent 10 min waveform. My current approach is to

  • What stack overflow through the line 26 and 27?

    Periodically, I get this message that I have an on line 26 and 27 stack overflow, and I can't do anything until I, stopping and restarting. What is c? I can solve this problem?

  • Interface not to ping

    Hello community, I have a problem that I just don't understand - although I do not know there is a simple explanation. R3 can not ping itself on 172.28.38.11/16 and I do not understand why. I appreciate R2 has of an interface on Eth 1/2 with ip addre