• Resolved leonardo1986

    (@leonardo1986)


    Hello I am having this issue. this is my robots.txt file.
    I am triying to block specific pages like “contact-v1” and “contact-v2” because I don’t want to be shown in google searches or any other. I uploades the new robot and tested it in search console and still allowed them. Google bot is not blocking them. I need to block 6 pages… contact, blog, terms of use, privacy policy and other. Google is indexing me pretty bad and not indexes my home page even Yoast SEO metrics are fully green..!!! Any ideas..??

    # Bloquear acceso a paginas secundarias
     # I tried to block them here
    
    User-Agent: *   
    Disallow: /contact-v1/
    Disallow: /contact-v2/
    
    # Bloquear o permitir acceso a contenido adjunto
    
    User-Agent: *   
    Allow: /wp-content/uploads/
    Disallow: *?replytocom
    Disallow: /cgi-bin
    Disallow: /wp-content/plugins/   
    Disallow: /wp-content/themes/   
    Disallow: /wp-includes/   
    Disallow: /wp-admin/
    
    # Desindexar carpetas que empiecen por wp-
    
    Disallow: /wp-
    
    # Permitir sitemap pero no las busquedas.
    
    Sitemap: https://falcopc.com/sitemap_index.xml
    Disallow: /?s=   
    Disallow: /search
    
    # Permitir Feed general para Google Blogsearch.   
    # Impedir que /permalink/feed/ sea indexado pues el feed de comentarios suele posicionarse antes de los post.
    # Impedir URLs terminadas en /trackback/ que sirven como Trackback URI (contenido duplicado).   
    
    Allow: /feed/$
    Disallow: /feed   
    Disallow: /comments/feed   
    Disallow: /*/feed/$   
    Disallow: /*/feed/rss/$   
    Disallow: /*/trackback/$   
    Disallow: /*/*/feed/$   
    Disallow: /*/*/feed/rss/$   
    Disallow: /*/*/trackback/$   
    Disallow: /*/*/*/feed/$   
    Disallow: /*/*/*/feed/rss/$   
    Disallow: /*/*/*/trackback/$
    
    # Evita bloqueos de CSS y JS.
    
    Allow: /*.js$
    Allow: /*.css$
    
    # Lista de bots que deberias permitir.
    
    User-agent: Googlebot-Image
    Allow: /wp-content/uploads/
    
    User-agent: Adsbot-Google
    Allow: /
    
    User-agent: Googlebot-Mobile
    Allow: /
    
    # Lista de bots que generan consultas abusivas aunque siguen las pautas del archivo robots.txt
    
    User-agent: MSIECrawler
    Disallow: /
    
    User-agent: WebCopier
    Disallow: /
    
    User-agent: HTTrack
    Disallow: /
    
    User-agent: Microsoft.URL.Control
    Disallow: /
    
    User-agent: libwww
    Disallow: /
    
    User-agent: Orthogaffe
    Disallow: /
    
    User-agent: UbiCrawler
    Disallow: /
    
    User-agent: DOC
    Disallow: /
    
    User-agent: Zao
    Disallow: /
    
    User-agent: sitecheck.internetseer.com
    Disallow: /
    
    User-agent: Zealbot
    Disallow: /
    
    User-agent: MSIECrawler
    Disallow: /
    
    User-agent: SiteSnagger
    Disallow: /
    
    User-agent: WebStripper
    Disallow: /
    
    User-agent: Fetch
    Disallow: /
    
    User-agent: Offline Explorer
    Disallow: /
    
    User-agent: Teleport
    Disallow: /
    
    User-agent: TeleportPro
    Disallow: /
    
    User-agent: WebZIP
    Disallow: /
    
    User-agent: linko
    Disallow: /
    
    User-agent: Xenu
    Disallow: /
    
    User-agent: larbin
    Disallow: /
    
    User-agent: libwww
    Disallow: /
    
    User-agent: ZyBORG
    Disallow: /
    
    User-agent: Download Ninja
    Disallow: /
    
    User-agent: wget
    Disallow: /
    
    User-agent: grub-client
    Disallow: /
    
    User-agent: k2spider
    Disallow: /
    
    User-agent: NPBot
    Disallow: /
    
    User-agent: WebReaper
    Disallow: /
    
    # Slurp (Yahoo!), Noxtrum y el bot de MSN que suelen generar excesivas consultas.
    
    User-agent: noxtrumbot
    Crawl-delay: 50
    
    User-agent: msnbot
    Crawl-delay: 30
    
    User-agent: Slurp
    Crawl-delay: 10
    
    # Previene problemas de recursos bloqueados en Google Webmaster Tools
    User-Agent: Googlebot
    Allow: /*.css$
    Allow: /*.js$

    The page I need help with: [log in to see the link]

Viewing 1 replies (of 1 total)
  • Plugin Author Sybre Waaijer

    (@cybr)

    Hi there!

    You’re posting in the wrong forum ?? This one is for The SEO Framework; you’re questioning for Yoast SEO. Nevertheless, I’ll explain what’s happening to your site.

    In robots.txt, duplicated user agent groups are generally seen as erroneous. You have doubled the * user agent group. It also takes time for Google to crawl and index the robots.txt file. You can learn more about your file via the link below:
    https://support.google.com/webmasters/answer/6062598

    Now, I generally condone adjusting the robots.txt file and suggest using meta tags and HTTP errors instead. This is because WordPress is a dynamic CMS, and one page may have multiple endpoints.

    As for your website ranking badly, there are two main reasons.

    Firstly, the load-time of your website makes it practically unusable; it’s horrendous. So, visitors bounce quickly, and Google picks up on that as a negative signal. I highly doubt the reported bounce rate is lower than 80% because of this. I recommend switching to a quality host, and you should remove all the bloat from your site.

    Secondly, you’ve killed your website by enabling nofollow on one of your main pages (/shop/), so robots won’t be able to find the products on your website. I believe you’ve been misinformed on some settings in the SEO plugin you’re using.

    In any case, I’m marking this topic as resolved as you’ve misplaced your question. You may get more useful answers to your questions regarding Yoast SEO here.

    Good luck. Cheers!

Viewing 1 replies (of 1 total)
  • The topic ‘Robots.txt is not blocking secondary pages’ is closed to new replies.