Setting up robots.txt for 1C-Bitrix

Our company is engaged in the development, support and maintenance of Bitrix and Bitrix24 solutions of any complexity. From simple one-page sites to complex online stores, CRM systems with 1C and telephony integration. The experience of developers is confirmed by certificates from the vendor.
Our competencies:
Development stages
Latest works
  • image_website-b2b-advance_0.png
    B2B ADVANCE company website development
    1175
  • image_bitrix-bitrix-24-1c_fixper_448_0.png
    Website development for FIXPER company
    811
  • image_bitrix-bitrix-24-1c_development_of_an_online_appointment_booking_widget_for_a_medical_center_594_0.webp
    Development based on Bitrix, Bitrix24, 1C for the company Development of an Online Appointment Booking Widget for a Medical Center
    564
  • image_bitrix-bitrix-24-1c_mirsanbel_458_0.webp
    Development based on 1C Enterprise for MIRSANBEL
    747
  • image_crm_dolbimby_434_0.webp
    Website development on CRM Bitrix24 for DOLBIMBY
    655
  • image_crm_technotorgcomplex_453_0.webp
    Development based on Bitrix24 for the company TECHNOTORGKOMPLEKS
    976

robots.txt Setup for Bitrix CMS

robots.txt controls which site sections search robots can crawl and which cannot. For typical Bitrix site, standard robots.txt out-of-the-box is usually absent or contains only basic bans. Proper setup saves crawl budget and excludes technical pages from index.

What to Block for Bitrix Site

Technical paths that shouldn't be indexed:

Disallow: /bitrix/
Disallow: /upload/iblock/
Disallow: /auth/
Disallow: /personal/
Disallow: /cart/
Disallow: /order/
Disallow: /search/?
Disallow: /catalog/*?*
Disallow: /?*

/bitrix/ — admin panel, core, cache. /upload/iblock/ — original images. Enough for search engine to index images via product pages, not direct URLs. /catalog/*?* — catalog pages with parameters (sorting, pagination). Prevents duplicate indexing.

Management via Administrative Panel

Bitrix has no built-in visual robots.txt editor. File is edited directly via Content → Site Structure → [root] → robots.txt, or via FTP/SSH.

The seo module adds sitemap link automatically when generating sitemap. Check that end of file has line:

Sitemap: https://your-site.ru/sitemap_index.xml

Directives for Different Search Engines

Yandex and Google support User-agent, Disallow, Allow, Crawl-delay and Sitemap directives. Different rules can be set for different robots:

User-agent: Yandex
Disallow: /catalog/*?*
Crawl-delay: 1

User-agent: Googlebot
Disallow: /catalog/*?*

User-agent: *
Disallow: /bitrix/
Disallow: /upload/tmp/
Disallow: /auth/

Crawl-delay: 1 for Yandex — 1 second delay between requests. Use for weak server sites to prevent crawler overloading hosting.

Verification

After changes, check file via webmaster tools:

  • Yandex.Webmaster → Tools → Analyze robots.txt
  • Google Search Console → Tools → Test file robots.txt

Tools show if specific URL is blocked for specific robot.

Implementation Timeline

Drafting and setting up robots.txt for typical Bitrix site — 30–60 minutes.