import languages from '../../lib/languages.js' import robotsParser from 'robots-parser' import { get } from '../helpers/e2etest.js' import { jest } from '@jest/globals' describe('robots.txt', () => { jest.setTimeout(5 * 60 * 1000) let res, robots beforeAll(async () => { res = await get('/robots.txt', { headers: { Host: 'docs.github.com', }, }) robots = robotsParser('https://docs.github.com/robots.txt', res.text) }) it('allows indexing of the homepage and English content', async () => { expect(robots.isAllowed('https://docs.github.com/')).toBe(true) expect(robots.isAllowed('https://docs.github.com/en')).toBe(true) expect( robots.isAllowed('https://docs.github.com/en/articles/verifying-your-email-address') ).toBe(true) }) it('allows indexing of generally available localized content', async () => { Object.values(languages).forEach((language) => { expect(robots.isAllowed(`https://docs.github.com/${language.code}`)).toBe(true) expect( robots.isAllowed( `https://docs.github.com/${language.code}/articles/verifying-your-email-address` ) ).toBe(true) }) }) it('disallows indexing of azurecontainer.io domains', async () => { const res = await get('/robots.txt', { headers: { host: 'docs-internal-preview-12345-asdfz.azurecontainer.io', }, }) expect(res.text).toEqual('User-agent: *\nDisallow: /') }) it('does not have duplicate lines', () => { const lines = new Set() for (const line of res.text.split('\n')) { if (/^\s*$/.test(line)) continue expect(lines.has(line)).toBe(false) lines.add(line) } }) })