# encoding: UTF-8 shared_examples 'WebSite::RobotsTxt' do let(:known_dirs) { WebSite::RobotsTxt.known_dirs } describe '#robots_url' do it 'returns the correct url' do expect(web_site.robots_url).to eql 'http://example.localhost/robots.txt' end end describe '#has_robots?' do it 'returns true' do stub_request(:get, web_site.robots_url).to_return(status: 200) expect(web_site.has_robots?).to be_truthy end it 'returns false' do stub_request(:get, web_site.robots_url).to_return(status: 404) expect(web_site.has_robots?).to be_falsey end end describe '#parse_robots_txt' do context 'installed in root' do after :each do stub_request_to_fixture(url: web_site.robots_url, fixture: @fixture) robots = web_site.parse_robots_txt expect(robots).to match_array @expected end it 'returns an empty Array (empty robots.txt)' do @fixture = fixtures_dir + '/robots_txt/empty_robots.txt' @expected = [] end it 'returns an empty Array (invalid robots.txt)' do @fixture = fixtures_dir + '/robots_txt/invalid_robots.txt' @expected = [] end it 'returns some urls and some strings' do @fixture = fixtures_dir + '/robots_txt/invalid_robots_2.txt' @expected = %w( /ÖÜ()=? http://10.0.0.0/wp-includes/ http://example.localhost/asdf/ wooooza ) end it 'returns an Array of urls (valid robots.txt)' do @fixture = fixtures_dir + '/robots_txt/robots.txt' @expected = %w( http://example.localhost/wordpress/admin/ http://example.localhost/wordpress/wp-admin/ http://example.localhost/wordpress/secret/ http://example.localhost/Wordpress/wp-admin/ http://example.localhost/wp-admin/tralling-space/ http://example.localhost/asdf/ ) end end context 'installed in sub directory' do it 'returns an Array of urls (valid robots.txt, WP installed in subdir)' do web_site_sub = WebSite.new('http://example.localhost/wordpress/') fixture = fixtures_dir + '/robots_txt/robots.txt' expected = %w( http://example.localhost/wordpress/admin/ http://example.localhost/wordpress/secret/ http://example.localhost/Wordpress/wp-admin/ http://example.localhost/wp-admin/tralling-space/ http://example.localhost/asdf/ ) stub_request_to_fixture(url: web_site_sub.robots_url, fixture: fixture) robots = web_site_sub.parse_robots_txt expect(robots).to match_array expected end end end describe '#known_dirs' do it 'does not contain duplicates' do expect(known_dirs.flatten.uniq.length).to eq known_dirs.length end end end