Skip to content

Latest commit

 

History

History
executable file
·
144 lines (131 loc) · 4.5 KB

README.md

File metadata and controls

executable file
·
144 lines (131 loc) · 4.5 KB

Behat SEO Contexts

Latest Version Build Status Quality Score

Behat extension for testing some On-Page SEO factors.

Includes contexts for testing:

  • title / meta description
  • canonical
  • hreflang
  • meta robots
  • robots.txt
  • indexation: tests meta robots + robots.txt + X-Robots-Tag header
  • redirects
  • sitemap validation (inc. multilanguage)
  • HTML validation
  • assets performance
  • accessibility
  • UX
  • more...

Installation

Basic requirements:

  • PHP 7.1+
  • Behat 3+
  • Mink + Mink extension

How to install it

  1. Install Composer
  2. Execute:
$ composer require marcortola/behat-seo-contexts --dev
  1. Add the Context you need to behat.yml:
# behat.yml
default:
    # ...
    suites:
        default:
          contexts:
            - MarcOrtola\BehatSEOContexts\Context\MetaContext
            - MarcOrtola\BehatSEOContexts\Context\LocalizationContext
            - MarcOrtola\BehatSEOContexts\Context\RobotsContext
            - MarcOrtola\BehatSEOContexts\Context\IndexationContext
            - MarcOrtola\BehatSEOContexts\Context\RedirectContext
            - MarcOrtola\BehatSEOContexts\Context\SitemapContext
            - MarcOrtola\BehatSEOContexts\Context\HTMLContext
            - MarcOrtola\BehatSEOContexts\Context\PerformanceContext
            - MarcOrtola\BehatSEOContexts\Context\SocialContext
            - MarcOrtola\BehatSEOContexts\Context\AccessibilityContext
            - MarcOrtola\BehatSEOContexts\Context\UXContext

Featured steps

MetaContext
Then the page canonical should not be empty
Then the page canonical should be :expectedCanonicalUrl
Then the page title should not be empty
Then the page title should be :expectedTitle
Then the page meta description should not be empty
Then the page meta description should be :expectedMetaDescription
Then the page meta robots should be noindex
Then the page meta robots should not be noindex
LocalizationContext
Then the page hreflang markup should be valid
RobotsContext
Given I am a :crawlerUserAgent crawler
Then I should not be able to crawl :resource
Then I should be able to crawl :resource
Then I should be able to get the sitemap URL
IndexationContext
Then the page should be indexable
Then the page should not be indexable
RedirectContext
Given I follow redirects
Given I do not follow redirects
Then I should be redirected to :url
SitemapContext
Given the sitemap :sitemapUrl
Then the sitemap should be valid
Then the index sitemap should be valid
Then the multilanguage sitemap should be valid
Then the index sitemap should have a child with URL :childSitemapUrl
Then /^the sitemap should have ([0-9]+) children$/
Then the multilanguage sitemap should pass Google validation
Then the sitemap URLs should be alive
Then /^(\d+) random sitemap URLs? should be alive$/
HTMLContext
Then the page HTML markup should be valid
Then /^the page HTML5 doctype declaration should (not |)be valid$
PerformanceContext
Then /^browser cache should be enabled for (.+\..+|external|internal) (png|jpeg|gif|ico|js|css) resources$/
Then /^Javascript code should load (async|defer)$/
Then HTML code should be minified
Then CSS code should be minified
Then Javascript code should be minified
Then CSS code should load deferred
Then critical CSS code should exist in head
SocialContext
Then /^the (Twitter|Facebook) Open Graph data should satisfy (minimum|full) requirements$/
AccessibilityContext
Then the images should have alt text
UXContext
Then the site should be responsive
Then the site should not be responsive

Examples

This library is self-tested, and you can find examples inside the features directory. Feel free to explore it to discover each step definition.

Useful tips