From d85895ebc1a91128aa479f14423170a78e3b6995 Mon Sep 17 00:00:00 2001 From: Alex Plischke Date: Mon, 9 Sep 2024 16:08:01 -0700 Subject: [PATCH] fix: performance anchors --- docs/performance/motion.md | 4 +++- docs/performance/transitions.md | 17 +++++++++++++++-- 2 files changed, 18 insertions(+), 3 deletions(-) diff --git a/docs/performance/motion.md b/docs/performance/motion.md index 4ea36c4dd0..6c6133e115 100644 --- a/docs/performance/motion.md +++ b/docs/performance/motion.md @@ -18,7 +18,9 @@ In addition to testing the efficiency of your app rendering, Sauce Labs also pro ### What You'll Need - Google Chrome (no older than 3 versions from latest) -- Test configuration must have performance enabled. See [Set Performance Capabilities](/performance/transitions#set-performance-capabilities) for instructions. +- Test configuration must have performance enabled. + See [Set Performance Capabilities](/performance/transitions/#setting-performance-capabilities) + for instructions. ## What is Jankiness? diff --git a/docs/performance/transitions.md b/docs/performance/transitions.md index 10ea7a10fa..ad1826ba52 100644 --- a/docs/performance/transitions.md +++ b/docs/performance/transitions.md @@ -100,7 +100,13 @@ options = {browser_name: browser_name, ## Implementing the Performance Command Assertion -The custom `sauce:performance` command measures the performance output against a baseline of previously accepted performance values. If no baseline has been set, the Performance test will create one by measuring performance output 10 times to get an aggregate baseline. The command returns `pass` when the current results are within the baseline allowances or `fail` when the results fall outside the baseline. A fail result gives you the option to handle [regressions](#handle-regressions). +The custom `sauce:performance` command measures the performance output against +a baseline of previously accepted performance values. If no baseline has been +set, the Performance test will create one by measuring performance output 10 +times to get an aggregate baseline. The command returns `pass` when the current +results are within the baseline allowances or `fail` when the results fall +outside the baseline. A fail result gives you the option to handle +[regressions](#handling-regressions). :::caution Enabling performance capturing can add up to 60 seconds per URL change in a test. We, therefore, advise separating your performance tests from your functional tests. See our [Performance Requirements and Recommendations](https://docs.saucelabs.com/performance/about/#sauce-performance-requirements-and-recommendations) for more advice on optimizing your performance test results. @@ -268,7 +274,14 @@ performanceLogs[metric] < value`metric ${metric} is over the performance budget` ## Handling Regressions -When one or more metric evaluations fail because the result falls outside the established baseline, it is considered a regression and the tester has an option to either troubleshoot and resolve the source of the regression to get the test back into the baseline range or [update the baseline](/performance/analyze#reset-baselines-for-a-failed-test) with the new performance values. If new baselines are accepted, the command will measure performance against those new values until another regression is detected, when you will again have the option to troubleshoot or update the baselines. +When one or more metric evaluations fail because the result falls outside the +established baseline, it is considered a regression and the tester has an option +to either troubleshoot and resolve the source of the regression to get the test +back into the baseline range or [update the baseline](/performance/analyze/#resetting-baselines-for-a-failed-test) +with the new performance values. If new baselines are accepted, the command will +measure performance against those new values until another regression is +detected, when you will again have the option to troubleshoot or update the +baselines. Since the command can be called throughout the test script, create tests that check for performance regressions across core business flows and screens. For example, evaluate pages that load following a successful login event or require multiple steps to trigger.