A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from https://docs.codecov.com/docs/automated-test-selection below:

[Beta] Automated Test Selection

📘

Automated Test Selection is in Open Beta for Python

Automated test selection currently only works for Python, but more languages are on the way!

Codecov's Automated Test Selection works within your CI pipeline to identify tests that are sufficient to execute the diff of your change. Use the following guides to get setup with Automated Test Selection in your pipeline of choice.

Automated Test Selection is a Codecov feature that ensures your CI pipeline only runs the tests needed to completely exercise the diff of a commit. Automated Test Selection is intended to greatly reduce the time needed to run test suites without compromising on test coverage or having to invest in creating the complex CI pipelines and tools (e.g., Bazel) that are typically required to facilitate test selection.

Automated Test Selection is comprised of 3 parts that work together to ensure your commit diff is properly exercised:

Testing tools typically have a stage of collecting tests followed by a stage of executing tests. The idea behind Automated Test Selection is to live exactly between those 2 stages as a filter to decide which tests necessarily need to be executed, considering the diff of the commit being tested.

The general flow of usage of Automated Test Selection is the following:

  1. The CI runs static analysis in the new commit and uploads that information to Codecov
  2. The tests names (labels) are collected and that list is uploaded to Codecov, along with the commit we’re comparing the current HEAD to (usually HEAD^). The commit “compared to” is referred to as BASE
  3. Codecov will run an analysis on the labels sent leveraging the git diff HEAD-BASE, static analysis information and previous test runs to generate a minimal subset of those labels that are enough to fully exercise the diff.
  4. Codecov sends this subset of labels back to the CI that then executes the tests
  5. Test coverage is uploaded to Codecov with labels context

Some things to note about the process outlined above:

Label-analysis is not perfect, so it might not be possible for us in some cases to be certain if a given tests needs to be run or not. We only tell you not to run a given test if we are 100% certain it doesn't need to be run.

This means that in some scenarios we may tell you to run more tests than you actually need to run.

Right now Automated Test Selection is still in active development, so things are changing fast and many questions arise. So we’re still building this portion of the documentation - if you have questions or comments you can suggest changes on this document or reach out to [email protected] to send feedback directly to Codecov's product team.

Tried integrating Automated Test Selection and is running into issues? You can try our 🔥Troubleshooting page.

Updated 9 months ago


RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4