|
Regardless of what tool you are using, there are several test automation concepts and best practices that are always the same when it comes to end-to-end tests. I am using examples in plain English supported by testRigor for illustrative purposes, but it could be literally any language you like.
Here is an overview of test automation best practices. I will get into the details below.
- Repeatability – if it ran once it should rerun again if nothing changes
- Data-driven testing – use datasets to run the same test on multiple rows of data
Test Automation Best Practices
- Test ends with a validation
- Test is stable enough to be used in CI/CD
- Test is very easy to understand
- Test requires as little maintenance as possible
- Test is independent and can run in parallel with all other tests
- Credentials are stored separately from the test
- Shared steps are grouped into functions
Test Automation Concepts
Repeatability
The difference between a test case for manual execution and an automated one is that when you are writing an automated one, you must be specific to the minute detail of how to execute things. This, in particular, means that once written, the test case must be able to be run repeatedly with consistent results. We as humans tend to be able to infer things and guess them on the fly. However, this is not acceptable for automation. In automation, you must be very explicit to do things the way that will allow them to be executed repeatedly.
For example, if you are testing a sign-up with email, like someone here, you must make sure that you are using a new email every time you do it to avoid your test not being successful because some email was already used before.
enter "[email protected]" into "Email" enter "password" into "Password" click "Sign Up" ...
generate unique email, then enter into "Email" and save as "newEmail" enter "password" into "password" click "Sign Up" ...
Unique test data generation
generate unique email, then enter into "Email" and save as "newEmail"
does exactly that. Now, there are several ways to generate data that hadn’t been used before. Let us consider three main types of unique data generation:- Ask your Engineers to provide you with an API
- Use sequential numbers/letters
- Use random sequences
You could, of course, ask your Engineers to provide such an API for you. However, that would just force them to use one of the following methods themselves, it will introduce the overhead and latency related to an API call, and, most importantly, it would be close to impossible to prioritize such a feature from Engineering in the first place.
Sequential numbers are almost infeasible as well since eventually you’d be forced to run your tests in parallel for performance reasons, and it is very hard to keep those numbers in sync, making them unusable for practical reasons.
Therefore, the best bet is to use long enough random sequences to make sure the data is unique and doesn’t overlap with your other tests running in parallel. Modern systems like testRigor have a built-in way of generating unique data out of the box. For other systems, you might consider using UUIDs and/or random number generator classes like this.
Just generating the unique data might not be enough. You often need to use this data later in the test. This brings us to the next point.
Use of variables
Getting back to the sign-up example above, once you generated the unique email, you need to be able to check that email to confirm it, as well as use it later in the test to make sure that you can actually log in using that exact email. This means that you need to store your generated email (or any of your generated data for that matter) somewhere to refer to it for use later. Those places you store your data for your test are called “variables”. You can think of a variable as a label on the box where you store particular data. The concept of variable is basically the same one as you learned in your algebra lessons in school.
"newEmail"
variable to check that email later like so:generate unique email, then enter into "Email" and save as "newEmail" enter "password" into "password" click "Sign Up" check that email to stored value "newEmail" was received click "Confirm email" ...
In this example, we used the variable "newEmail"
to check the email and render it in the browser and then click “Confirm email” link/button in that email to proceed with the registration flow.
See how we used check that email to stored value "newEmail" was received
referring to variable “newEmail” to refer to the email we generated before instead of checking a predefined email like check that email to "[email protected]" was received
.
You can see a running example of generating data and using variables here.
Data-driven testing
This concept can be considered a best practice by some, but since it is a built-in feature of virtually all of the testing systems, we’ll consider it as a “concept”.
- Empty email
- Email with no domain specified like “user”
- Email with only one level domain specified like “[email protected]”
- Email with multiple “@” signs in it
- etc.
enter stored value "myTestEmail" into "Email" click "Sign Up" check that page contains "Error" ...
myTestEmail |
---|
user |
[email protected] |
[email protected]@domain |
You can see the example of a full sign-up test here.
Test Automation Best Practices
Test ends with a validation
Imagine that your test won’t end with a validation. Then it ends with either a click or entering data. In both of these cases, you do not know if your last action was successful or not. Therefore, you always want to add validation at the end of every test. This rule applies to all kinds of tests, including end-to-end tests, integration tests, and unit tests.
Some people believe that tests should only contain validation at the end and not in the middle of the test. I think this only applies to unit tests and not necessarily end-to-end tests, since an end-to-end test might take some time to execute, and it might be wasteful to duplicate a test just to stick to the rule.
Test is stable enough to be used in CI/CD
If you built your tests and can run them yourself locally, that’s great. However, you’ll get far more value if they are run automatically on every change. You can relatively easily add them to CI/CD. However, if the tests are not stable enough – they will fail, constantly wasting your team’s time on figuring out why they failed. So, clearly, you want to build your tests in a way that makes them as stable as possible. And yes, tools can help here. For instance, since testRigor doesn’t rely on details of implementation – it won’t fail when these details change.
Test is very easy to understand
Imagine that you wrote a test 3 months ago, and now functionality changed, and you need to go back and update it. I can guarantee that you most probably forgot how you wrote this test and why you did it this way. Now imagine that someone else needs to look into your test. This is the effort multiplied by two since this person had never been familiar with your code in the first place.
The difference between readable and unreadable code can be pretty substantial when people are trying to change things. Multiplied by the number of people on the team, it can amount to a sizable part of everyone’s day.
Also, if your test cases are readable by non-engineers (like product managers) you get the benefit of being able to share your code and get feedback if you are testing the right thing.
Test requires as little maintenance as possible
This is very straightforward. Test maintenance is what eats the time. You want to minimize it. Again, modern frameworks like testRigor will get you covered here. With tools like Selenium, you are pretty much on your own in front of the least stable system you can imagine. The problem is not Selenium per se, but rather the fact that it encourages you to use details of implementation like XPath to refer to elements on the screen.
Test is independent and can run in parallel with all other tests
As much as you think that you just need only several tests, at some point, your team will most probably end up with a lot of them running longer than several hours (or whatever threshold you can tolerate), and you’d want them to run faster. Tweaking will get you only as far, you still bare the cost of network calls, server responses, and browser rendering times. At some point, the only way to speed things up will be parallelization. And, if you built your tests to be able to run independently from the get-go, you’ll be able to enjoy the fruits of parallelization and speeding things up.
What does it take to keep the tests independent? It is mostly about the data. And keeping the test stable is actually a spectrum, not a binary. On one side, all your tests will use the same data, passing results from one to another. With these dependencies, parallelization would be challenging. On another side is the way where each test creates all the data it needs within the test. The problem with it, as you can imagine, is that it takes a lot of time. In real life, you’d probably be somewhere in the middle, with some data being shared among the tests and some created for each individual test. What you certainly want to avoid at all costs is dependency, where the next test depends on the results of the previous test.
Credentials are stored separately from the test
- They do tend to change regularly
- It is insecure to keep credentials in the open like this
Shared steps are grouped into functions
- In case the system under test changes, you now have just one place to adapt instead of multiple
- You can make your code far more readable and easy to understand, as well as closer to the literal specification described in your test case, if you give sequences of steps with readable names
There are more best practices to cover, it would probably require a full book to list them all. However, we believe that the ones above are the top most important ones to use in automation.
Here is a webinar we had some time ago on the easiest way to learn test automation where I talk about these concepts: