Connecting Test Automation to Specific App Features and Devices

Having the right Test Automation plan helps bridge gaps and fragmentations in the complex mobile environment.

Figuring out the best Test Automation plan is one of the biggest frustrations for today’s digital teams.

Organizations struggle to develop cross-platform Test Automation that can fit with their Continuous Integration cadence, their regression cycles and other elements of their software development lifecycle (SDLC).

On top of the known complexities associated with mobile environments such as the network, the various OS versions, competing apps, there’s also the growing disconnect between the application being tested, the Test Automation scripts and the device.

When an application vendor develops its Android or iOS app, usually the mobile app development team would mark in the property files what features it uses and supports (i.e. GPS, TouchID, Doze, etc.) and what permissions this app requires, such as HTTP connection, GPS and File system.

For the app developer, once these property files are developed, his role in defining the app capabilities is done — which now pushes the test responsibility to the testing teams.

The problem that now arises is, how do these testing teams associate the right test cases based on the app capabilities that are only supported by certain mobile devices?

The screenshot below is taken from a Google Android Developer guide. It’s an example of a line that would usually be in an application Manifest.MF file within the App.APK file. What it’s showing here is that this specific app would require the use of the phone’s camera – this will trigger the user to allow camera access in the app’s first launch post-installation.

While that’s a common case, what would a tester do in a case where the requirement is for TouchID, ForceTouch, Android Doze or Multi-Window (coming in next Android N release)?

These features would complicate the test scenarios and Test Automation because only specific devices and/or OS versions would support these new OS capabilities. Without automatically planning and knowing which test should be executed against which device/OS combination, there is simply no way to manage these tests in today’s mobile app landscape. As a result, many organizations will lean on manual testing as a way to overcome the test delta for newer features that are unique to device/OS combinations. But when this gap grows, test cycles slow down and if not addressed quickly, the balance between manual and automation tests weighs toward manual as more enhancements are introduced in the app.

Break the App-Device-Test Disconnect

The main point here is to continuously make sure that the Test Automation suite supports all app capabilities, and is able to be executed against the right mobile platforms based on all the supported capabilities. So how can you overcome this challenge? Continuing to add to your “manual testing debt” is definitely not a sustainable solution.

Development and testing teams should work together closely, especially when a new supported feature, permission request or other capability is added to the Android app Manifest.MF file (see example to the left) or Apple Info.plist files. Whenever this happens there should be a call to action that requires updating the test execution rules and redirecting existing or newly developed tests for the right devices and platforms.

There are a few ways of addressing this call-to-action:

  • Testing teams receive a new developer build with a change log (pointing to the manifest.mf and info.plist changes)
  • Testing teams validate that there are existing tests for the above changes

⇒ Exists –> OK

⇒ Doesn’t exist –> Implement delta related cases (see the Facebook example above)

  • Testing teams, including developers, are synced on the right target devices (smartphones, tablets, wearables) that match the new and older test cases
  • Schedule test execution with proper rules and redirections to the right devices

In the case of an Android app that supports TouchID within the app Manifest.MF

(uses-permission android:name=”android.permission.USE_FINGERPRINT” /)

the dedicated Test Automation script will only be targeted against Samsung Galaxy S7 as a supported device.

For an iOS application that would use ForceTouch or 3D Touch through the UIApplicationShortcutItems, the developer would typically identify which of the menu options will be exposed to the user for testing to make the test plan easier to frame and develop.

Keeping Test Automation, Apps and Devices Aligned

As mobile operating systems evolve and introduce new functionalities, Test Automation suites are expected to become more fragmented. In order to not lose any existing investment in automation and at the same time increase Test Automation coverage and reduce Manual Testing, it’s important to identify the test suite gaps in your automation and pinpoint the devices that support your unique app features.

By doing this, you’ll be able to synchronize your test elements, your apps being tested and the devices in your test lab.

It’s important to note that “test coverage” will be defined differently based on a company’s industry, its organizational maturity in the digital space and its application types (mobile web, responsive web and/or native apps).

Eran Kinsbruner
Eran Kinsbruner is the mobile evangelist at Perfecto and the author of The Digital Quality Handbook, as well as a monthly columnist at InfoWorld.com. He is a software engineering professional with nearly twenty years of experience at companies such as Matrix, Sun Microsystems, General Electric, Texas Instruments and NeuStar. He holds various industry certifications from ISTQB, CMMI, and others. Eran is a recognized mobile testing influencer and thought leader, as well as an experienced speaker in the major software engineering conferences. He is also a patent-holding inventor for a test exclusion automated mechanism for mobile J2ME testing, public speaker, researcher, and blogger. He can be found all over social media, including on Facebook, Twitter (@ek121268), LinkedIn, and his professional mobile testing blog.

The Related Post

Test Strategy A test strategy describes how the test effort will reach the quality goals set out by the development team. Sometimes called the test approach, test strategy includes, among other things, the testing objective, methods and techniques of testing and the testing environment.
I recently came back from the Software Testing & Evaluation Summit in Washington, DC hosted by the National Defense Industrial Association. The objective of the workshop is to help recommend policy and guidance changes to the Defense enterprise, focusing on improving practice and productivity of software testing and evaluation (T&E) approaches in Defense acquisition.
The 12 Do’s and Don’ts of Test Automation When I started my career as a Software Tester a decade ago, Test Automation was viewed with some skepticism.
An Overview of Four Methods for Systematic Test Design Strategy Many people test, but few people use the well-known black-box and white-box test design techniques. The technique most used, however, seems to be testing randomly chosen valid values, followed by error guessing, exploratory testing and the like. Could it be that the more systematic test ...
In recent years, much attention has been paid to setting up Test Automation frameworks which are effective, easy to maintain, and allow the whole testing team to contribute to the testing effort. In doing so, we often leave out one of the most critical considerations of Test Automation: What do we do when the Test ...
Regardless of your current state of tools, building an effective Continuous Integration suite of significant automated regression tests is the key to moving to a higher level of confidence in today’s development world. In the evolution timeline of software development tools, new tools have recently proliferated. We have all been sold on collaboration, transparency and ...
“Happy About Global Software Test Automation: A Discussion of Software Testing for Executives” Author: Hung Q. Nguyen, Michael Hackett, and Brent K. Whitlock Publisher: Happy About (August 1, 2006) Finally, a testing book for executives!, November 17, 2006 By Scott Barber “Chief Technologist, PerfTestPlus” Happy About Global Software Test Automation: A Discussion of Software Testing ...
June Issue 2019: Testing the Software Car
I’ve been teaching a lot lately, was in India for one week, and I’m off to Seattle in two weeks to teach on performance topics. I thoroughly enjoy teaching, it allows me to stay sharp with current trends, and provides a nice break from the “implementation focus” that I generally have day to day.
All too often, software development organizations look at automating software testing as a means of executing existing test cases faster. Very frequently there is no strategic or methodological underpinning to such an effort. The approach is one of running test cases faster is better, which will help to deliver software faster. Even in organizations that ...
“Testing Applications on the web” – 2nd EditionAuthors: Hung Q. Nguyen, Bob Johnson, Michael HackettPublisher: Wiley; edition (May 16, 2003) This is good book. If you test web apps, you should buy it!, April 20, 2001By Dr. Cem Kaner – Director of Florida Institute of Technology’s Center for Software Testing Education & Research Book Reviews ...
LogiGear Magazine – April 2013 – Test Automation

Leave a Reply

Your email address will not be published. Required fields are marked *

Stay in the loop with the lastest
software testing news

Subscribe