In a recent article in the Enterprisers Project, Kevin Casey wrote, “If it’s repetitive and manual, it’s probably a good fit for RPA.”
People make mistakes. Human behavior so often falls short of ‘expected standards’, it begs the question why we hold ourselves to such standards at all. Too often, we build systems and processes on the implicit assumption that the people using them will be rational, infallible, and consistent. Of course, the truth is that most of us are anything but.
Our general fallibility is obviously closely tied to AI and test automation. Automated testing is immune to the unintentional biases and lapses in concentration that affect human testers.