Cassandra Leung and Pete Walen join us today in a discussion about Requirements. What are they? Do we get enough of them? Do we understand the ones we do get? Can we make them better, and if so, how can we help that process? If you’ve ever struggled with trying to make sense of a story, or fear that programmers are just implementing things for the sake of implementing them, and there’s no rhyme or reason, you may not be alone. It’s possible that you really are dealing with a severe case of Requirements Deficiency. Fortunately, we are here to help, or at the very least, give it a spirited try.
Also, in the news, Unified Windows Platform and Software Verification Competition. Yes, apparently, these are both “things”, and we pontificate on both of them.
For many testers, Selenium is a well known tool and a familiar friend. For others, it may be something you are curious about but haven’t had a chance to do much with yet. All of our panel has had some level of experience with Selenium, and Brian Van Stone visits us again to tell us what he’s been up to and how his Adventures with Selenium have informed his automation processes and overall testing.
Also on the Selenium and testing tools front, what is up with the VC community making big bets on software testing tools? Is it Silicon Valley business as usual, or is there something else going on here? We investigate, or pontificate, or at least we offer an opinion or four.
This week we are joined by Kim Knup, who is with Songkick and tells us a tale of intrigue and guile, and the behavior of concert attendees. Wait, what? OK, not quite that juicy, but she does work with Songkick, she does test and monitor performance, and it turns out that different audiences and different fans of different performers have distinctly different approaches to how the source and buy tickets through Songkick, and Kim shares some of those examples with us. Also, in our news segment, when Apple Support is down, do we care as much as when AWS is down? In other words, do we grade quality on a curve?
Sometimes, you can find experts on topics in unusual places. This week we discuss security and privacy with Doug Traser, an Information Security Manager with Five9. He’s also the guitar player for Michael’s band, Ensign Red (or is Michael Doug’s singer? We’re never entirely sure). Regardless, if you have questions about security, OWASP, polities that drive you crazy and wondering if any of this makes any sense, Doug has some answers, and maybe raises a few more questions.
Also, in our news segment: what happens when Amazon Echo might hold the key to a murder trial. Can your personal digital home assistant testify against you in a court of law?
Have you found yourself looking at deals and services online that seem too good to be true? Wondering “where’s the catch?” You’re not alone. There are lots of ways that software uses and manipulates us to give up details about ourselves, or to somehow get us to pay for services that we either didn’t want, or to provide information about ourselves and our habits to others that we don’t really want to be known. These practices are grouped together under the phrase ‘Dark Patterns” and Emma Keaveny has made it a point to learn about and warn about them. We discuss several varieties of Dark Patterns and debate where on the spectrum they fall, whether they be nuisances, poor design or an outright breach of ethics.
Also, where were you when Amazon’s E3 services went down on February 28, 2017? Did it affect you? It affected some of us, and at the time we recorded this episode it was a very fresh memory, so we had plenty to say about it.
How well do we know the work that we do as testers? Do we understand what it is we do? Really understand it? Jon Bach thinks we can do better at figuring out what it is that we do in our roles as testers and in the roles that support and offer service to people in our organizations. Much of what we do is implicit, and carries responsibilities, expectations and even contracts for what we do and how we act. In today’s episode, Jon helps us break down both traditional and not so traditional roles that we may find ourselves in, and ways that we can leverage both explicit and implicit knowledge of what we do, and maybe what we can stop doing.
Also, in an unconventional news segment this go around, friend of the show Anna Royzman tells us about Test Masters Academy and a fresh take on testing conferences geared towards testing leaders (the Testing Leadership Conference is May 1-3, 2017 in New York City) and emerging topics and technologies at the New Testing Conference coming this fall to New York City). It’s a wild ride!
Have you been to a testing conference? Wanted to go? Wondered which ones you should attend? Matt Heusser, Jessica Ingrassellino, and Michael Larsen have been to more than a few as participants and presenters. We discuss our favorites, the pros and cons of various conferences, and what makes each of them worthy destinations to consider.
Also, putting a different spin on the News Segment this time around, Michael shares his enthusiasm for and about “The Privacy Paradox”.
The truth is, no one will care as much or be as interested in your developing testing career as you are. There are studies that say that Software Testing is one of the Happiest Jobs there is. Does that sound weird, or does that sound spot on? In this episode of “The Testing Show” we welcome back Alex Schladebeck and welcome for the first time QualiTest’s Elle Gee to discuss software testing careers and how they differ or are similar depending on the organization in question. Regulars Perze Ababa and Justin Rohrman also riff along with Matt Heusser on the unique challenges in developing and sustaining a career in software testing.
Also, in our news segment, what happens when Automation and a software glitch makes it impossible to do a task we often take for granted? 900 Shell stations in Malaysia discovered exactly that, and we certainly have opinions about that, too.
This is Part Two of our discussion with Alex Schladebeck and Joel Montvelisky. We tackle regression testing and share a few stories from the trenches (and a ghostly Michael even makes a contribution to this topic), discuss the idea that perhaps continuous testing is a concept that’s time really has come, and look to see if we can possibly break out of the “hardening” process at the end of sprints in favor of more testing up front, so that discoveries can actually be addressed sooner rather than pushed off to later.
It seems that 2017 is shaping up to be the year of the two-parter, as we are back with another two part episode. This is Part One, in which the Testing Show regulars chat with Alex Schladebeck and Joel Montvelisky about the way that testing is practiced globally. Joel has some insights on this in that he steers the State of Testing Questionnaire that runs in January and February each year, and gathers statistics about how testers actually work. We look at some issues that were discovered with the survey, such as how many organizations claim to do automation versus how many actually are making a solid go at it, as well as where those organizations choose to, or choose not to, apply their efforts. Also, in the news, what happens when TSA’s computers go out on one of the busiest travel days of the year (the day after New Years)? The Testing Show panel and their guest weigh in, and they have plenty to say, both on the outage and the process in general.
We continue our conversation with Angie Jones about ways that automation can be put first in stories (yes, really) and ways that she has been able to get team buy in and cooperation to make that process effective. Also, we have a mailbag question that we answer in depth, or as much as we can… is it possible to be paid as much as a developer or an SDET if you are just a manual tester? The answer is “it depends”, but we go into a lot more about why that is the case.
Have you wondered how your team could better utilize its automation resources? Does your definition of “Done” include new automation efforts for stories that are in flight? How about when changes to functionality (or new additions) cause your old tests to stop working? Do we play continuous catch up, or is there a better way to applying automation efforts?
Angie Jones of Lexus Nexus joins us to talk about better ways to have those automation discussions, who should be responsible for what, and how everyone on the team can contribute to automation efforts (hint, you don’t need to be a coder to help make great automation, but it certainly helps).
Also, this week we delve into Spotify taking over hard drives with continuous writes that could shave years off of their operation life, and are Uber’s autonomous vehicles even close to ready for prime time?
This is part one of a two part series. Come back in two weeks when we continue our conversation with Angie.
It’s a new year, and it’s that classic time for people to make New Year’s Resolutions, as well as quickly run out of steam trying to actually succeed at them. We discuss ways in which we have set goals, or not set them, how we have been stymied in the past, or how we have pushed on regardless of failing, and the fact that failing is often the key element that helps us progress and ultimately succeed.
Also, in the news, we look back on the fifth anniversary of the death of Christopher Hitchens, how a typo may well have been the root cause of “The Russian Hack” regarding the U.S. Presidential election, and will all exploratory testers be replaced in five years by AI, neural networks and machine learning? We have opinions on all of those.