Skip to content
QualityLogic logo, Click to navigate to Home

The Quality Trail: September 2025 QA News

Home » Blogs/Events » The Quality Trail: September 2025 QA News

From the Desk of the Editor

Welcome to a mid-September edition of the Quality Trail, your one stop shop for the latest news and information in the QA space. It’s been a super busy month for us, so we’ve opted to cover the happenings in the month of August and the first half of September together this time around. 

– The QualityLogic Editorial Team

What’s Inside

As always, let us know if you think we’ve missed something, or share the link with your colleagues or partners who may benefit from some or all of this information. You can also sign up to receive these testing updates via email.


Conferences & Events 

  • STARWEST 2025 (Anaheim, CA): Sunday September 21 – Friday September 26 
  • KWSQA Targeting Quality 2025: Cambridge, Ontario Canada, Monday September 22 – Tuesday September 23 
  • TestBash Brighton is the UK’s largest software testing conference, and it’s put on by the Ministry of Testing (so you know it’ll be good). This year it will take place from October 1 – 2 at the Brighton Dome in Brighton. 

Becoming a Better QA Leader

This newsletter often touches on the technical or trendy aspects of software quality. While we think that’s largely a good thing (you can’t understate the importance in staying ahead these days), we’d also be remiss for failing to talk about the people element from time to time too. 

Gary Hawkes, a seasoned QA leader, recently wrote an article titled 6 leadership mistakes that made me a better leader. It caused us to reflect a bit on the attributes that make-or-break management in our industry. After all, you can have the best testers and tools out there, but without effective leadership, it won’t matter. If you have the time, the article is a good one to read in full. Regardless, here are the lessons with some added commentary: 

  • Support your team’s growth as professionals and as people proactively so that you won’t have to defend them reactively. One thing we have found works wonders is providing reports with the time for protected learning blocks tied to our current risks and/or objectives. 
  • Be clear on the principles underlying the kind of leader you want to be and play that role. Make sure to regularly ask yourself: “Did failure lead to blame, or did it lead to inquiry and a lesson learned?” 
  • Don’t impose metrics on teams. Instead, understand what success and failure look like to them. When you impose the metrics you want to focus on without asking for input, it feels forced and passes up the chance for a good conversation. When everyone mutually agrees on performance indicators, they will be more likely to want to meet them. If we can’t discuss metrics without weaponizing them, we shouldn’t be using them. 
  • Protect your culture. When one of your reports is damaging it and mentoring fails, you may need to make tough decisions to protect the culture. This is obviously much easier said than done. However, it’s worth remembering the stats: one that sticks with me, MIT found that toxic culture was 10 times more of a predictor of industry-adjusted attrition than compensation. Put another way, avoiding one toxic worker can literally yield more value than hiring a “superstar.” 
  • Give recruitment candidates an enjoyable interview experience. Recruit for your culture and diverse team dynamics first, skills second. Skills are significantly easier to train than culture anyway. 
  • Be a leader when it comes to decisions about your career. Never be led by your emotions or your current situation. 

Of course, all of this would be moot without talking about a good set of values. As a bonus, check out this article Want Strategic QA? Start with What Every Great Company Has: Core Values – MarinaJordao (Medium)

AI in QA

The Rise of AI Specific QA Roles 

A few months back we predicted that, as LLMs become more widespread in the workplace, new roles and methodologies would begin to crop up. That seems to be well under way now. 

A simple LinkedIn search will surface a bunch of people with the title of “Vibe code cleanup specialist.” 

To level set, the term “vibe code” was coined by Andrej Karpathy (one of the founders of OpenAI) on X. In his own words: “There’s a new kind of coding I call “vibe coding”, where you fully give in to the vibes, embrace exponentials, and forget that the code even exists. It’s possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like “decrease the padding on the sidebar by half” because I’m too lazy to find it. I “Accept All” always, I don’t read the diffs anymore. When I get error messages I just copy paste them in with no comment, usually that fixes it. The code grows beyond my usual comprehension, I’d have to really read through it for a while. Sometimes the LLMs can’t fix a bug so I just work around it or ask for random changes until it goes away. It’s not too bad for throwaway weekend projects, but still quite amusing. I’m building a project or webapp, but it’s not really coding – I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works.” 

If you’re reading this as a quality specialist, that paragraph has probably got you seeing stars. Imagine knowing that there exists a subset of software engineers who do this on the job every day. A colleague recently told me a story about walking into a coffee shop in SF and watching a developer playing Chess. Occasionally, every few minutes, he’d switch over to Claude Code and almost absent mindedly give it an instruction, before going right back to his game. 

Companies are loving the fact that AI has allowed them to spin up demos and prototypes in a matter of days where it used to take weeks… and herein lies the problem. We as QA practitioners know that getting code to run is actually the easy part. The edge cases that users are likely to encounter in the wild only start to present themselves with rigorous automated and manual testing. When developers let LLMs write the majority of their code, the importance of the tester is magnified tenfold. Enter the “Vibe Code Cleanup Specialist,” who goes in and refines the existing code to make it more maintainable prior to production. I originally learned about this from a post on LinkedIn. Since then, others have been posting about it as well. 

Testing LLM-based Applications 

It seems like everyone has a chat bot or LLM integration now. Yet, the nondeterministic nature of GenAI presents a million-dollar question, how are we supposed to effectively test them? In a post titled Stop “vibe testing” your LLMs. It’s time for real evals, developers at Google captured the sentiment brilliantly: “If you’re building with LLMs, you know the drill. You tweak a prompt, run it a few times, and… the output feels better. But is it actually better? You’re not sure. So you keep tweaking, caught in a loop of “vibe testing” that feels more like art than engineering.” Here are a few more resources on the topic: 

What We’ve Been Reading


Interested in More Information About QualityLogic?

Let us know how we can help out – we love to share ideas! (Or click here to subscribe to our monthly newsletter email, free from spam.)