Improving Test Efficiency with AWS Lambda

May 10, 2018
0 minute read

In end-to-end (E2E) automated testing systems, there are several challenges that keep automation teams busy. One major challenge is the ability to combine tests with a continuous integration platform. E2E tests can be slow, difficult to maintain, and easily broken on the long path through the twists and bends of the tested system (front-end, back-end,... (more...) The post Improving Test Efficiency with AWS Lambda appeared first on Duda Blog.

In end-to-end (E2E) automated testing systems, there are several challenges that keep automation teams busy. One major challenge is the ability to combine tests with a continuous integration platform. E2E tests can be slow, difficult to maintain, and easily broken on the long path through the twists and bends of the tested system (front-end, back-end, database, interservices communication, etc.).

It’s no surprise, then, that many teams struggle to keep their test suite execution time in the minutes zone, often drifting into hours. Delaying the development process until all the tests end is counterproductive. So, how do companies solve this problem? They split the tests into a “sanity check” that includes minimal coverage, and push most tests to a once-a-day execution.

At Duda, we’ve developed an infrastructure that enables us to execute the full automation suite in 10 to 15 minutes (and even faster). We achieved this by carrying out tests on AWS Lambda .

Before we dive into how we execute tests inside a Lambda, let’s address the foundation upon which this is even feasible. If your tests are not stronghold stable, this solution does not help. In order to go quickly, you have to go well. See this post about how to get your tests super stable and unbreakable with the WWHB design pattern. Only strong tests can withstand the speed of Lambda.

First, a few words about Lambdas

As defined in Wikipedia, “AWS Lambda is an event-driven, serverless computing platform provided by Amazon as a part of the Amazon Web Services.”

Let’s break this down bit by bit.

Serverless : A Lambda is a container running Amazon Linux that can execute code that we write. The advantage over traditional servers is that we don’t maintain the server. We just ask it to execute the code and Amazon takes care of the rest. It handles all the administration and scalability. From our point of view, we treat a Lambda as a snippet of code that we can call as part of our system.

Event-driven: As Lambdas are part of AWS, they connect with Amazon S3 buckets, Amazon DynamoDB tables, and other Amazon cloud services. This enables a Lambda to trigger after a database update or a file uploaded to S3.

The code running in the Lambda is referred to as a Lambda Function . A term taken from programming languages that can treat code function as a first class object. This architectural approach can create a more decoupled system and thus one that is more modular and robust.

AWS Lambdas can act in a similar way by extracting your functionality to an external computing platform with minimal connections to your core product. By extracting your code, Amazon can easily invoke the Lambda, returning its result back to your code and shutting down. If you need to scale this code, Amazon takes care of it by invoking more and more Lambdas simultaneously, like an army of working ants.

A Lambda does have has its limitations. Its disk space is limited and the time it takes the Lambda to execute your code is also limited. So, we have to keep the code small and quick.

Now, let’s talk about Selenium in Lambdas

At Duda, we write Selenium tests using Java and TestNG as a test executing framework. We have more than 1000 tests in our suite, and each test can be autonomously executed (as discussed in the post I mentioned above).

In order to execute the test inside a Lambda, we have to get the test code into the Lambda. We need to have Chrome and Chromedriver here, and finally, we have to execute the test. Let’s split this topic; first, to solving the Chrome and Chromedriver; then, to considering how to solve the test code itself.

The easy part is getting Chrome browser and the Chromedriver executable into the Lambda before execution. As mentioned above, you can write code to be executed in Lambda, aka Lambda function. You’ll need to upload the latest version of Chromedriver for Linux and a special version of headless Chrome for AWS Lambda (available in a dedicated GitHub project ) to a location in the AWS S3 storage. Your Lambda function will begin by downloading these artifacts to the local storage.

To get the test code into the Lambda, you also have to have it in the S3 storage. You can achieve this by compiling all the test code infrastructure into a single executable (at Duda, a single JAR (Java ARchive)). The entry point must get a test name to execute as an argument. In our case, we built a Main class as an entry point and get the test name with TestNG framework. It is the single test executed in the Lambda. This means the whole JAR contains all the tests, and only one is executed in each Lambda. By utilizing the power of parallel execution with Lambdas, we can decrease the total suite time dramatically.

Now we need to invoke the Lambda function, which in turn will pull the Chrome, Chromedriver and test JAR from S3 and finally execute the test. We implement it by a class we call ExecuteLambdaTest. It sends a JSON object using the AWS API kit for Java, invoking the Lambda with a single test name to execute. We can run this functionality with multiple threads, thus executing multiple Lambdas at once. This class will invoke the Lambda and wait until it’s finished returning a JSON object as a response.

All that is left is to gather all the test reports from all the Lambdas back to the S3 storage to show the final status of the suite.

How the system works

The three components for orchestrating all this: S3 storage, ASW Lambda and Jenkins

After implementing the basic mechanism, you can add many more improvements, such as a test retry after failure, unified reporting system, reuse of the Lambda container, and more.

Highlight of the testing improvements

This system resolves several issues, and leads to many testing improvements, including:

1. Short distance between Selenium & Browser

With Selenium Grid, you execute the test on a CI or corresponding server to execute the test code. It sends its actions using RemoteWebDriver to the Hub server, which tries to share the load over nodes holding the browser. It’s a very long distance for each web action to go.

Text executed inside the Lambda

In traditional testing, the distance each web action must cover is long and takes time.

With our system, the test is executed inside the Lambda, and the browser sits on the same machine, giving a much better response for each action and speeding up each test.

By using the Lambda, both the time and the distance are much shorter.

2. Parallel scalability

With Selenium Grid, you are limited to the amount of nodes you manage. With Lambda, you can invoke as many tests you want at once. The bottleneck is the server on which you are running the tests, which must be able to handle the amount of simultaneous users.

3. Parallel precision

With Selenium Grid, the hub manages a queue of tests and you cannot intervene. After some testing, we found that the Grid does not cope with high loads, and we don’t get the parallel tests we expect. Because we’ve implemented our own parallel test executor, we know exactly what holds us back, and how to break barriers.

At Duda, we’ve implemented our system writing all the code ourselves for better control of execution speed. Our implementation of Selenium in Lambda significantly improved execution time without damaging stability, which is exactly what we intended.

I hope this blog post inspires your team to try it out and helps your automation play a more significant role in the dev process.


Did you find this article interesting?


Thanks for the feedback!
By Shawn Davis April 1, 2026
Core Web Vitals aren't new, Google introduced them in 2020 and made them a ranking factor in 2021. But the questions keep coming, because the metrics keep changing and the stakes keep rising. Reddit's SEO communities were still debating their impact as recently as January 2026, and for good reason: most agencies still don't have a clear, repeatable way to measure, diagnose, and fix them for clients. This guide cuts through the noise. Here's what Core Web Vitals actually measure, what good scores look like today, and how to improve them—without needing a dedicated performance engineer on every project. What Core Web Vitals measure Google evaluates three user experience signals to determine whether a page feels fast, stable, and responsive: Largest Contentful Paint (LCP) measures how long it takes for the biggest visible element on a page — usually a hero image or headline — to load. Google considers anything under 2.5 seconds good. Above 4 seconds is poor. Interaction to Next Paint (INP) replaced First Input Delay (FID) in March 2024. Where FID measures the delay before a user's first click is registered, INP tracks the full responsiveness of every interaction across the page session. A good INP score is under 200 milliseconds. Cumulative Layout Shift (CLS) measures visual stability — how much page elements unexpectedly move while content loads. A score below 0.1 is good. Higher scores signal that images, ads, or embeds are pushing content around after load, which frustrates users and tanks conversions. These three metrics are a subset of Google's broader Page Experience signals, which also include HTTPS, safe browsing, and mobile usability. Core Web Vitals are the ones you can most directly control and improve. Why your clients' scores may still be poor Core Web Vitals scores vary dramatically by platform, hosting, and how a site was built. Some of the most common culprits agencies encounter: Heavy above-the-fold content . A homepage with an autoplay video, a full-width image slider, and a chat widget loading simultaneously will fail LCP every time. The browser has to resolve all of those resources before it can paint the largest element. Unstable image dimensions . When an image loads without defined width and height attributes, the browser doesn't reserve space for it. It renders the surrounding text, then jumps it down when the image appears. That jump is CLS. Third-party scripts blocking the main thread . Analytics pixels, ad tags, and live chat tools run on the browser's main thread. When they stack up, every click and tap has to wait in line — driving INP scores up. A single slow third-party script can push an otherwise clean site into "needs improvement" territory. Too many web fonts . Each font family and weight is a separate network request. A page loading four font files before rendering any text will fail LCP, especially on mobile connections. Unoptimized images . JPEGs and PNGs served at full resolution, without compression or modern formats like WebP or AVIF, add unnecessary weight to every page load. How to measure them accurately There are two types of Core Web Vitals data you should be looking at for every client: Lab data comes from tools like Google PageSpeed Insights, Lighthouse, and WebPageTest. It simulates page loads in controlled conditions. Lab data is useful for diagnosing specific issues and testing fixes before you deploy them. Field data (also called Real User Monitoring, or RUM) comes from actual users visiting the site. Google collects this through the Chrome User Experience Report (CrUX) and surfaces it in Search Console and PageSpeed Insights. Field data is what Google actually uses as a ranking signal — and it often looks worse than lab data because it reflects real-world device and connection variability. If your client's site has enough traffic, you'll see field data in Search Console under Core Web Vitals. This is your baseline. Lab data helps you understand why the scores are what they are. For clients with low traffic who don't have enough field data to appear in CrUX, you'll be working primarily with lab scores. Set that expectation early so clients understand that improvements may not immediately show up in Search Console. Practical fixes that move the needle Fix LCP: get the hero image loading first The single most effective LCP improvement is adding fetchpriority="high" to the hero image tag. This tells the browser to prioritize that resource over everything else. If you're using a background CSS image for the hero, switch it to anelement — background images aren't discoverable by the browser's preload scanner. Also check whether your hosting serves images through a CDN with caching. Edge delivery dramatically reduces the time-to-first-byte, which feeds directly into LCP. Fix CLS: define dimensions for every media element Every image, video, and ad slot on the page needs explicit width and height attributes in the HTML. If you're using responsive CSS, you can still define the aspect ratio with aspect-ratio in CSS while leaving the actual size fluid. The key is giving the browser enough information to reserve space before the asset loads. Avoid inserting content above existing content after page load. This is common with cookie banners, sticky headers that change height, and dynamically loaded ad units. If you need to show these, anchor them to fixed positions so they don't push content around. Fix INP: reduce what's competing for the main thread Audit third-party scripts and defer or remove anything that isn't essential. Tools like WebPageTest's waterfall view or Chrome DevTools Performance panel show you exactly which scripts are blocking the main thread and for how long. Load chat widgets, analytics, and ad tags asynchronously and after the page's critical path has resolved. For most clients, moving non-essential scripts to load after the DOMContentLoaded event is a meaningful INP improvement with no visible impact on the user experience. For websites with heavy JavaScript — particularly those built on frameworks with large client-side bundles — consider breaking up long tasks into smaller chunks using the browser's Scheduler API or simply splitting components so the main thread isn't locked for more than 50 milliseconds at a stretch. What platforms handle automatically One of the practical advantages of building on a platform optimized for performance is that many of these fixes are applied by default. Duda, for example, automatically serves WebP images, lazy loads below-the-fold content, minifies CSS, and uses efficient cache policies for static assets. As of May 2025, 82% of sites built on Duda pass all three Core Web Vitals metrics — the highest recorded pass rate among major website platforms. That baseline matters when you're managing dozens or hundreds of client sites. It means you're starting each project close to or at a passing score, rather than diagnosing and patching a broken foundation. How much do Core Web Vitals actually affect rankings? Honestly, they're a tiebreaker — not a primary signal. Google has been clear that content quality and relevance still dominate ranking decisions. A well-optimized site with thin, irrelevant content won't outrank a content-rich competitor just because its CLS is 0.05. What Core Web Vitals do affect is the user experience that supports those rankings. Pages with poor LCP scores have measurably higher bounce rates. Sites with high CLS lose users mid-session. Those behavioral signals — time on page, return visits, conversions — are things search engines can observe and incorporate. The practical argument for fixing Core Web Vitals isn't just "because Google said so." It's that faster, more stable pages convert better. Every second of LCP improvement can reduce bounce rates by 15–20% depending on the industry and device mix. For client sites that monetize through leads or eCommerce, that's a revenue argument, not just an SEO argument. A repeatable process for agencies Audit every new site before launch. Run PageSpeed Insights and record LCP, INP, and CLS scores for both mobile and desktop. Flag anything in the "needs improvement" or "poor" range before the client sees the live site. Check Search Console monthly for existing clients. The Core Web Vitals report surfaces issues as they appear in field data. Catching a regression early — before it compounds — is significantly easier than explaining a traffic drop after the fact. Document what you've improved. Clients rarely see Core Web Vitals scores on their own. A monthly one-page performance summary showing before/after scores builds credibility and makes your technical work visible. Prioritize mobile. Google uses mobile-first indexing, and field data shows that mobile CWV scores are almost always worse than desktop. If you only have time to optimize one version, do mobile first. Core Web Vitals aren't a one-time fix. Platforms change, new scripts get added, campaigns bring in new widgets. Build the audit into your workflow and treat it like any other ongoing deliverable, and you'll stay ahead of the issues before they affect your clients' rankings. Duda's platform is built with Core Web Vitals performance in mind. Explore how it handles image optimization, script management, and site speed automatically — so your team spends less time debugging and more time building.
By Ilana Brudo March 31, 2026
Vertical SaaS must transition from tools to an AI-powered Vertical Operating System (vOS). Learn to leverage context, end tech sprawl, and maximize retention.
By Shawn Davis March 27, 2026
Automate client management, instant site generation, and data synchronization with an API-driven website builder to create a scalable growth engine for your SaaS platform.
Show More

Latest posts