The LEAP

Automation insights and productivity tips from LEAPWORK.

All Posts

Release Announcement: Big Performance Improvement in Data Serialization

Usually, in this blog, we tend to look at the LEAPWORK Automation Platform from a user perspective. We talk about how to automate different kinds of work processes and we also give you a few tips and tricks every now and then. However, we’ve recently made some changes to the core engine, so we thought it would be interesting to look into the engine room for a change. Buckle up because we’re about to get a little technical.

For a long time, LEAPWORK has been using the JSON.Net code library to serialize and deserialize data we send between Studio and the Controller. We’ve also used it to serialize complex structures we send to the Agent during both preview and scheduled runs.

JSON.Net is a very powerful and extremely popular code library, but when working with large and complex data structures at scale, it falls a bit short. The result is high pressure on CPU and large memory consumption.

At LEAPWORK we strive for continuous improvement so, for that reason, we decided to fix both.

After researching different solutions, we ended up loving a component of the open source Azos framework for building scaled-up business applications. It handles JSON and BSON serialization, both of which are needed in LEAPWORK.

We had to make some minor improvements and adjustments to the component, which we contributed back to the Azos project, but the results were very good.

We experienced a significant impact on LEAPWORK’s performance when working with large volumes of keyframes, that is the individual steps that happen when flows run. Due to this success, we are happy to announce that this new component has already been included into our upcoming service release.

Actions speak louder than words, so let us show you some of the results you will get to experience with the upcoming release. The following graph shows how much memory was consumed by LEAPWORK for gathering and communicating approximately 300.000 keyframes between an Agent and the Controller. In here you can see the difference between the current and the upcoming release:

Chart Memory Usage for 300.000 Keyframes

This change has impacted not just memory consumption but also CPU pressure. You can see a significant performance boost in the following graph,  where we show the difference in time spent serializing and deserializing large volumes of keyframes:

Chart: Elapsed Time for 60.000 Keyframes

Together, this means a much faster execution time of large automation flows, and an overall better experience working with LEAPWORK Automation Platform.

We always strive for the best user experience and, for that reason, we will continue to improve how data travels through LEAPWORK. This includes serialization and deserialization, as these account for a large amount of the CPU pressure and memory consumption in the Controller.

If you would like to know more about the LEAPWORK Automation Platform, or how this upcoming release can help you improve your automation efforts, book a demo in the link down below.

Book Demo

Claus Topholt
Claus Topholt
CTO and co-founder of LEAPWORK.

Related Posts

Test Automation and RPA Tools: What's the Difference?

If you’re using test automation and recently began looking into RPA, or if you are new to test automation but using RPA as a means to achieve efficient operations, you might at some point have wondered what the difference is between the two often interchanged terms, as well as if you can use the same tool for both.

[WHITEPAPER] Automating Workday® with LEAPWORK

An increasing number of companies are seeing the benefits of RPA in Workday®. Unfortunately, developing RPA functions is costly and time-consuming, requiring new personnel with specialized and expensive skills. So how can a company get the most out of automation without incurring massive implementation and maintenance costs?

Data-Driven Testing With Codeless Selenium

Data is found everywhere. Whether you are testing an application’s functionality with different data parameters or automating the process of moving data from A to B, you have probably realized by now that data is found to be a key component in most automation cases.