Optimizing Client Savings and Reducing Stress: Expert Strategies from a DevOps Engineering Leader

Vasilii Angapov
Vasilii Angapov

As artificial intelligence continues its upward trajectory, the core transformations in major IT firms remain anchored in human intelligence and the dedication of skilled professionals.

Vasilii Angapov, a Senior DevOps Engineer with an American consulting firm and a distinguished Red Hat Certified Architect, boasts a decade of invaluable experience in the global IT arena.

Throughout his career, Vasilii has achieved remarkable feats, from generating millions in savings for the creators of the renowned video game "World of Tanks" to fortifying the US healthcare system against data loss vulnerabilities. His expertise is further validated by certifications from IT behemoths like Google, Amazon, IBM, Microsoft, and the Linux Foundation.

In our conversation, Vasilii delves into the pivotal role of contemporary cloud solutions, cutting-edge technologies, and an unwavering commitment to professional advancement in enhancing client operations.

Challenge Yourself

Vasilii, with four years in the US market and seven in the CIS countries, which project has pushed your limits and fostered the most professional growth?

Every project I've undertaken has its distinct challenges and rewards. But the pinnacle of technical complexity for me was my tenure at Wargaming, a global gaming giant behind titles like World of Tanks and World of Warships. I came on board in 2016 as a systems administrator for roughly two years.

My primary mission was to roll out a corporate private cloud platform using OpenStack. Back then, OpenStack was an emerging open-source cloud platform, with limited expertise available on the market. Its setup, configuration, and maintenance were intricate.

Fortunately, I had dabbled with this technology at GDC (ICL Services), a key ally of the Japanese IT powerhouse, Fujitsu. Out of sheer curiosity, I had explored and even tried implementing it in a test environment.

At Wargaming, my role expanded to deploying OpenStack clusters across various regions, from North America to China. The challenge wasn't just in the deployment but in enhancing its capabilities.

I focused on bolstering the clusters' resilience and swift recovery from potential failures. My efforts bore fruit, with the system now capable of rapid auto-recovery. My proficiency in Python, the language of OpenStack, enabled me to introduce new plugins and tweak the source code.

During that period, I was deeply engaged in the project, dedicating almost every waking hour to it. The outcomes were immensely rewarding. Beyond enhancing the system's resilience, we made significant strides in our Time To Market metric. What used to be a process spanning weeks or even months for new product releases to reach consumers was streamlined to mere hours, and occasionally, just minutes.

It's crucial to highlight the cost-effectiveness of this transition. The preceding virtualization platform, VMware vSphere, had a hefty price tag, costing the company millions. In stark contrast, our shift to the OpenStack cloud platform resulted in 90% savings.

In the grand scheme of things, I spearheaded the migration of approximately 70% of Wargaming's web services to this new platform. Today, they operate seamlessly.

Can you recall moments in your career where you've made a notable impact on a company's cost-efficiency?

Absolutely, cost-efficiency has been a recurring theme throughout my career. I've always gravitated towards process optimization and the adoption of innovative technologies. This not only benefits the company in terms of reduced expenses but often extends savings to our clients as well.

For instance, I had the privilege of working with the American branch of Juniper Networks, a global frontrunner in networking technologies. A vast segment of the world's internet operates on Juniper's hardware. My role was pivotal in the software development for Network Function Virtualization (NFV) and Software Defined Networking (SDN), with Python being the primary development language.

In layman's terms, I was instrumental in transitioning the company's physical network infrastructure to a virtual one. This shift allowed for the shared use of network equipment among various clients, eliminating the need for individualized, costly setups. This initiative not only led to significant savings on equipment but also expanded the company's network service offerings and reach.

Furthermore, NFV solutions are integral to the architecture of modern 5G networks. By leveraging NFV and SDN, mobile network operators can drastically cut down on both capital and operational costs associated with network equipment. Plus, these technologies pave the way for breakthroughs in Edge Computing, which would be challenging to achieve without the foundation of NFV and SDN. I'm proud to have been at the forefront of such transformative developments.

Faster, Stronger, More Resilient

Can you shed some light on the projects you're currently immersed in?

Certainly. At present, I'm navigating three primary domains: applications resilience, automation, and IT security. I'm associated with Thryv Holdings Inc., a renowned American corporation recognized globally for its software solutions tailored for small and medium businesses. In fact, they've secured a spot in the top 10 of Selling Power's 50 Best Companies to Sell For in 2023. For the past two years, I've been contributing as a Senior DevOps Engineer.

My mandate is to usher in a new era of digital transformation for the company. This entails integrating state-of-the-art technologies and the newest cloud platforms to elevate their services.

A significant portion of my work revolves around Kubernetes, a cutting-edge container orchestration platform known for its scalability and resilience. It's a game-changer in managing containerized applications at scale. To put it simply, contemporary corporations package their applications in containers, which encapsulate both the application code and its essential dependencies. Kubernetes ensures that such applications can be seamlessly deployed across diverse infrastructures, independent of the base operating system. Moreover, should any hiccups arise, Kubernetes facilitates swift resolutions, often without the need for manual intervention.

Can you highlight the key outcomes you've accomplished in your current role?

Certainly. Spearheading the containerization initiative has yielded remarkable results. We've slashed application delivery times to the market by over 70%. Moreover, containerization has bolstered our application security, offering enhanced vulnerability monitoring and a fortified deployment platform. As a result, we've seen a reduction of over 50% in potential data breaches and leaks.

When I first joined, Kubernetes was only being utilized for select internal services, and there was a noticeable lack of enthusiasm about it among the development teams. Recognizing its potential, I championed its broader adoption. Through demonstrations and discussions, I elucidated the myriad benefits of Kubernetes. This advocacy bore fruit: we've now achieved complete automation of the application release lifecycle, eliminating the previously time-consuming manual processes.

This automation has streamlined operations to such an extent that tasks which once required extensive teams can now be managed by a handful of individuals. This efficiency translates to substantial savings in human resource costs.

Additionally, as the company's primary authority on public clouds like AWS and Microsoft Azure, I've been pivotal in cost optimization strategies. To date, I've facilitated savings of 30-40% in cloud-related expenses, which is a significant financial milestone.

These achievements have contributed to my professional experience and impacted my reputation within the company. I'm consulted on critical architectural decisions for our services. I've gained the opportunity to share my architectural vision more actively and substantially influence the final product.

Absolutely, juggling multiple responsibilities and diving deep into various facets of the job is indeed demanding. But I thrive in such dynamic environments. I firmly believe that the depth of one's commitment to their work directly correlates with the quality of the outcome.

You've highlighted your involvement in system reliability and disaster resilience. How do you measure success in these areas?

To gauge system reliability, we primarily rely on the Service Level Objective (SLO) metric. It provides a snapshot of the system's uptime, essentially reflecting the proportion of time it remains operational throughout the year. For instance, in our collaboration with Veeva Systems, a prominent company in healthcare software development, we rolled out a robust GitLab development platform designed for fault tolerance and disaster resilience. Our objective was clear: the platform should remain functional with negligible downtime, even if an entire region faces disruptions. Through rigorous testing, we ensured that the GitLab system could autonomously bounce back after a complete regional outage. Before our intervention, the SLO hovered around 90%. Post our enhancements, it soared to an impressive 99.99%.

Another pivotal metric is the Recovery Time Objective (RTO), which represents the duration required for a system to recover after a disaster. My team and I managed to trim down the RTO from a range of 6-10 hours to a mere 10 minutes. This resilience, ensuring minimal downtime even amidst significant infrastructural failures, is particularly vital for systems in sectors like healthcare.

Lastly, we monitor the Recovery Point Objective (RPO), which quantifies potential data loss during system disruptions. It's gauged in terms of time. For example, an RPO of 30 minutes implies that in a system failure, the latest half-hour's data might be lost. This was the scenario before our team stepped in. Our concerted efforts brought this down to just 10 minutes, marking a threefold enhancement. In the context of the GitLab platform, this translates to practically zero data loss, ensuring data integrity at all times.

Security is Paramount

Given the evolving landscape of cloud services, which trends do you believe are pivotal? How do you envision the future trajectory of the industry, and which innovations excite you the most?

One of the most transformative trends I've observed is the integration of machine learning to bolster IT service efficiency. ChatGPT's advancements have been a testament to this evolution, leaving a profound impact on cloud technologies. We're witnessing a concerted effort by cloud providers to weave AI-driven algorithms into their offerings.

During my tenure at Bell Integrator, I delved into machine learning technologies about five years ago. Even before it became a buzzword in the industry, I was deeply involved in an internal project centered around it, sensing its transformative potential early on.

Another significant shift is the gravitation towards hybrid infrastructure. This model entails a blend of cloud-based infrastructure complemented by traditional data centers. Such a configuration offers a win-win for both cloud providers and their clientele. Consider scenarios like peak user activity on an e-commerce platform during a sale. The sudden surge in traffic necessitates additional resources. A hybrid setup empowers companies to nimbly tap into the cloud's resources to cater to such spikes.

Lastly, but by no means least, is the emphasis on security. The frequency of data breaches, including personal data leaks, has been on an alarming rise. The pandemic-induced digital shift has made the security of online services and applications paramount. Protecting user data isn't just about compliance; it's about retaining user trust. Companies are acutely aware of the stakes, and cloud services are at the forefront, harnessing the latest in technology to ensure their clients' data remains uncompromised.

How do you weave these evolving trends into your professional endeavors?

I integrate them quite proactively. At Thryv Holdings, I'm in charge of over 50 diverse products, spanning applications, websites, and portals. Some cater to our internal teams, while others serve our broader audience. Through the automation frameworks I've instituted, we've managed to eliminate a plethora of manual interventions, streamlining operations. Machine learning plays a pivotal role in this efficiency drive.

Data protection, especially personal data, is a cornerstone of my strategy. With the burgeoning complexity of our data storage infrastructure, there's a pressing need for innovative management solutions. My focus is trained on harnessing automation and AI-driven algorithms. These tools not only enhance our vulnerability detection capabilities but also expedite the rectification process.

It's imperative to understand that my commitment isn't just to the company but extends to our users. We're routinely subjected to rigorous security audits centered on data storage. I'm proud to say that, time and again, we've met and often exceeded the industry's stringent standards.

IT Pro's Trap

With over 25 certifications under your belt, including accolades from the esteemed Red Hat professional certification program, and achieving the pinnacle with the Red Hat Certified Architect (RHCA) in 2022, how would you say these have influenced your professional trajectory?

Certifications, in essence, validate the depth and breadth of knowledge I've acquired over the years, acting as a testament to my expertise in the field.

Earning certifications, especially from globally renowned programs like Red Hat, is a significant milestone in any IT professional's journey. In the CIS region, the Red Hat Certified Architect distinction is held by a select group of around thirty individuals, and I'm honored to be part of this elite cohort.

Such recognitions undoubtedly capture the attention of potential employers, amplifying one's marketability in the competitive IT landscape.

Beyond the tangible benefits, certifications serve as a catalyst for personal and professional growth. They push me to continually expand my horizons, delve into emerging trends, and validate my knowledge. While experience lends confidence, it's crucial to avoid complacency. The tech world is ever-evolving, and staying updated is not just a choice but a necessity.

Given your extensive experience, what would you say are your standout professional strengths? And what advice would you offer to budding professionals aiming for success in this field?

One of my defining strengths is my adeptness at troubleshooting. When faced with challenges, I don't just apply a quick fix; I delve deep to identify and rectify the root cause. Often, my analysis reveals that the core issue lies within the service's architecture. In such instances, I collaborate with the development team to ensure a comprehensive solution, eliminating the problem at its source.

Years of hands-on experience have honed my ability to grasp intricate service architectures. I can intuitively map out how systems should function, understanding the broader picture rather than just a segmented view.

I also believe in continuous self-refinement. I'm candid about areas I need to bolster and take proactive steps to bridge any knowledge gaps. Keeping abreast of emerging technologies is crucial to remain relevant in this ever-evolving domain.

Engaging with professional communities, like Stack Overflow, has been invaluable. For reference, I'm proud to rank within the top 3% of its active contributors. Platforms like Upwork, where I'm recognized as a top-rated professional, are also excellent avenues for newcomers to gain exposure and experience.

In essence, a blend of relentless self-improvement and practical application of one's skills is the key to carving out a competitive edge in the industry.

ⓒ 2024 TECHTIMES.com All rights reserved. Do not reproduce without permission.
Join the Discussion
Real Time Analytics