Sai Kaushik
Development
Texas, United States
Skills
DevOps
About
SAI KAUSHIK's skills align with Programmers (Information and Communication Technology). SAI also has skills associated with Consultants and Specialists (Information and Communication Technology). SAI KAUSHIK has 7 years of work experience.
View more
Work Experience
Sr. DevOps Engineer/ Site Reliability Engineer (SRE)
C-HIT Maryland
December 2022 - Present
- Description: As a DevOps engineer/SRE at C-HIT, a forward-thinking Health Information Technology (HIT) software development firm, I am a part of Cloud Architecture and Management Team to implement advanced data analytics tools and developed a singleview dashboard integrating data from 60+ applications and 10 data centers. Improved in collaboration, operational visibility, and predictive analytics. Successfully prevented outages during peak times without extra resources. Responsibilities: • Integrated monitoring tools to enhance MTTD, MTTA, MTTR. Improved observability for platform efficiency. • Led Root Cause Analysis, proposed enhancements for applications, tools, and processes. Streamlined issue pinpointing. • Developed diagnostic tools, test harnesses for proactive monitoring. Enhanced troubleshooting for system issues. • Reviewed and ensured seamless integration of new releases into production environment. • Leveraged AWS Cloud Services such as EC2, auto-scaling and VPC to build secure highly scalable and flexible systems that handled expected and unexpected load bursts. • Configured storage on S3 buckets, utilizing Amazon SDK wrapper libraries to define Lifecycle Configuration Rules for S3, enabled actions like Transition and Expiration for groups of objects within the bucket. Deployed webpages into S3 buckets. • For ensuring the secure zone inside company's used Network ACLs, Internet Gateways, NAT instances and Route Tables, for setting up new servers on AWS and for setting up the lifecycle policies to back the data from AWS S3 to AWS Glacier, utilized Terraform Migrate Legacy and monolithic systems to AWS. • For making the Virtual firewalls which supports the traffic flow for one or more AWS EC2 instances, have defined the AWS Security Groups and ACL's. Used the AWS VPC, for configuring the VPN tunnel to communicate inside network. • For creation, management and deploying of VMs on Azure used Azure Infrastructure, Azure Web Roles, Worker Roles, VM Role, AzureSQL, Azure Storage, Azure AD Licenses and Virtual Machine Backup. • Composed Terraform templates and designed, configured Azure Virtual Networks (VNets), Subnets, Azure network settings, DHCP address blocks, DNS settings, and security policies. • Used Azure Auto-scaling, API management and REST APIs for deploying the Azure IaaS virtual machines (VMs) and PaaS role instances into secure VNets and subnets, Utilized Azure Kubernetes service (AKS), for cluster management. • Utilized Azure Kubernetes Service (AKS) for container-based applications, managing multi-container services, monitoring cluster and pod health, and implementing Azure service mesh atop the Kubernetes platform. • Used Virtual Network to deploy the agent nodes, Ingress API Gateway, MySQL Databases and Cosmo DB for stateless storage of external data, and by managing the Machine Learning Lifecycle. • Developed Infrastructure as Code (IaC) templates using Terraform to construct staging and production environments, defined reusable Terraform modules like Compute, Network, Operations, and Users for versatility across various setups. • Developed Python and Go automation scripts for creating resources in OpenStack using Terraform modules. Utilized Python for Kubernetes-based container orchestration. • Leveraged Helm for Kubernetes chart management, apps for builds, and manifests for templating. Orchestrated deployments with Helm packages, provisioning parameters. Established and configured Kubernetes clusters for pods, controllers, sets, services, deployments, labels, health checks, and ingress using YAML files. • Used Kubernetes to deploy, scale, load balance and manage Docker containers with different names, spaced versions and developed test environment on Docker containers. • Led the implementation of Docker to containerize applications, enhancing their portability and scalability, created Docker images and Docker files to package applications and their dependencies in optimizing deployment. • Used Docker snapshots and volumes to ensure data integrity and efficient backup strategies within the containerized environment, managed Docker registries to store, distribute, and version Docker images, improving deployment process. • Collaborated with the team to integrate Docker technology into the CI/CD pipelines using the Docker-maven-plugin. Orchestrated Jenkins jobs and pipelines to automate the build, test, and deployment processes of containerized applications. • Spearheaded the setup and configuration of Kubernetes for orchestrating Docker containers, allowing for efficient management of application clusters. Engineered a node cluster using Kubernetes, optimizing resource utilization. • Contributed to the development of the Minimum Viable Product (MVP) in the cloud environment, showcasing the successful migration and deployment of critical applications. Collaborated with cross-functional teams to ensure the seamless integration of Docker containers with Kubernetes, streamlining deployment and scaling processes. • Used Jenkins for the weekly build, test, and deploy chain, Integrated Jenkins with GIT for Dev, Test and Prod branching models for weekly releases and implemented continuous build based on check-in for applications. • Utilized Maven, Nexus, Jenkins, Git, and JIRA, storing build artifacts in Nexus. Developed custom Jenkins jobs and pipelines incorporating Bash/Shell scripts for infrastructure provisioning while implementing CI/CD framework using Jenkins and Maven. • Used GIT source control to establish and apply the appropriate branching, labeling/naming conventions while coordinating with the developers, Resolved the conflicts and analyzed in merging the source code of GIT. • Utilized AppDynamics in creating and managing the service availability, also deployed the AppDynamics for mission critical applications, alert configuration in AppDynamics for Service and Server unavailability. • Designed and deployed ELK clusters including Elasticsearch, Logstash, Kibana, Beats, Kafka, and ZooKeeper. Utilized Elasticsearch for search functionality and ELK stack for logging and system monitoring with Beats. Proficient in managing and optimizing Spark and Hadoop clusters for large-scale data processing. • Used scripting languages such a Shell, Perl, Python for build deployment, build scripts, and automated solutions, also scripted in Groovy to automate the collection analysis of Cassandra. • Used JIRA in deploying and configuring along with that hosted and local instances for issue tracking, workflow collaboration, and tool-chain automation. Environment: AWS, Azure, Terraform, Docker, Kubernetes GIT, GitHub, Bash, Python, Maven, SonarQube, Nexus, Jenkins, Chef, Linux, Unix, Apache Tomcat, Spark, Hadoop, Redshift, GO, .Net, Java, Python Unix, Shell, React, ELK Stack, AppDynamics, VMware, Windows, Shell, Pearl, Jira etc.
DevOps Engineer
Verizon Communications Inc
April 2022 - December 2022
- Description: Verizon is a trailblazing telecommunications powerhouse, revolutionizing connectivity, and digital experiences through relentless innovation. As a DevOps Engineer, I was in Enterprise Cloud Platform Team, where I was responsible for building the infrastructure for other teams. In the part of Billing Team, I have worked on an application where we have provided clients detailed insights into cloud resource consumption and control over their cloud service expenditures. Responsibilities: • Experienced in managing and configuring the Google Cloud Platform (GCP) services like Compute Engine, Cloud Load Balancing, Cloud Storage, Cloud SQL, Stack driver monitoring and Cloud Deployment Manager etc. • Used GCP cloud services like Compute, Auto-scaling and VPC (Virtual Private Cloud) for building and securing the highly scalable and flexible systems which handles expected and unexpected outbursts. • Set-up and build AWS Infrastructure using resources VPC, EC2, S3, RDS, Dynamo DB, IAM, EBS, Route53, SNS, SES, SQS, CloudWatch, CloudTrail, Security-Group, Autoscaling & Amp; RDS using CloudFormation templates. • Set up IAM policies for Users, SubNets, NAT to ensure successful deployment of Web applications, database templates and security groups by creating a custom sized VPC, designed by Cloud formation templates. • Used CloudWatch in AWS for implementing and maintaining the monitoring and alerting of the production servers, stored in AWS, assigned the elastic IP addresses to work around host and availability zone failures by remapping instance addresses. • Architected and executed Terraform configurations to automate the provisioning and management of multi-environment infrastructure, encompassing Developing, Staging, Pre-production, Production, and Disaster Recovery setups in Verizon. • Leveraged Terraform modules, variables, and state management to ensure scalability, maintainability, and adherence to industry standards, facilitating seamless environment deployment and configuration across cloud platforms. • Installed, configured, and managed Docker containers and images for webservers and applications. Implemented the Dockermaven-plugin to build Docker images for all microservices and utilized Docker files to build Docker images from Java JAR files. • Created Docker images using a Docker file and worked on Docker container snapshots, removing the images and managed the Docker Volumes, virtualized the servers in docker QA, configured automation using the Docker containers. • Orchestrated the deployment, scaling, and management of Docker containers using Google Kubernetes Engine (GKE), Managed local deployments in Kubernetes, created local clusters, and deployed application containers. • Utilized GKE clusters in rollback and rollout methods within deployment strategies. Integrated Istio and configured packages with GKE clusters for service mesh functionality. • Proficient in using Horizontal Pod Auto-scalar (HPA) and Vertical Pod Autoscaler (VPA) to scale the available resources in the container, for scaling the nodes inside of the cluster used Cluster Auto-scaler. • By writing YAML files in GKE, created clusters, pods, Replica Sets, Services, deployments, labels, and health checks. Managed multiple Kubernetes pod containers for scaling and auto-scaling. • Implemented Secret Manager for securely managing API keys, passwords, and certificates. Used Audit logging with Cloud Audit for logs to generate an audit entry to interact with Secret Manager and used those logs for abnormal access patterns. • Leveraged Ansible to spearhead streamlined configuration management processes, orchestrating the automated deployment of web applications, and meticulously overseeing environment management tasks. • Developed customized Ansible Roles to effectively configure, deploy, and maintain software components. Engineered Ansible Playbooks to enhance existing processes, elevating AWS Server Provisioning and configuration. • Expertly orchestrated end-to-end DevOps workflows by configuring Jenkins pipelines, seamlessly integrating code from GitHub. Leveraged Maven to drive build processes within these orchestrated workflows. • Utilized SonarQube for code analysis, ensuring code quality, and streamlined artifact management through Nexus artifactory to facilitate efficient development and deployment lifecycles. • Orchestrated Bitbucket pipelines to streamline automated builds and deployments, in managing pre, post, and release stages. • Seamlessly integrated Bitbucket with essential plugins like SonarQube for code analysis and Maven for efficient application transformation into .jar and .war files. • Implemented artifact integrity management within Nexus Artifactory, enhancing the efficiency of development workflows. • Engineered and tailored Datadog reports, dashboards, and visualizations to furnish actionable insights into system performance and application behavior, elevating the team's decision-making capabilities. • Expertise in using the Bash, Python and Ruby for creating, maintaining, and automating the scripts. Developed Cron jobs and Shell Scripts for automatic installations to extract logs and administrations tasks like file system management, process management, backup and restore. • Used ServiceNow as a ticketing tool to create workflows for the IT service management. Created reports in ServiceNow for better tracking and process improvements. Environment: GCP, AWS, Terraform, Maven, Nexus, Docker, Kubernetes, Ansible, Bitbucket, GitHub, SonarQube, Datadog, Dynatrace, MYSQL, SQL, MongoDB, WebLogic, JBoss, NodeJS, Bash, Python, Ruby, Shell, ServiceNow etc.
DevOps Engineer
AVIVA HEALTH
September 2018 - August 2021
- Description: I was a part of the Data, Analytics Platforms (DAP) infrastructure team, where I was responsible for contributing to the development of centralized design patterns that can be used across Life, GI, and other Aviva business areas. I was experienced in integrating various cloud (SaaS and PaaS) solutions as well as on-premises applications. Responsibilities: • Spearheaded the implementation of sophisticated continuous deployment pipelines leveraging AWS technologies like EC2, AWS S3, AWS VPC, AWS ELB, AWS RDS, DynamoDB, AWS Lambda, and Auto Scaling, significantly expediting software releases and reducing manual interventions. • Orchestrated seamless integration of deployment tools such as Chef, CloudFormation, and Bamboo, enabling efficient infrastructure provisioning, configuration management, and automated build deployments. • For infrastructure, Designed and managed AWS-based architectures, utilizing VPCs, subnets, NACLs, and security groups to establish robust network segmentation and security controls, while optimizing traffic routing with Route53. • Leveraged CloudFormation and Chef to code the infrastructure as code (IaaC), for consistency across environments and enabling repeatable deployments of resources such as EC2 instances, RDS databases, and Lambda functions. • For Containerization and orchestration, Utilized Docker to create portable and isolated containerized environments by expertly crafted Docker images and Docker files to encapsulate applications dependencies and runtime environments. • Used Chef for the configuring the EC2 instances, by creating Chef Cookbooks, Recipes for configuring the infrastructure. • For Automation and scripting, developed automation scripts using Python and Boto3 to interact with AWS services, Automated routine tasks using Bash scripts, optimizing system administration processes on diverse Linux distributions. • Employed GIT for version control, ensuring organized code management and effective collaboration within cross functional teams by utilizing branching strategies and pull requests, also collaborated seamlessly with development teams, using tools like JIRA for issue tracking and Agile methodologies for project execution and provide Software releases. • Configuration Management and monitoring using the Chef to ensure consistent system configurations and application deployments across multiple environments by enhancing stability and reliability. • For database management, managed MySQL databases including provisioning RDS instances and optimizing query performance, contributing to data integrity and application efficiency. • Automated build processes using Bamboo and ANT (pom.xml), promoting continuous integration and reliable build artifacts, for building automation, For Automation Management and for Artifact Management leveraged JFrog as a central repository for storing and managing build artifacts, streamlining application delivery and version control. Environment: AWS, Docker, Chef, Bamboo, CloudFormation, Bash, Python, Java, Maven, Groovy, Git, Kubernetes, BigPanda, ANT, Jira, RedHat Linux, JFrog, MySQL etc.
DevOps Engineer/ System Administrator
Santander Bank
March 2016 - September 2018
- Description: Santander Bank is an American Bank and offer an array of financial services and products. I have worked in the service team as the DevOps Engineer and subsided with the duties as a system admin as well. I am responsible for configuring the build and release activities on web applications and responsible for maintaining and deploying builds in UAT and Production environments. Responsibilities: • Responsible for maintaining the IT infrastructure, including servers, networks, and software systems, installing, configuring, and updating OS, managing user accounts, ensuring data security through firewalls and encryption, and monitoring system performance to identify and address any issues promptly. • Configured diverse Azure Virtual Machine images to align with the project requirements. • Utilized Azure DevOps for streamlined project management, encompassing source code management, continuous integration, and deployment pipelines, enhancing collaboration and efficiency within the development team. • Implemented Azure Load Balancers with Virtual Machine Scale Sets (VMSS) to ensure robust application availability and scalability. Utilized Azure Monitor extensively to monitor and manage system performance, • Utilized Docker to containerize microservices, streamlining deployment and consistency across various environments. • Employed Puppet to automate configuration management tasks, maintenance of virtual machines and cloud resources. • Integrated Grafana with Prometheus to establish a robust monitoring framework, providing real-time insights into key metrics and facilitating proactive issue identification. Environment: Microsoft Azure, Azure DevOps, Azure Virtual Network Setup, Subnet Configuration, Network Security Groups (NSGs), Route Tables, Docker, Ansible, Python scripts, Grafana, Prometheus, Grafana dashboards.