Ravi A - Cloud DevOps Engineer |
[email protected] |
Location: Charlotte, North Carolina, USA |
Relocation: |
Visa: |
Ravi
Sr. Cloud / DevOps Engineer Email Id: [email protected] Contact: 475-777-2465 Professional Summary: Sr. Cloud & DevOps specialist with around 10 years of IT experience, including expertise in Linux and System Administration. Proficient in AWS, Azure, GCP, OpenStack, CI/CD, Configuration Management, Build/Release Management, Virtualization, and performance optimization, with a strong focus on troubleshooting. Adept at Implementing IAC principles using tools like Terraform and Ansible, enabling automated provisioning and configuration management of infrastructure. Skilled in container technologies (Docker) and orchestration platforms (Kubernetes), with hands-on experience in containerizing applications and scaling them efficiently. Proven expertise in building and maintaining CI/CD pipelines, automating testing, and enabling rapid and reliable software releases. Experienced in setting up and configuring monitoring, logging, and observability tools such as Prometheus, Grafana, ELK Stack, Datadog, and New Relic to ensure system health and performance. Proficient in designing and implementing hybrid cloud solutions, seamlessly integrating on-premises infrastructure with cloud environments to optimize resource utilization. Adept at formulating comprehensive cloud migration strategies tailored to meet organizational goals, ensuring a smooth transition to the cloud environment. Proficient in designing, implementing, and managing SaaS, PaaS, and IaaS solutions, delivering scalable and user-friendly software applications to meet client needs. Proficient in incident management, root cause analysis, and automation of incident response workflows, reducing downtime and improving system resilience. Strong focus on implementing security best practices, including vulnerability scanning, access control, and compliance checks in cloud and on-premises environments. Proven ability to scale infrastructure resources dynamically based on workload demands and optimize resource utilization for cost savings. Proficient in scripting languages like Python, Bash, and Ruby, with the ability to write custom automation scripts and tools to enhance operational efficiency. Experience in designing and implementing disaster recovery and high availability strategies to ensure business continuity. Skilled in optimizing cloud costs by leveraging cloud provider cost management tools and strategies, resulting in significant savings for the organization. Collaborative and adept at bridging the gap between development, operations, and other departments, promoting a holistic approach to software development and infrastructure management. Knowledgeable in preparing for and participating in compliance audits (e.g., SOC 2, PCI DSS) and working to address findings and implement necessary controls. Experience implementing effective change management processes to minimize disruptions during system updates and changes while maintaining system stability. Consistently participated in on-call rotations, contributing to a balanced workload, ensuring effective incident response across the team, and leveraging automation to streamline incident detection and resolution processes. Strong focus on creating comprehensive runbooks and standard operating procedures (SOPs) to guide team members during incidents, resulting in efficient incident response and reduced mean time to resolution (MTTR). Technical Skills: Cloud Platforms AWS, Azure, OpenStack, GCP Configuration Management Tools Terraform, Ansible, Chef, Puppet CI/CD Tools Jenkins, CloudBees, Bamboo, Harness, GitHub Actions, Azure DevOps, GitLab CI/CD Build Tools Maven, ANT, Gradle Containerization Tools Docker, Docker Swarm, Kubernetes, Mesos, OpenShift, AWS ECS, Azure AKS Version Control Tools Git, GitHub, Bitbucket, Subversion (SVN), GitLab Logging & Monitoring Tools New Relic, Datadog, Loggly, Nagios, Splunk, ELK Stack, CloudWatch, Azure Monitor, Prometheus, Grafana, Dynatrace Scripting & Programming Languages Bash Shell, Ruby, C, C++, XML, SQL, Java/J2EE, HTML, Perl, Power Shell, Python, .Net, Json, Yaml, Groovy, Golang Databases My SQL, MS SQL, Oracle, Dynamo DB, Cassandra, Mongo DB, AWS Aurora Application/Web Servers Web logic, Web sphere, Apache Tomcat, Nginx, Oracle application server Operating Systems Unix, Linux, Windows, Solaris, CentOS, Debian, Ubuntu and RHEL. Virtualization Platforms Oracle VirtualBox, VMware Workstation, Vagrant, VMware vSphere ESXi 5.x/4.X, ESX /3.x, Hyper-V Bug Tracking Tools Jira, Confluence, Bugzilla, Remedy, HP Quality Center, IBM Clear Quest, Mingle Repositories Antifactory, Nexus Web Technologies HTML, CSS, JavaScript, jQuery, Bootstrap, XML, JSON, XSD, XSL, XPATH. Certifications: AWS Solutions Architect Associate AWS Cloud Practitioner Foundational Professional Experience: Client: Arvest Bank, Lowell, AR March 2021 Till date Role: AWS DevOps/SRE Engineer Responsibilities: Automated provisioning of infrastructure building AWS CloudFormation stacks from the resources VPC, EC2, S3, RDS, Dynamo DB, IAM, EBS, Route53, SNS, SES, SQS, CloudWatch, Security Group, Auto Scaling. Configured Amazon Elastic Container Service (ECS) to use scale cluster size and adjust its desired count up or down in response to CloudWatch alarms. Worked on integrating AWS DynamoDB with AWS Lambda to store the values of the items and back up the DynamoDB streams. Reduced mean time to resolution (MTTR) by 30% through efficient incident response strategies and cut down incident recurrence by 50% by conducting detailed postmortems. Implemented server-side logic, database interactions, and API development in Python. Identified security issues and risks, and developed mitigation plans from Nessus (Tenable Security Center) reports. Implemented cluster services using Docker and Kubernetes to manage local Kubernetes deployments by creating a self-hosted Kubernetes cluster with Terraform and Ansible and deploying application containers. Setting up and configuring New Relic, enhanced monitoring, and application performance with New Relic. Analyzed Datadog data to optimize application and infrastructure performance. Designed and configured real-time alerts and monitoring in Splunk to proactively detect issues and security breaches. Improved system uptime to 99.9% by automating routine tasks and implementing SRE best practices, leading to a 40% reduction in manual interventions. Developed reusable Terraform modules, streamlining infrastructure provisioning and promoting best practices in module design. Configured Harness workflows, environments, and integrations to streamline application delivery. Developed Ansible Playbooks with Python SSH as the Wrapper to Manage Configurations of OpenStack Nodes and Test Playbooks on AWS instances using Python. Created and maintained Helm charts for packaging and deploying applications on Kubernetes, simplifying application lifecycle management. Integrated Artifactory seamlessly into Jenkins CI/CD pipelines to automate artifact deployment and retrieval processes. Configured Artifactory to proxy and cache external repositories to improve build and deployment times. Chef Recipes were created to configure, deploy, and maintain software components of the existing infrastructure. Developed various cookbooks in Chef using the Ruby scripting language. Set up and configured API gateways to manage API traffic, routing, and security. Implemented API security measures, including OAuth 2.0, API keys, JWT, and TLS/SSL encryption. Proficient in deploying and managing virtual machines, networks, and storage resources on Google Cloud Platform (GCP) using services like Compute Engine, Virtual Private Cloud (VPC), and Cloud Storage. Additionally, developed, and deployed serverless applications using services such as Cloud Functions and Cloud Run, effectively optimizing resource utilization and scalability. Utilized Dynatrace APM to monitor application response times, error rates, and transaction traces. Designed and built custom dashboards and alert policies in Dynatrace to monitor specific key performance indicators. Configured additional Docker Slave Nodes for Jenkins using custom Docker Images and pulled them to the Cloud. Worked on all major components of Docker like Docker Daemon, Hub, Images, Registry, and Swarm. Orchestrated complex release pipelines with dependencies, approvals, and rollback strategies in Harness. Utilized AWS Beanstalk to specify and track system configurations, prevent drift, and uphold software compliance for EC2 and on-premises configurations. Designed and implemented complex contact flows in AWS Connect to automate customer interactions, including IVR menus, call routing, and dynamic customer data retrieval. Automated build and deployment process utilizing Bash, Python, and Shell scripts. Successfully deployed and configured Adobe Experience Manager (AEM) instances, designed and configured Adobe Cloud Services (ACS) for high availability and scalability, ensuring optimal performance during traffic spikes and peak usage periods. Defined and negotiated Service Level Agreements (SLAs) with stakeholders, established measurable Service Level Objectives (SLOs) for critical services, and implemented Service Level Indicators (SLIs) to monitor and ensure compliance, improving customer satisfaction and service reliability. Implemented and maintained security controls to align with SOC 2 criteria, ensuring the protection of customer data and system availability. Assumed a key role in ensuring PCI DSS compliance by implementing and maintaining security controls related to payment card data protection. Serve as part of an on-call rotation, escalation contact, and respond to system and application availability incidents. Environment: AWS (VPC, EC2, S3, RDS, Dynamo DB, IAM, EBS, Route53, SNS, SES, SQS, Connect, CloudWatch, and Aurora), Google Cloud (GCP), Terraform, Packer, Bitbucket, Git, Jenkins, Harness, Opsgenie, FreeIPA, OpenVPN, Tenable.io, Jira, Confluence, Chef, Ansible, Docker, Splunk, Python, Datadog, New Relic, Loggly, Dynatrace, JFrog Artifactory, Bash, Adobe, Unix/Linux. Client: DaVita, Denver, CO Aug 2019 Feb 2021 Role: Sr. Azure Cloud/DevOps Engineer Responsibilities: Experience working with Azure Web applications, Azure Data Factory & Storage, Azure Media & Content Delivery, Azure Networking, Azure Hybrid Integration, and Azure Identity & Access Management. Created U-SQL script for transform activities and developed complex queries to transform the data from multiple sources and outputted the data into Azure Data warehouse. Configured Azure Blob storage services for storing massive amounts of unstructured object data that can be accessed over HTTP and HTTPS. Experience working with Azure PaaS and IaaS solutions, such as Azure Web Apps, Web Roles, Worker Roles, SQL Azure, and Azure Storage. Utilized Kubernetes to deploy, load balance, scale, and manage Docker containers with multiple namespace versions. Implemented synthetic monitoring tests in Dynatrace to simulate user interactions and monitor application availability. Successfully implemented and administered Splunk environments, including managing data ingestion, indexers, and search heads. Employed GCP's CI/CD tools like Cloud Build and Deployment Manager to automate the software delivery process, reducing deployment time and minimizing errors. Implemented infrastructure as code (IaC) practices using Terraform to manage GCP resources. Implemented Azure Databricks to other Azure services (such as Azure Storage) in a more secure manner using service endpoints. Published and consumed artifacts from Artifactory repositories as part of the build and deployment workflow. Integrated Prometheus and Grafana with alerting channels such as Slack, PagerDuty, or email for timely notifications. Integrated automated testing frameworks (e.g., Selenium, JUnit) into Harness pipelines to ensure application quality. Utilized CloudBees Artifactory integration to manage and version artifacts, ensuring consistency and traceability in the deployment process. Analyzed Datadog data to optimize application and infrastructure performance. Defined and instrumented custom metrics within applications and services for Prometheus. Implemented dashboard templating and variable usage in Grafana to create dynamic and flexible dashboards. Deployed web apps (web and worker roles) on Azure, using PowerShell. Worked on orchestration tools like Terraform, leveraging modern tools like Vault, Kubernetes, Docker, and Kafka. Experience partitioning Kafka messages and configuring replication factors in a Kafka cluster. Implemented Azure Active Directory (AD) to enable single sign-on for thousands of SaaS cloud services including Dropbox and Office 365. Managed major architectural changes from a single server large software system to a distributed system with Kubernetes orchestration. Worked on Chef as a configuration management tool and created Chef Cookbooks and modules to automate system operations using Ruby programming language. Configured servers to host Team Foundation Server (TFS) instance to set up and manage continuous integration using TF Build Service & deployed JAVA applications to JBoss and Oracle Tuxedo servers. Utilized ARM templates with Azure Data Factory (ADF) to implement CI/CD pipelines. Environment: Azure (VM, Resource Manager, Active Directory, Azure Monitor, ARM, Azure DevOps, IoT Hub, Data Factory), Agile Methodology, AVN s, TFS, IBM Clear Case, GIT, Maven, Jenkins, Chef, Datadog, Docker, Kubernetes, JBoss, Splunk, Harness, CloudBees, Terraform, Dynatrace, Nagios, JIRA, Confluence, Kafka, Jfrog Artifactory. Client: Verisk Analytics, Jersey City, NJ Dec 2017 July 2019 Role: AWS Cloud/DevOps Engineer Responsibilities: Developed and supported key pieces of the company's AWS cloud infrastructure. With Opscode, a large deployment of Ubuntu Linux instance systems was built and managed. Experience working on Identity Access Management (IAM), creating users, groups, and assigning roles based on their duties. Involved in SSL Security certificate management for enterprises, maintaining certificates across multiple SSL providers, and integrating certificates into products such as Apache, tomcat, and AWS-ELB. Developed Chef Cookbooks for multiple DB configurations to modularize and optimize the configuration of the end product. Converted production support scripts into Chef Recipes to provision AWS servers. Performed development and version control of Chef Cookbooks, testing of Cookbooks using Test Kitchen, and running recipes on nodes managed by on-premises Chef Server. Integrated SonarQube with Jenkins using Maven to get the Quality Analysis for all the Projects pre-deployment. Discussed the report with developers to clarify the SonarQube Report and aid in improving code quality. Used security groups, network ACLs, Internet Gateways, NAT instances, and Route tables to ensure a secure zone for organizations in the AWS public cloud. Creating and managing NAT and Proxy instances in AWS, as well as route tables, EIPs, and NACLs. Configuration of a Virtual Private Cloud (VPC) with networking of server subnets. Developed Lambda functions in Python for AWS Lambda and invoked Python scripts for data transformations and analytics on large data sets in EMR clusters and AWS Kinesis data streams and configuration management platforms such as Kafka. Developed a Continuous Delivery pipeline using Docker, Jenkins, and GitHub. Responsible for installation & configuration of Jenkins to support various Java builds and Jenkins plugins to automate continuous builds and publishing Docker Images to the Nexus Repository. Docker engine and Docker Machine environments were used to deploy micro services-oriented environments for scalable applications, Docker swarm for cluster and container scheduling, and Docker compose for defining numerous container applications. Leveraged New Relic distributed tracing to gain insights into distributed systems and microservices architectures. Docker Swarm and Kubernetes were used for container orchestration, and containers were spun up by writing Docker files and configuring the automated build on Docker HUB. Kubernetes Replication controllers have been configured to allow multiple pods, such as the Jenkins master server, to run in multiple minions. Managed Kubernetes charts using Helm, managed Kubernetes manifest files, and created reproducible builds of the Kubernetes application. Created Kubernetes (k8s) Clusters for running microservices and deployed microservices using infrastructure powered by Kubernetes. Creation of Ansible playbooks for automating Kubernetes clusters. Virtualized servers in Docker under the needs of development and test environments and set up automation using Docker containers. Integrated Prometheus with Grafana for creating informative dashboards and visualization of metrics. Worked on a Splunk VMware application to schedule components that manage data collecting tasks for API data. The collection nodes and the collection configuration dashboard work together to coordinate the data flow. Automated Java builds with Maven and implemented numerous plugins for code analysis, Junit, code coverage, PMD, SonarQube, and other tools. Installed and managed the Artifactory repository for deploying Maven artifacts. Environment: AWS, GitHub, SonarQube, Jenkins, Maven, Nexus, Ansible, Chef, LVM, Splunk, Nagios, DynamoDB, Python, shell scripting, Linux, Prometheus, Grafana, New Relic, Docker, Kubernetes, Terraform. Client: College Board, Henrico, Virginia Mar 2016 Nov 2017 Role: Azure Cloud/DevOps Engineer Responsibilities: Deployed OpenShift Enterprise v3.4/3.6 on RedHat 7 environment and integrated with private Docker Registry. Managed OpenShift master, upgraded nodes, and decommissioned them from active participation by evacuating and upgrading the nodes. Used OpenShift to increase application security by placing admin consoles on several Docker containers that can only be accessed via unusual port numbers. Had extensive experience with ansible automation in replacing various OpenShift components such as master, compute, infra node, and glusterFS. Assisted developers in creating and implementing proper branching, labeling/naming conventions through the use of GIT source control. Worked on Agile Methodology and was responsible for monitoring and managing development, production, and working with Azure Portal to provide IAAS resources to clients. Experience working with Azure PaaS and IaaS solutions like SQL Azure, Azure Storage, Web Roles, Worker Roles, and Azure Web Apps. Created the Dashboards in Splunk & Grafana for logging and monitoring metrics for OpenShift Applications. Configured applications that run multi-container Docker applications using the Docker-Compose tool and a YAML-configured file. Application containers were deployed as Pods and managed using Kubernetes' nodes, Config-Maps, selectors, and Services. Set up Docker to automate container deployment through Jenkins and dealt with Docker Hub, creating Docker Images and taking care of numerous Images primarily for middleware establishments. Docker was used to successfully set up a Dev & Test environment for Developers and Testers using Docker Build & Docker Compose. Developed Terraform scripts to deploy platform standard services such as the CI/CD tool Jenkins, the Configuration Management Tool Chef, and the Secrets Management Tool Vault. Experience working with Azure Web applications, Azure Data Factory & Storage, Azure Media & Content Delivery, Azure Networking, Azure Hybrid Integration, Azure Identity & Access Management. Utilized service endpoints to integrate Azure Databricks more securely with other Azure services (like Azure Storage). Implemented cluster services using Docker and Kubernetes to manage local Kubernetes deployments by creating a self-hosted Kubernetes cluster with Terraform and Ansible and deploying application containers. Designed, planned, and deployed Azure virtual machines, as well as implemented and managed virtual networking within Azure and connected it to on-premises systems. Worked on Azure Databricks to use custom DNS and establish network security group (NSG) rules to enable egress traffic limitations. Designed and implemented RBAC policies in Azure AD to control access to Azure resources. Used Ansible and Ansible Tower as Configuration management tools, to automate repetitive tasks, quickly deploy critical applications, and manage changes proactively. Utilized Ansible for Configuration Automation and Centralized Management. Implemented Ansible to automate the build and configuration of new servers as well as the management of existing servers. Integrated Automated Build with Deployment Pipeline. Installed Chef Server and clients in order to get the Build from the Jenkins repository and deploy it in the target environments (Integration, QA, and Production). Installing and configuring the Jenkins slave and master nodes. Chef was used to build a CI/CD pipeline and manage the infrastructure as code. Environment: Azure, OpenShift, Office 365, Maven, Jenkins, Chef, Ansible, Azure Artifacts, Docker, Kubernetes, Git, PowerShell, Jira, RHEL, ELK, PowerShell scripting, python script, Terraform. Client: CredenceIS Computing Pvt Ltd, INDIA June 2013 Dec 2015 Role: Linux Administrator Responsibilities: Day-to-day responsibilities included Linux server maintenance and support for the developer team's application, tuning, troubleshooting, and software operating issues. Installed Oracle's most recent fixes on servers running Red Hat Linux, Linux environment configuration, and management of Send mail, Samba, and Squid servers. Set up Linux Cron jobs to automate different build-related and application data synchronization jobs. Responsible for the installation of Linux OS servers utilizing the kickstart automation tool. Configured Kickstart and Jumpstart servers to initiate installation of RedHat Linux and Solaris on several machines at once. Previous LDAP tools were updated to operate with the latest Ruby Rails version. Involved in installing, configuring, and upgrading RedHat Linux AS 4/5, and Solaris 9/10 OS. Performed automated OS installations using Kickstart for Red Hat Enterprise Linux5/6 and Jumpstart for Solaris 9/10 Linux. Administration and support of Linux distributions, including Linux Enterprise Desktop, SUSE Linux Enterprise Server, RedHat, and Solaris. Install, maintain, and upgrade Drupal and Word press on LAMP stack and Configured LAMP Stack on Unix/Linux servers. Configured the NIS, NIS+, and DNS on RedHat Linux 5.1 and update NIS maps and organized the RHN Satellite Servers in combination with RHN Proxy Server. Worked on Linux package installation using RPM and YUM, and system provisioning with LVM. Developed, modified, and built packages on Solaris and rpms on Linux for deployment on various servers through the Software Development Life Cycle. Environment: Oracle on Red hat Linux, Samba, Squid, RedHat Linux AS 4/5, Solaris 9/10, Linux Enterprise Desktop, SUSE Linux Enterprise Server, RedHat and Solaris, LDAP. Keywords: cprogramm cplusplus continuous integration continuous deployment quality analyst sthree database active directory information technology hewlett packard microsoft Arkansas Colorado Idaho New Jersey |