Kiran - DevOps Engineer |
[email protected] |
Location: Dallas, Texas, USA |
Relocation: |
Visa: |
Kiran Budde
DevOps Engineer Phone#: Email ID: [email protected] LinkedIn: DevOps/Site Reliability Engineer with 8+ years working background in Development, Automation, Support, SME and Designing and building Continuous Integration and Continuous Deployment pipelines and frameworks using tools, languages and technologies such as Python, Scala, SQL, Shell Script, Docker, Snowflake, Git, Jenkins, SonarQube, Ansible, Spark, Hadoop using cloud Infrastructure like AWS Services such as Elastic Map Reduce (EMR) and EC2, S3, SNS, SQS, ECS, Lambda, etc. and exposure to Azure and GCP cloud technologies using SDLC policies. CAREER SUMMARY: Experience in working as an L3 operations support and SRE (site reliability engineer). Created CICD pipelines by integrating Git for code checkouts, SonarQube for code quality, and maven/sdist/bdist as build tools, and other tools such as Nexus/Jfrog antifactory, ServiceNow, PagerDuty, Mend along with the AWS services. Strong Knowledge on DevOps tools such as Ansible, Terraform, Jenkins, Maven, GIT, Nexus and Docker. Strong Knowledge as SRE for maintaining application and infra by staying on-call rotations and debugging issues for production issues. As an SRE, I provided 100% of my contribution to stay on top of all issues and meet the SLAs on time. As an SRE, performed Root Cause Analysis and provided reports and confluence blueprints on issue root cause and resolutions. Experience in working with ITIL and ITSM tools such as Service Now and HPSM. Experience working on migrating applications from internal data centers to AWS instances by setting up the Security measures, Load Balancer and Virtual machines, Auto Scalers to design cost effective, fault tolerant and highly available systems using both AWS services. Experienced with both framework and Infrastructure as Code deployment (to automate Cloud environment creation. Ability to deploy on AWS and Azure, using build scripts (Boto 3 & CLI) and automates solutions using Shell/BASH and Python scripts. Strong knowledge and experience on AWS Cloud services like EC2, S3, EBS, RDS, VPC, and IAM also familiar with Cloud watch, Elastic IPs on AWS and GCP services such as Compute engines, auto scalers, VM Managers, App engines, IAM, Cloud Storage, Cloud Tasks, Data Proc, Big Query, Cloud storage. Built the AWS cloud formation and GCP cloud deployment manager templates to deploy the AWS cloud stacks and GCP cloud resources. Managed Linux and Windows virtual servers by configuring and monitoring distributed and multi-platform servers using Ansible. Strong Knowledge of Splunk and Splunk query language for analyzing the machine data and reporting. Developed automation to quickly and rapidly deploy instances from hardened machine images. Provisioned the highly available resources using Terraform and wrote new plugins to support new functionality in Terraform. Managed AWS, Azure and GCP infrastructure as code using Terraform. Responsible for working on several Docker components like Docker Engine, Docker Hub, Docker Compose, Docker Registry and Docker Swarm. Responsible for Creating and Managing User accounts, password, security, rights, disk space and process monitoring in CentOS, Red Hat Linux and Solaris. Experience in working with package manager and command line tools such as Apt Get, DPKG, YUM, RPM and GRUB. Experience in configuring scalability for web applications using Infrastructure as code tools such as Terraform and Ansible and Cloud Formation. Experience in Troubleshooting and Backup and Recovery, Monitoring in different environments such as cloud, VMware and local machines. A highly motivated, energetic individual, a team player with excellent communication and interpersonal skills Good understanding of Software models and Life Cycle. Ability to learn new technologies and work with new best practices. EDUCATION: Silicon Valley University, San Jose, CA - master s in computer science Graduation: Dec 2016 JNTUH university, Hyderabad, India Bachelor of Technology in Information Technology Graduation: May 2015 TECHNOLOGIES: Programming and Scripting Languages Python, Scala, Java, SQL, SnowSQL, Shell Databases Snowflake, PostgreSQL, MySQL Containerized/Virtualization Docker, AWS ECS, VMware Cloud Services AWS EMR, EC2, Lambda, ECS, S3, Route53, Load Balancer, SNS, SQS, EFS, RDS, Cloud Formation, Security Groups, GCP Cloud Compute engines, IAM, Cloud Storage and Azure VM CI/CD Tools Jenkins, SonarQube, Ansible, Puppet. Version control Git, Bit Bucket Monitoring PagerDuty, Splunk ITIL and Other ServiceNow, JIRA, HPQC, Spark, Hadoop, JFrog, Nexus Web services Nginx, Gunicorn, WSGI, Uvicorn, Fastapi, Flask, Apache Tomcat. CERTIFICATIONS: AWS Certified Developer Associate PROFESSIONAL EXPERIENCE: EQUIFAX ( REMOTE ) JULY 2024 - NOW DEVOPS/SRE ENGINEER PROJECT DESCRIPTION: As a DevOps and Site Reliability Engineer working on migration projects to merge from Legacy systems to AWS cloud and supporting development teams on issues and mitigating them along with permanent automation tasks and solutions using terraform, AWS cloud services. RESPONSIBILITIES: Experience in working with terraform and AWS. Experience in working with Jenkins and CICD pipelines. Strong knowledge in debugging and automation using Shell/PowerShell, Python. Strong knowledge in CICD and SDLC along with understanding of Jira methodologies. Experience in working with tools such as HashiCorp Vault, GCP and Kubernetes. Experience in working to migrate infra from legacy to AWS cloud using Jenkins CICD and Terraform. Experience in debugging application issues and infra issues for both Linux and Windows environments. Experience in Windows patch management and Fixing vulnerabilities. Experience in installing and maintaining Security and Vulnerability tools such as Qualys and Sentinel One and Resolve CVE vulnerable issues. Created automation to create/update/move Vault secrets. Created automation to fix day to day issues observed during migrations. Created automation for 60 days Image rotation to keep AMI and images in Latest and Vulnerability free. WALMART ( REMOTE ) MAY 2024 - JULY 2024 DEVOPS ENGINEER PROJECT DESCRIPTION: As a DevOps Engineer worked on a migration project for secrets management from Terraform Vault to A keyless that supports critical applications responsible for critical business transactions. Created automation scripts and pipelines using Jenkins, Python and several internal CICD tools such as concord and cloud services such as WCNP ( Walmart internal cloud ) built on top of Kubernetes platform using azure services. RESPONSIBILITIES: Strong knowledge of Python, Jenkins, Terraform vault and CICD tools services. Strong knowledge on building REST API s. Strong knowledge on creating Jenkins pipelines using languages such as python, bash shell scripting and PowerShell. Knowledge in Kubernetes, node deployments and tools such as Kubectl to maintain and configure pods and services within the Kubernetes platform. Experience in working with Jira services and creating documentation pages such as confluence pages. Experience in working with SDLC methodologies in a day-to-day work environment. Team interactions and building strong relations with the DEV and SRE teams and working together to support and maintain the critical applications. Experience in deploying and managing Kubernetes clusters in production environments. Experience in using Kubernetes primitives like pods, services, deployments, and stateful sets. Experience in automating Kubernetes operations with tools like Helm and Ansible. Experience with troubleshooting Kubernetes-related issues and optimizing cluster performance. Experience with Troubleshoot issues using Kubectl and check reports in k9 dashboards for sys logs and other activities. CAPITAL ONE, PLANO, TX SEPTEMBER 2018 MARCH 2024 DEVOPS ENGINEER PROJECT DESCRIPTION: As a Site Reliability Engineer (SRE), DevOps our goal is to support the Capital One Internal applications, frameworks and Infrastructure hosted in multi cloud environments such as AWS providing zero downtime and zero tolerance support for our customers. Automating the day-to-day activities and providing the data within SLA and monitoring the applications 24/7 along with developing and maintaining the microservices and frameworks using Python, Scala, Java and Snowflake. Build, maintain and support pipelines using Jenkins, SonarQube, Docker in AWS Cloud Services. These frameworks help the Data Analysts/Engineers develop the business metrics and generate business insights. RESPONSIBILITIES: Strong Knowledge as SRE for maintaining application and infra by staying on-call rotations and debugging issues for production issues. As a SRE, I provided 100% of my contribution to stay on top of all issues and meet the SLA s on time. As a SRE, performed Root Cause Analysis and provided reports and confluence blueprints on issue root cause and resolutions. Developed streamline frameworks for Data processing and maintained the application support and production support. As a part of the support team, worked in the 60-day retention program by recreating a complete production environment to maintain enterprise policies using enterprise tools. Involved and worked in Production deployment every week for critical applications. Developed application and infrastructure design and architecture for running more reliable and cost efficient. Provided extensive support and solutions to the product users on questions related to Product and helped them achieve their business needs from the product. Provided extensive support and solutions to the data analysts and data engineers on questions related to Framework and helped them achieve their business needs from the framework. Provided solutions to the business teams on developing metrics much more efficiently and meet their business needs. Provided solutions for the infrastructure and helped in debugging problems and resolve them in time and provided solutions to fix them. Developed applications using REST api s for data pipelines for easy executions and data movements. Developed frameworks with api hosting services like Flask, FastAPI and Uviron. Developed a high availability framework using Nginx, Gunicorn asd WSGI on top og Flask using Python. Used Terraform to map more complex dependencies and identify network issues. Developed SQL scripts for business logics in snowflake databases. Developed Data pipelines for build, test and deployment process. As a DevOps Engineer, Built Data pipelines to transform the data and store them into the database like snowflake. Created and maintained Docker Images for ECS Cluster and worked on creating Service and Task Definitions. Worked on developing the CloudFormation templates to use spot instances for EC2, EMR, ECS by following the High Availability and achieved the reduction of cost. Built an Internal tool using python to use the EMR cluster for big data applications in an efficient and optimized way. As a DevOps Engineer, Automated the daily data pipelines by building the tools to schedule and execute in time. As a DevOps Engineer, Migrated the application frameworks to microservices and containerized using docker and deployed as services with the auto scaling as high availability applications. Built robust systems with an eye on the long-term maintenance and support of the application. Worked with Amazon Web Services such as EC2, EMR, ECS, S3, SNS, SQS and Lambda etc. for Data pipeline applications As a DevOps Engineer, developed pipelines to build, test and deploy in Unix environments. Worked on support activities and helped customers get the issues resolved before they reach the SLA s. Worked closely with the Architecture teams to build a reliable and secure environment for data teams to develop their metrics in a guided process. TECHNOLOGY SPA, DALLAS, TX MAY 2018 AUG 2018 DEVOPS ENGINEER Project Description: Working for a solution and service-based company for a short-term project with a team of DevOps along with Enterprise Architect and DevOps engineers. Our goal is to provide a valid, secure, and efficient solution for the clients who are struggling to make use of DevOps. Proof of concepts (POC) for Logging, Monitoring and service discovery tool (Consul) with the main objectives we successfully completed and provided a well-organized solution to run these in their environment. RESPONSIBILITIES: Working with Docker EE using Swarm, deployed the applications as a stack into the nodes running in Azure. Worked with a Logging application named Logz.io which is SaaS based ELK stack and offered a solution to deploy into their production environment. Worked with FluentD to configure, filter, and ship their logs from docker daemon, containers, and nodes to logz.io using a API Token and Logz.io container along with fluentd container running as a single stack. Worked with the team in creating config files for FluentD in assigning and using rightful filters to obtain their goal in using the application. Worked with monitoring tools such as Prometheus, Grafana, Alert Manager and deployed these applications in docker swarm and created visual dashboards in Grafana to visualize the data obtained from Prometheus. Used Alert manager to configure the notification channel to trigger alerts and created the alert rules to notify to channel whenever the specified threshold exceeds and recovers. Created different kinds of dashboards in Grafana to visualize Memory usage, I/O usage to and from the network, CPU usage, Disk usage, Swap memory usage and so on in host level, node level and swarm level. Developed Docker Images in using tools like Docker Compose, Kubernetes, for Orchestrating and deploying the services related to the Containers and with container-based deployments using Docker, working with Docker images, Docker hub. Implemented the branching, tagging, develop, manage Pre-commit, post-commit hook scripts and maintaining the versions across different Source Code Management (SCM) tools like GIT, Subversion (SVN) on Linux and Windows platforms and SCM client tools like GitLab, GitHub, GitGUI, and Bitbucket. Work on AWS CloudWatch to monitor the performance environment instances for operational and performance metrics during load testing. Experience in designing a Terraform and deploying it in cloud deployment manager to spin up resources like cloud virtual networks, Compute Engines in public and private subnets along with Auto Scaler in Google Cloud Platform. Developed the CloudFormation to create EMR, EC2, Lambda and other variety of resources Implemented the Blue-green deployments with Route 53 and Elastic Load Balancer to have high-availability and Regional Isolation. Creating and managing users and groups, installation of packages and services on RedHat Enterprise Linux. Implemented the Prometheus Integration on all EC2 instances and created dashboards to monitor Infrastructure and Container Metrics CAPITAL INFO GROUP INC, DALLAS, TX DEC 2016 MAY 2018 DEVOPS ENGINEER RESPONSIBILITIES: Worked for the Migration and Datacentre Exit (DCE) activities and decommissioning the services and Data Migration (Data Center Exit) from HDFS to s3 and decommissioning unwanted data as part of cost saving. Working with Amazon Web Services such as Ec2, EMR, S3 buckets, CloudFormation and SNS and Lambda for Data pipeline applications developing and managing. Working with the ITSM tool such as HP Service Management and Service Now for monitoring the Incident tickets for the Application and Infrastructure Issues and troubleshooting and debugging the issues. Involved and worked in Production deployment every week for critical applications. Responsible for Continuous Integration (CI) and Continuous Delivery (CD) process implementation-using Jenkins along with scripts to automate routine jobs to speed up deployments. Automate provisioning and repetitive tasks using Terraform and Python, Docker container, Service Orchestration. Developed the CloudFormation to create EMR, EC2, Lambda and other variety of resources. Implemented the Blue-green deployments with Route 53 and Elastic Load Balancer to have high-availability and Regional Isolation. Creating and managing users and groups, installation of packages and services on RedHat Enterprise Linux. Implemented the Prometheus Integration on all EC2 instances and created dashboards to monitor Infrastructure and Container Metrics. Installation, Configuration, and Management of Data Dog agent on instances and building the dashboards for monitoring and reporting with real-time metrics. Creating S3 buckets and maintaining and utilizing the policy management of S3 buckets and Glacier for storage and backup on AWS and providing role-based access and maintaining the bucket policies to give a user the least privileges as needed. Keywords: continuous integration continuous deployment sthree information technology hewlett packard California Idaho Texas |