Ajay Reddy
Sr DevOps/Cloud Engineer
About
● Overall, 9 years of experience in DevOps Environment with Continuous Integration, and Continuous Delivery (CICD), Aws and Azure cloud Linux System Administration and Cloud Management comprising Build and Release Management, and Containerization having automation, build and releasing in Agile Methodologies and Software Development Life Cycle (SDLC). ● Experience in AWS, Azure, and GCP Cloud Services like EC2, Auto Scaling, VPC, Route53, RDS, S3, IAM, ELB, Lambda, CloudTrail, Cloud Routing, Security Principles, Cloud Arch, Config, Elastic Cache, CloudWatch, Snowflake, SNS, Security Groups, Elastic Beanstalk, Glacier, SQS, DynamoDB. ● Experience in administration and troubleshooting of different operating systems Windows, Ubuntu, Red Hat Linux, and CentOS. ● Experienced in System Administration, System builds, Server builds, Upgrades, Patching, Migration, Troubleshooting, Backups, Disaster Recovery, and performance monitoring. ● Expertise in implementing, merging, branching strategies, defect fixes, and configuration of version control tools like SVN, GIT, and Bitbucket for smooth release management into production environments. ● Ability to build deployment, scripts, and automated solutions using scripting languages such as Shell, Bash, and Python. ● Experience in Designing and implementing infrastructure as code using tools such as Terraform, Ansible, and Chef to automate the provisioning and configuration of resources. ● Solid hands-on experience in implementing and managing containerization technologies such as Docker and Kubernetes to enable efficient deployment and scaling of applications. ● Experience with container-based deployments using Docker, working with Docker images and Docker Hub. ● Experience in using Kubernetes to orchestrate the deployment, load balance, scaling, and management of Docker containers with multiple-spaced versions using Helm Charts. ● Good Experience in monitoring, logging, and analyzing system performance using tools such as Grafana, Splunk, Prometheus, Dynatrace, and Nagios to detect issues proactively and optimize resource utilization and cost. ● Experience in implementing and managing configuration management tools such as Puppet, Salt, and Ansible to automate the configuration of servers and other infrastructure components. ● Experienced in designing and implementing continuous integration and delivery (CI/CD) pipelines using tools such as Jenkins, Bamboo, CircleCI, and ArgoCD to automate the build, test, and deployment of applications. ● Good experience in configuring and managing Maven and Gradle builds in Jenkins to manage dependencies and ensure consistent build results. ● Experience in setting up and managing artifact repositories using Nexus and JFrog Artifactory to store and manage binary artifacts and dependencies. ● Having good knowledge in automated testing using tools such as Selenium, Postman, and JUnit to ensure that the software is of high quality and reliable. ● Experience in setting up and configuring code quality analysis tools such as SonarQube and Checkmarx to enforce code quality and security standards. ● Good experience in configuring and managing environment variables and secrets using tools such as HashiCorp Vault and AWS Secrets Manager to ensure secure storage and access to sensitive data. ● Experienced with AWS CLI, AWS Python SDK (Boto3), AWS lambda, and other serverless microservices used for better management of cloud environments. ● Hands-on experience in deploying applications on Apache Tomcat, JBoss, WebSphere, and WebLogic servers. ● Profound experience in programming, deploying, and configuring JAVA/J2EE Applications using Apache Tomcat and JBoss. ● Good knowledge of Databases like Oracle, SQL Server, MySQL, NoSQL, PostgreSQL, MongoDB.
Skills & Expertise (14)
Work Experience
SRE/DevOps Engineer
United Airlines – Chicago, IL
1-2022 - Present
● Designed and implemented scalable and highly available infrastructure solutions on AWS using tools such as EC2, S3, ELB, RDS, CloudFormation, and Terraform. ● Developed and maintained custom scripts and tools using Shell/Bash, Python, and scripting languages to automate routine operational tasks and improve system resilience. ● Implemented monitoring and alerting systems using tools such as Splunk, Grafana, Prometheus to detect issues proactively and ensure high availability and performance of the system. ● Designed and implemented disaster recovery and business continuity plans to minimize downtime and data loss in case of disasters or outages. ● Participated in incident management and post-mortem activities, using tools such as JIRA or ServiceNow to track and resolve issues, and to identify and implement improvements to prevent future incidents. ● Conducted capacity planning and performance analysis to ensure that the system can handle expected workloads and to optimize resource utilization and cost. ● Worked closely with development teams to ensure that applications are designed and deployed with operational excellence in mind, and to provide guidance on best practices for monitoring, logging, and error handling. ● Implemented security best practices such as access control, network segmentation, encryption, and vulnerability management to protect the system from cyber threats. ● Created automated CI/CD pipelines using Jenkins, Git, and Ansible to build, test, and deploy Java applications to Kubernetes clusters. ● Managed source code repositories and implemented automated CI/CD pipelines using tools such as Git, Bitbucket, Jenkins, and Docker. ● Designed and implemented infrastructure as code using tools such as Terraform, Ansible, and Chef to automate the provisioning and configuration of resources. ● Implemented and managed Kubernetes clusters and container orchestration using tools such as Kops, Kubeadm, and Rancher. ● Designed and implemented networking solutions such as VPC, VPN, DNS, and load balancers to enable secure and reliable communication across the system. ● Performed etcd upgrades to ensure that the distributed data store is running the latest version and to prevent security vulnerabilities or performance issues. ● Participated in on-call rotation and responded to critical incidents, using technical and analytical skills to quickly diagnose and resolve issues. Environment: AWS, Linux, Java, GIT, Bitbucket, Jenkins, Docker, JIRA, Splunk, Grafana, Ansible, Service Now, PagerDuty, Shell/Bash scripts, Chef, Kubernetes, Networking, SSO, Terraform. Client: Office Depot – Boca Raton, FL Role: DevOps/Cloud Engineer Duration: Nov’19 – Dec’21 Responsibilities: Created and maintained Helm charts for deploying applications and managed the release process. ● Built and configured EC2 instances on the AWS cloud platform and configured Elastic Load Balancer for traffic control for the EC2 and S3 buckets. ● Worked with AWS S3 and IAM services, created S3 buckets, and configured them with permissions, logging, versioning, and tagging, and created new IAM users and groups, defined roles, and policies, and identified providers. Configured AWS IAM and security Groups in Public and Subnets in VPC. ● Created S3 buckets and bucket policies and utilized S3 buckets and Glacier for storage and backup on AWS. ● Implemented Argo CD for managing and automating the deployment of applications in a Kubernetes environment. ● Designed and developed GitOps workflows using Argo CD, ensuring that all application changes were version-controlled and auditable. ● Collaborated with development and operations teams to define and implement CI/CD pipelines using Argo CD. ● Created and maintained Helm charts for deploying applications and managed the release process. ● Defined AWS Security Groups which acted as virtual firewalls that controlled the traffic allowed to reach one or more AWS EC2 instances. ● Implemented Elastic Load Balancer (ELB) and Auto-scaling groups in AWS on production EC2 instances to build Fault-Tolerant and highly available applications. ● Implemented a serverless architecture using API Gateway, Lambda, and DynamoDB and deployed AWS Lambda code from Amazon S3 buckets. ● Launched database for MySQL and Oracle using RDS in multiple Availability Zones, set up monitoring for CPU utilization and limited DB connections by implementing alarms using Cloud Watch. ● Registered domain names and configured Route 53 to route internet traffic for our domains and finally configured Route 53 to check the health of our resources. ● Fine-tuned existing code for optimum performance using data virtualization, and integration techniques on both (UNIX, Java, and SQL scripts) & tool-based (Informatica, Denodo, OBIEE) techniques. ● Designed AWS Cloud Formation templates to create custom-sized VPC, subnets, and NAT to ensure successful deployment of web applications and database templates. ● Used Ansible as a configuration management tool to automate repetitive tasks and to quickly deploy critical applications, by using Ansible playbooks, modules, and roles. ● Worked on creating inventory, job templates, and scheduling jobs using Ansible tower and writing Python modules for Ansible customization. ● Used Docker to virtualize the servers for the Dev and Test environment needs, implemented Docker containers to configure automation, and created various environments to deploy the application which reduces cost for the enterprise. ● Configured Docker containers by creating Docker Compose files and pushed the Docker Images onto EC2 instances from Docker-Registry to deploy the applications using both Kubernetes and Docker Swarm. ● Integrated Docker container orchestration framework using Kubernetes by creating pods, config Maps, and deployments. ● Managed Kubernetes charts using Helm created reproducible builds of the Kubernetes applications and managed Kubernetes manifest files. ● Used Kubernetes for creating new Projects and services for load balancing and adding to Routes by accessing from outside, created Pods through the new application, and controlled, scaled, and troubleshooted pods through SSH. ● Written templates for AWS infrastructure as a code using Terraform to build staging and production environments. ● Developed automation scripting in Python (core) using Chef to deploy and manage Java applications across Linux servers. ● Set up Jenkins server and built jobs to provide continuous automated builds based on polling the GIT and integrated GIT with Jenkins to automate the code checkout process with the help of Jenkins DSL Plugin. ● Used Maven as a build tool on Java projects to build artifacts on the source code, and defined dependencies and plugins in POM.xml for various required activities to be done. ● Deployed and configured Elastic search, Logstash, and Kibana (ELK) for log analytics, full-text search, and application monitoring in integration with AWS Lambda and CloudWatch. ● Migrated the workload from on-prem to GCP kubernetes via helm package managenment. Environment: AWS, Redshift, Python, GIT, Bitbucket, Cloud Formation Templates, Jenkins, Docker, Red Hat Linux, WebLogic Servers, JFrog, Shell scripts, Chef, Kubernetes, Networking, Shibboleth, SSO, CIC
Education
Master’s: Mechanical Engineering - Oklahoma Christian Universit
2014 - 2017 · United States