profile-pic
Vetted Talent

Saikumar k

Vetted Talent

A passion driven and versatile individual, prepared to add value to the organization by aligning to its vision and core values.

Having around 5+ years of experience in IT industries with Professional Development, Automation of Build, Deployment, Release Management and Change Management activities on AWS as a DevOps Engineer.

  • Role

    DevOps Engineer

  • Years of Experience

    5.8 years

Skillsets

  • Automation of build
  • Creating incidents
  • Shellscripts
  • Bitbucket
  • Creating/managing volume snapshots
  • Coordinate application releases
  • Iam role
  • Long running applications
  • Troubleshooting build/compile/configuration
  • Service Requests
  • Professional Development
  • Splunk
  • MFA
  • Kibana
  • Github
  • Ci/Cd Pipelines
  • Shell - 4 Years
  • Ansible - 4 Years
  • AWS EC2
  • Lambda - 4 Years
  • Identity and Access Management - 5 Years
  • Troubleshooting
  • SonarQube
  • Jenkins
  • Helm
  • AWS S3
  • AWS Lambda
  • Git - 4 Years
  • infrastructure as code - 4 Years
  • Kubernetes
  • Terraform modules
  • Quality Gates
  • Running analysis for sonarqube
  • Deploying the war into tomcat
  • Bug fixing of devops tools
  • change requests
  • Maven
  • Auto Scaling
  • RDS
  • Containerization - 4 Years
  • automation
  • SNS
  • Security Groups
  • cloud watch
  • DevOps - 5 Years
  • ELB
  • Unix
  • Release Management
  • Load Balancing
  • Change Management
  • EC2
  • Python - 4 Years
  • Kubernetes - 4 Years
  • Terraform - 4 Years
  • Volume snapshots
  • Docker - 4 Years
  • AWS - 5 Years
  • VSTS
  • Security policies
  • Virtualization
  • build
  • Incident Management
  • Public git repositories
  • CI/CD - 5 Years
  • EBS
  • ServiceNow
  • Cloud Formation
  • AWS Cloud Services
  • Deployment
  • VPC
  • S3
  • IAM

Vetted For

15Skills
  • Roles & Skills
  • Results
  • Details
  • icon-skill_image
    Senior Software Engineer, DevOpsAI Screening
  • 61%
    icon-arrow-down
  • Skills assessed :infrastructure as code, Terraform, AWS, Azure, Docker, Kubernetes, 組込みLinux, Python, AWS (SageMaker), gcp vertex, Google Cloud, Kubeflow, ml architectures and lifecycle, pulumi, seldon
  • Score: 55/90

Professional Summary

5.8Years
  • Jan, 2019 - Dec, 20212 yr 11 months

    Cloud Engineer

    Blueberry Digital Labs Pvt Ltd
  • Senior Cloud Engineer

    HCL Technologies
  • Cloud Engineer

    ValueLabs Global Solutions Pvt Ltd

Applications & Tools Known

  • icon-tool

    Maven

  • icon-tool

    AWS services

  • icon-tool

    Git

  • icon-tool

    Jenkins

  • icon-tool

    Docker

  • icon-tool

    Tomcat

  • icon-tool

    ServiceNow

  • icon-tool

    SonarQube

  • icon-tool

    Jira

  • icon-tool

    BMC Remedy

  • icon-tool

    Nexus

  • icon-tool

    Apache Tomcat

  • icon-tool

    Websphere

  • icon-tool

    Apache

  • icon-tool

    Terraform

  • icon-tool

    Kubernetes

  • icon-tool

    Ansible

  • icon-tool

    Grafana

  • icon-tool

    ELK Stack

  • icon-tool

    Amazon EC2

  • icon-tool

    Amazon EKS

  • icon-tool

    AWS EC2

  • icon-tool

    VPC

  • icon-tool

    RDS

  • icon-tool

    ELB

  • icon-tool

    S3

  • icon-tool

    SNS

  • icon-tool

    CloudWatch

  • icon-tool

    EBS

  • icon-tool

    IAM

  • icon-tool

    CloudFormation

  • icon-tool

    GitHub

  • icon-tool

    BitBucket

  • icon-tool

    Terraform

  • icon-tool

    Helm

  • icon-tool

    VSTS

  • icon-tool

    Splunk

  • icon-tool

    Kibana

Work History

5.8Years

Cloud Engineer

Blueberry Digital Labs Pvt Ltd
Jan, 2019 - Dec, 20212 yr 11 months

Senior Cloud Engineer

HCL Technologies
    Managing AWS infrastructure, automation of build, deployment, release management, and change management activities.

Cloud Engineer

ValueLabs Global Solutions Pvt Ltd
    Handling AWS infrastructure, CICD pipelines and build automation.

Major Projects

3Projects

Data Monitoring

    AWS infrastructure setup and maintenance, handling maven build issues, creating RFCs and automating CI/CD pipelines for Paypal.

Data Visualization (Data Viz)

    Creation and configuration of AWS infrastructure including serverless computing, deployment on multiple environments for Caterpillar, USA.

Dealer Service Portal (DSP)

    Extensive AWS setup and maintenance including VPC, EC2, Load Balancer, Auto Scaling for Caterpillar, USA.

Education

  • B.E. (Bachelor of Engineering) in Computer Science & Engineering

    Saveetha School of Engineering

AI-interview Questions & Answers

Yeah. Hi, my name is Sai Kumar, I have 5 years of experience in AWS and DevOps. Coming to AWS, I work within a couple of services like compute, storage, database, networking, management and governance, application integration, security identity, and transfer containerization. In DevOps tools, I work within Bitbucket administration, Jenkins, Docker, Kubernetes, and Terraform. In compute sessions, I work within EC2 instances, launch templates, AMIs, volumes, snapshots, and target groups, security groups, and load balancers, and autoscaling. I also work within the Lambda services, where we use to create serverless processes by using API Gateway with Lambda services, and I also work within Elastic Beanstalk. Whenever our DevOps teams require any development activities, we use Elastic Beanstalk as part of a POC, like POC is a proof of concept. Coming to the storage part, I have worked on S3 buckets, objects, and different types of storage classes, where we integrate with EFS and FSX for data sharing across all missions using Jump Host, and I also work within Elastic File System and AWS backup services, where we take frequent backups of deployments. Coming to the database, I've been working with MySQL, which is available in RDS, and network identity, I have created automated networks using Terraform, where we create VPCs, route tables, and public and private subnets, and internet gateways, and use Cloudflare for the content delivery network. I also have good knowledge of establishing a VPN from on-premises to AWS using IPsec, and troubleshooting NAT gateways and transit gateways. I also work on cloud monitoring tools, I have worked on CloudWatch. For auditing purposes, we use CloudTrail to track network activity from VPCs. I have worked on IAMs and creating roles, users, and policies, and established multi-factor authentication for all users across the project. I also work within secret manager, key manager, and certificate manager for SSL certifications, where we provide details like CSR, key files, and certificates. I have worked on SHA-0 and SHA-256. I also worked on Docker and Kubernetes, like ECR, EKS, and the orchestration level. I also worked on Terraform, where we create resources like VPCs, subnets, and different types of storage, like stages, pull source code, and stages, testing, and deploying. We use different types of deployments, like SSS deployments and containerized deployments.

And coming to Kubernetes, where we use creating of bots, deployment sets, and stateful sets, as well as remote sets. We'll be using stateful sets for banking-related things, like database-related issues. We use stateful sets for those issues. We'll also be using services, such as cluster IP and load balancers in Kubernetes. For security reasons, we use commands like, and we use configuration management tools like Ansible, where we write playbooks and scripts to patch Linux and OS, and schedule automated jobs to apply operators regularly. Coming to root with kernel updates, we'll be using cluster deployments, including daemon sets for OS-level patching, unattended updates, node drains, and uncordant during patching work. We take one node into uncordant level, and we use cluster autoscaling. We'll also be using CSED pipeline integration, where we use Jenkins and Git to create jobs and trigger OS patches and other things. We'll be using monitoring and alerts like Prometheus and Grafana. If you have any patch status, or security vulnerabilities, we'll be using those tools. We'll also be using backup and patching, ensuring regular backups for vulnerabilities, that's it.

Automated approach to scale Kubernetes deployments in response to increase the number of adjustments, such as CPU and memory usages, and configure the HPA to the scale-based and custom metrics like the web traffic we will be using, and the set of resources and requests and limits we will be using, like if we are having in the deployments the manifest ensures that whether the resource request is limited to the CPU and memory issues and the setting on it, and the creating the HPA object and the defining the horizontal pod autoscaler based on the CPU utilizations and the custom metrics, like writing often in YAML files like API and the API and the kind metadata and the specs, the specifications details and applying those, applying the HPA, we use a command, that YAML file, we will run the YAML files and the cluster auto scaling also available, where we automatically adjust the sizes of the Kubernetes and cluster based on the resources requested by the pods and pending due to the lack of resources, we will be using to step configuration auto scaler, installing the cluster auto scaler, cluster auto scaler in the EKS, and the auto scaler might already be available, we just need to enable it, and configuring the clusters like setting the parameters such as the minimum and the maximum nodes we will be using, and the node auto scaling behavior like when HPA increases the number of pods, it is like if the existing nodes are insufficient, clusters may be having like HPA decreases the number of pods in the cluster, automated scale removes unused nodes, and use the custom metrics with the HPA also like the web traffic and scaling up, scaling up the things and the HTTP requests per second, like RPS will be having or the latency, and to check this, we use Prometheus and Grafana for the graphical view, and configuring the Kubernetes metrics adapter, we will be using, for setting up of the Prometheus metrics, we will be having like such as an HTTP request total, like auto scaling decision we will be taking, and the ingress controller, ingress controller where we have the multiple scale up, scale out of pods, and monitoring the number of requesters we are using for the ingress, like NGINX ingress we will be using for the controller.

We propose a CI-CD pipeline design for an application deployed across AWS and Azure that integrates end-to-end testing. The design involves the following stages: For continuous integration, we will be using a pipeline that triggers and builds when code is pushed to specific branches, such as main or master. This pipeline will: - Build Docker images and applications - Store Docker images in the Elastic Container Registry (ECR) on AWS - Run unit tests during the CI stage using JUnit and PyTest to ensure basic application correctness - Perform static code analysis and security scanning - Store artifacts, test reports, and other outputs in a central location, such as AWS S3 For continuous deployment, we will have a pipeline with stages that deploy containers to Elastic Kubernetes Services (EKS) or Amazon Elastic Container Service (ECS). The pipeline will: - Deploy infrastructure code using AWS CloudFormation or Terraform - Deploy and run end-to-end (E2E) testing on EC2 instances to validate application behavior in a real-time environment - Use frameworks and tools to execute E2E testing after deployments - Have approvals for workflows, sending notifications for manual approval if all tests pass We will be using manual and automated testing to ensure the quality and reliability of the application.

For restoring of stateful applications in Kubernetes, we have components involved in stateful applications like persistent volumes, persistent volume claims, and stateful sets, and the Kubernetes configuration. The system overview will be like a persistent volume application where we take the data backup and configure the backup into YAML or manifest, and secrets will be placed, and monitoring and automation will be used, and restoring procedures like restoring the data and configuration in the case of failure, while taking off on the backups. For the backup of stateful applications, we mostly use persistent volumes where we use cloud-native tools like storing snapshots, like snapshots, and EBS volumes creating EBS snapshots of the VP, like persistent volumes, using the AWS CLI, and also we will be having backups like creating the snapshots and the volume ID will be provided in the code, and the NFS volume snapshots by Kubernetes, like providing Kubernetes APIs that can be used to take the snapshots of persistent volumes if it is stored, and provided by supported providers. Backup applications will be using the data via sidecar containers running the backup into the running those backup into saving the data for external backup storage, we use like S3, and these things. For the backup of the Kubernetes configuration, we use config maps and secrets for any sensitive information, we use secrets, like secrets keys, and keys in the YAML file. Automating the configuration backup, we have the entire cluster or specific resources, like having any PVCs, but also the backup of the entire Kubernetes resources, like stateful sets, and secrets will be using the backup of the CACD file configuration, if you have stateful applications, and use external tools, like we could say, like Jenkins or GitLab, using ensure that the pipeline and the secrets are all getting backed up. For automated and monitoring backups, we use cron jobs, and for the automatic backup process, and also use the tools for scheduling the backups, also may be a particular timestamp, may be using. For the monitoring backups, we use Prometheus and Grafana for the backup of the cron jobs and the volume snapshots, and the restoring process of like stateful applications from the cloud providers may be using EBS.

I've got a number of trips. For deploying, for deploying the Lambda functions using CICD pipelines in AWS, we use having the function update efficiency, CICD pipeline in the AWS like having the AWS CodePipeline. CICD pipeline like build stages and test stages will be there, and the deployment stages. We use SCM for the build stages, this is used like a package, packaging of the Lambda function and the code build runs this. For testing, it is an optional way; we can say unit testing and JUnit tests will be used. For the deployments, we are using AWS Lambda after the build is completed and the test passes, deploying the updated Lambda function by updating the functional code by using AWS CLI or AWS CloudFormation also, and AWS CloudPipeline, which will have a source stage, build stage, and deploying stages. For AWS Lambda, the support for versioning allows you to rollback to previous versions if issues are faced during deployment, and canary and linear deployments will take place. For monitoring and alerts, we are having CloudWatch and CloudMetrics, with SNS notifications integrated into it.

three four five six seven eight nine ten eleven twelve thirteen fourteen fifteen sixteen seventeen eighteen nineteen twenty-two eighteen nineteen twenty-three twenty-four twenty-five twenty-six twenty-seven twenty-eight twenty-nine twenty-seven twenty-seven twenty-six twenty-five twenty-seven twenty-eight twenty-nine fifty-two fifty-three thirty-four fifty-three fifty-four twenty-five twenty-six twenty-seven twenty-seven twenty-eight twenty-nine twenty-one twenty-nine twenty-eleven twenty-eleven twenty-two twenty-nine twenty-two twenty-one twenty-five twenty-six twenty-seven twenty-four twenty-five twenty-nine twenty-eight twenty-nine twenty-nine twenty-one twenty-nine twenty-two twenty-four twenty-three twenty-four twenty-five twenty-six twenty-seven twenty-eight twenty-nine twenty-nine twenty-nine

As SSH key and the name were hard-coded in it, so if you're having a lack of encryption method, we'll be using and like what we could say avoiding of the security risk will be happening. Hard-coding of SSH key names or the paths will be placed in it and we need to provide the required SSH key to be passed and the variable without default values and securely managed by secrets we'll be using.

How would you design a system to autoscale the concurrent simulation, learning workloads in a hybrid way, set up using the new parameters? I would design a system to autoscale the concurrent simulation, learning workloads in a hybrid way, set up using the new parameters by utilizing a combination of machine learning algorithms and cloud-based autoscaling technologies. The system would consist of a monitoring component that continuously tracks the performance of the simulation and learning workloads, a prediction component that uses machine learning algorithms to forecast future workload demands, and an autoscaling component that adjusts the number of resources allocated to the workloads based on the predicted demands and current performance metrics.

Like the methodologies we will be using, having the SOC System Organization Control and General Data Protection Regulator. We use security by design and threat modeling. We will be using compliance-first design and data classification to identify classified data. For example, sensitive data will be applying the appropriate security measures, encrypting that data to its base, using ELS, and for data transit protection, we will be using SOC 2 and GDP requirements. The security development practices we will be using include dynamic application security testing, code reviews, and DevSecOps integration, such as CACD pipelines with security gates, adding security checks in the CACD pipeline, and using tools like Trivy for container vulnerabilities. Open policies agents will be placed for data handling and privacy controls, and data retention policies will be used. We will also be using data subject access requests and audit logging, and having continuous monitoring and auditing, with security information and event management tools like Splunk. We will be using role-based access controls and multi-factor authentications to ensure MFA is enforced for all users, reducing the risk of unauthorized access. We will also be having vulnerability management, regular patching, and vulnerability scanning, penetration testing, and incident response planning, with data feature notifications. We will be having data governance and third-party vendor management, performing S4C and data processing agreements, regular auditing, and compliance reviews. We will be following a secure SDLC process and following these methods and standards during the deployment of applications.

Let's put it on the porch. I'm going to zoom in closer. I'm going to zoom this up a little bit. So you can see it better. For approaches like optimizing a Kubernetes cluster deployment, like computer vision model, we're having a Kubernetes cluster optimizing node configuration method for ensuring that the Kubernetes nodes and the GPU, for example, NVIDIA, and GPU scheduling, and the resources allocations, like enabling the device plugins, for example, NVIDIA device plugins, will be a clustering of the points, and they use the Kubernetes GPU scheduling point scheduling and containerization of the computer vision model. We optimize the Python docker images, and the multi-stage building will be placed, and ensuring that the necessary Python dependencies will be installed. And modeling the model packages, like open network exchange, will be having the cluster level optimization, having horizontal auto-scaling for which we have horizontal bar auto-scalings, for which we have, and vertical bar scalings, we have, automatically adjusting the CPU and memory requests of the parts based on the actual usages, and the optimization level of resources usage, and avoiding over-provisioning, and efficient in the deployment. Efficient in the data model, and the storages, and the persistent storage for large database sets, like Amazon S3, using distributed storage solutions, and implementing catching models, catching mechanisms in the parts, and reducing, and optimizing the model servicing, like load balancing, and interface scaling. And latencies, like node affinity, we have setting of a set of rules on the nodes, and tolerance will be placed, and monitoring, and logs, where we have GPU utilization set up of the monitoring, and performing whether it is like centralized logging, and implementing centralized usage of Elasticsearch. Security compliances, well, coming to security compliances, network policies, applying the Kubernetes, and secret management, and data privacy approaches, and bot security policies will be having, and the model versioning of continuous deployments, we use the CSCD for the model updates, and the canary and blue and green deployment methods, and the model versioning, and cost optimization, coming to like the spot of smart instances for the noncritical tasks, and right sizing the resources, and efficiently load scaling.