profile-pic
Vetted Talent

VIKASBALIYAN

Vetted Talent

As a data scientist, I am passionate about solving complex problems using cutting-edge technologies and data-driven insights. I have a strong background in computer science, information technology, and software engineering, with a B.Tech degree from the College of Engineering Roorkee and three years of professional experience at Accenture.


At Accenture, I worked as an application development analyst and associate, where I developed, tested, and deployed various software applications for clients across different industries. I also leveraged my skills in Python, SQL, Excel, and Power BI/Tableau to perform data analysis, visualization, and reporting, delivering high-quality solutions that met the client's requirements and expectations. Additionally, I obtained multiple certifications in Enterprise Design Thinking and SQL for Data Analysis, demonstrating my commitment to continuous learning and improvement.


Currently, I am working as a data scientist at TCS, where I continue to apply my expertise in data science and machine learning to drive impactful results and contribute to the success of innovative projects. My role involves utilizing advanced analytical techniques and state-of-the-art tools to extract meaningful insights from complex datasets, helping organizations make informed decisions and achieve their strategic goals.

Skillsets

  • Python
  • Statistics
  • SQL
  • Git
  • NLP
  • ML/DL
  • GenAI

Vetted For

10Skills
  • Roles & Skills
  • Results
  • Details
  • icon-skill_image
    AI Chatbot Developer (Remote)AI Screening
  • 67%
    icon-arrow-down
  • Skills assessed :CI/CD, AI chatbot, Natural Language Processing (NLP), AWS, Azure, Docker, Google Cloud Platform, Kubernetes, machine_learning, Type Script
  • Score: 60/90

Professional Summary

3Years
  • Sep, 2023 - Present2 yr 6 months

    Data Science Intern

    MN Squarred Inc.
  • Nov, 2019 - Aug, 20222 yr 9 months

    Application Development Analyst

    Accenture

Applications & Tools Known

  • icon-tool

    Power BI

  • icon-tool

    Tableau

  • icon-tool

    Excel

Work History

3Years

Data Science Intern

MN Squarred Inc.
Sep, 2023 - Present2 yr 6 months
    Spearheaded the development of an innovative image-to-text project, showcasing advanced skills in computer vision and NLP. Leveraged state-of-the-art models from Hugging Face to enhance project efficiency and accuracy.

Application Development Analyst

Accenture
Nov, 2019 - Aug, 20222 yr 9 months
    Utilized advanced programming skills in Python, leveraging libraries such as NumPy and Pandas, to manipulate and analyze large datasets. Implemented machine learning models with scikit-learn.

Achievements

  • Led the development of an innovative image-to-text project, utilizing computer vision and NLP
  • Increased model accuracy by 10% through adoption of new algorithms
  • Reduced customer attrition by 20% through predictive modeling in customer churn

Major Projects

1Projects

LLAMA-based chatbot

    Implemented in Python, leveraging Meta's LLAMA model, with custom training of LORA and PEFT techniques for enhanced performance.

Education

  • Applied Artificial Intelligence (AI) Solutions Development

    George Brown College (2023)
  • Bachelor of Technology (IT)

    College of Engineering Roorkee (2019)

AI-interview Questions & Answers

Hello. My name is Vikas Balyan. I'm from Delhi. I did my bachelor's in information technology from College of Engineering, Erudwig. After that, I worked for 3 years at Accenture as a application development analyst. There, I have worked with multiple technologies like Python, SQL, NLP machine learning, and deep learning. After that, I pursued my post graduation diploma in applied AI from Toronto, Canada, and I just completed my, uh, post grad in 2023 in December.

Well, handling concurrency chatbot, uh, when multiple users are interacting, I would, um, implement threading or asynchronous synchronous programming techniques where each user can interact and can be processed concurrently, ensuring that one user's action do not block or interfere with another. Also, um, it is very essential to manage shared resource carefully, uh, like data storage or any external API calls to prevent these raised conditions and ensure data integrity.

So, uh, preventing our shooting neural networks can we can use several techniques such as regularization. Uh, we can use, like, l one 2 regularization to penalize the large weights in a neural network, preventing it from, uh, fitting noise in the training data. Also, we could use, uh, dropout where, uh, we could, uh, introduce the layers during training and, uh, randomly selecting neurons are ignored, which, uh, helps us preventing the network from relying too heavily on any particular, uh, subsets of neurons. Also, we could use, uh, data augmentation techniques as well as the earliest topping where, uh, in the early stopping, we monitor the validation loss during training and stop the training process when the validation loss starts to increase. And, uh, also the other, uh, you I would say is cross validation. Uh, it is, uh, uh, we I could use techniques, uh, like k fold cross validation to evaluate the model's performance on different subsets of data, ensuring it generalize well to unseen, uh, like, examples.

I can implement a thread safety for accessing, uh, shared database using various techniques like, um, synchronized methods, uh, uh, to ensure that only one thread can access a critical section of the code at at a time. And, uh, the other is synchronized blocks, uh, which, uh, is specifically for blocks of code instead of the entire method to reduce contention. And, uh, the other one, uh, I would recall is thread safe data access objects. Uh, it it ensures that the DAOs are designed to be thread safe. This may involve encapsulating databases access within these synchronized methods. Also, the, um, other thing is, uh, collection pooling, uh, collection pooling libraries, uh, like, uh, Apache, uh, or Hikari CP to manage database connection efficiently. So these libraries are typically handle threat safety concerns internally.

Um, posted deployment, uh, monitoring the performance of the, uh, AI chatbots. It's, uh, I would say to to make it optimal by, uh, like, monitoring the response time, uh, by we can measure the time it takes for the chatbot to respond the, uh, user query and the number of interactions or messages processed by the chatbot per unit of time. Uh, it, uh, you know, increasing increasing in the in the time may indicate improved efficiency, uh, while it decrease would signify for the signify the performance. And, um, I would say the monitor the error rate, uh, to track the percentage of, uh, user interactions that result in errors or failed responses. So, um, high error rates may indicate issues with the chatbot's understanding or the response generation capabilities of chatbot. Also, uh, the accuracy, uh, we could measure the accuracy of chatbot responses, uh, and, uh, by regularly when you're doing the accuracy against a test dataset, it could ensure consistent performance. And, uh, then, uh, we have user satisfaction. Uh, we could collect feedback from users through our survey or any sentiment analysis to gauge their satisfaction. So, uh, that's how, uh, we could monitor the performance, uh, for the chatbot. And, also, uh, we could, uh, utilize our resources, uh, like the, uh, system resources such as the CPU memory and, uh, network to ensure efficient use. And, uh, also by checking the retention rate, I mean, the percentage of users who continue to engage, uh, with the chatbot all the time. So, um, these all parameters, I would say, uh, we could use to monitor our chatbot's performance.

We could implement entity recognition in a chat bot, uh, flow by, uh, choosing a natural language processing library, um, like NLTK or, uh, spaCy, or we could use the Hugging Faces, uh, Transformer models. Also, um, reprocess, uh, the user input, tokenize that, and, uh, reprocess the user's input to prepare it for entity recognition, which may involve removing, like, stop words or punctuation and special characters. Also, uh, we've chosen NLP library, uh, to apply entity recognition model to the proper preprocessed user input. The model will identify the and, uh, classify entities such as names, dates, locations, organizations. And, uh, we could I mean, uh, by handling the ambiguity and errors, So, uh, implement logic to handle cases where entity recognition may be ambiguous or incorrect. And by, like, um, iterate and refine the, uh, by, uh, I would say, uh, uh, by continuously evaluating the performance of the entity, uh, recognition model and refine it as as needed. So this main mode retraining the model with additional label data or adjusting parameters to improve the accuracy and coverage. And at the last, uh, by monitor and analyzing the performance.

Well, as, uh, I have any less experience in JavaScript, so I'm not able to answer it correctly. And because I, uh, more into Python and, uh, in AIN machine learning.

Um, by, uh, I would say accessing records, uh, assessing, uh, individual reports by their index or key is, uh, and, uh, I would say adding new reports to the database can be done by appending a new dictionary to the list and updating the reports and removing reports from the databases using the detail, uh, the statement.

The critical aspects to consider when scaling an AI chatbot, uh, for handling, uh, millions of user would be the infrastructure scalability. Uh, so we could design a scalable infrastructure capable of handling the anticipated load. Um, like, we could utilize cloud services like, uh, AWS, Azure, or GCP to dynamically scale our resources based on the demand. And implementing a distributed architecture that can scale horizontally by adding more instances of the chatbot the instances. And, uh, by the, I would say, asynchronous processing, we could off, like, offload time consuming tasks such as intensive computations or externally play calls. So to background workers or, uh, asynchronous processing frameworks, uh, we could use messages queue like, um, a Apache Kafka to, uh, decouple components and then the request as in. And, um, at last, I would say, uh, optimize our algorithms and data structure.

So how would an understanding of graph databases benefit the development of an AI channel? Okay. Uh, let's say Um, like, uh, it could, uh, I would say in several ways, uh, like the complex relationship modeling graph databases, uh, excellent modeling complex relationships between entities and, uh, a flexible schema. Graph databases have a flexible schema that allows for easy adaption to changing end requirements. So this flexibility is particularly useful for AI chatbots. And efficiency querying graph databases provide efficient query capabilities for traversing relationships.

Discuss how would you implement voice recognition and processing capabilities in a chatbot. Okay. I would implement the voice recognition and processing capabilities in chatbot by, uh, using techniques like Google Cloud, speech to text, or Microsoft Azure speech service. Uh, I would say, uh, some other like Mozilla, DeepSpeech, open source. And, uh, also, audio input, uh, we capture audio input from users using a microphone. Microphone. And, uh, depending on the platform, uh, we, uh, use different APIs or libraries to access audio input streams. Uh, and, yeah, that is it. We can integrate with our chatbot platform, integrate the speech recognition component with the chatbot platform. And, uh, also, uh, by using natural language understanding, process the transcribed text using NL, uh, understanding techniques to extract intents, entities, and other relevant information from the user's utterance. And we could, uh, use, like, audio output. We play the synthesized audio response to the user through speakers or any audio output devices ensuring, uh, the compatibility with different platforms.