profile-pic
Vetted Talent

Yogesh Joshi

Vetted Talent
I possess the role of developer proficient in Python and AI, having hands-on expertise in crafting, programming, and validating intricate systems and models. My skill set encompasses deep understanding of Descriptive and Predictive Modelling. I showcase robust analytical, logical, and statistical capabilities, driven by an innate passion for continuous learning. My proficiency extends to Python, R, C#, C++, and the realm of Data Visualization. AREAS OF EXPERTISE Research and Development of AI Software Web, Software Development Machine Learning Techniques Deep Learning Methodology Prompt Engineering with open api, Tensorflow, Keras, Pytorch, Sparks etc. Web Frameworks: Flask, Django etc. Agile Methodology Database Modelling UI/UX Design and Development Micro-Computing and IOT Cloud Integration and Cloud Deployment Utilization of Hugging face, Wandb, OpenAI.
  • Role

    AI Developer & Data Scientist

  • Years of Experience

    5 years

  • Professional Portfolio

    View here

Skillsets

  • cloud deployment
  • Deep learning methodology
  • Prompt Engineering
  • OpenAI
  • Machine learning techniques
  • Hugging Face
  • Docker containers
  • Database modeling
  • Data Visualization
  • PyTorch
  • Ai software development
  • Agile Methodology
  • UI/UX Design
  • Python
  • Keras
  • Django
  • Flask
  • TensorFlow

Vetted For

18Skills
  • Roles & Skills
  • Results
  • Details
  • icon-skill_image
    Senior Generative AI EngineerAI Screening
  • 55%
    icon-arrow-down
  • Skills assessed :BERT, Collaboration, Data Engineering, Excellent Communication, GNN, GPT-2, graphs, Large Language Models, Natural Language Processing, Sagemaker, Deep Learning, neural network architectures, PyTorch, TensorFlow, machine_learning, Problem Solving Attitude, Python, Vertex AI
  • Score: 55/100

Professional Summary

5Years
  • Nov, 2021 - Present4 yr 2 months

    Associate Data Scientist

    Reflexion.ai
  • Aug, 2020 - Nov, 20211 yr 3 months

    AI Developer

    Dataviv Technologies
  • Feb, 2019 - Jul, 20201 yr 5 months

    Python Developer (R&D-AI Team)

    Biosense Technologies

Applications & Tools Known

  • icon-tool

    Python

  • icon-tool

    Docker

  • icon-tool

    AWS (Amazon Web Services)

  • icon-tool

    Google Cloud Platform

  • icon-tool

    Django

  • icon-tool

    FastAPI

  • icon-tool

    Hugging Face

  • icon-tool

    TensorFlow Hub

  • icon-tool

    PyTorch

  • icon-tool

    Tensorflow

  • icon-tool

    OpenCV

Work History

5Years

Associate Data Scientist

Reflexion.ai
Nov, 2021 - Present4 yr 2 months
    Working on existing AI/Deep learning-based video meta-data generation product. Improving and adding newer modules to video meta-data generation product using various deep learning architectures.

AI Developer

Dataviv Technologies
Aug, 2020 - Nov, 20211 yr 3 months
    Handle and manage development lifecycle for entire project. Develop and design efficient algorithms for ML and DL for client requirements. Leading team for meeting close deadlines and finding innovative solutions.

Python Developer (R&D-AI Team)

Biosense Technologies
Feb, 2019 - Jul, 20201 yr 5 months
    Started as ML Intern for Research and Development Team. Grabbed permanent position as Python Developer in June. Worked with team to build AI incorporated software for healthcare diagnostics, and made ML/AI projects on Raspberry Pi.

Achievements

  • Awarded Mr.Reliable from Reflexion.ai for steady development and timely completions
  • Placed Best Host of the Year from Late G N Sapkal College for Communication Skills and Improvisation
  • Placed Avid Readers award from High School for communication and story telling skills

Major Projects

4Projects

Blood Counter (AutoMicroscope)

    Part of a team working on Deep learning-based solution for counting microscopic blood cell from images. Worked as Python ML developer.

Retail Analytics based on CCTV feed

    Carrying out analysis based on retail consumer information gathered with CCTV cameras. Modules include Walk-in Counter, Facial Attendance, Social Distancing Monitor, Mask Detection, Threat Detection.

Traffic and Quality of Traffic Prediction

    Predicting traffic and quality of traffic at any location using Census data and application ping data by implementing multi-output regression.

Dental Treatment and Diagnosis

    Support AI Developer to build and deploy diagnosis solutions for dental issues using Mask CNN. X-ray images for teeth were used to predict the treatment plan.

Education

  • Bachelor in Computer Engineering

    KCT Late G N Sapkal College, SPPU (2017)
  • HSC in Science Stream with Computer Science

    G D Sawant College, Nashik (2013)
  • SSC

    St. Francis High School, Nashik (2011)

Certifications

  • Data Science and R Programming

    ETLHive (Pune) (Oct, 2017)
  • TensorFlow Developer Certificate

    Udemy (Aug, 2023)
    Credential ID : UC-4dd17338-7d01-4e48-9d61-d5b5ceacaefa
    Credential URL : Click here to view
  • AWS Certified Machine Learning Specialty

    Udemy (Aug, 2023)
    Credential ID : UC-96ab07af-8255-46b6-874a-a1645a680f47
    Credential URL : Click here to view
  • TRAINING Deep Reinforcement Learning - Hugging Face [2023]

  • Deep Reinforcement Learning

    Hugging Face (Aug, 2023)
    Credential ID : yogjoshi14
    Credential URL : Click here to view

AI-interview Questions & Answers

Uh, myself, Yuki Tushy. I have about nearly 5 years of experience in, uh, data science, AI, ML, as well as, a Python development role. I have a versatile portfolio where I worked on many different types of projects, including machine learning, mostly deep learning and computer vision. My expertise lies within deep learning and computer vision, but as the trend has more diverted towards LLM and the large language models. I have also contributed myself to getting familiarized with Tuuk's technologies. Uh, about, uh, in a year ago, I have started working on NLMs. One of the projects that I am working on is based off of visual transformers as well as, uh, the latest, uh, video, which has been released, uh, with the code. So I have, uh, expertise in Python language as well. Apart from that, I have good understanding of core principles such as, uh, database managements, handling pipelines of data, handling deployment. Uh, I'm also eager to learn new technologies. So I have recently gained exposure to MLflow and other such deployable uh, technologies. I also have, uh, expertise in the latest upcoming post post PostgreSQL, this database.

The techniques depend on the and the size of data which we need to train. Uh, most preferably, I'll use continuous development and continuous integration techniques where I would, uh, have a base model prepared on the large dataset. And depending on the base model Accuracies and base model results performance benchmark on different sets of, uh, dataset. I'll Build up on that by hyperparameter tuning. I can do manual hyperparameter tuning depending on the experimentation as well as I can choose To opt for optioner, which is an automated hyperparameter tuning tool, there are also some different libraries such as the weights and biases and, uh, TensorBoard which can help in logging and Monitoring the different types of parameter used for experimentation.

Uh, there are different types of understanding of loss functions. Uh, primarily, uh, the Problem that we are, um, we are tackling all the algorithm that we are using, uh, depending on what Loss function has worked the best in the research paper. That is one of the mechanism that I'll use, uh, while approaching the deep learning and choosing the loss function, but in case there is some Other resources or other material which suggest better, I can opt for that particular loss question as well, But, primarily, uh, the research paper, the published conferences suggest the loss function which has been performing better in this case, and my approach would be to take that particular loss function. In case I have, uh, to switch loss function, my ideology would be to have a, basically have a, uh, experimentation approach where I would consider opting for multiple, uh, similar type of loss functions. Example, if I'm choosing, uh, cross entropy loss function, then it might be Also beneficial to make sure that, uh, it works well on the classification type of data.

Assigning task is, I think, what is primarily important when we are working on cross functionalities Depending on any manner they might be. And as the latest trends and technologies have been evolving, I think There have been incorporations to implement multiple sources or multiple types of technologies to get an better output. Uh, one of those example, uh, might be a Postgres ML where now the, uh, you can have a model, Uh, model inferencing through a basic SQL script. So I think that is something which I take as a lesson, And I would prioritize the individual tasks, uh, then the team task because once the individual tasks have been fulfill and met, then only the dependencies towards the other team members task can be functionally fulfilled. So while implementation in any kind of implementation of AI feature in a cross functional development, I would prioritize on an individual task than the team manner because in the cross functional platform or cross functional Environment, individual, uh, ideologies or individual work is more important than the teamwork because The individual's work is dependent on the cross functional team's

Uh, there are multiple approaches, uh, to solve, uh, the to tackle this particular problem of implementing a with continuously training and NLP model with new incoming data. Uh, one of the approach Or the most efficient approach is to have a, basically, have a, uh, storage mechanism where we have incoming traffic then the storage mechanism is attached to a data pipeline. The data pipeline job is to clean the data, have the data properly established or properly ready for the model, then there will be a training pipeline. The training pipeline's job will be to take that data which has been cleaned by the data pipeline and train the model whichever it may be to perform the NLP task. Once the training task is done, then there will be 2 types of pipelines. 1 can be a deployment pipeline and another is an inference Now within the deployment pipeline, there is again we need to add a scheduler or a scheduling mechanism where it would periodically depending on the solution that, uh, the depending on the solution that is required it would train it retrain itself, but also the doesn't forget, uh, doesn't have an catastrophic forget mechanism where Totally forget the previous, uh, rates and ready and have an some kind of different output. So that is something we need to monitor. Uh, to tackle those problems, I think newer, uh, trends and newer technologies such as vector databases can be implemented to basically have a, uh

Given the scenario that the model trained Nice. Q. The mitigation strategies are basically, there are multiple mitigation strategies to avoid biases in the data. One of which is to incorporate, uh, and representative data where it basically represents all type of data where Whichever problem we need to solve, one of one of the approaches by going in such a direction. Another way is, uh, that Demographically, we need the data to show all the regions and all the countries. That is also one of the approach that we can use to mitigate the skewness and the bias in the data. Uh, there are also some other, There are also some other imputation and augmentation techniques that can be used to have the data normalized. Yeah. That's it. I think these are the strategies that we can use. Apart from that, there are also some other strategies That can be researched and implemented depending on the data and the scenario that we are working on.

Self attention mechanism wouldn't take 3 inputs of x. I think it would be a single input. Uh, the basic transformable lock has the attention layer with forward feedforward neural network, so I think that is something which is, uh, missing. And also before the attention layer, there should be some additional layers such as embedding layer In case we are incorporating it within the transformer block, uh, so we can add, uh, the positional embedding vector layer also within the transformer block. I think that is something. The input shape for the self attention mechanism, I think that is something which is

Uh, absolute mean mechanism is somewhat often, uh, nonstandard for Loss mechanism in LLM. The most frequently used or most promiscuously used, uh, loss mechanism for Any kind of generative model is mostly the root mean squared errors. So that, uh, so that the steps so that the step that they are taking towards gradient is higher, and it reaches, uh, this it reaches its, uh, global minima faster. So I think that is something we can implement here.

There are multiple, uh, readily available architectures which we can use for text generation. Uh, one of which is using the But mechanism, which is bidirectional encoding bidirectional encoding and coding representation of transformers, uh, where basically it has only the decoder layer And, uh, stacking the decoder layers together can be helpful index summarization problems, uh, whereas The output layer can be a soft match layer.

Uh, for ingestion pipeline, there are Multiple cleaning and preprocessing steps that can be done, uh, beforehand to identify the anomaly. One of the steps while condition of data is to clean the data and have a, uh, preprocessing done on the data. This, uh, particular step can be useful in avoiding the anomalies. I think that is a particular step that can be done to avoid anomalies in the data. Apart from that, we can also have a cloud watch mechanism where it, uh, checks the data. It has the crawlers. We can use, uh, different types of crawlers even before having the data ingested into the pipeline To basically, uh, sort and, uh, eliminate any kind of

There are different types of benchmarks to validate output, uh, based on the LLM. So The sentiment analysis part, I think there are different benchmarks readily available. I don't, uh, remember, uh, on the top of my head, the benchmarks to validate the particular particular process. But, uh, apart from The readily available benchmark, we can also have a human in loop mechanism where the human itself can go through that result output and validate it if the if the output is somewhat acceptable

Um, depending on the use case, we can choose either TensorFlow or PyTorch. In both cases, either or or, uh, I think the, Uh, programmatically and logical consensus between TensorFlow and PyTorch are somewhat similar. So either or or either choosing TensorFlow flow or choosing PyTorch won't hamper the overall development, but for the simplicity's sake, I would prefer PyTorch as it is open source, Although TensorFlow is also open source, but it has its own limitations and works well with TPU TPU, uh, format arch architecture, whereas PyTorch, it can handle almost all types of architecture. The conversion of TensorFlow and PyTorch to optimized version, uh, to an quantized or optimized situation are somewhat similar. Both can be condensed into smaller forms. Both has batch fetching mechanisms which can be useful. Uh, both have their own, uh, limitations and enhancements, but I think either or are Fine. Fine. Uh, while choosing development, uh, transformer based texture mostly most of the examples of transformers have been made available by PyTorch. So I think PyTorch is more preferable Then TensorFlow as the back end for the transformers hugging v c p I uses, uh, PyTorch. But, uh, again, TensorFlow, if we are choosing the, uh, choosing to develop Google based architecture like Bard, then I think TensorFlow is much better than Python. So depending on which you choose for the alternative model. It can, uh, it can be beneficial to have a uh