Sanjay Vishwakarma(He/Him)

Sanjay Vishwakarma(He/Him)

Austin, Texas, United States
12K followers 500+ connections

About

I build AI systems that billions of people interact with every day — without knowing…

Articles by Sanjay

Activity

Join now to see all activity

Experience

  • Meta Graphic

    Meta

    Austin, TX

  • -

    Seattle, WA

  • -

    Greater Seattle Area

  • -

    India

  • -

    Bengaluru Area, India

  • -

    Bangalore

  • -

    Bangalore

  • -

Education

  • Stanford University Graduate School of Business Graphic

    Stanford University Graduate School of Business

    -

    -

    • Neuroscience and the Connection to Exemplary Leadership
    • Strategic Leadership and Decision Making
    • Financial Analysis for Engineering Leaders
    • Building Power to Lead
    • Critical Analytical Thinking
    • Innovation and Entrepreneurship in Technology
    • Financing Innovation: The Creation of Value
    • Persuasion: Principles and Practice

  • -

    -

  • -

Licenses & Certifications

Volunteer Experience

Publications

  • PSP-HDRI+: A Synthetic Dataset Generator for Pre-Training of Human-Centric Computer Vision Models

    ICML 2022 - First Workshop on Pre-training: Perspectives, Pitfalls, and Paths Forward

    Co-authored research on synthetic data generation for human-centric computer vision published at ICML 2022.

    Key Contributions:
    • Developed PeopleSansPeople, a synthetic data generator for human pose estimation and segmentation
    • Achieved 60.37 keypoint AP on COCO test-dev2017, outperforming ImageNet pre-training (57.50 AP) and training from scratch (55.80 AP)
    • Demonstrated that synthetic data pre-training improves few-shot transfer learning by +38% keypoint AP on limited…

    Co-authored research on synthetic data generation for human-centric computer vision published at ICML 2022.

    Key Contributions:
    • Developed PeopleSansPeople, a synthetic data generator for human pose estimation and segmentation
    • Achieved 60.37 keypoint AP on COCO test-dev2017, outperforming ImageNet pre-training (57.50 AP) and training from scratch (55.80 AP)
    • Demonstrated that synthetic data pre-training improves few-shot transfer learning by +38% keypoint AP on limited real-world data
    • Open-sourced project with 28 parameterized 3D human assets, 39 animation clips, and 21,952+ unique clothing textures
    • Generated >3M training instances for privacy-preserving computer vision model development

    Impact: This research enables data-efficient training for CV models in privacy-sensitive applications where real-world data is scarce or expensive.

    Authors: Salehe Erfanian Ebadi, Saurav Dhakad, Sanjay Vishwakarma, Chunpu Wang, You-Cyuan Jhang, Maciek Chociej, Adam Crespi, Alex Thaman, Sujoy Ganguly

    Additional Resources:
    • Paper: https://arxiv.org/abs/2207.05025
    • GitHub: https://github.com/Unity-Technologies/PeopleSansPeople
    • Demo Video: https://youtu.be/m9Kb_UewuVk

    Keywords: Computer Vision, Synthetic Data, Human Pose Estimation, ICML, Machine Learning, Deep Learning, Domain Randomization, Transfer Learning

    See publication

Courses

  • Analysis of Algorithms

    -

  • Artificial Intelligence

    -

  • Data Structures

    -

  • Deep Learning

    -

  • Linear Algebra

    -

  • Machine Learning

    -

  • Natural Language Processing

    -

  • Predictive Modeling

    -

  • Probability and Statistics

    -

  • Reinforcement Learning

    -

  • Statistics for Business Analytics

    -

Projects

  • Avito Demand Prediction Challenge

    -

    When selling used goods online, a combination of tiny, nuanced details in a product description can make a big difference in drumming up interest.
    The target variable is the likelihood that an ad actually sold something. It's not possible to verify every transaction with certainty, so this column's value can be any float from zero to one.

    Data: https://www.kaggle.com/c/avito-demand-prediction/data

    Company site: https://www.avito.ru/rossiya?verifyUserLocation=1

    -…

    When selling used goods online, a combination of tiny, nuanced details in a product description can make a big difference in drumming up interest.
    The target variable is the likelihood that an ad actually sold something. It's not possible to verify every transaction with certainty, so this column's value can be any float from zero to one.

    Data: https://www.kaggle.com/c/avito-demand-prediction/data

    Company site: https://www.avito.ru/rossiya?verifyUserLocation=1

    - Analyzed user product’s data and extracted features from text and image to build a model to predict demand of used product. Implemented various models such as LGB, Inception net, ResNet, VGG. The best model was stacking of all three models, gave the RMSE value 0.2237

    See project
  • Digit Recognizer

    -

    Project description
    Data Source: https://www.kaggle.com/c/digit-recognizer
    Git: https://github.com/sanjayuconn/Digit_Recognizer_MNIST

    Implemented SVM to recognize handwritten digits on MNIST dataset with an accuracy of 99.4%. Accuracy further improved to 99.77% using 6 layers architecture of CNN with data augmentation and decay learning rate.

    Below are the configurations:
    SVM:
    C=5
    Gamma=0.05
    kernel=RBF(Radial basis function)
    Accuracy- 99.40%

    Deep…

    Project description
    Data Source: https://www.kaggle.com/c/digit-recognizer
    Git: https://github.com/sanjayuconn/Digit_Recognizer_MNIST

    Implemented SVM to recognize handwritten digits on MNIST dataset with an accuracy of 99.4%. Accuracy further improved to 99.77% using 6 layers architecture of CNN with data augmentation and decay learning rate.

    Below are the configurations:
    SVM:
    C=5
    Gamma=0.05
    kernel=RBF(Radial basis function)
    Accuracy- 99.40%

    Deep CNN1:
    Layer: 5
    Layer activation- Relu, SoftMax
    Regularization: Dropout
    Gradient optimizer: Adam
    Accuracy- 99.57%

    Deep CNN2 with data augmentation:
    Layer: 6
    Layer activation- Relu, SoftMax
    Regularization: Dropout
    Gradient optimizer: RMP prop with decreasing learning rate as per epoch.
    Accuracy- 99.77%

    See project
  • Google Stock Pricing Prediction

    -

    Data Source: https://www.kaggle.com/shivinder/googlestockpricing/kernels

    Implemented ARIMA to predict the stock price for the next day with the RMSE value of 44.5 which is further improved by 3 layers RNN LSTM on sliding window of 60 days to RMSE of 14.8.

    Below are the other models and RMSE value which we have tried:

    1. ARIMA(0,1,0): 123
    2. Prophet: 87
    3. Two layers RNN GRU on Sliding window of 60: 68.2
    4. ARIMA(1,1,1)(0,1,0): 65.7
    5. ARIMA(1,1,1)(0,1,0)+…

    Data Source: https://www.kaggle.com/shivinder/googlestockpricing/kernels

    Implemented ARIMA to predict the stock price for the next day with the RMSE value of 44.5 which is further improved by 3 layers RNN LSTM on sliding window of 60 days to RMSE of 14.8.

    Below are the other models and RMSE value which we have tried:

    1. ARIMA(0,1,0): 123
    2. Prophet: 87
    3. Two layers RNN GRU on Sliding window of 60: 68.2
    4. ARIMA(1,1,1)(0,1,0): 65.7
    5. ARIMA(1,1,1)(0,1,0)+ Events: 44.5
    6. One Layer RNN LSTM: 28
    7: Two-layer RNN LSTM: 16.2
    8. Three-layer RNN LSTM: 14.8
    9. Naive Bay: 51.2%
    10. SVM: 55%
    11. Only RNN LSTM: 64%
    12: Only 2D CNN on word embedding: 69%.
    13. Hybrid model(Inception CNN and RNN LSTM): 82%

    Other creators
    See project
  • Mercedes-Benz Greener Manufacturing

    -

    This dataset contains an anonymized set of variables(400+ columns), each representing a custom feature in a Mercedes car. For example, a variable could be 4WD, added air suspension, or a head-up display.
    The ground truth is labeled ‘y’ and represents the time (in seconds) that the car took to pass testing for each variable.
    Data: https://www.kaggle.com/c/mercedes-benz-greener-manufacturing/kernels

    - Extracted the important features out of 400 features.
    - Reduced the dimension…

    This dataset contains an anonymized set of variables(400+ columns), each representing a custom feature in a Mercedes car. For example, a variable could be 4WD, added air suspension, or a head-up display.
    The ground truth is labeled ‘y’ and represents the time (in seconds) that the car took to pass testing for each variable.
    Data: https://www.kaggle.com/c/mercedes-benz-greener-manufacturing/kernels

    - Extracted the important features out of 400 features.
    - Reduced the dimension using PCA.
    - Applied LGBM, random forest, XGBoost on important features.
    - Ensable model resulted private score 359 although stacking improved score to 154.

    See project

Honors & Awards

  • Awarded for flawless rollout of 100% Biz-logging and data collection for Risk predictive models at PayPal

    PayPal

    Flawless rollout of 100% Biz-logging of Billing Agreement data Via Risk-Specific Kafka Cluster to Hadoop. These data is used by PayPal Risk Models to adjudicate Reference based transactions.

  • Spot award for evincing focus and ownership for PayPal Risk transaction checkpoint development and release in Q1.

    PayPal

    Awarded for evincing focus and ownership for checkpoint development and release in Q1

  • Awarded for immense passion to learn and deliver team commitments before deadline.

    Paypal

    Awarded for immense passion to learn and deliver team commitments.

  • 1st Runner Up - Product Fair Hackathon at Target 2015

    -

  • 2nd Runner Up - CODE-MANIA-3# Hackathon at Bangalore 2014

    -

Organizations

  • Toastmasters

    -

    - Present

Recommendations received

6 people have recommended Sanjay

Join now to view

More activity by Sanjay

View Sanjay’s full profile

  • See who you know in common
  • Get introduced
  • Contact Sanjay directly
Join to view full profile

Other similar profiles

Explore top content on LinkedIn

Find curated posts and insights for relevant topics all in one place.

View top content

Add new skills with these courses