Filter by Skill:

Boston Scientific

Nov 2022 – Present | MN, USA

Global Talent Analytics Platform Owner

Delivered global HR data analytics initiatives resulting in improving analytical capabilities by 40% and realized $1.4M annual time and cost savings within 12 months.

  • Subject matter expert (SME) in fostering a high-performance data culture, managing and mentoring cross- functional teams (8+ analysts, researchers, domain experts, engineers, scientists, and leaders), increasing team productivity and reducing data science and analytics project backlog by 40% within 12 months.
  • Launched self-service analytics and predictive solutions (Tableau, Notifications, data automation, Forecasting, Insights Recommendations) across the org (North America, EMEA, APAC, LATAM), improved executive decision- making speed and reduced ad-hoc reporting requests, totaling $1.4M in time savings.
  • Analytical tools: AWS, SQL, Python, Tableau, Snowflake, ETL, Storage, AI/ML, RPA, Automated data pipelines.
  • Facilitated strategic prioritization meetings with Steerco, Program, HR, Finance, and IT partners to assess the effort, value, and urgency of projects focused on utilizing data, machine learning, and AI techniques to improve the organization's talent analytics capabilities.
  • Advised leaders, providing data-backed solutions and partnered with executives at all levels and HR leadership to deliver actionable insights, analyzing organizational health (headcount trends, growth forecast, hiring, internal mobility, DE&I, and employee experience) to support strategic business decisions.
  • Owned the governance of HR data and reporting methodologies, ensuring consistent data definitions, metrics alignment, and 100% compliance with privacy policies.
  • Technologies: Tableau, AWS, Snowflake, Python, SQL

Sr. Data Scientist/Engineer (Consultant)

  • Led the strategic development of analytics and data platforms for Cardiology Sales Operations(Cardiac Rhythm Management, Atrial Fibrillation, and Interventional Cardiology Therapies), establishing a governance framework that improved data integrity, drove innovation, and made insights accessible to businesses, improving decision-making by 40%. (AWS, Alteryx, Snowflake, Python, SQL, and Tableau)
  • Investigate and resolve data inconsistencies, ensuring accuracy and reliability across several data sources.
Leadership Strategy Data Governance AI/ML Mentorship

Crowd Capital

Mar 2022 – Oct 2022 | Remote

Sr. Data Scientist/Engineer

Led analysts and business partners to introduce a new business-innovated Automated Property Scoring Module on AWS; accelerated bidding strategy, scaled decision-making from 20 to 1000+ bids, and increased profits by $300K in quarterly recurring revenue within twelve weeks of joining.

  • Communicating with stakeholders to study domain and build data product requirements.
  • Supervised a team of two data engineers, helping them learn and grow.
  • Collaborated on Building robust Property Scoring Module Analysis and Developing Bid Optimization by querying big data. (CloudTech/AWS/Azure, BigData/Databricks, Python, Spark/Distributed Computing, Containers/Docker/Serverless, SQL REST APIs, CI/CD, Windows/UNIX Systems).
  • Implemented big data solutions that extract, transform, validate and load (ETL) real-state mortgage industry data and house industry analytics.
  • Illustrated property investigation team to engage, retain, and acquire new properties.
  • Assessed current processes and implemented data engineering and application development track for the organization.
  • Investigation of data inconsistencies for the team to understand the root cause and fix the issue.
  • Technologies: Python, Spark, SQL, AWS
Mentorship Analytics Big Data AI

Emeritus

Oct 2021 – Apr 2022 | Remote

Subject Matter Expert - Data Engineering

Supported Data Engineering curriculum and academic content delivery. Worked on tool integration and module troubleshooting with Professors and content management teams.

  • Data Engineering Content development, modules review, and troubleshooting workflows.
  • Weekly Meetings, Videos reviews, Data Engineering Tools Troubleshooting, Team Engagement, and Support.
  • Manage and help deliver learning Module Content.
  • Technologies: SQL, Python, Educational Tools
Mentorship Data Engineering Big Data

Global Overview

Oct, 2020 - Feb,2022 | Minnetonka, MN

Lead Data Engineer

  • Mentored analysts, engineers and business partners, Achieved $6M+ in cost savings and revenue enablement through improved retail, operations and marketing analytics for 100+ Amazon partners and business stakeholders. [Team Size: 6].
  • Demonstrating leadership skills, evaluating current processes, and implementing data engineering and application development track for the organization.
  • Accomplished robust Marketing Spend Model (DSP, OTT, Search compared to Sales) Analysis and Developing Ecommerce Customer Acquisition Funnel by querying big data. MTA, cross channel optimization.
  • Integrated marketing team performs engagement, retention, acquisition, marketing efficiency, A/B testing

Data Engineer

  • Created robust analytical infrastructure: connecting structured and unstructured data sources from data sources such as AWS RDS, REST API, AWS S3 to Postgres Database for ease of use by different analysts using Python, SQL, and Alteryx. 100+ Vendor and Seller accounts plus approximately 150+ advertising brands. Python, Alteryx
  • Unified legacy data into an analytics-ready data warehouse, reducing costs by $2M annually within 2 quarters.
  • Designed Data Warehouse: Design, Implementation, and Management - Built Single Version of Truth for the analytical and reporting needs. AWS RDS
  • Optimized Reporting Performance: Deep dive into performance issues with analysts improving External Marketing/Sales Report performance from dead slow to 2 seconds refresh leveraging report user experience interacting with reporting. The same thing goes for most other reports. #Tableau, Datorama
  • Analytical study of big data and extracting insights: from the big data lake. AMC Visualization, Development, Coding, Execution. Navigating the project and drive communications with Amazon and internal teams. Optimizing marketing channel presentation through automating reporting solutions to gain marketing spend clarity. #AmazonMarketingCloud, Python, SQL, AWS
  • Datorama: Help support Salesforce Datorama development, execution, and building feedback for Professional Service Provider. Understand datorama backend and development.
  • Identify business priorities and reply timely to keep up with the demands. Introducing ways of tracking high-priority projects, compiling agile methods to introduce and track team progress better. Jira, Confluence, Github
  • Supporting technical development aspects of the Data engineering/Goalkeeper with external data vendor
  • Utilizing technical skills to guide team and support major visualization product developments and transitions
  • Data Tools evaluation to see if that will cost less and give better performance. In-housing data engineering projects, Coming with project and development plans and sharing them with the team to build cost-savings measures.
  • Chime in wherever needed and assist the team resolve high-priority issues - data engineering, validations, and visualizations. Continuous coordination with analysts, business users, and stakeholders to make data and systems available on time.
  • Working with team in-housing share of voice analysis development, integration, and automation. Planning to capture buyer experience when they look for a product on amazon.com. Involves guiding new engineers to get acquainted with building data engineering solutions.
  • Tableau Server Setup and Management - Optimization and Improvements. Alteryx Server Setup and Management - Access and Environment Management.
  • Interviewing technical data engineering candidates and helping human resources with the hiring decisions.
  • Shared thoughts with leadership making Decisions and made myself available as needed.

ETL Developer

  • Fixed Database (data swamp) by building data warehouse system in about 3 months - Worked on 560+ Tables each having different complexity simplifying Global Overviews Backend System to 20-30 valuable database tables. (AWS RDS – PostgreSQL, ALteryx). Evaluating and Performing Grain Evaluation: Identified the granularity of each table and business process to drill down to very important dimensions and measures. Various scenarios and Questions got brainstormed to improve the intelligent system. (SQL, Database Postgres)
  • Tableau Server Crash – help restore the server from the crash and defined a backup strategy around it to avoid future crashes. (Tableau Server, AWS EC2)
  • DWE Server Setup and Management: Build Database architecture from scratch, also responsible for Access and Environment Management. (AWS Cloud, Security Groups and Networking)
  • Build Healthy Data Delivery Pipeline supporting Advertising, Vendor Central, and Seller Central Reporting (24x7 with less Human Intervention) #Reason Automation Migration and implementation.
  • Involved with leadership(Operations, Retail, and Marketing Teams) share knowledge in decisions making.
  • Worked on documenting all tables created to ensure all transactions are drafted properly.
Leadership Strategy Data Governance Mentorship Data Engineering Big Data

Cytilife Inc.

Mar 2022 – Oct 2022 | Minneapolis, MN

Data Scientist

Analytical Pipeline (AWS Web Services Stack: Built robust analytical infrastructure connecting structured and unstructured data sources from data sources such as (RDS, Click Stream and logs) to AWS Redshift, AWS EMR for ease of use by different analysts using Pandas, AWS Athena and QuickSight. (Amazon Kinesis Data Firehouse, AWS S3, AWS Lamda, Glue, DMS, Athena).

Migrated Analytical Artifacts from Tableau to AWS Quicksight with a team of four, leading to $500K in cost savings annually.

Cloud Computing: The AWS infrastructure for the real-time data analysis that involves visualizing and Forecasting Resource Utilization (100K to Millions of rows) of processed data on AWS S3. This helps the administration to investigate business insights from campus resource usage from IoT and Applications Data to determine space and equipment’s demands to take strategic budgeting and resource requirements decisions. (Python, SQL, Visualization).

Data Warehousing: Evaluating and Performing Grain Evaluation: Identified the granularity of each table and business process to drill down to very important dimensions and measures. Various scenarios and Questions got brainstormed to improve the intelligent system. (SQL, MySQL, Visualization).

ML Insights and Anomaly Detection: Python, QuickSight Visualization, Tableau.

Computer Vision: Building real-time people counting application Dashboard for University Administration. OpenCV, Python, AWS.

Usage Forecasting: End to End, Designed Forecasting App with machine learning pipeline and hosted model on Rest API to utilize it making predictions. Pytorch, Python, MySQL, Django, AWS. Deployed machine learning model as a REST API using AWS services like ECR, Sagemaker and Lambda for the Mobile App Development.

  • Continuous coordination with QA team, production support team and deployment team.
  • Worked on documenting all tables created to ensure all transactions are drafted properly.
Strategy Data Governance AI/ML Mentorship

University of St. Thomas, St Paul.

Mar 2022 – Oct 2022 | Remote

Research Assistant

ML/AI Projects

Research Assistant Machine Learning: Unsupervised Contextual Clustering of Abstracts. Used NSF abstract data (300K to Millions of rows) for the last 34 years producing document context through Gensim Doc2Vec Model which suggests similar abstracts based on given abstract. (Deep Learning, ML, NSF, NLP, Python, Visualization, SAS). The submitted paper is being selected at SAS Global Forum 2020 and My Team ranked 1st in the Nationwide SAS Competition. Working on BERT/GPT Models to perform transfer learning/document embeddings.

#Publication #YoutubeVideo

AWSMktPlaceApp: AWS Implementation of the machine learning model as a REST API using Docker(with pre-requisite libraries) and AWS services like ECR, Sagemaker and Lambda for the AWS Marketplace Competition.


Computer Vision - Semantic Image Segmentation Fall 2019: Developed Computer Vision Project. This is Artificial Intelligence Class project was meant to familiarize and apply understanding from DL and ML methods solving Semantic Image Segmentation Problem. (Deep Learning, Pytorch Python, OpenCV, GPU Programming, Deep Learning).

Classification: ML Project: Kickstarter projects to predict the crowdfunded project would be successful, cancelled or unsuccessful.(Scikit Learn, RandomForest, XGBosst, LightGBM, Python, SNS, Pandas, Machine Learning etc..).

Classification- AIM Consulting Hackathon: Won the competition. Designed and developed machine learning classifier on a highly imbalanced dataset. (Scikit Learn, RandomForest, XGBoost, Python, SNS, Pandas, Machine Learning)

Kaggle.com and Hackerrank.com (Some of my Recent Projects Listed Below) – Since 2015:

  • Jigsaw Unintended Bias in Toxicity Classification NLP – created classifying text comment model based on toxicity sentiment considering gender bias into the picture, BERT, LSTM – Top 8% (225/3165 competitors) (NLP, Python, Pytorch, Deep Learning).
  • Built Image Classifier(Diabetic Retinopathy)– Fastai, Pytorch, Machine Learning, Deep Learning, OpenCV - Top 11% (298/2943 competitors, Deep Learning).

Big Data Projects

AWS Data Lake Automation (Serverless): Developed automated cloud formation template to create data lake which ingests the customer’s data, maps it into an analysis-ready structure and makes this data and dashboards available to end-users. Tools Used: AWS Services which includes Cloud Formation, S3, Kinesis, Redshift cluster, Tableau, EMR(Spark/Hadoop), SageMaker, Amazon Athena, Amazon Kinesis, Amazon QuickSight, Amazon Simple Storage Services (S3), AWS Big Data, AWS Database Migration Service, AWS Glue, AWS Lambda.

BigData Architecture: Hands-on experience, working on Apache Hadoop ecosystem components like MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Oozie, Zookeeper, Flume, Spark, Python and EC2 cloud computing with AWS. Build Architecture from scratch to store the dataset from twitter to Hadoop and then perform data analysis using spark/pyspark. (Hadoop, Spark, NiFi, Oozie).

  • Used Sqoop to import data from different RDBMS systems like MySQL, Oracle and loaded into HDFS. Developed Map-Reduce programs to clean and aggregate the data.
  • Developed workflow in Oozie to manage and schedule jobs on Hadoop cluster to trigger daily, weekly and monthly batch cycles. Extensive knowledge of working on NiFi.

Software Development Projects

Software Development- UST MicroGrid Monitoring and Controller: Software Development Project Work. Identifying requirements and developing MicroGrid Devices Controlling System. Angular CLI, ReactJS, Java Spring, MySQL, Linux, Ruby, InfluxDB. #ResearchWork

  • Responsible for understanding the scope of project and requirement gathering.
  • Used Tomcat web server for development purpose.
  • Involved in creation of Test Cases for Unit Testing.
  • Developed application using Eclipse and used to build and deploy a tool like Maven.
  • Used Log4J to print logging, debugging, warning, info on the server console.

Django Web Application: Created Graduate Program in Software department-wide Web-based Software Request System. (Python, Django, MySQL)

Audio Tagging Application: Implemented Audio Tagging application using various Natural Language Processing techniques and Django Web Framework. (BERT, Google Speech to Text, Google Content Classifier, Python, Django, MySQL, GPUs).

Research Strategy Mentorship

Minnemudac 2019

Mar 2022 – Oct 2022 | Remote
This competition was hosted by Minneanalytics and Mudac is well known nonprofit organization serving the data science and emerging technology community in Minnesota, the Upper Midwest. Datasets were sponsored by Farm Femme(Soybean Commodity Data Sponsor) and advised participant to freely use openly available dataset to optimize model predictions. Challenge was to predict 3 different Soybean Contract End of the Day prices for the month of November 2019.

Multivariate Time Series LSTM Predictor: Learned and perfomed detailed analysis from scratch to produce best performing model. Used ideas from Stats, ML,Deep learnig and Commodity market to produce Multivariate Prediction Model. Our Student Team from St. Thomas GPS participated in two consecutive MinneAnalytics competitions (MinneMUDAC and FastCon 2019) and our work was one of the most competitive among the top performers leading most of the teams behind in predictions and technical Design Approach. We were ranked second based on our 15 days prediction for 3 different Soybean's future contract price.

For more information. Click Here

Research Strategy Competition

Semantic Image Segmentation

Mar 2022 – Oct 2022 | Remote
Designed and Developed Computer Vision Project to complete Arificial Intelligence Credit Class. This project was meant to familiarize and apply understanding from DL and ML methods solving Semantic Image Segmentation Problem.Our team worked on different applications of computer vision techniques and chose Semantic segmentation to learn granualrities of designing Neural Network Architecture using Pytorch.

Semantic Segmentation: Identify the object category of each pixel for every known object within an image. Here, Labels are class-aware.

Thumbnail [200x250]

Unet Architecture: U-Net is Fully Connected Network that consists of a contracting path (left side, learns classification) and an expansive path (right side, learns segmantation masks).

Pytorch: PyTorch is an open source deep learning framework built to be flexible and modular for research, with the stability and support needed for production deployment. It enables fast, flexible experimentation through a tape-based autograd system designed for immediate and python-like execution.

For more information. Click Here

Research Strategy Competition

Unsupervised Contextual Clustering of Abstracts

Mar 2022 – Oct 2022 | Remote
This competition is hosted by SAS called SAS Global Forum 2020. The Sponsors wish to award outstanding student innovators with the opportunity to attend SAS Global Forum 2020 where they will have an opportunity to learn, network, and exchange ideas and experiences. This study utilizes publicly available data from the National Science Foundation (NSF) Web Application Programming Interface (API). We are in process of producing context through BERT Language Model.

In this paper, various machine learning techniques are demonstrated to explore, analyze and recommend similar proposal abstracts to aid the NSF or Awardee with the Merit Review Process. These techniques extract textual context and group it with similar context. The goal of the analysis was to utilize the unsupervised learning algorithms to embed NSF funding proposal abstracts text into vector space. Once vectorized, the abstracts were grouped together using K-means clustering. These techniques together proved to be successful at grouping similar proposals together and could be used to find similar proposals to newly submitted NSF funding proposals.

BERT Language Model: BERT builds upon recent work in pre-training contextual representations — including Semi-supervised Sequence Learning, Generative Pre-Training, ELMo, and ULMFit. However, unlike these previous models, BERT is the first deeply bidirectional, unsupervised language representation, pre-trained using only a plain text corpus.

For more information. Click Here

Research Strategy Competition
This lists my additional projects which were done by me to build statistics and machine learning skills.

Recent Works (Esc to view a list of slides and 125% zoom browser to view material clearly):

For more information. Click Here

Research Strategy Competition

Amphora Inc.

Mar 2022 – Oct 2022 | Remote

Worked as a backend automation engineer and applications developer. This involved


  • Responsibilities consisted of Requirements gathering, estimating change requests, Designing and developing code, scripting database stored procedures and functions, Participate in client meetings. Primary technologies involved Java, .NET, Web Services, SQL, Jira, Git, DevOps, CI/CD, Worked with oil trading clients like Mercuria Energy, Noble Group, Arcadia, Eni..
  • Notifications: Developed service for sending an email, push and in-app notifications. Involved in features such as delivery time optimization, tracking, queuing and A/B testing. Built an internal app to run batch processing for software delivery etc.(Java, MS SQL)
  • Trade Aggregator: Simplified bulk data processing and injection service from global exchanges to CTRM and provides preprocessed data for application users. (Unix/Linux, Java, MS SQL)
  • Workflows: Outlined and improved Apache Ant workflow to create and manage build and testing pipelines leveraging automation to expedite development productivity. (JavaScript)
  • Remote Application Services Management Tool: Improved multi-server application service control by building a new application for software product allowing easy handling over multiple remote server environments including DB Scripts Automation. (C#, MS SQL)
  • Technology Migration: Lead team and Effectively implemented GitHub/SCRUM migration process for the development team, which helped developers smoothly transition code repository from TFS to Git. (Git, Subversion, TFS)

Research Strategy Competition