Home

Aravind B - Gen AI/AI/ML Developer
[email protected]
Location: Tampa, Florida, USA
Relocation: yes
Visa: H1B
Resume file: Aravind_AIML_Engineer-10_1745415394763.docx
Please check the file(s) for viruses. Files are checked manually and then made available for download.
Professional Summary:
AI/ML Engineer with 10 years of expertise in Data Science, Machine Learning, Deep
Learning, and Large Language Models (LLMs), specializing in building, fine-tuning,
and deploying scalable AI-driven solutions.
Expertise in developing, fine-tuning, and deploying LLMs and Generative AI using GPT-
4, Gemini AI, LLaMA, Falcon, and BLOOM, with proficiency in diffusion models,
GANs, VAEs, and parameter-efficient fine-tuning (LoRA, QLoRA, PEFT).
Designed and deployed personalized treatment recommendation systems using
reinforcement learning and patient history analysis to optimize medication plans and
improve healthcare outcomes.
Implemented Retrieval-Augmented Generation (RAG) using vector databases (FAISS,
Pinecone, Weaviate, ChromaDB) for enterprise AI solutions; hands-on experience with
Python, TensorFlow, PyTorch, and Hugging Face Transformers.
Developed and optimized NLP models (e.g., BERT, GPT, T5, RoBERTa) with
advanced tokenization techniques (Word2Vec, FastText, BPE, Sentence
Transformers) for chatbots, document summarization, and text analytics using
NLTK, SpaCy, and OpenAI APIs.
Deployed AI/ML models on AWS, Azure, and GCP using SageMaker, Vertex AI, and
Azure ML; skilled in serverless architectures (AWS Lambda, Cloud Run) and
distributed model training with PySpark and Databricks.
Built scalable AI-driven data pipelines using Apache Airflow, Kafka, and Snowflake,
ensuring efficient data processing and real-time inference with integrated feature stores
(Feast, Databricks Feature Store).
Proficient in processing large-scale structured and unstructured data with Python,
PySpark, Snowflake, BigQuery, and Redshift, enhancing real-time model inference
and retraining workflows.
Hands-on experience with containerized deployments using Docker, Kubernetes,
and Helm, optimizing model serving with KServe, TensorFlow Serving, and Triton
Inference Server.
Utilized MLflow, Kubeflow, and TensorFlow Extended (TFX) for automated model
versioning, tracking, and lifecycle management, ensuring continuous improvement and
governance.
Integrated AWS SageMaker Model Monitor, GCP Vertex AI, and Azure ML for model
drift detection, performance evaluation, and anomaly detection, maintaining reliable
AI performance.
Built HIPAA- and FHIR-compliant AI/ML models for secure healthcare data processing
and EHR integration, enhancing predictive analytics for patient outcomes.
Proficient in API development using FastAPI, Flask, and Uvicorn, delivering
asynchronous AI inference services and scalable API solutions for AI-powered
applications.
Skilled in evaluation metrics like precision, recall, F1-score, and AUC-ROC,
ensuring 95%+ accuracy and 90% precision in classification and regression tasks.

Skills:
Programing & Scripting
Languages

Python, Pyspark, Java, SQL, Bash, Perl, YAML, Groovy
Machine Learning Supervised Learning, Unsupervised Learning, Feature Engineering,
Model Optimization, Time Series Analysis, Recommendation
Systems, Sentiment Analysis, Evaluation Metrics.

Deep Learning TensorFlow, PyTorch, Keras, CNNs, RNNs, LSTMs, GANs,
Transformer Models, ANN, Transfer Learning, Ensemble Models,
Computer Vision.

Large Language
Models (LLMs)

GPT, BERT, LLaMA, Falcon, Hugging Face Transformers, LoRA,
DeepSpeed.

Retrieval-Augmented
Generation (RAG) FAISS, ChromaDB, Pinecone, Weaviate, Llama Index.
Generative AI Fine-tuning LLM s, Langchain, Llama index, RAG, AI Agents,
CrewAI, Auto Gen, Prompt Engineering, Hugging face, Phi, OpenAI,
Llama, FIASS.

Big Data & Distributed
Computing HDFS, MapReduce, PySpark
Model Serving FastAPI, Flask, Streamlit, Kserve, TensorFlow Serving, TorchServe,
Cloud Run, Vertex AI Endpoints, SageMaker Endpoints, Azure ML
endpoints.

Cloud Platforms

AWS (SageMaker, Bedrock, Lambda, Lex, CloudWatch, CloudTrail,
Redshift ML, DynamoDB, CodeBuild, CodeDeploy, S3, EC2, IAM,
AMIs). GCP (Vertex AI, AutoML, Cloud Vision API, Dialogflow,
NVIDIA GPUs, BigQuery ML, VM Instance, VPC).
Azure (Azure ML, Azure AI & OpenAI, Blob Storage, Azure
Functions, Azure Cognitive Services)

MLOps & CI/CD GitHub Actions, GitLab CI/CD, Bitbucket Pipelines, Jenkins,
CircleCI, Cloud Build, CodePipeline, Azure Pipelines, MLflow,
Kubeflow, DVC, DagsHub

Monitoring & Logging AWS CloudWatch, GCP Cloud Monitoring, Azure Monitor, CloudTrail
Logs, Prometheus, Grafana, MLflow, Weights & Biases,
TensorBoard, Evidently AI, Vertex AI Model Monitoring, Cloud
Logging
MLOps & Workflow
Orchestration

Kubeflow, MLflow, Airflow, TFX Pipelines, Vertex AI Pipelines,
Supervised/Unsupervised Algorithms, ANN, CNN, NLP, Computer
Vision, GAN s, LSTM, Feature Engineering, Transfer Learning,
ensemble models, Time series, Recommendation systems,
Sentiment Analysis, Evaluation Metrics.
.

Frame Works PySpark, Tensorflow, Keras, PyTorch, Scikit-Learn, openCV, NLTK,
Pandas, Transformers, Flask, Celery, FastAPI, Streamlit, Gradio,
Pickle, Pydantic, Anaconda, Jupyter Notebook.
Version Control GIT, GitHub, Azure Repos, AWS Code Commit.
Data Warehousing Snowflake, Data Lake Storage, Google Big Query.
Container
Orchestration and
Infrastructure as Code

Docker, Kubernetes, Terraform.

Operating Systems Windows, Ubuntu, Linux.

Certifications:
Azure Certification - Link
Professional Experience:
Client: Motive MI, SFO, CA(Remote) Jun
2023 - Jan 2025
Role: Senior AI/ML GenAI Engineer
Responsibilities:
Developed Generative AI-powered chatbots using AWS Bedrock and fine-tuned LLMs
(e.g., GPT-4, Falcon, T5) for real-time patient symptom analysis and personalized
treatment recommendations.
Implemented Retrieval-Augmented Generation (RAG) with FAISS and AWS
OpenSearch to enhance chatbot accuracy by integrating real-time EHR and clinical trial
data.
Leveraged Deep Learning and NLP models (e.g., BERT, T5) with SpaCy for clinical text
processing, enabling the chatbot to extract vital health indicators from unstructured medical
documents.
Designed and deployed LLM-powered chatbot APIs using TensorFlow Serving,
PyTorch, FastAPI, and Google Vertex AI, ensuring low-latency inference and scalable
deployment for real-time medical queries.
Developed feature engineering pipelines using PySpark, Pandas, NumPy, SciPy, and
scikit-learn for efficient data ingestion from AWS S3, Google Vertex AI, and Snowflake to
optimize training data.
Built secure, scalable, and compliant chatbot solutions with SageMaker endpoints, Google
Vertex AI, Docker, and Kubernetes, ensuring HIPAA-compliant inferencing.
Integrated AWS CloudWatch, Prometheus, and Grafana with Google Vertex AI for real-
time chatbot monitoring, model drift detection, and automated retraining.
Deployed fine-tuned LLMs on AWS SageMaker, AWS Bedrock, and Google Vertex AI for
high-performance inference, powering the chatbot s natural language processing
capabilities.
Applied TF-IDF, BERT embeddings, and PCA for text feature extraction and
dimensionality reduction, improving chatbot accuracy in understanding medical queries.
Developed AI/ML models for secure healthcare data processing and predictive analytics
using Python, TensorFlow, PyTorch, and NLP libraries, ensuring compliance with
healthcare regulations like HIPAA and GDPR.
Preprocessed medical images using OpenCV and wavelet transforms, enabling multi-
modal capabilities in the chatbot to process and analyze medical visuals, all through
Python.
Implemented Named Entity Recognition (NER) using SpaCy and BioBERT in Python,
enabling the chatbot to extract important medical entities such as symptoms and treatments
from unstructured clinical texts.
Fine-tuned GPT-4, LLaMA, and BioBERT models for clinical text analysis and AI-driven
diagnosis within the chatbot, optimizing performance using LoRA, quantization,

TensorRT with TensorFlow and PyTorch to reduce inference latency and improve
conversational accuracy.
Client: OPTUM, Albany, NY
Dec 2021 May 2023
Role: AI/ML Engineer
Responsibilities:
Developed a predictive healthcare analytics system to assess patient readmission
risks using Azure ML, Snowflake, and Python.
Designed and implemented machine learning models (Logistic Regression, Random
Forest, XGBoost, and LSTMs) using TensorFlow and PyTorch for time-series patient
history analysis.
Processed and cleaned structured & unstructured medical data from Azure Data Lake
& Snowflake, handling missing values, feature engineering, and text-to-numeric
transformations.
Built an NLP pipeline using TensorFlow and PyTorch for analyzing doctor s notes and
extracting clinical insights with Named Entity Recognition (NER).
Deployed ML models as RESTful APIs using FastAPI and TensorFlow Serving, enabling
real-time predictions for hospitals and medical staff.
Optimized model inference performance with ONNX and TensorRT, reducing latency by
40% for real-time decision-making.
Implemented dimensionality reduction techniques (PCA, t-SNE) to enhance feature
selection and improve model efficiency on large-scale patient datasets.
Integrated real-time monitoring with Azure Monitor, Prometheus, and MLflow, tracking
model drift and performance metrics in production.
Ensured data security and compliance by implementing RBAC (Role-Based Access
Control), HIPAA, and GDPR standards with Azure Key Vault for encrypted storage.
Developed a self-learning pipeline for continuous model retraining using new patient data,
automating updates through Azure ML Pipelines and Databricks Notebooks.
Designed and maintained an end-to-end CI/CD pipeline for ML model deployment using
GitHub Actions, Docker, and Kubernetes in Azure Kubernetes Service (AKS).
Collaborated with cross-functional teams including data scientists, cloud engineers, and
healthcare professionals to improve patient risk prediction accuracy by 30%.

Client: British American Tobacco, Albany, NY Mar 2020 -
Nov 2021
Role: ML Engineer
Responsibilities:
Designed and implemented a predictive analytics solution using Azure Machine
Learning and Databricks to forecast both cost changes and future sales volumes of
tobacco products.
Built ML models using Linear Regression, Gradient Boosting, SVM, ElasticNet,

Random Forest, and K-Means clustering to analyze cost trends, demand patterns, and
optimal pricing strategies.
Developed large-scale data pipelines in PySpark within Databricks Notebooks to
process past 3 years of historical sales data from Snowflake and predict sales amounts
and costs for the next 2 years.
Implemented feature engineering and data transformation techniques to extract
meaningful insights from sales data, customer trends, and external market factors.
Integrated Snowflake as a scalable data source for both batch and real-time ingestion,
enabling automated data pipelines using Azure Data Factory.
Trained and optimized ML models to minimize forecasting errors and improve cost
efficiency and sales planning accuracy by leveraging Azure ML AutoML and
Hyperparameter tuning.
Deployed predictive models as endpoints in Azure ML and exposed them via FastAPI
and Azure Functions for real-time inference and integration with business applications.
Implemented CI/CD pipelines with Azure DevOps, GitHub Actions, and MLflow to
automate model retraining, deployment, and versioning for continuous improvement.
Monitored model performance, accuracy, and drift using Azure Monitor, MLflow,
Prometheus, and Grafana, ensuring reliability in sales and cost predictions.
Developed dashboards and visualizations in Power BI to provide stakeholders with
insights into future sales forecasts, pricing strategies, and demand planning.
Reduced forecasting errors by 25% and improved sales planning efficiency, enabling
optimized production and pricing decisions for British American Tobacco.

Maveric Systems, Chennai, India Apr
2018 - Feb 2020
Role: Data Scientist & DevOps Engineer
Responsibilities:
Developed predictive models using Python (pandas, NumPy, SciPy, scikit-learn) for
credit risk and fraud detection.
Designed feature engineering pipelines, handling missing values, outliers, and categorical
encoding, improving model accuracy by XX%.
Designed and implemented end-to-end machine learning pipelines using PySpark and
Databricks Notebooks for sales forecasting.
Built time-series forecasting models (ARIMA, Holt-Winters, Exponential Smoothing) to
predict customer default risks.
Developed ETL workflows using AWS Glue and PySpark, processing structured and semi-
structured data.
Deployed trained models as Flask APIs on AWS EC2, integrating them into real-time risk
analysis systems.
Automated CI/CD pipelines with Jenkins, AWS CodePipeline, and Docker, ensuring
seamless model deployment.
Configured AWS S3 for data storage with lifecycle rules for cost optimization.
Monitored and logged model performance using AWS CloudWatch and Prometheus.

Integrated models into business intelligence tools like Tableau and Power BI for reporting
and analytics.

Client Name: ACE Technologies, Hyderabad, India Jul
2015 - Mar 2018
Role: DevOps Engineer (Data Scientist)
Responsibilities:
Conducted customer segmentation using K-Means Clustering, Hierarchical Clustering,
and Principal Component Analysis (PCA).
Developed Python-based data pipelines using pandas, NumPy, and PySpark, processing
large financial datasets for risk modeling.
Designed Python-based data preprocessing modules, handling missing values, feature
scaling, and categorical encoding.
Built demand forecasting models using Linear Regression, Decision Trees, and
Random Forest with Python.
Created ETL pipelines with Azure Data Factory, automating data ingestion and
transformation from multiple sources.
Managed data in Azure SQL Database and Azure Blob Storage, optimizing performance
with indexing and partitioning.
Deployed REST APIs with FastAPI and Flask on Azure App Services for real-time
analytics.
Automated model retraining and deployment using Azure DevOps pipelines, reducing
deployment time by XX%.
Set up monitoring and logging with Azure Monitor and Application Insights, tracking
API usage and errors.
Developed interactive Power BI dashboards, providing stakeholders with insights on
customer behavior and sales trends.
Secured data access with Azure RBAC, Managed Identities, and encryption, ensuring
compliance with data security policies.
Education:
Bachelors in Computer Science JNTUK(2011-2015)
Keywords: continuous integration continuous deployment artificial intelligence machine learning business intelligence sthree California Michigan New York

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];5336
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: