π Master's Student in Applied Data Science @ University of Chicago
π Aspiring Data Scientist | Machine Learning Engineer | LLM Enthusiast
π Currently in Chicago | Fluent in Python, SQL, and cloud-based deployment
I am passionate about transforming data into strategic decisions and scalable systems. With cross-industry experienceβfrom automotive forecasting to SaaS product analyticsβI build end-to-end machine learning pipelines, deploy predictive models in production, and translate data insights into business outcomes.
Currently, Iβm exploring Large Language Models (LLMs) and multimodal learning, integrating NLP pipelines with real-world applications such as recommendation systems and customer behavior modeling.
- π¬ Research-driven with a strong foundation in deep learning, time series forecasting, and recommendation systems
- π§ Recently optimized a DNN-based credit default model
- π Interned at Tencent, Roland Berger, Deloitte, GM, and Continental Automotive across business analytics, risk modeling, and automation
- βοΈ Proficient in cloud deployment using AWS Lambda, Docker, and Google Cloud Platform
- π Strong focus on interpretable AI, using SHAP/LIME to bridge models and decision-makers
- Fine-tuning open-source LLMs for Medical bilingual Q&A and enterprise knowledge retrieval
- Building LangChain-powered pipelines integrated with vector databases
- Creating an AI-driven movie recommendation engine using embedding and Wide & Deep architectures
- Contributing to GitHub repos on AutoML, time series modeling, and multi-task learning
-
π Customer Default Prediction (Verizon)
Custom DNN and XGBoost models with financial cost-based loss function
β +9% business impact, deployed via Docker & AWS Lambda -
π₯ Netflix Movie Recommender System
Combined CF, deep learning(Wide&Deep), and GCP SQL backend for real-time recommendations -
π Cryptocurrency Quant Trading Strategy (MIT Collaboration)
Applied GARCH, LSTM, and momentum strategies using live data from Binance -
π§ LLM Finetuning for Bilingual Medical Application (Ongoing)
Finetuning LLaMA-2 using LoRA and AlpacaEval for Chinese-English knowledge synthesis and applying contrastive learning and data augmentation for model distillation
Letβs build smarter systems and drive decisions through data & AI! π