-

Trusted Partner in Talent Search If you don’t hire — you don’t pay.
Axiom Pro Your Trusted Partner in Talent Search We pre-qualify candidates so you don’t have to. If you don’t hire — you don’t pay. Let’s build your dream team, risk-free.
-
Intelligence Developer
09/01/2024 - 05/14/2025• Designed and maintained Power BI dashboards to track KPIs across organ donation and finance, improving executive visibility and operational decisions. • Built and optimized complex SQL queries to extract, clean, and transform CRM data, increasing reporting accuracy by 30%. • Collaborated with business, technical, and clinical teams to define KPIs, automate reports (SSRS to Power BI), and improve stakeholder engagement. • Used Power Query, DAX, and Excel to shape data for self-serve analytics and scalable reporting. • Led a migration to Power BI Service, enhancing scalability, version control, and real-time data delivery. • Contributed to insight-driven strategies for retention and process optimization through BI tools and data storytelling.
• Designed and maintained Power BI dashboards to track KPIs across organ donation and finance, improving executive visibility and operational decisions. • Built and optimized complex SQL queries to extract, clean, and transform CRM data, increasing reporting accuracy by 30%. • Collaborated with business, technical, and clinical teams to define KPIs, automate reports (SSRS to Power BI), and improve stakeholder engagement. • Used Power Query, DAX, and Excel to shape data for self-serve analytics and scalable reporting. • Led a migration to Power BI Service, enhancing scalability, version control, and real-time data delivery. • Contributed to insight-driven strategies for retention and process optimization through BI tools and data storytelling.
• Designed D3.js and Python-based data visualizations for bi-annual cellular dataset releases, improving visual clarity and usability for research partners • Preprocessed large biological datasets using Excel and Python for integration into visualization pipelines, accelerating insight generation. • Maintained a Git-based CI/CD workflow to ensure collaboration, reduce errors, and streamline visualization deployment. • Collaborated with a cross-functional research team to translate biological data into interpretable insights for scientific publications. • Contributed to scalable, reusable visualization templates aligned with data storytelling and reproducibility best practices
• Led instruction for 100+ students on Python, Power BI, and data storytelling, improving data visualization skills by 20%. • Designed hands-on modules using real-world datasets to teach dashboarding, network analysis (Gephi), and exploratory analytics. • Mentored students in interpreting complex data using SQL, R, and statistical methods, reinforcing analytical thinking. • Communicated technical concepts to diverse learners, strengthening clarity and cross-functional communication skills.
• Developed Python scripts for ETL automation and Oracle/Redshift optimization, improving data quality by 20% and retrieval speed by 30%. • Created and maintained Tableau and QuickSight dashboards to visualize performance metrics and support executive reporting. • Managed structured datasets using AWS S3, enabling secure storage and access for daily analytics workflows. • Streamlined ETL workflows and scalable data models using SQL, R, and Unix, while automating reporting via Control-M for cross-functional analytics delivery. • Conducted A/B testing and applied statistical methods (chi-square, t-tests) to evaluate business strategies and inform product decisions..
Data Science Student Teacher
05/01/2024 - 01/01/2025• Led machine learning and statistical analysis labs, ensuring students mastered theoretical concepts and real-world applications. • Designed interactive labs on model evaluation, feature selection, and bias-variance tradeoff, fostering analytical problem-solving. • Created teaching materials on neural networks, computer vision, and statistical analysis, ensuring clarity and practical application.
• Led machine learning and statistical analysis labs, ensuring students mastered theoretical concepts and real-world applications. • Designed interactive labs on model evaluation, feature selection, and bias-variance tradeoff, fostering analytical problem-solving. • Created teaching materials on neural networks, computer vision, and statistical analysis, ensuring clarity and practical application.
• Designed and evaluated classification models using cross-validation, ROC-AUC, and SHAP to guide marketing campaign targeting. • Applied Natural Language Processing (NLP) using spaCy and BERT to extract trends from customer feedback and online reviews. • Conducted A/B testing, measured KPIs, and analyzed experiment lift using Python to support iterative product improvements. • Performed churn analysis using logistic regression to identify at-risk customers, driving retention strategies that reduced attrition. • Identified inefficiencies in workflows using process mining and business rules modeling, delivering actionable recommendations. • Developed requirement documents and functional specs by translating market trends and performance analysis into business needs. • Analyzed CRM data to uncover underperforming segments and recommended territory realignment strategies that improved sales. • Measured ROI across digital channels using UTM tagging and conversion tracking, optimizing campaign spend allocation. • Created audience segmentation models using clustering algorithms to enhance personalization in marketing automation platforms. • Built Tableau dashboards to track product KPIs (activation, engagement, retention), enabling real-time visibility for product teams. • Worked with stakeholders across operations and finance to define KPIs and analyze cost-saving opportunities using SQL models.
Data Scientist
02/01/2022 - 05/05/2025At Chubb Insurance Company, as a Data Scientist, I significantly improved a Modified GLM model’s F1-score by 22% through Programmatic Labeling, enhancing classification accuracy and reducing manual annotation efforts. I collaborated with Underwriters to design A/B tests for evaluating the impact of machine learning-driven quote prioritization strategies. Additionally, I implemented a Graph-based Community Detection solution, leading to a ~24% increase in business volume by providing instant quotes to small businesses. I took ownership of the end-to-end machine learning pipeline for predicting propensity to bind, resulting in a positive income impact of $4 million. My work involved developing models using XGBoost, GLM, and Decision Trees for tabular data, and BERT for unstructured data like webtext on Azure ML. I also created visualizations using Matplotlib and Plotly to present complex data insights to non-technical stakeholders, ensuring clear communication of the project's benefits.
At Chubb Insurance Company, as a Data Scientist, I significantly improved a Modified GLM model’s F1-score by 22% through Programmatic Labeling, enhancing classification accuracy and reducing manual annotation efforts. I collaborated with Underwriters to design A/B tests for evaluating the impact of machine learning-driven quote prioritization strategies. Additionally, I implemented a Graph-based Community Detection solution, leading to a ~24% increase in business volume by providing instant quotes to small businesses. I took ownership of the end-to-end machine learning pipeline for predicting propensity to bind, resulting in a positive income impact of $4 million. My work involved developing models using XGBoost, GLM, and Decision Trees for tabular data, and BERT for unstructured data like webtext on Azure ML. I also created visualizations using Matplotlib and Plotly to present complex data insights to non-technical stakeholders, ensuring clear communication of the project's benefits.
At Avanade, I designed and developed dynamic Power BI dashboards for a major retail client, utilizing data from Dynamics CRM to provide actionable insights. I also played a key role in developing ETL pipelines, transforming and moving data from various legacy systems to Azure DataLake for a leading insurance firm, streamlining data processes and enabling more efficient data access and analysis.
At Cognizant Technology Solutions, I developed a chatbot using AWS Lex and AWS Lambda to automate the search for over 50,000 parts for a major manufacturing firm. I also automated the ticket creation process using NLP and the MS Azure Bot Framework, reducing customer service time by 30%. I wrote SQL queries for data extraction from MS Azure Cosmos DB to optimize an autocomplete service, improving latency by 40%. Additionally, I used Pandas to transform raw data and developed an Ibex dashboard for sentiment analysis, improving virtual agent performance and increasing customer satisfaction by 50%. My efforts also included semi-automating chatbot training, reducing training time by 40%. I was awarded the Mountain Mover award for outstanding performance in the second quarter of 2019.
Data Scientist
01/01/2024 - 05/02/2025• Designed and deployed scalable ETL pipelines using Python, SQL, Apache Airflow, and AWS Glue to automate ingestion, transformation, and loading of high-volume customer and transaction data into Snowflake and Redshift, improving data accessibility across departments. • Engineered real-time data streaming workflows using Apache Kafka and AWS Kinesis to support fraud detection, Smart Search recommendations, and credit risk analytics, enabling sub-second decision-making. • Developed modular, production-grade ELT workflows with dbt (data build tool) on Snowflake, improving data modeling efficiency and reducing transformation latency by 35% for financial dashboards. • Built and maintained centralized data lakes on AWS S3, establishing a robust data foundation for cross-functional ML, BI, and reporting use cases while optimizing storage costs using lifecycle policies. • Led integration of GenAI-powered Smart Search on Synchrony Marketplace using OpenAI's GPT models and FAISS vector stores, enabling personalized product recommendations and enhancing customer engagement. • Operationalized LLM-based analytics pipelines using Hugging Face Transformers and LangChain for entity extraction, document summarization, and semantic search, accelerating insights from unstructured data by 60%. • Developed and deployed ML pipelines on AWS SageMaker for credit scoring, churn prediction, and sentiment classification models, integrating with downstream workflows via Lambda and API Gateway. • Implemented data quality frameworks using Great Expectations and custom validation layers within Airflow DAGs, ensuring high data trust and compliance with regulatory reporting standards. • Optimized complex analytical queries on Snowflake and Redshift using clustering keys, partition pruning, and query profiling techniques, reducing report generation time by 50%. • Designed and version-controlled infrastructure with Terraform and AWS CloudFormation, automating provisioning of scalable, fault-tolerant components like Lambda, Glue Jobs, IAM, and S3 buckets. • Led schema design and dimensional modeling using Star and Snowflake schemas for customer analytics and financial data marts, boosting reporting performance and simplifying end-user consumption. • Built and maintained self-service data marts and APIs for Power BI, Tableau, and internal microservices, empowering business teams to access curated data with minimal engineering dependencies. • Collaborated with cross-functional teams (AI/ML, risk, customer experience) to deploy NLP solutions using spaCy, BERT, and TensorFlow for feedback classification, sentiment scoring, and call center automation. • Instrumented CI/CD pipelines for data and ML deployments using GitHub Actions and SageMaker Model Registry, reducing release cycles and ensuring model reproducibility in production. • Supported Synchrony's AI governance framework, ensuring responsible use of GenAI and ML tools through metadata tracking, model monitoring, and alignment with internal compliance protocols.
• Designed and deployed scalable ETL pipelines using Python, SQL, Apache Airflow, and AWS Glue to automate ingestion, transformation, and loading of high-volume customer and transaction data into Snowflake and Redshift, improving data accessibility across departments. • Engineered real-time data streaming workflows using Apache Kafka and AWS Kinesis to support fraud detection, Smart Search recommendations, and credit risk analytics, enabling sub-second decision-making. • Developed modular, production-grade ELT workflows with dbt (data build tool) on Snowflake, improving data modeling efficiency and reducing transformation latency by 35% for financial dashboards. • Built and maintained centralized data lakes on AWS S3, establishing a robust data foundation for cross-functional ML, BI, and reporting use cases while optimizing storage costs using lifecycle policies. • Led integration of GenAI-powered Smart Search on Synchrony Marketplace using OpenAI's GPT models and FAISS vector stores, enabling personalized product recommendations and enhancing customer engagement. • Operationalized LLM-based analytics pipelines using Hugging Face Transformers and LangChain for entity extraction, document summarization, and semantic search, accelerating insights from unstructured data by 60%. • Developed and deployed ML pipelines on AWS SageMaker for credit scoring, churn prediction, and sentiment classification models, integrating with downstream workflows via Lambda and API Gateway. • Implemented data quality frameworks using Great Expectations and custom validation layers within Airflow DAGs, ensuring high data trust and compliance with regulatory reporting standards. • Optimized complex analytical queries on Snowflake and Redshift using clustering keys, partition pruning, and query profiling techniques, reducing report generation time by 50%. • Designed and version-controlled infrastructure with Terraform and AWS CloudFormation, automating provisioning of scalable, fault-tolerant components like Lambda, Glue Jobs, IAM, and S3 buckets. • Led schema design and dimensional modeling using Star and Snowflake schemas for customer analytics and financial data marts, boosting reporting performance and simplifying end-user consumption. • Built and maintained self-service data marts and APIs for Power BI, Tableau, and internal microservices, empowering business teams to access curated data with minimal engineering dependencies. • Collaborated with cross-functional teams (AI/ML, risk, customer experience) to deploy NLP solutions using spaCy, BERT, and TensorFlow for feedback classification, sentiment scoring, and call center automation. • Instrumented CI/CD pipelines for data and ML deployments using GitHub Actions and SageMaker Model Registry, reducing release cycles and ensuring model reproducibility in production. • Supported Synchrony's AI governance framework, ensuring responsible use of GenAI and ML tools through metadata tracking, model monitoring, and alignment with internal compliance protocols.
• Developed robust ETL pipelines using PySpark and Azure Data Factory to ingest and transform high-volume transactional data into Azure SQL Data Warehouse, reducing batch processing times by 40%. • Engineered real-time monitoring systems using Isolation Forests and Autoencoders to detect anomalies in supplier lead times and logistics delays, achieving 92% accuracy and reducing vendor penalties by 15%. • Built scalable data pipelines on Azure Databricks for structured and unstructured data, enabling seamless downstream ML model training and analytics consumption. • Implemented NLP-based sentiment analysis models using BERT and spaCy to analyze financial news and customer reviews, improving investment insights and decision-making frameworks. • Designed and deployed cloud-native data workflows on Azure, leveraging Data Lake Gen2, Azure Functions, and Event Hub for end-to-end data orchestration and stream processing. • Developed dynamic dashboards in Power BI and Tableau, visualizing supplier risk metrics and operational KPIs, resulting in a 25% improvement in procurement decisions. • Led development of Monte Carlo simulation models to assess supply chain risk and optimize sourcing strategies under demand variability, reducing inventory holding costs by 20%. • Automated A/B testing frameworks using statistical methods and Python libraries (Statsmodels, Scipy), enabling rapid experimentation for product and pricing strategies. • Collaborated with ML engineers to deploy forecasting models using XGBoost and Random Forest for demand prediction, increasing forecast accuracy by 20% and streamlining logistics. • Optimized big data pipelines with Spark and Delta Lake for scalable machine learning workloads, reducing compute costs and improving time to-insight for executive stakeholders. • Built speech-to-text and text-to-speech models using Azure Cognitive Services to automate customer service voice interactions, improving accessibility and reducing call center load. • Designed and implemented reusable data validation checks using Pytest and Great Expectations, enhancing trust in data and reducing manual QA efforts by 30%. • Streamlined DevOps for data workflows using Azure DevOps, Git, and automated CI/CD pipelines for faster deployment of analytics and data engineering releases. • Created reusable templates and modules for Databricks notebooks, enabling faster onboarding and consistent code practices across engineering teams. • Led cross-functional efforts to scale GenAI prototypes into production, supporting large-scale testing of NLP pipelines in a secure and distributed Azure environment.
Data Scientist
01/01/2025 - 05/02/2025• Designed and fine-tuned a customer lifetime value (CLV) model using LightGBM, boosting marketing ROI by 18% quarter-over quarter. • Led migration of predictive analytics workflows to Azure Databricks, improving model training speeds by 45% and enabling large-scale data processing (~10M+ rows). • Built ML APIs using FastAPI and Docker, facilitating real-time integration of risk scoring models into customer service portals. • Conducted model monitoring via MLflow, setting drift thresholds and retraining alerts that decreased model performance decay by 30%. • Collaborated with cross-functional teams (engineering, product, marketing) to deliver scalable ML pipelines end-to-end.
• Designed and fine-tuned a customer lifetime value (CLV) model using LightGBM, boosting marketing ROI by 18% quarter-over quarter. • Led migration of predictive analytics workflows to Azure Databricks, improving model training speeds by 45% and enabling large-scale data processing (~10M+ rows). • Built ML APIs using FastAPI and Docker, facilitating real-time integration of risk scoring models into customer service portals. • Conducted model monitoring via MLflow, setting drift thresholds and retraining alerts that decreased model performance decay by 30%. • Collaborated with cross-functional teams (engineering, product, marketing) to deliver scalable ML pipelines end-to-end.
• Developed a churn prediction system using XGBoost and PySpark, improving early detection of high-risk clients with 22% higher recall than previous heuristic methods. • Automated EDA workflows and feature generation for 12+ business use cases, cutting manual preprocessing efforts by 40%. • Built Power BI dashboards tracking feature drift, model health KPIs, and customer behavior clusters for strategic planning. • Participated in internal AI guild initiatives to pilot LLM-based document summarization using Hugging Face models and Azure Cognitive Services. • Deployed batch inference pipelines, serving daily scoring outputs to CRM and campaign teams.
• Conducted predictive modeling research on urban mobility datasets, developing a graph neural network (GNN) prototype for optimizing traffic flow predictions. • Processed 50GB+ GPS trajectory data using PySpark, resulting in a 35% reduction in training time compared to legacy pandas based approaches. • Published findings in an internal research report, highlighting the potential of temporal graph analytics for smart city initiatives. • Assisted faculty on curriculum enhancement, adding real-world MLOps content (MLflow, Airflow) into the university's ML engineering courses.
• Built and optimized ETL pipelines for marketing analytics datasets using AWS S3, Airflow, and PostgreSQL, handling daily ingestion of 5M+ records. • Designed and deployed Star and Snowflake schemas to improve analytical query speeds by 50%, enabling near real-time reporting. • Automated metadata tracking with Python scripts, improving audit readiness and data lineage visibility across systems. • Implemented SQL-based feature stores supporting churn prediction and fraud detection models, reducing feature engineering rework by 30%. • Collaborated with ML teams to ensure production data pipelines met model freshness SLAs.
• Built initial churn prediction model using logistic regression and random forests, increasing early warning system precision by 15% . • Conducted feature selection experiments using recursive feature elimination (RFE) to optimize model interpretability. • Delivered quarterly customer insights reports using Python visualizations, improving executive stakeholder engagement with data driven decisions. • Presented findings to marketing and sales leadership, influencing campaign re-targeting strategies for at-risk segments. • Drafted ML pipeline blueprints, laying groundwork for future predictive analytics roadmap.
Junior Data Scientist
02/01/2024 - 05/02/2025• Perform end-to-end ML/AI modeling processes for data science projects to help improve marketing and campaign performance, including: • Understanding business objectives and translating them into data-driven solutions. • Accessing and extracting data from diverse and large-scale sources, such as sales, service records, client demographics, marketing campaigns, facilities, infrastructure, staffing, financials, and more. • Performing data cleaning, integration, and quality control to ensure data accuracy and usability. • Conducting exploratory data analysis (EDA) using summary statistics, aggregations, feature engineering, and random sampling techniques to find data insights and prepare data for modeling. • Building and developing ML/AI models tailored to the project goals. • Validating and selecting the optimal model by evaluating robustness and stability. This includes calculating key performance metrics such as accuracy, ROC, AUC, capture rate, lift, decile analysis, scoring, and overall population performance. • Providing actionable recommendations based on model insights to project managers and clients. • Documenting the full workflow and delivering clear, comprehensive reports. • Deploying models and creating implementation codes to support automation. • Projects in the past year: • Nissan EV re-purchase predictive model. • Brookdale senior living center monthly campaign predictive models (for assisted living, independent living, and memory care units). • KIA re-purchase recommendation model. • Hyundai multi-classifier re-purchase recommendation model. • KIA multi-label service recommendation model. • Brookdale senior living community forecast model. • KIA service feedback sentiment analysis, service theme clustering and voice of customer model. • Nissan open-retail re-purchase predictive model. • BMW Teleservice inactive user service predictive model.
• Perform end-to-end ML/AI modeling processes for data science projects to help improve marketing and campaign performance, including: • Understanding business objectives and translating them into data-driven solutions. • Accessing and extracting data from diverse and large-scale sources, such as sales, service records, client demographics, marketing campaigns, facilities, infrastructure, staffing, financials, and more. • Performing data cleaning, integration, and quality control to ensure data accuracy and usability. • Conducting exploratory data analysis (EDA) using summary statistics, aggregations, feature engineering, and random sampling techniques to find data insights and prepare data for modeling. • Building and developing ML/AI models tailored to the project goals. • Validating and selecting the optimal model by evaluating robustness and stability. This includes calculating key performance metrics such as accuracy, ROC, AUC, capture rate, lift, decile analysis, scoring, and overall population performance. • Providing actionable recommendations based on model insights to project managers and clients. • Documenting the full workflow and delivering clear, comprehensive reports. • Deploying models and creating implementation codes to support automation. • Projects in the past year: • Nissan EV re-purchase predictive model. • Brookdale senior living center monthly campaign predictive models (for assisted living, independent living, and memory care units). • KIA re-purchase recommendation model. • Hyundai multi-classifier re-purchase recommendation model. • KIA multi-label service recommendation model. • Brookdale senior living community forecast model. • KIA service feedback sentiment analysis, service theme clustering and voice of customer model. • Nissan open-retail re-purchase predictive model. • BMW Teleservice inactive user service predictive model.
• Managing and developing data quality assurance for multiple projects and data systems, including educational entity, financial, student/staff demographic, and student performance data. E-Mail: xinliu.co@gmail.com Cell Phone: 720-695-8191 • Building and writing programming codes for application data elements and business rules. • Implementing common data standards to map from the raw data resources, ensuring compliance with regulatory requirements and data governance practices. • Creating and using standard macro programs in SAS to support data quality checks and analysis. • Conducting comprehensive data analysis, generating data tables, listings, and summary reports and responding to inquiries from internal stakeholders and external clients.
• Designing automotive accessory products using AutoCAD • Managing the associated manufacturing processes, monitoring product quality, and generating detailed performance and production reports.
• Conducting field, area, and basin studies, and managing geological data, including cleaning, verifying data sources, and reviewing datasets to ensure accuracy, consistency, and integrity. • Providing petrophysical log data management and technical support to team members, including editing and normalizing log data, working with tops and maps, calibrating core-to-log data and well test data, and validating and integrating fracture and production data. • Developing and maintaining data-driven quantitative petrophysical models and managing the resulting datasets to support integrated workflows across the team. • Preparing presentations and reports that offer recommendations for operational strategies and business decisions, with a focus on enhancing long-term production and reducing total costs.
• Providing academic and administrative committee services. • Delivering lectures, including Electronic Circuits, Linear System, Pattern Recognition (Data Mining). • Conducting research focused on data modeling and statistical analysis using C++, MATLAB, SAS.
• Leading research, managing data, coordinating laboratory operations and contributing to academic publishing. • Applying the model-based POD (Probability of Detection) techniques to improve the reliability of steam generator tube inspection.
Successfully created
AXIOM PRO is your strategic partner for hiring top global talent.
We specialize in recruiting and outstaffing solutions that help you scale efficiently — without the overhead.
Our team rigorously pre-screens candidates, verifies English fluency and technical proficiency, and presents only job-ready professionals who can hit the ground running.
Risk-free hiring — pay only when you succeed!