DEV Community

Sourav Sarangi
Sourav Sarangi

Posted on

My Journey Through the Kaggle Google 5-Day Intensive ML Sprint

This is a submission for the Google AI Agents Writing Challenge: [Learning Reflections OR Capstone Showcase]

My Learning Journey / Project Overview

Key Concepts / Technical Deep Dive

Reflections & Takeaways

Over the past week, I completed the Kaggle Γ— Google 5-Day Intensive Program β€” a fast-paced, hands-on sprint that helped me dive into Python for Data Science, Machine Learning basics, and Kaggle-style workflows. Below, I’m sharing the full structure of the course, how I experienced each day, what I built, and the skills I gained. If you’re starting out in ML or thinking of trying Kaggle, this might help you decide if this path is for you.


πŸ“… Course Structure & My Daily Experience

Day 1 β€” Getting Started: Python Basics + Kaggle Environment

βœ”οΈ Introduction to the Kaggle environment: Notebooks, datasets, competitions.

βœ”οΈ Brushed up on Python essentials β€” lists, dictionaries, loops, conditionals, functions.

βœ”οΈ First hands-on task: loaded a dataset using Pandas and performed basic exploration (head, shape, info).

My takeaway: Kaggle Notebooks are beginner-friendly, and running code live makes experimentation very straightforward.


Day 2 β€” Data Cleaning & Exploratory Data Analysis (EDA)

βœ”οΈ Learned data cleaning: handling missing values, removing duplicates, filtering outliers.

βœ”οΈ Explored data using Pandas: .describe(), grouping, filtering, summary statistics.

βœ”οΈ Performed preliminary visualization to observe data distributions and relationships.

My takeaway: Investing time in clean, well-explored data is critical β€” it lays the foundation for good ML results.


Day 3 β€” First Machine Learning Models (Baseline)

βœ”οΈ Understood the ML workflow: splitting data into training and test sets, fitting models, evaluating performance.

βœ”οΈ Built baseline models using Scikit-Learn:

Linear Regression (for regression tasks)

Decision Trees

Random Forests

βœ”οΈ Ran a quick mini-competition/prediction task on a real dataset.

My takeaway: Even baseline models β€” with minimal tuning β€” can deliver surprisingly decent results on real-world data.


Day 4 β€” Enhancing Models: Feature Engineering & Hyperparameter Tuning

βœ”οΈ Practiced feature engineering: generating new features, encoding categorical variables, scaling when required.

βœ”οΈ Applied hyperparameter tuning and cross-validation strategies to improve model performance.

βœ”οΈ Learned about the importance of model interpretation and avoiding overfitting.

My takeaway: Often, smarter features and better validation improve performance more than choosing a more complex model.


Day 5 β€” Final Project: End-to-End Pipeline + Submission

βœ”οΈ Built a complete ML pipeline: Data loading β†’ cleaning β†’ exploration β†’ feature engineering β†’ model training β†’ evaluation β†’ prediction.

βœ”οΈ Generated submission.csv and submitted to a real competition on Kaggle.

βœ”οΈ Witnessed the model’s score and placement on the leaderboard β€” first β€œreal” ML submission.

My takeaway: Going from zero to a full submission in 5 days is possible β€” and hugely motivating. It turns theory into a tangible outcome.

Top comments (0)