Understanding Decision Trees: Learning Through Structured Thinking
In our recent Machine Learning session, we explored one of the most intuitive, structured, and powerful supervised learning algorithms — Decision Trees.
But this session was not just about understanding an algorithm. It was about understanding how machines replicate logical human thinking, breaking down complex decisions into smaller, structured steps.
Decision Trees are not just models — they are frameworks for systematic reasoning.
Introduction to Decision Trees
A Decision Tree is a supervised learning algorithm used for:
Unlike many complex algorithms that act like black boxes, decision trees are highly interpretable. Every prediction can be traced through a sequence of logical conditions.
At its core, a decision tree works by repeatedly asking questions about data and splitting it into subsets based on feature values.
Structural Components of a Decision Tree
We explored the architecture of a decision tree in depth:
Root Node
The starting point of the tree. It represents the most important feature chosen to split the dataset first.
This feature is selected based on a splitting criterion (such as Information Gain).
Internal (Decision) Nodes
These represent conditional tests on features.
Example:
Each internal node further divides the dataset.
Branches
Branches represent outcomes of conditions.
Each branch moves us closer to a final decision.
Leaf Nodes
Leaf nodes represent the final prediction:
This hierarchical structure mirrors structured reasoning.
Real-Life Example 1: Student Academic Performance Prediction
To predict whether a student will Pass or Fail, the model evaluates multiple features step-by-step:
Each answer splits the dataset into smaller groups.
This example demonstrated:
We also discussed that relying on only one feature (like marks) is insufficient. Better predictions require multiple contributing variables.
Real-Life Example 2: Daily Routine Analysis
We extended the concept further using behavioral parameters:
Each parameter becomes a decision node in the tree.
The key understanding was:
A decision tree evaluates conditions sequentially until sufficient certainty is achieved.
This example showed how structured logical evaluation leads to clear outcomes.
Core Theoretical Concepts Discussed
To understand how trees decide the best split, we explored foundational ideas:
Recommended by LinkedIn
Feature Selection
Not all features are equally important.
The model selects the feature that best separates the data at each stage.
Splitting Criteria
The decision of “which feature to split on” is not random.
It is based on mathematical measures such as:
Entropy (Basic Understanding)
Entropy measures impurity or randomness in data.
A pure node contains mostly one class.
Information Gain
Information Gain measures how much uncertainty is reduced after splitting.
Formula conceptually: Information Gain = Entropy (before split) − Entropy (after split)
Key takeaway from class:
Higher Information Gain → Better Split → Reduced Uncertainty
This explains how the model systematically reduces randomness at each step.
Learning Outcomes Using ASK Model
The session was structured to develop Knowledge, Skill, and Attitude holistically.
Knowledge Developed
Students gained clarity on:
Students moved beyond memorizing definitions and began understanding the reasoning mechanism.
Skill Development
Students developed the ability to:
This improved analytical and structured problem-solving capability.
Attitude Development
Perhaps the most important outcome was mindset development.
Students cultivated:
The class emphasized that AI should not just predict — it should be explainable and logical.
Overall Impact of the Session
By the end of the session:
Students were able to connect real-world decision-making with machine learning logic
-> Students understood how machines split data intelligently
Students gained clarity on reducing uncertainty through structured reasoning ✔ Students strengthened multi-parameter analytical thinking
Decision Trees are not merely an algorithm. They are a representation of structured intelligence and explainable reasoning.