AI6 min read
Random Forests
Improve predictions using multiple decision trees.
Robert Anderson
December 18, 2025
0.0k0
Many trees = Better predictions.
What is Random Forest?
Collection of many decision trees working together!
Like asking 100 experts instead of 1.
How It Works
- Create many decision trees
- Each tree votes on answer
- Majority vote wins
Example
Predicting if customer will buy product in Austin:
- Tree 1: Yes
- Tree 2: No
- Tree 3: Yes
- Tree 4: Yes
- Tree 5: Yes
Final answer: Yes (3/5 voted yes)
Python Code
from sklearn.ensemble import RandomForestClassifier
# Customer data: [age, income_k]
X = [[25, 40], [30, 60], [35, 80], [40, 100]]
y = [0, 0, 1, 1] # 0=No buy, 1=Buy
# Train with 100 trees
model = RandomForestClassifier(n_estimators=100)
model.fit(X, y)
# Predict
customer = [[32, 70]]
prediction = model.predict(customer)
print("Will buy!" if prediction[0] == 1 else "Won't buy")
Why Better Than Single Tree?
- Less overfitting
- More stable
- More accurate
- Handles noise better
Disadvantages
- Slower to train
- Harder to interpret
- Uses more memory
Feature Importance
Random Forest shows which features matter most!
importances = model.feature_importances_
print(f"Age importance: {importances[0]}")
print(f"Income importance: {importances[1]}")
Remember
- Very powerful algorithm
- Great for most problems
- Trade-off: accuracy vs speed
#AI#Intermediate#Ensemble