_gp_singh Profile Banner
Gurinder Pal (GP) Profile
Gurinder Pal (GP)

@_gp_singh

Followers
11
Following
57
Media
3
Statuses
266

Master's in IT Student @swinburne 🎓 | Aspiring Machine Learning & Deep Learning Engineer 💻

Melbourne, Victoria
Joined August 2023
Don't wanna be here? Send us removal request.
@_gp_singh
Gurinder Pal (GP)
2 years
4️⃣ Application in Machine Learning: - Used in algorithms like Decision Trees, Random Forests, and Gradient Boosted Trees. - Helps in feature selection by identifying the variables that best split the data.
0
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
- Difference: While both aim for similar outcomes, Gini Impurity is slightly faster to compute due to its simpler mathematical formulation, making it the default in some algorithms. Entropy can produce slightly more balanced trees.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
3️⃣ Gini Impurity vs. Entropy: - Similarity: Both are used to calculate the homogeneity of a dataset. Decision trees use these measures to decide the best splits; nodes with lower Gini Impurity or Entropy are preferred for making decisions.
1
0
1
@_gp_singh
Gurinder Pal (GP)
2 years
2️⃣ Entropy: Entropy, on the other hand, measures the amount of information disorder or uncertainty. In the context of decision trees, a lower entropy point towards a subset with more homogenous labels, with 0 being completely homogenous.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
1️⃣ Gini Impurity: It measures the likelihood of incorrect classification of a randomly chosen element if it were randomly labeled according to the distribution of labels in the subset. A Gini score of 0 indicates perfect purity, where all elements belong to a single class.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
🚀 Day 32 of #100DaysofMachineLearning - Gini Impurity vs. Entropy in Decision Trees! Exploring the heart of decision-making in trees: How do they decide where to split? #AI #MachineLearning #DataScience #100DaysOfCode
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
📊Understanding and utilizing the ROC Curve and AUC metric is essential for developing and evaluating predictive models, ensuring we make informed decisions based on the model's true performance. Stay tuned as we continue to uncover more tools and techniques in the realm of ML!
0
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
5️⃣ Interpreting AUC Values: - 0.5: No discriminative power (equivalent to random guessing). - 0.7 to 0.8: Acceptable. - 0.8 to 0.9: Excellent. - >0.9: Outstanding.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
4️⃣ Why Use ROC and AUC? - They provide insight into the balance between sensitivity and specificity. - Useful for comparing the performance of multiple classifiers. - AUC is scale-invariant and classification-threshold-invariant, making it a robust metric.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
3️⃣ What is AUC? AUC stands for Area Under the ROC Curve. It provides a single measure of how well a model can distinguish between positive and negative classes. The higher the AUC, the better the model is at predicting 0s as 0s and 1s as 1s.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
2️⃣ True Positive Rate (TPR) and False Positive Rate (FPR): - TPR (Sensitivity): The proportion of actual positives correctly identified by the model. - FPR: The proportion of actual negatives incorrectly classified as positive by the model.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
1️⃣ What is the ROC Curve? The ROC curve is a tool used to evaluate the performance of a binary classification model by plotting the True Positive Rate (TPR) against the False Positive Rate (FPR) at various threshold levels.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
🚀 Day 31 of #100DaysofMachineLearning - Deciphering the ROC Curve and AUC! Today, we're exploring how to evaluate binary classification models using the Receiver Operating Characteristic (ROC) Curve and the Area Under the Curve (AUC) metric. #AI #DataScience #100DaysOfCode
2
0
1
@akshay_pachaar
Akshay 🚀
2 years
Precision, Recall clearly explained:
11
78
474
@_gp_singh
Gurinder Pal (GP)
2 years
5️⃣ Looking Ahead: Focusing on data quality sets the stage for advanced topics in machine learning and deep learning. As we continue, we'll explore how to handle complex datasets, implement advanced models, and tackle real-world machine learning challenges.
0
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
4️⃣ Improving Data Quality: - Data Cleaning: Identifying and correcting inaccuracies or inconsistencies. - Data Augmentation: Enhancing data through techniques like oversampling or synthetic data generation. - Feature Engineering: Selecting, modifying, or creating new features
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
3️⃣ Common Data Quality Issues: - Missing values. - Duplicate entries. - Outliers and anomalies. - Irrelevant features.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
2️⃣ Characteristics of High-Quality Data: - Accuracy: The data correctly reflects real-world conditions. - Completeness: No missing values or gaps in the data. - Consistency: Uniform formats and scales across the dataset. - Relevance: Data is pertinent to the problem at hand.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
1️⃣ Why Data Quality Matters: The adage "garbage in, garbage out" is particularly true in machine learning. The quality of your input data directly influences your model's accuracy and reliability. High-quality data leads to meaningful insights and predictions.
1
0
0
@_gp_singh
Gurinder Pal (GP)
2 years
🚀 Day 30 of #100DaysofMachineLearning - Emphasizing The Importance of Data Quality! As we progress in our learning journey, understanding the critical role of data quality in machine learning is essential for the development of effective models. #MachineLearning #100DaysOfCode
2
0
0