SeldonResearch Profile Banner
Seldon Research Profile
Seldon Research

@SeldonResearch

Followers
79
Following
11
Media
35
Statuses
174

Seldon Data Science and Research. Developers of Alibi Explain https://t.co/SvvkZ4vbwo and Alibi Detect https://t.co/3zKwz9sXwQ. Slack: https://t.co/pZo6GwIt4v

Joined July 2021
Don't wanna be here? Send us removal request.
@SeldonResearch
Seldon Research
4 years
There is an increasing awareness among practitioners that data drift poses a challenge to the robust deployment of machine learning models. But what precisely is meant by β€œdrift” and how can we protect ourselves against it? πŸ‘‡πŸ“½οΈ 🧡 https://t.co/Bru4FaCcGe
1
10
17
@SeldonResearch
Seldon Research
3 years
Check out Alex's (@oblibob) blog post on generative modelling using vector-quantized VAEs!
0
0
1
@seldon_io
Seldon
3 years
We have a paper accepted into the R2HCAI workshop titled, Model-agnostic and Scalable Counterfactual Explanations via Reinforcement Learning. πŸ™Œ Very excited to be a part of the conversation around the advances of responsible AI. πŸ’ͺ Learn more: https://t.co/GiG9pjA6Hx #AAAI
0
1
4
@SeldonResearch
Seldon Research
3 years
For more details, check out our example benchmarking drift detectors with the KeOps backend here:
0
0
0
@SeldonResearch
Seldon Research
3 years
This drastically speeds and scales up the detectors to large dataset sizes, with dataset sizes in the order of 100,000’s easily achievable on a single consumer grade GPU.
1
0
0
@SeldonResearch
Seldon Research
3 years
Alibi Detect v0.11.0 introduces a new backend for the MMD and learned kernel MMD detectors. Internally, these detectors use the KeOps library, developed by @FeydyJean and @JoanGlaunes. This allows the kernel matrices to represented by symbolic tensors.
1
0
0
@SeldonResearch
Seldon Research
3 years
The sensitivity of a drift detector scales with dataset size. However, the memory and computational costs of a number of convenient and powerful kernel-based drift detectors, such as the MMD detector, do not scale favourably with increasing dataset size.
1
0
0
@SeldonResearch
Seldon Research
3 years
We are excited to announce the release of Alibi Detect v0.11.0, featuring widened serialisation support and a new backend that allows drift detection to be rapidly performed on large datasets.
1
1
6
@SeldonResearch
Seldon Research
3 years
Much more information on Permutation Importance and Partial Dependence Variance, including worked examples, can be found on our documentation pages: https://t.co/ZYgYz76vum https://t.co/d1NUeK8yQ6
0
0
0
@SeldonResearch
Seldon Research
3 years
Both of these insights are complementary as PI captures not only main feature effects but also interactions, and we recommend considering both, when possible, for a thorough analysis of model behaviour.
1
0
0
@SeldonResearch
Seldon Research
3 years
When to use PI vs PDV? The key lies in the interpretation of the importance values. Whilst PDV quantifies how much of the model's output variance is explained by each feature, PI measures how much model performance degrades when a feature is noised.
1
0
0
@SeldonResearch
Seldon Research
3 years
Furthermore, PDV can be extended to also quantify pairwise feature interaction strengths, allowing a deeper understanding which features interact with each other inside the model.
1
0
0
@SeldonResearch
Seldon Research
3 years
Partial Dependence Variance (PDV) derives from Partial Dependence (PD) plots. Intuitively, calculating PD for a feature, the resulting points on the plot will collectively have higher variance if the feature is more discriminative wrt the model. PDV formalizes this calculation.
1
0
0
@SeldonResearch
Seldon Research
3 years
The metric/loss can be customized. The plot below shows the feature importance wrt to accuracy and F1 metrics of a random forest predicting whether employees are likely to leave a company. The feature "satisfaction_level" is the most important one regardless of the metric.
1
0
0
@SeldonResearch
Seldon Research
3 years
Permutation Importance (PI) works by selecting a feature of interest, shuffling the values of that feature across the dataset and then measuring the effect on some metric or loss function on this new dataset with respect to the original.
1
0
0
@SeldonResearch
Seldon Research
3 years
Both Permutation Importance (PI) and Partial Dependence Variance (PDV) assign a scalar value to each feature to quantify their importance with respect to the model. Both methods are model-agnostic, but PI requires ground-truth labels, so will be more useful during development.
1
0
0
@SeldonResearch
Seldon Research
3 years
We are pleased to announce the release of Alibi Explain v0.9.0 with support for calculating global feature importance via Permutation Importance or Partial Dependence Variance.
Tweet card summary image
github.com
Algorithms for explaining machine learning models. Contribute to SeldonIO/alibi development by creating an account on GitHub.
1
0
3
@SeldonResearch
Seldon Research
3 years
For a more extensive discussion of the method, its usage and examples please visit our documentation page: https://t.co/ntscZBUGC5.
0
0
0
@SeldonResearch
Seldon Research
3 years
Our PD implementation in Alibi v0.8.0 has the following advantages over other implementations: - Applies to any black-box model - Full support for 1-way, 2-way and higher order PD for numerical and categorical variables - Flexible plotting functionality for 1-way and 2-way PD
1
0
0
@SeldonResearch
Seldon Research
3 years
There is an improvement upon PD plots called Accumulated Local Effects (ALE) which take feature correlations into account. This is implemented in Alibi, but only applies to numerical features:
Tweet card summary image
docs.seldon.ai
1
0
0
@SeldonResearch
Seldon Research
3 years
Note that underlying PD computation is the assumption of feature independence (i.e. features are not correlated) which usually does not hold in practice and has to be taken into account when interpreting PD plots.
1
0
0