_danielmas Profile Banner
Daniel Mas Montserrat Profile
Daniel Mas Montserrat

@_danielmas

Followers
495
Following
4K
Media
37
Statuses
173

Building AI at @GalateaBio @Stanford đź§®

Joined March 2018
Don't wanna be here? Send us removal request.
@_danielmas
Daniel Mas Montserrat
19 days
Excited to share our preprint on iLTM: an Integrated Large Tabular Model! arxiv: https://t.co/oTGRYGy3wy No single technique consistently excels across all tabular tasks. iLTM addresses this by integrating distinct paradigms in a single architecture: - Gradient Boosted
1
6
9
@_danielmas
Daniel Mas Montserrat
19 days
Despite being meta-trained exclusively on classification, iLTM transfers effectively to regression tasks with light fine-tuning, matching or surpassing strong baselines on both tasks. iLTM achieves top rankings on TabZilla Hard, TabReD, and more benchmarks, outperforming
1
1
4
@_danielmas
Daniel Mas Montserrat
19 days
Excited to share our preprint on iLTM: an Integrated Large Tabular Model! arxiv: https://t.co/oTGRYGy3wy No single technique consistently excels across all tabular tasks. iLTM addresses this by integrating distinct paradigms in a single architecture: - Gradient Boosted
1
6
9
@_danielmas
Daniel Mas Montserrat
19 days
We’re releasing code + pre-trained weights so anyone working with large-scale tabular data can get stronger baselines and build on iLTM. We’d love feedback and comparisons on your own datasets: Paper: https://t.co/oTGRYGy3wy Code: https://t.co/WLMhfKr73y Weights:
Tweet card summary image
github.com
iLTM: Integrated Large Tabular Model. Contribute to AI-sandbox/iLTM development by creating an account on GitHub.
0
0
5
@_danielmas
Daniel Mas Montserrat
19 days
From small tables to real industry-grade datasets with >1M rows and >10k features, our benchmarks show how iLTM scales across sizes. In our labs at @Stanford and @UCSC, we’re already exploring applications of iLTM to genomic data, where dimensionality is even higher. (5/N)
1
0
4
@_danielmas
Daniel Mas Montserrat
19 days
Unlike standard neural network, iLTM's hypernetwork explicitly parametrizes the relationship between dataset features and network weights. This allows us to visualize how the model “understands” different tasks, for example, clustering similar datasets and learning diverse
1
0
4
@_danielmas
Daniel Mas Montserrat
19 days
Despite being meta-trained exclusively on classification, iLTM transfers effectively to regression tasks with light fine-tuning, matching or surpassing strong baselines on both tasks. iLTM achieves top rankings on TabZilla Hard, TabReD, and more benchmarks, outperforming
1
1
4
@_danielmas
Daniel Mas Montserrat
19 days
iLTM is meta-trained on >1,800 heterogeneous real-world classification datasets. Instead of training from scratch for every new table, the hypernetwork learns to generate dataset-specific weights of a neural network, which can then be optionally fine-tuned and ensembled. (2/N)
1
0
5
@predict_addict
Valeriy M., PhD, MBA, CQF
9 months
How was this paper even accepted to ICLR? The commercial promoters of TabPFN are now trying to discredit one of the best open repositories, OpenML. Utterly unacceptable, how did this paper pass ethics board at ICLR?
1
3
7
@arturolp
Arturo
2 years
Excited to share our latest PRS work! Our @GalateaBio and @genomelink team performed a comprehensive analysis of published @PGSCatalog models along with locally trained models using LDPred2, PRS-CSx, and SNPnet, across diverse populations using @UKBIOBANK and our own data
@medrxivpreprint
medRxiv
2 years
Polygenic risk score portability for common diseases across genetically diverse populations https://t.co/Mgw0jjT7Tf #medRxiv
0
3
8
@medrxivpreprint
medRxiv
2 years
Polygenic risk score portability for common diseases across genetically diverse populations https://t.co/Mgw0jjT7Tf #medRxiv
0
3
7
@ykilcher
Yannic Kilcher 🇸🇨
2 years
No son of a construction worker is just going to randomly start doing ML research if they never hear of it and don't get told that it could be important for their future career, no matter how intelligent the kid is
16
9
276
@_danielmas
Daniel Mas Montserrat
2 years
Introducing "HyperFast: Instant Classification for Tabular Data" at @RealAAAI, which received the Best Paper Award at @NeurIPSConf Table rep. workshop @TrlWorkshop! We provide easy-to-use sklearn-like code: https://t.co/qrMF6XStAA Some insights of the work below 👇🧵(1/N)
Tweet card summary image
github.com
HyperFast : Instant Classification for Tabular Data - AI-sandbox/HyperFast
1
3
14
@_danielmas
Daniel Mas Montserrat
2 years
This work has been led by @d_bonet with the supervision of @DocXavi and @alexGioannidis! Code available at: https://t.co/qrMF6XStAA #AI #AAAI #AAAI2024 #NeurIPS2023 #NeurIPS (5/5)
0
2
4
@_danielmas
Daniel Mas Montserrat
2 years
Hyperfast provides competitive results in several tabular classification datasets, even matching boosting-tree-based accuracies! While still far from solving tabular data classification, we believe Hyperfast provides a step forward in NN-based tabular applications! (4/N)
1
2
5
@_danielmas
Daniel Mas Montserrat
2 years
Hyperfast provides multiple mechanisms to scale to both large and high-dimensional datasets and can be easily applied to real-world applications! (3/N)
1
1
4
@_danielmas
Daniel Mas Montserrat
2 years
Hyperfast replaces the slow process of training MLPs with gradient-based methods (e.g. Adam) with a fast hypernetwork that directly predicts the weights of the MLP. The generated MLP typically matches (or even surpasses) the accuracy of those trained with gradient descent. (2/N)
1
1
6
@_danielmas
Daniel Mas Montserrat
2 years
This work has been led by @d_bonet with the supervision of @DocXavi and @alexGioannidis! Code available at: https://t.co/qrMF6XStAA #AI #AAAI #AAAI2024 #NeurIPS2023 #NeurIPS (5/5)
0
2
4
@_danielmas
Daniel Mas Montserrat
2 years
Hyperfast provides competitive results in several tabular classification datasets, even matching boosting-tree-based accuracies! While still far from solving tabular data classification, we believe Hyperfast provides a step forward in NN-based tabular applications! (4/N)
1
2
5
@_danielmas
Daniel Mas Montserrat
2 years
Hyperfast provides multiple mechanisms to scale to both large and high-dimensional datasets and can be easily applied to real-world applications! (3/N)
1
1
4