Explore tweets tagged as #residual_learning
@Grad62304977
Grad
10 months
43% of the speedup in the new NanoGPT record is due to a variant of value residual learning that I developed. Value residual learning (recently proposed by https://t.co/uQhoUuCk7n) allows all blocks in the transformer to access the values computed by the first block. The paper
Tweet media one
Tweet media two
12
40
390
@vxnuaj
vxnuaj
11 months
DEEP LEARNING DAY 81 (i think) | An Update 🫡 Covered: > "Rethinking the Inception Architecture for Computer Vision" by Szegedy et al. > Implemented InceptionV3. > Began reading "Deep Residual Learning for Image Recognition" by He et al. > Reviewing Calculus: Definition of
Tweet media one
Tweet media two
Tweet media three
Tweet media four
4
16
254
@Almorgand
Alexandre Morgand
4 months
MAC-VO: Metrics-Aware Covariance for Learning-based Stereo Visual Odometry TL;DR: learning-based stereo; learned metrics-aware matching uncertainty for dual purposes: selecting keypoint and weighing the residual in pose graph optimization.
5
48
286
@vxnuaj
vxnuaj
11 months
DEEP LEARNING DAY 73 Time - 8 hrs. Worked on: > FInished "Identity Mappings in Deep Residual Networks" > Implemented ResNet V2 > Begun "Aggregated Residual Transformations for Deep Neural Networks" > Learnt ab Highway Gating Mechanism via Sigmoid Act. > Learnt ab Depthwise
Tweet media one
Tweet media two
Tweet media three
Tweet media four
3
35
480
@vxnuaj
vxnuaj
11 months
DEEP LEARNING DAY 72 (got yesterday's count wrong) Time - 8.6 hrs. > Finished reading "Deep Residual Learning for Image Recognition" > Derived Backpropagation through Residual Connections. > Implemented ResNet > Begun reading "Identity Mappings in Deep Residual Networks" >
Tweet media one
Tweet media two
Tweet media three
Tweet media four
4
25
319
@siyuanhuang95
Siyuan Huang
6 months
📢📢📢 Excited to release ManipTrans: Efficient Dexterous Bimanual Manipulation Transfer via Residual Learning (CVPR25). 🤏🤙✌️With ManipTrans, we can transfer dexterous manipulation skills into robotic hands in simulation and deploy them on a real robot, using a residual policy
1
45
203
@programmingocea
أكاديمية محيط البرمجة لعلوم الذكاء الاصطناعي
17 days
👨‍🔬 Kaiming He – The Architect of Residual Networks ✅ Co-author of ResNet (2016), introducing deep residual learning and revolutionizing computer vision ✅ Developed Faster R-CNN and Mask R-CNN, setting benchmarks for detection and segmentation ✅ Pioneered Focal Loss and Feature
Tweet media one
0
1
1
@itsdannyb_dev
bams
11 months
Reading the "Deep Residual Learning for Image Recognition" paper ☕ #deeplearning
Tweet media one
0
0
0
@vxnuaj
vxnuaj
11 months
DEEP LEARNING DAY 74 🧵 Worked on: > Finished "Aggregated Residual Transformations for Deep Neural Networks" > Implemented ResNeXt > Wrote a Blog Post on ∂'s of Residual Connections > Final Calc I Review: Chain Rule, Implicit Differentiation, Higher Derivatives, Logarithmic
Tweet media one
Tweet media two
Tweet media three
3
17
229
@Bioeng_MDPI
Bioengineering MDPI
13 days
🔬 Excited to share the publication "Lightweight Super-Resolution Techniques in Medical Imaging: Bridging Quality and Computational Efficiency"👉 https://t.co/ZpCcgj0WR6 #medical_imaging #super_resolution #lightweight_model #residual_learning
Tweet media one
0
1
1
@zhanchaozhou
Zhanchao Zhou
3 months
"Value residual learning" was accepted to ACL2025 main conference #ACL2025 ! In the camera-ready version, we: 1) add results for the 1.6B models trained on 200B tokens, and 2) introduce Learnable ResFormer Plus, which significantly outperforms Learnable ResFormer by using
Tweet media one
1
9
60
@Alafiade1
Alafiade
1 year
Computer Vision update: ° Studied Convnet in detail with the help of CS23 resource ° Understood some of the Advanced CNN architectures ° Went further to learn how the VGG and ResNet architecture works ° Read Deep Residual Learning for Image Recognition article
Tweet media one
0
0
0
@victor_explore
Victor
4 months
Dive into ResNet, the neural network architecture behind modern deep learning. This lecture covers: 1️⃣ First and second-order logic in neural networks 2️⃣ Skip connections and their mechanism 3️⃣ Residual learning - learning what to add vs complete transformations 4️⃣ Solving
Tweet media one
1
35
197
@vxnuaj
vxnuaj
10 months
DEEP LEARNING DAY 77 > Finished Implementing DenseNet > Review on Depthwise Separable Convolutions > Wrote Blog Post on Dense Connections vs Residual Learning > Begun reading "MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications" > Calculus Review:
Tweet media one
Tweet media two
Tweet media three
Tweet media four
2
14
184
@eugeneyan
Eugene Yan
2 months
tools for learning and building • zotero: papers, annotations, citations • obsidian: notes, writing, connecting the dots • claude: thinking partner, improve understanding • claude code: implementation, code review (images from work on semantic IDs and residual-quantized
Tweet media one
Tweet media two
Tweet media three
Tweet media four
7
16
161
@_akhaliq
AK
6 months
ManipTrans is out on Hugging Face Efficient Dexterous Bimanual Manipulation Transfer via Residual Learning
5
25
153
@vxnuaj
vxnuaj
10 months
DEEP LEARNING DAY 79. Worked on: > Reading "MobileNetV2: Inverted Residuals and Linear Bottlenecks" > Inverted Residual Blocks > 1x1 Convolutions as Dimensionality Reduction > Understanding the role of ReLU6 > Begun MobileNetV2 Implementation Shorter day. Long tomorrow.
Tweet media one
Tweet media two
2
9
159
@GuanyaShi
Guanya Shi
7 months
ASAP learns diverse, agile, whole-body humanoid motions via learning a residual action model from the real world to align sim and real physics, enabling motions that were previously difficult to achieve. It has two stages: Stage 1 pretrains a phase-based motion tracking policy
Tweet media one
@TairanHe99
Tairan He
7 months
🚀 Can we make a humanoid move like Cristiano Ronaldo, LeBron James and Kobe Byrant? YES! 🤖 Introducing ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills Website: https://t.co/XQga7tIfdw Code: https://t.co/NpEeJtVxpp
5
24
133
@_micmico
Michael Joseph Panaga
8 months
Daily #AI / #MachineLearning Study Log Entry: 102 - CNN Architectures -- LeNet-5 -- AlexNet -- GoogleNet --- Inception module -- VGGNet -- ResNet --- Skip Connection / Shortcut Connection --- Residual Learning Key takeaways: Today, I tried implementing some of the famous CNN
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
0
2
@its_vayishu
Vishal S Pandey
2 months
Residual connections aren’t just skip links. They’re discretized differential equations: This one equation changed deep learning forever. Here’s why ResNets, and what most miss: [1/5]
Tweet media one
6
4
51