Sebastian Loeschcke's Avatar

Sebastian Loeschcke

@sloeschcke

Working on Efficient Training, Low-Rank Methods, and Quantization. PhD at the University of Copenhagen ๐Ÿ‡ฉ๐Ÿ‡ฐ Member of @belongielab.org, Danish Data Science Academy, and Pioneer Centre for AI ๐Ÿค– ๐Ÿ”— sebulo.github.io/

371
Followers
202
Following
25
Posts
14.11.2024
Joined
Posts Following

Latest posts by Sebastian Loeschcke @sloeschcke

Post image Post image

A factor of 10 billion since 2010 ๐Ÿ˜ฎ

A couple of eye-opening slides form @sloeschcke.bsky.social's presentation at todayโ€™s @belongielab.org meeting (1/2)

30.01.2026 13:58 ๐Ÿ‘ 184 ๐Ÿ” 39 ๐Ÿ’ฌ 22 ๐Ÿ“Œ 0
Post image

๐Ÿ‡ณ๐Ÿ‡ฑ ๐—ค๐˜‚๐—ฎ๐—น๐—ฐ๐—ผ๐—บ๐—บ ๐—”๐—œ ๐—ฅ๐—ฒ๐˜€๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต ๐—œ๐—ป๐˜๐—ฒ๐—ฟ๐—ป๐˜€๐—ต๐—ถ๐—ฝ ๐Ÿ‡ณ๐Ÿ‡ฑ
Excited to join @qualcomm.bsky.social in Amsterdam as a research intern in the Model Efficiency group, where Iโ€™ll be working on quantization and compression of machine learning models.
Iโ€™ll return to Copenhagen in December to start the final year of my PhD.

13.08.2025 18:42 ๐Ÿ‘ 6 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

๐Ÿ“ฏ Best Paper Award at CVPR workshop on Visual concepts for our (@doneata.bsky.social + @delliott.bsky.social) paper on probing vision/lang/ vision+lang models for semantic norms!

TLDR: SSL vision models (swinV2, dinoV2) are surprisingly similar to LLM & VLMs even w/o lang ๐Ÿ‘€
arxiv.org/abs/2506.03994

13.06.2025 15:15 ๐Ÿ‘ 12 ๐Ÿ” 4 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Preview
TensorGRaD: Tensor Gradient Robust Decomposition for Memory-Efficient Neural Operator Training Scientific problems require resolving multi-scale phenomena across different resolutions and learning solution operators in infinite-dimensional function spaces. Neural operators provide a powerful fr...

Thanks to my co-authors David Pitt, Robert Joseph George, Jiawwei Zhao, Cheng Luo, Yuandong Tian, Jean Kossaifi, @anima-anandkumar.bsky.social, and @caltech.edu for hosting me this spring!
Paper: arxiv.org/abs/2501.02379
Code: github.com/neuraloperat...

03.06.2025 03:16 ๐Ÿ‘ 5 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

We also show strong results on other PDE benchmarks, including ๐ƒ๐š๐ซ๐œ๐ฒ ๐Ÿ๐ฅ๐จ๐ฐ and the ๐๐ฎ๐ซ๐ ๐ž๐ซ๐ฌ equation, demonstrating TensorGRaDโ€™s broad applicability across scientific domains.

03.06.2025 03:16 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

We test TensorGRaD on large-scale Navierโ€“Stokes at 1024ร—1024 resolution with Reynolds number 10e5, a highly turbulent setting. With mixed-precision and 75% optimizer state reduction, it ๐ฆ๐š๐ญ๐œ๐ก๐ž๐ฌ ๐Ÿ๐ฎ๐ฅ๐ฅ-๐ฉ๐ซ๐ž๐œ๐ข๐ฌ๐ข๐จ๐ง ๐€๐๐š๐ฆ while cutting overall memory by up to 50%.

03.06.2025 03:16 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

We also propose a ๐ฆ๐ข๐ฑ๐ž๐-๐ฉ๐ซ๐ž๐œ๐ข๐ฌ๐ข๐จ๐ง ๐ญ๐ซ๐š๐ข๐ง๐ข๐ง๐  strategy with weights, activations, and gradients in half precision and optimizer states in full precision, and empirically show that storing optimizer states in half precision hurts performance.

03.06.2025 03:16 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

We extend low-rank and sparse methods to tensors via a ๐ซ๐จ๐›๐ฎ๐ฌ๐ญ ๐ญ๐ž๐ง๐ฌ๐จ๐ซ ๐๐ž๐œ๐จ๐ฆ๐ฉ๐จ๐ฌ๐ข๐ญ๐ข๐จ๐ง that splits gradients into a low-rank Tucker part and an unstructured sparse tensor. Unlike matricized approaches, we prove our tensor-based method converges.

03.06.2025 03:16 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Recent methods reduce optimizer memory for matrix weights. This includes Low-rank and sparse methods from LLMs that work on matrices. But to use them for Neural Operators, weโ€™d need to flatten tensors, which destroys their spatial/temporal structure and hurts performance.

03.06.2025 03:16 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

These Neural Operators use tensor weights. However, optimizers like Adam store two full tensors per weight, making memory the bottleneck at scale.
TensorGRaD reduces this overhead by up to 75% (๐‘‘๐‘Ž๐‘Ÿ๐‘˜ ๐‘”๐‘Ÿ๐‘’๐‘’๐‘› ๐‘๐‘Ž๐‘Ÿ๐‘ ), without hurting accuracy.

03.06.2025 03:16 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Scientific computing operates on multiscale, multidimensional (๐ญ๐ž๐ง๐ฌ๐จ๐ซ) ๐๐š๐ญ๐š. In weather forecasting, for example, inputs span space, time, and variables. Neural operators can capture these multiscale phenomena by learning an operator that maps between function spaces.

03.06.2025 03:16 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Check out our new preprint ๐“๐ž๐ง๐ฌ๐จ๐ซ๐†๐‘๐š๐ƒ.
We use a robust decomposition of the gradient tensors into low-rank + sparse parts to reduce optimizer memory for Neural Operators by up to ๐Ÿ•๐Ÿ“%, while matching the performance of Adam, even on turbulent Navierโ€“Stokes (Re 10e5).

03.06.2025 03:16 ๐Ÿ‘ 30 ๐Ÿ” 7 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 2
Preview
NeurIPS participation in Europe We seek to understand if there is interest in being able to attend NeurIPS in Europe, i.e. without travelling to San Diego, US. In the following, assume that it is possible to present accepted papers ...

Would you present your next NeurIPS paper in Europe instead of traveling to San Diego (US) if this was an option? Sรธren Hauberg (DTU) and I would love to hear the answer through this poll: (1/6)

30.03.2025 18:04 ๐Ÿ‘ 280 ๐Ÿ” 160 ๐Ÿ’ฌ 6 ๐Ÿ“Œ 12
Post image Post image

Visited the beautiful UC Santa Barbara yesterday.

08.03.2025 17:41 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

Thrilled to announce "Multimodality Helps Few-shot 3D Point Cloud Semantic Segmentation" is accepted as a Spotlight (5%) at #ICLR2025!

Our model MM-FSS leverages 3D, 2D, & text modalities for robust few-shot 3D segmentationโ€”all without extra labeling cost. ๐Ÿคฉ

arxiv.org/pdf/2410.22489

More details๐Ÿ‘‡

11.02.2025 17:49 ๐Ÿ‘ 25 ๐Ÿ” 7 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

While Pasadena will be my home, Iโ€™ll also be making trips to Austin, the Bay Area, and San Diego. If youโ€™re nearby and up for a chat, reach outโ€”letโ€™s meet up!

28.01.2025 15:57 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
View from the office building

View from the office building

โ˜€๏ธ Moved to Pasadena, California! โ˜€๏ธ
For the next five months, Iโ€™ll be a Visiting Student Researcher at Anima Anandkumar's group at Caltech, collaborating with her team and Jean Kossaifi from NVIDIA on Efficient Machine Learning and AI4Science.

28.01.2025 15:57 ๐Ÿ‘ 8 ๐Ÿ” 0 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0
Screenshot of the course website for "SSL4EO: Self-Supervised Learning for Earth Observation"

Screenshot of the course website for "SSL4EO: Self-Supervised Learning for Earth Observation"

Recordings of the SSL4EO-2024 summer school are now released!

This blog post summarizes what has been covered:
langnico.github.io/posts/SSL4EO...

Recordings: www.youtube.com/playlist?lis...

Course website: ankitkariryaa.github.io/ssl4eo/
[1/3]

24.01.2025 15:32 ๐Ÿ‘ 36 ๐Ÿ” 15 ๐Ÿ’ฌ 3 ๐Ÿ“Œ 2

New Starter Pack: Pioneer Centre for AI researchers

16.12.2024 17:42 ๐Ÿ‘ 33 ๐Ÿ” 5 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

Come by our poster session tomorrow!
๐Ÿ—“๏ธ West Ballroom A-D #6104
๐Ÿ•’ Thu, 12 Dec, 4:30 p.m. โ€“ 7:30 p.m. PST
@madstoftrup.bsky.social and I are presenting LoQT: Low-Rank Adapters for Quantized Pretraining: arxiv.org/abs/2405.16528
#Neurips2024

12.12.2024 05:02 ๐Ÿ‘ 7 ๐Ÿ” 3 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image Post image Post image

Copenhagen University and Aarhus University meet-up in Vancouver ๐Ÿ‡ฉ๐Ÿ‡ฐ๐Ÿ‡จ๐Ÿ‡ฆ
#NeurIPS2024

11.12.2024 07:27 ๐Ÿ‘ 6 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

On my way to NeurIPS in Vancouver ๐Ÿ‡จ๐Ÿ‡ฆ
Looking forward to reconnecting with friends and meeting new people. Let me know if you are interested in efficient training, quantization, or grabbing a coffee!
#NeurIPS2024

08.12.2024 04:08 ๐Ÿ‘ 18 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Check out the work our lab in Copenhagen will be presenting at #NeurIPS2024 ๐ŸŒŸ
@neuripsconf.bsky.social @belongielab.org

03.12.2024 11:49 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
Belongie Lab Join the conversation

Hereโ€™s a starter pack with members of our lab that have joined Bluesky

25.11.2024 10:42 ๐Ÿ‘ 13 ๐Ÿ” 4 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image Post image

Pre-NeurIPS Poster Session in Copenhagen.
Thanks to the Pioneer Centre for AI and @ellis.eu for sponsoring.
@neuripsconf.bsky.social
#neurips2024

22.11.2024 19:00 ๐Ÿ‘ 13 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
ELLIS Pre-NeurIPS Fest 2024: Celebrate, Connect, Collaborate The ELLIS mission is to create a diverse European network that promotes research excellence and advances breakthroughs in AI, as well as a pan-European PhD program to educate the next generation of AI...

Check out the ELLIS Pre-NeurIPS Fest event today in...๐Ÿ‡ฉ๐Ÿ‡ฐCopenhagen!

ELLIS Unit Copenhagen is holding their event at the Pioneer Center for AI showcasing #NeurIPS posters and other Denmark-affiliated papers in #AI and #ML.

More info: bit.ly/4fRFrAh

22.11.2024 07:56 ๐Ÿ‘ 24 ๐Ÿ” 5 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 2
A photo of Boulder, Colorado, shot from above the university campus and looking toward the Flatirons.

A photo of Boulder, Colorado, shot from above the university campus and looking toward the Flatirons.

I'm recruiting 1-2 PhD students to work with me at the University of Colorado Boulder! Looking for creative students with interests in #NLP and #CulturalAnalytics.

Boulder is a lovely college town 30 minutes from Denver and 1 hour from Rocky Mountain National Park ๐Ÿ˜Ž

Apply by December 15th!

19.11.2024 10:38 ๐Ÿ‘ 303 ๐Ÿ” 136 ๐Ÿ’ฌ 9 ๐Ÿ“Œ 12

Thanks to the ๐—ฃ๐—ถ๐—ผ๐—ป๐—ฒ๐—ฒ๐—ฟ ๐—–๐—ฒ๐—ป๐˜๐—ฟ๐—ฒ ๐—ณ๐—ผ๐—ฟ ๐—”๐—œ for organizing this event as part of the ๐—˜๐—Ÿ๐—Ÿ๐—œ๐—ฆ ๐—ฃ๐—ฟ๐—ฒ-๐—ก๐—ฒ๐˜‚๐—ฟ๐—œ๐—ฃ๐—ฆ ๐—™๐—ฒ๐˜€๐˜! ๐ŸŽ‰

#NeurIPS2024

19.11.2024 08:58 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Join us for the ๐—ฃ๐—ฟ๐—ฒ-๐—ก๐—ฒ๐˜‚๐—ฟ๐—œ๐—ฃ๐—ฆ ๐—ฃ๐—ผ๐˜€๐˜๐—ฒ๐—ฟ ๐—ฆ๐—ฒ๐˜€๐˜€๐—ถ๐—ผ๐—ป in Copenhagen!
๐Ÿ—“๏ธ ๐—ช๐—ต๐—ฒ๐—ป: 16:00โ€“18:00, Nov. 22, 2024
๐Ÿ“ ๐—ช๐—ต๐—ฒ๐—ฟ๐—ฒ: Entrance Hall, Gefion, ร˜ster Voldgade 10, 1350 Kรธbenhavn K.
Present or explore European contributions to NeurIPS 2024 and connect with colleagues.
๐Ÿ‘‰ ๐—œ๐—ป๐—ณ๐—ผ & ๐˜€๐—ถ๐—ด๐—ป-๐˜‚๐—ฝ: www.aicentre.dk/events/pre-n...

19.11.2024 08:58 ๐Ÿ‘ 12 ๐Ÿ” 3 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

LoQT will be presented at NeurIPS 2024! ๐ŸŽ‰

This research was funded by @DataScienceDK, and @AiCentreDK and is a collaboration between @DIKU_Institut, @ITUkbh, and @csaudk

18.11.2024 09:28 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0