<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Article | Tobias Nauen</title><link>https://nauen-it.de/publication_types/article/</link><atom:link href="https://nauen-it.de/publication_types/article/index.xml" rel="self" type="application/rss+xml"/><description>Article</description><generator>HugoBlox Kit (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Mon, 23 Feb 2026 00:00:00 +0000</lastBuildDate><image><url>https://nauen-it.de/media/icon.svg</url><title>Article</title><link>https://nauen-it.de/publication_types/article/</link></image><item><title>When Pretty Isn't Useful: Investigating Why Modern Text-to-Image Models Fail as Reliable Training Data Generators</title><link>https://nauen-it.de/publications/when-pretty-isnt-useful/</link><pubDate>Mon, 23 Feb 2026 00:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/when-pretty-isnt-useful/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;</description></item><item><title>PRISM: Diversifying Dataset Distillation by Decoupling Architectural Priors</title><link>https://nauen-it.de/publications/prism/</link><pubDate>Thu, 13 Nov 2025 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/prism/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;</description></item><item><title>HyperCore: Coreset Selection under Noise via Hypersphere Models</title><link>https://nauen-it.de/publications/hypercore/</link><pubDate>Fri, 26 Sep 2025 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/hypercore/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;</description></item><item><title>SubZeroCore: A Submodular Approach with Zero Training for Coreset Selection</title><link>https://nauen-it.de/publications/subzerocore/</link><pubDate>Fri, 26 Sep 2025 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/subzerocore/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;</description></item><item><title>ForAug: Recombining Foregrounds and Backgrounds to Improve Vision Transformer Training with Bias Mitigation</title><link>https://nauen-it.de/publications/foraug/</link><pubDate>Wed, 12 Mar 2025 00:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/foraug/</guid><description>&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;&lt;img alt="gif"
src="https://nauen-it.de/publications/foraug/images/foraug-gif.gif"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;h1 id="introduction"&gt;Introduction&lt;/h1&gt;
&lt;p&gt;Image classification – teaching computers to label images – is a cornerstone of AI vision, powering everything from medical diagnosis to autonomous driving.
Datasets like ImageNet have been crucial, especially with the rise of powerful models like Vision Transformers (ViTs).&lt;/p&gt;
&lt;p&gt;However, unlike older Convolutional Neural Networks (CNNs), ViTs don&amp;rsquo;t inherently understand that an object remains the same regardless of its position in an image (they lack &amp;ldquo;translation equivariance&amp;rdquo;).
Standard data augmentation techniques (like flipping or cropping) help, but they weren&amp;rsquo;t specifically designed for this trait of ViTs.&lt;/p&gt;
&lt;p&gt;To tackle these problems, we propose &lt;strong&gt;ForAug&lt;/strong&gt;, a novel data augmentation for ViTs.
The core idea?
Make the spatial relationships explicit in the training data.
ForAug achieves this by:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;Separating foreground objects from their backgrounds in the dataset.&lt;/li&gt;
&lt;li&gt;Recombining these objects with different backgrounds on-the-fly during training.&lt;/li&gt;
&lt;li&gt;Controlling the object&amp;rsquo;s size and position during this recombination.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;The results?
Training ViTs with ForAug instead of standard ImageNet boosts accuracy by up to 4.5 percentage points on ImageNet classification and significantly cuts error rates (up to 39.3% reduction) on downstream tasks.&lt;/p&gt;
&lt;p&gt;Furthermore, ForAug provides powerful new ways to analyze model biases.
Researchers can now precisely measure:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Background Robustness: How much does the background influence the prediction?&lt;/li&gt;
&lt;li&gt;Foreground Focus: Does the model correctly focus on the main object?&lt;/li&gt;
&lt;li&gt;Center &amp;amp; Size Bias: Is the model overly reliant on objects being centered or a specific size?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Training with ForAug demonstrably reduces these biases, leading to more robust models.&lt;/p&gt;
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;This post is just a short overview over ForAug. For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;h1 id="foraug-method"&gt;ForAug (Method)&lt;/h1&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="ForAug Flowchart"
srcset="https://nauen-it.de/publications/foraug/images/fig-2_hu_2f0955c903beee0e.webp 320w, https://nauen-it.de/publications/foraug/images/fig-2_hu_5313138d6618e3b9.webp 480w, https://nauen-it.de/publications/foraug/images/fig-2_hu_70ceae1cbeb25b52.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/fig-2_hu_2f0955c903beee0e.webp"
width="760"
height="291"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
So, how does ForAug actually build these dynamic training images?
The process, visualized above, involves two main stages: an offline &lt;strong&gt;Segmentation&lt;/strong&gt; stage and an online &lt;strong&gt;Recombination&lt;/strong&gt; stage.&lt;/p&gt;
&lt;h2 id="segmentation"&gt;Segmentation&lt;/h2&gt;
&lt;p&gt;The process kicks off with the Segmentation stage, a one-time, offline preparation step performed before model training even begins.
Think of it as carefully prepping the visual ingredients.
Here, we leverage the state-of-the-art Grounded SAM segmentation model, guiding it with the known class label of each image (e.g., instructing it to specifically find the &amp;lsquo;golden retriever&amp;rsquo;) to precisely isolate the main subject.
Once the foreground object is digitally &amp;lsquo;cut out&amp;rsquo;, an object removal or &amp;lsquo;inpainting&amp;rsquo; model intelligently fills the resulting hole in the original background, ensuring the backdrop looks natural and plausible.
Crucially, not all generated assets make the cut; a filtering step employs other pre-trained AI models to assess quality.
This ensures only clearly defined foregrounds and clean backgrounds – ones that don&amp;rsquo;t inadvertently give away the object&amp;rsquo;s identity or look overly artificial – are selected.
This meticulous preparation yields the core assets for ForAug: a collection of ready-to-use foreground objects (with transparency) and a diverse pool of cleaned-up backgrounds.&lt;/p&gt;
&lt;h2 id="recombination"&gt;Recombination&lt;/h2&gt;
&lt;p&gt;With the assets prepared, the real action unfolds during the Recombination stage, which happens dynamically online while the Vision Transformer is training.
This is where ForAug truly comes alive, creating new training examples on the fly.
For every foreground object prepared in the first stage, the system randomly selects a background to pair it with.
This background might be the object&amp;rsquo;s original one, perhaps one from another image belonging to the same object class, or even a completely unrelated background drawn from the entire dataset to maximize contextual variety.
The chosen foreground object is then randomly resized (within sensible limits based on its original appearance) and placed at a random position onto this background canvas.
To create a more seamless integration, a subtle smoothing effect is applied to the object&amp;rsquo;s edges where it meets the new background.
Only after this dynamic composition is complete does the resulting image undergo the standard data augmentation techniques commonly used in AI training, like random color shifts or minor flips.
This constant mixing-and-matching means that each time the AI cycles through the training data, it encounters familiar objects in entirely new visual contexts.
This directly forces the ViT to learn robust features that identify the object itself, effectively teaching it the spatial invariance that doesn&amp;rsquo;t come built-in, by demonstrating repeatedly that appearance, not specific placement or background, is what defines the object.&lt;/p&gt;
&lt;h1 id="experiments"&gt;Experiments&lt;/h1&gt;
&lt;h2 id="image-classification-results"&gt;Image Classification Results&lt;/h2&gt;
&lt;p&gt;We compare training on ImageNet with and without ForAug for 10 different models:
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="ImageNet results"
srcset="https://nauen-it.de/publications/foraug/images/foraug-imagenet-results_hu_d7381cc18cb8227f.webp 320w, https://nauen-it.de/publications/foraug/images/foraug-imagenet-results_hu_9a346107ce9df01c.webp 480w, https://nauen-it.de/publications/foraug/images/foraug-imagenet-results_hu_851c625e34531f5c.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/foraug-imagenet-results_hu_d7381cc18cb8227f.webp"
width="760"
height="647"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
We find that training with ForAug increases the accuracy of every model by up to 4.5%.
It also combats the overfitting problem of larger models.&lt;/p&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Downstream Results"
srcset="https://nauen-it.de/publications/foraug/images/foraug-downstream-results_hu_bc13a772bd62ba95.webp 320w, https://nauen-it.de/publications/foraug/images/foraug-downstream-results_hu_4093f7149dcda2cb.webp 480w, https://nauen-it.de/publications/foraug/images/foraug-downstream-results_hu_67cdcd2209a42509.webp 671w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/foraug-downstream-results_hu_bc13a772bd62ba95.webp"
width="671"
height="760"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
When finetuning these models on 5 fine-grained down-stream datasets, we find that the ForAug-pretrained models consistently outperform the ImageNet-pretrained ones.
Especially when looking at the transformer-based models.&lt;/p&gt;
&lt;h2 id="model-robustness"&gt;Model Robustness&lt;/h2&gt;
&lt;p&gt;We also evaluate multiple robustness metrics.&lt;/p&gt;
&lt;h3 id="background-robustness"&gt;Background Robustness&lt;/h3&gt;
&lt;p&gt;We check the background robustness of models, by inspecting the accuracy-change when evaluating with ForAug using backgrounds from the &lt;em&gt;same&lt;/em&gt; class compared to backgrounds from &lt;em&gt;all&lt;/em&gt; classes:&lt;/p&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Background Robustness Scores"
srcset="https://nauen-it.de/publications/foraug/images/foraug-background-robustness_hu_2489c03de26af797.webp 320w, https://nauen-it.de/publications/foraug/images/foraug-background-robustness_hu_dc3b8b16d1a5fc3a.webp 480w, https://nauen-it.de/publications/foraug/images/foraug-background-robustness_hu_289f170ed90e6dd3.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/foraug-background-robustness_hu_2489c03de26af797.webp"
width="760"
height="224"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
Training with ForAug reduces the &lt;em&gt;Background Gap&lt;/em&gt; for all transformer models.&lt;/p&gt;
&lt;h3 id="foreground-focus"&gt;Foreground Focus&lt;/h3&gt;
&lt;p&gt;Since we have the foreground segmentation masks, we can also investigate the foreground focus of the trained models.
For this, we utilize different input-importance metrics like GradCAM and IntegratedGradients (IG).
We define a models foreground focus, by how much more it focuses on the foreground object compared to a uniform distribution:
&lt;/p&gt;
$$
\text{FG Focus}(M; \text{img}) = \frac{\text{Area}(\text{img}) \hspace{5pt} \text{Importance}_M(\text{fg})}{\text{Area}(\text{fg}) \hspace{5pt} \text{Importance}_M(\text{img})}
$$&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Foreground Focus Scores"
srcset="https://nauen-it.de/publications/foraug/images/foraug-foreground-focus_hu_1afeb71ecf915471.webp 320w, https://nauen-it.de/publications/foraug/images/foraug-foreground-focus_hu_4efb61ce3baf1f8b.webp 480w, https://nauen-it.de/publications/foraug/images/foraug-foreground-focus_hu_467c7e1ebb906659.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/foraug-foreground-focus_hu_1afeb71ecf915471.webp"
width="760"
height="204"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
We find that training with ForAug mostly significantly improves the foreground focus of all models.&lt;/p&gt;
&lt;h3 id="center-bias"&gt;Center Bias&lt;/h3&gt;
&lt;p&gt;Since we can freely change the object&amp;rsquo;s position and size, we can evaluate the model bias when the position changes.
For this, we subdivide the image into $3 \times 3$ sections (nonants) and place each object only in one nonant.
We then compare the accuracy of a model when an object is in a specific nonant to when it&amp;rsquo;s in the center nonant.&lt;/p&gt;
&lt;p&gt;Our center-bias score is defined at the mean of (1) the worst accuracy in a corner and (2) the worst accuracy on an edge, relative to the accuracy in the center.
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Center Bias Table"
srcset="https://nauen-it.de/publications/foraug/images/foraug-center-bias_hu_4292720d3de7ff2c.webp 320w, https://nauen-it.de/publications/foraug/images/foraug-center-bias_hu_86b74bbd53280813.webp 480w, https://nauen-it.de/publications/foraug/images/foraug-center-bias_hu_1c2231712f30f607.webp 483w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/foraug-center-bias_hu_4292720d3de7ff2c.webp"
width="483"
height="760"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;p&gt;We visualize the center bias for 3 instantiations of each model.
Training with ForAug significantly reduces the center bias; especially of larger transformers.
We also find that when training on ImageNet, model consistently perform better when an object is on the right side of an image compared to the left side (even though we use 50% random flipping during training of all models).&lt;/p&gt;
&lt;h3 id="size-bias"&gt;Size Bias&lt;/h3&gt;
&lt;p&gt;We vary the object size by an additional factor of $f_\text{size}$ to see how the model accuracy changes relative to $f_\text{size} = 1$.
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Size Bias Plot"
srcset="https://nauen-it.de/publications/foraug/images/foraug-size-bias_hu_6d10bfb297ada5cb.webp 320w, https://nauen-it.de/publications/foraug/images/foraug-size-bias_hu_7f5b922edf83b92b.webp 480w, https://nauen-it.de/publications/foraug/images/foraug-size-bias_hu_13974a67c0bc523a.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/foraug/images/foraug-size-bias_hu_6d10bfb297ada5cb.webp"
width="760"
height="377"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;p&gt;Using ForAug significantly reduces the accuracy drop-off when going towards smaller objects.
These gains come on top off the overall better accuracy (at $f_\text{size} = 1$).&lt;/p&gt;
&lt;h1 id="conclusion"&gt;Conclusion&lt;/h1&gt;
&lt;p&gt;So, what&amp;rsquo;s the big takeaway from ForAug?
This research introduces a genuinely novel data augmentation scheme designed specifically to enhance how Vision Transformers learn to classify images.
By cleverly separating foreground objects from their backgrounds and dynamically recombining them during training, ForAug tackles a key characteristic of Transformer models head-on.&lt;/p&gt;
&lt;p&gt;As the results clearly demonstrate, this dynamic approach pays off significantly.
Training models with ForAug leads to substantial performance boosts on the standard ImageNet benchmark and translates to impressive gains on related fine-grained classification tasks downstream.&lt;/p&gt;
&lt;p&gt;But the impact of ForAug extends beyond just improving accuracy scores.
It also provides a powerful and much-needed framework for analyzing model behavior and uncovering hidden biases.
Crucially, the experiments show that training with ForAug doesn&amp;rsquo;t just highlight these biases – it actively reduces them.
This results in models that are not only more accurate but also more robust, reliable, and generalizable to varied real-world conditions.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Associated Projects:&lt;/strong&gt;
,
,
&lt;/p&gt;</description></item><item><title>A Study in Dataset Distillation for Image Super-Resolution</title><link>https://nauen-it.de/publications/dataset-distillation-sr/</link><pubDate>Wed, 05 Feb 2025 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/dataset-distillation-sr/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Associated Projects:&lt;/strong&gt;
,
&lt;/p&gt;</description></item><item><title>Distill the Best, Ignore the Rest: Improving Dataset Distillation with Loss-Value-Based Pruning</title><link>https://nauen-it.de/publications/distill-best-ignore-rest/</link><pubDate>Mon, 18 Nov 2024 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/distill-best-ignore-rest/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Associated Projects:&lt;/strong&gt;
,
,
&lt;/p&gt;</description></item><item><title>Just Leaf It: Accelerating Diffusion Classifiers with Hierarchical Class Pruning</title><link>https://nauen-it.de/publications/just-leaf-it/</link><pubDate>Mon, 18 Nov 2024 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/just-leaf-it/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Associated Projects:&lt;/strong&gt;
,
,
&lt;/p&gt;</description></item><item><title>A Low-Resolution Image is Worth 1x1 Words: Enabling Fine Image Super-Resolution with Transformers and TaylorShift</title><link>https://nauen-it.de/publications/taylor-shift-super-resolution/</link><pubDate>Fri, 15 Nov 2024 13:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/taylor-shift-super-resolution/</guid><description>
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;This work builds on the
attention mechanism.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;div class="callout flex items-baseline gap-2 px-3 py-2 mb-4 rounded-md border-l-4 bg-primary-50 dark:bg-primary-900/30 border-primary-500"
data-callout="note"
data-callout-metadata=""&gt;
&lt;span class="callout-icon shrink-0 translate-y-0.5 text-primary-600 dark:text-primary-400"&gt;
&lt;svg height="20" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24"&gt;&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/&gt;&lt;/svg&gt;
&lt;/span&gt;
&lt;div class="callout-content text-base dark:text-neutral-300"&gt;
&lt;div class="callout-body"&gt;&lt;p&gt;For more information, see the
.&lt;/p&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Associated Projects:&lt;/strong&gt;
,
,
&lt;/p&gt;</description></item></channel></rss>