paper

an archive of posts with this tag

Apr 11, 2026 FlashAttention-4: Algorithm and Kernel Pipelining Co-Design for Asymmetric Hardware Scaling
Apr 09, 2026 FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision
Dec 28, 2024 LoRA vs Full Fine-tuning: An Illusion of Equivalence
Dec 11, 2024 Pretraining Data Detection for Large Language Models: A Divergence-based Calibration Method 설명
Sep 19, 2024 META-REWARDING LANGUAGE MODELS: Self-Improving Alignment with LLM-as-a-Meta-Judge 설명
Nov 19, 2023 What Makes Multi-modal Learning Better than Single (Provably)
Aug 06, 2023 FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning
Jun 28, 2023 TinyViT
Jun 28, 2023 EdgeViT
Jun 21, 2023 Integral Neural Network
Apr 29, 2023 Invariant Representation for Unsupervised Image Restoration
Apr 16, 2023 DINE: Domain Adaptation from Single and Multiple Black-box Predictors
Apr 16, 2023 MobileOne: An Improved One millisecond Mobile Backbone
Apr 08, 2023 Proper Reuse of Image Classification Features Improves Object Detection
Apr 01, 2023 Meta Pseudo Labels
Mar 29, 2023 MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer
Mar 28, 2023 FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Mar 21, 2023 Cross-Domain Adaptive Teacher for Object Detection
Mar 14, 2023 Rethinking “Batch” in BatchNorm
Mar 07, 2023 Convolutional Character Network
Feb 18, 2023 Simple Baselines for Image Restoration
Jan 05, 2023 Bootstrap your own latent
May 09, 2022 FitNet
Jan 27, 2021 [AutoML] NASNet
Jan 10, 2021 학부생이 본 SENet
Jan 03, 2021 [네트워크 경량화] EfficientNet