Deprecated : The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456
Papers - Attention - Cross - a matlok Collection
Papers - Attention - Cross updated Dec 25, 2024
Vid2Robot: End-to-end Video-conditioned Policy Learning with
Cross-Attention Transformers Paper
• 2403.12943
• Published Mar 19, 2024 • 15
Masked Audio Generation using a Single Non-Autoregressive Transformer Paper
• 2401.04577
• Published Jan 9, 2024 • 44
Cross-Attention Makes Inference Cumbersome in Text-to-Image Diffusion
Models Paper
• 2404.02747
• Published Apr 3, 2024 • 13
InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image
Generation Paper
• 2404.02733
• Published Apr 3, 2024 • 22
Prompt-to-Prompt Image Editing with Cross Attention Control Paper
• 2208.01626
• Published Aug 2, 2022 • 3
Paper
• 2404.07821
• Published Apr 11, 2024 • 13
HSIDMamba: Exploring Bidirectional State-Space Models for Hyperspectral
Denoising Paper
• 2404.09697
• Published Apr 15, 2024 • 1
TextHawk: Exploring Efficient Fine-Grained Perception of Multimodal
Large Language Models Paper
• 2404.09204
• Published Apr 14, 2024 • 11
Long-form music generation with latent diffusion Paper
• 2404.10301
• Published Apr 16, 2024 • 27
GLIGEN: Open-Set Grounded Text-to-Image Generation Paper
• 2301.07093
• Published Jan 17, 2023 • 4
MultiBooth: Towards Generating All Your Concepts in an Image from Text Paper
• 2404.14239
• Published Apr 22, 2024 • 9
XC-Cache: Cross-Attending to Cached Context for Efficient LLM Inference Paper
• 2404.15420
• Published Apr 23, 2024 • 11
InstantFamily: Masked Attention for Zero-shot Multi-ID Image Generation Paper
• 2404.19427
• Published Apr 30, 2024 • 74
Unveiling Encoder-Free Vision-Language Models Paper
• 2406.11832
• Published Jun 17, 2024 • 55
TokenFormer: Rethinking Transformer Scaling with Tokenized Model
Parameters Paper
• 2410.23168
• Published Oct 30, 2024 • 24
HAT: Hybrid Attention Transformer for Image Restoration Paper
• 2309.05239
• Published Sep 11, 2023 • 1
Byte Latent Transformer: Patches Scale Better Than Tokens Paper
• 2412.09871
• Published Dec 13, 2024 • 108