VISION TRANSFORMERS NEED REGISTERS | #ai #2024 #genai #meta
AI Today30 Joulu 2024

VISION TRANSFORMERS NEED REGISTERS | #ai #2024 #genai #meta

Paper: https://arxiv.org/pdf/2309.16588 This research paper examines artifacts in vision transformer feature maps, specifically high-norm tokens appearing in non-informative image areas. The authors propose adding "register" tokens to the input sequence as a solution. This simple addition eliminates the artifacts, improves performance on dense prediction tasks and object discovery, and results in smoother feature and attention maps. The findings apply to both supervised and self-supervised vision transformer models, significantly enhancing their interpretability and effectiveness. Experiments across various models and tasks validate the approach's efficacy and generalizability. ai , artificial intelligence , arxiv , research , paper , publication , llm, genai, generative ai , large visual models, large language models, large multi modal models, nlp, text, machine learning, ml, nividia, openai, anthropic, microsoft, google, technology, cutting-edge, meta, llama, chatgpt, gpt, elon musk, sam altman, deployment, engineering, scholar, science, apple, samsung, anthropic, turing

Jaksot(30)

SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with MotionAware Mem | #2024

SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with MotionAware Mem | #2024

Paper: https://arxiv.org/pdf/2411.11922 Github: https://github.com/yangchris11/samurai Blog: https://yangchris11.github.io/samurai/ The paper introduces SAMURAI, a novel visual object tracking method...

27 Marras 202414min

Adding Error Bars to Evals: A Statistical Approach to LM Evaluations | #llm #genai #anthropic #2024

Adding Error Bars to Evals: A Statistical Approach to LM Evaluations | #llm #genai #anthropic #2024

Github: https://arxiv.org/pdf/2411.00640 This research paper advocates for incorporating rigorous statistical methods into the evaluation of large language models (LLMs). It introduces formulas for c...

27 Marras 202414min

Marco-o1: Towards Open Reasoning Models for Open-Ended Solutions | #ai #llm #alibaba #genai #2024

Marco-o1: Towards Open Reasoning Models for Open-Ended Solutions | #ai #llm #alibaba #genai #2024

Paper: https://arxiv.org/pdf/2411.14405 Github: https://github.com/AIDC-AI/Marco-o1 The Alibaba MarcoPolo team introduces Marco-o1, a large reasoning model designed to excel in open-ended problem-sol...

27 Marras 202414min

FLUX.I TOOLS | #ai #computervision #cv #BlackForestLabs #2024

FLUX.I TOOLS | #ai #computervision #cv #BlackForestLabs #2024

Github: https://github.com/black-forest-labs/... Black Forest Labs announced FLUX.1 Tools, a suite of four open-access and API-based models enhancing their FLUX.1 text-to-image model. FLUX.1 Fill exc...

27 Marras 202414min

Tülu 3 opens language model post-training up to more tasks and more people | #ai #llm #allenai #2024

Tülu 3 opens language model post-training up to more tasks and more people | #ai #llm #allenai #2024

Blog: https://allenai.org/blog/tulu-3 Summary The Allen Institute for Artificial Intelligence (Ai2) has released Tülu 3, an open-source family of post-trained language models. Unlike closed models fr...

27 Marras 202414min

Multimodal Autoregressive Pre-training of Large Vision Encoders | #ai #computervision #apple #2024

Multimodal Autoregressive Pre-training of Large Vision Encoders | #ai #computervision #apple #2024

Paper: https://arxiv.org/pdf/2411.14402 Github Link: https://github.com/apple/ml-aim This research introduces AIMV2, a family of large-scale vision encoders pre-trained using a novel multimodal auto...

27 Marras 202414min