View in Telegram
Forwarded from the last neural cell (Aleksandr Kovalev)
transformers-october-2024.png
2 MB
[2/4] tasty transformer papers - october 2024 Differential Transformer what: small modification for self attention mechanism. - focuses on the most important information, ignoring unnecessary details. - it does this by subtracting one attention map from another to remove "noise." link: https://arxiv.org/abs/2410.05258 Pixtral-12B what: good multimodal model with simple arch. - Vision Encoder with ROPE-2D: Handles any image resolution/aspect ratio natively. - Break Tokens: Separates image rows for flexible aspect ratios. - Sequence Packing: Batch-processes images with block-diagonal masks, no info “leaks.” link: https://arxiv.org/abs/2410.07073 Fluid: Scaling Autoregressive Text-to-image Generative Models with Continuous Tokens what: maskGIT with continual tokens. - get vae with quantized loss but do not use quantization in decoder ( stable diffusion) - propose BERT-like model to generate in random-order. - ablation shows that bert-like better than gpt-like for images(tbh small improvements) link: https://arxiv.org/abs/2410.13863 UniMTS: Unified Pre-training for Motion Time Series what: one model to handle different device positions, orientations, and activity types. - use graph convolution encoder to work with all devices - contrastive learning with text from LLMs to “get” motion context. - rotation-invariance: doesn’t care about device angle. link: https://arxiv.org/abs/2410.19818 my thoughts I'm really impressed with the Differential Transformer metrics. They made such a simple and clear modification. Basically, they let the neural network find not only the most similar tokens but also the irrelevant ones. Then they subtract one from the other to get exactly what's needed. This approach could really boost brain signal processing. After all, brain activity contains lots of unnecessary information, and filtering it out would be super helpful. So it looks promising. Mistral has really nailed how to build and explain models. Clear, brief, super understandable. They removed everything unnecessary, kept just what's needed, and got better results. The simpler, the better!
Love Center - Dating, Friends & Matches, NY, LA, Dubai, Global
Love Center - Dating, Friends & Matches, NY, LA, Dubai, Global
Find friends or serious relationships easily