Tag Index

 LLM (4) Multimodality (2) charts and graphs (1) efficiency (2) image2text (2) long input transformers (4) test (1)

 LLM (4)

Paper Review 6: Mixtral of Experts
Paper Review 5: Mistral 7B
Paper Review 4: Self-attention Does Not Need O(n^2) Memory
Paper Review 1: LLaMA - Open and Efficient Foundation Language Models

 Multimodality (2)

Paper Review 3: Pix2Struct is an image-encoder-text-decoder based on the Vision Transformer (ViT)
Paper Review 2: MATCHA : Enhancing Visual Language Pretraining with Math Reasoning and Chart Derendering

 charts and graphs (1)

Paper Review 2: MATCHA : Enhancing Visual Language Pretraining with Math Reasoning and Chart Derendering

 efficiency (2)

Paper Review 7: FlashAttention - Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper Review 4: Self-attention Does Not Need O(n^2) Memory

 image2text (2)

Paper Review 3: Pix2Struct is an image-encoder-text-decoder based on the Vision Transformer (ViT)
Paper Review 2: MATCHA : Enhancing Visual Language Pretraining with Math Reasoning and Chart Derendering

 long input transformers (4)

Paper Review 7: FlashAttention - Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper Review 6: Mixtral of Experts
Paper Review 5: Mistral 7B
Paper Review 4: Self-attention Does Not Need O(n^2) Memory

 test (1)

Sample blog post to learn markdown tips