Train Scene Graph Generation for Visual Genome and GQA in PyTorch >= 1.2 with improved zero and few-shot generalization [BMVC 2020, ICCV 2021]
-
Updated
Jun 18, 2023 - Jupyter Notebook
Train Scene Graph Generation for Visual Genome and GQA in PyTorch >= 1.2 with improved zero and few-shot generalization [BMVC 2020, ICCV 2021]
Decoding Attention is specially optimized for MHA, MQA, GQA and MLA using CUDA core for the decoding stage of LLM inference.
Predicting a subgraph alongside the answer in a graph based VQA model
Vision-Language, Solve GQA(Visual Reasoning in the Real World) dataset.
A RAG-based question-answering system that processes user queries using local documents. It extracts relevant information to answer questions, falling back to a large language model when local sources are insufficient, ensuring accurate and contextual responses.
249M-param MoE transformer built from scratch in PyTorch. GQA, RoPE, SwiGLU, sparse MoE with 3 aux losses, AMP training loop no Trainer abstractions. Architecture mirrors LLaMA/Mistral/Mixtral decisions, fully inspectable.
LaTeX files for my honours thesis: "Graph Attention Networks for Compositional Visual Question Answering"
PyTorch implementation of the Transformer architecture (“Attention Is All You Need”) for English–Italian text-to-text translation, featuring encoder–decoder layers, multi-head attention, and training on a subset of the OPUS Books dataset.
Source code for my honours thesis: "Graph Attention Networks for Compositional Visual Question Answering"
A code deep-dive on one of the key innovations from Deepseek - Multihead Latent Attention (MLA)
This repository contains an implementation of Group Query Attention (GQA), an efficient variant of multi-head attention used in modern transformer models like LLaMA.
A toolkit for vision-language processing to support the increasing popularity of mulit-modal transformer-based models
Simple Llama architecture LLM in pytorch
Reference Flash Attention implementation in PyTorch with V1/V2, GQA/MQA, Triton kernels, benchmark and docs.
Modern LLM Attention from Scratch — MHA, GQA, MQA, RoPE, and KV-Cache implemented in pure PyTorch.
Add a description, image, and links to the gqa topic page so that developers can more easily learn about it.
To associate your repository with the gqa topic, visit your repo's landing page and select "manage topics."