Awesome Open Source
Search
Programming Languages
Languages
All Categories
Categories
About
Search results for attention mechanism
attention-mechanism
x
649 search results found
Vit Pytorch
⭐
16,298
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Ml Nlp
⭐
10,874
此项目是机器学习(Machine Learning)、深度学习(Deep Learning)、NLP面试中常考到的知识点和代码实现,也是作为一个算法工程师必会的理论基础知识。
Rwkv Lm
⭐
10,705
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Text_classification
⭐
7,628
all kinds of text classification models and more with deep learning
Palm Rlhf Pytorch
⭐
7,496
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
Dalle Pytorch
⭐
5,477
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
Awesome Graph Classification
⭐
4,703
A collection of important graph embedding, classification and representation learning papers with implementations.
Awesome Transformer Attention
⭐
3,895
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Transformer
⭐
3,882
A TensorFlow Implementation of the Transformer: Attention Is All You Need
X Transformers
⭐
3,840
A simple but complete full-attention transformer with a set of promising experimental features from various papers
Awesome Speech Recognition Speech Synthesis Papers
⭐
2,869
Automatic Speech Recognition (ASR), Speaker Verification, Speech Synthesis, Text-to-Speech (TTS), Language Modelling, Singing Voice Synthesis (SVS), Voice Conversion (VC)
Keras Attention
⭐
2,794
Keras Attention Layer (Luong and Bahdanau scores).
Musiclm Pytorch
⭐
2,686
Implementation of MusicLM, Google's new SOTA model for music generation using attention networks, in Pytorch
Audiolm Pytorch
⭐
2,112
Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Research, in Pytorch
A Pytorch Tutorial To Image Captioning
⭐
2,084
Show, Attend, and Tell | a PyTorch Tutorial to Image Captioning
Gat
⭐
2,078
Graph Attention Networks (https://arxiv.org/abs/1710.10903)
Reformer Pytorch
⭐
1,917
Reformer, the efficient Transformer, in Pytorch
Pytorch Gat
⭐
1,815
My implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Toolformer Pytorch
⭐
1,788
Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI
Pygat
⭐
1,684
Pytorch implementation of the Graph Attention Network model by Veličković et. al (2017, https://arxiv.org/abs/1710.10903)
Make A Video Pytorch
⭐
1,449
Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch
Hopfield Layers
⭐
1,258
Hopfield Networks is All You Need
Whisper Timestamped
⭐
1,217
Multilingual Automatic Speech Recognition with word-level timestamps and confidence
Sockeye
⭐
1,190
Sequence-to-sequence framework with a focus on Neural Machine Translation based on PyTorch
Lambda Networks
⭐
1,110
Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less compute
Alphafold2
⭐
1,086
To eventually become an unofficial Pytorch implementation / replication of Alphafold2, as details of the architecture get released
Soundstorm Pytorch
⭐
1,054
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch
Awesome Transformer Nlp
⭐
1,022
A curated list of NLP resources focused on Transformer networks, attention mechanism, GPT, BERT, ChatGPT, LLMs, and transfer learning.
Textclassifier
⭐
1,003
Text classifier for Hierarchical Attention Networks for Document Classification
Perceiver Pytorch
⭐
980
Implementation of Perceiver, General Perception with Iterative Attention, in Pytorch
Show Attend And Tell
⭐
907
TensorFlow Implementation of "Show, Attend and Tell"
Coca Pytorch
⭐
900
Implementation of CoCa, Contrastive Captioners are Image-Text Foundation Models, in Pytorch
Pointer_summarizer
⭐
859
pytorch implementation of "Get To The Point: Summarization with Pointer-Generator Networks"
Chatbot_cn
⭐
850
基于金融-司法领域(兼有闲聊性质)的聊天机器人,其中的主要模块有信息抽取、NLU、NLG、知识图谱等
Retro Pytorch
⭐
784
Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch
Performer Pytorch
⭐
777
An implementation of Performer, a linear attention-based transformer, in Pytorch
Muse Maskgit Pytorch
⭐
739
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch
Yolo_research
⭐
715
based on yolo-high-level project (detect\pose\classify\segment\):include yolov5\yolov7\yolov8\ core ,improvement research ,SwintransformV2 and Attention Series. training skills, business customization, engineering deployment C
Tf Rnn Attention
⭐
703
Tensorflow implementation of attention mechanism for text classification tasks.
Awesome Attention Mechanism In Cv
⭐
686
Awesome List of Attention Modules and Plug&Play Modules in Computer Vision
Phenaki Pytorch
⭐
674
Implementation of Phenaki Video, which uses Mask GIT to produce text guided videos of up to 2 minutes in length, in Pytorch
Keras Attention
⭐
656
Visualizing RNNs using the attention mechanism
Seq2seq Pytorch
⭐
653
Sequence to Sequence Models with PyTorch
Longnet
⭐
613
Implementation of plug in and play Attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens"
Tab Transformer Pytorch
⭐
609
Implementation of TabTransformer, attention network for tabular data, in Pytorch
Transformer Tts
⭐
599
A Pytorch Implementation of "Neural Speech Synthesis with Transformer Network"
Moran_v2
⭐
593
MORAN: A Multi-Object Rectified Attention Network for Scene Text Recognition
Text Summarization Tensorflow
⭐
586
Tensorflow seq2seq Implementation of Text Summarization.
Keras Self Attention
⭐
570
Attention mechanism for processing sequential data that considers the context for each timestamp.
Eeg Dl
⭐
563
A Deep Learning library for EEG Tasks (Signals) Classification, based on TensorFlow.
Memorizing Transformers Pytorch
⭐
556
Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and retrieval of memories using approximate nearest neighbors, in Pytorch
Self Attention Cv
⭐
550
Implementation of various self-attention mechanisms focused on computer vision. Ongoing repository.
Flamingo Pytorch
⭐
549
Implementation of 🦩 Flamingo, state-of-the-art few-shot visual question answering attention net out of Deepmind, in Pytorch
Simgnn
⭐
540
A PyTorch implementation of "SimGNN: A Neural Network Approach to Fast Graph Similarity Computation" (WSDM 2019).
Bottleneck Transformer Pytorch
⭐
523
Implementation of Bottleneck Transformer in Pytorch
Deeplearning.ai Natural Language Processing Specialization
⭐
523
This repository contains my full work and notes on Coursera's NLP Specialization (Natural Language Processing) taught by the instructor Younes Bensouda Mourri and Łukasz Kaiser offered by deeplearning.ai
Nmt Keras
⭐
514
Neural Machine Translation with Keras
Transformer
⭐
505
A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
Wama_modules
⭐
501
A PyTorch Computer Vision (CV) module library for building n-D networks flexibly ~
Parti Pytorch
⭐
487
Implementation of Parti, Google's pure attention-based text-to-image neural network, in Pytorch
Timesformer Pytorch
⭐
483
Implementation of TimeSformer from Facebook AI, a pure attention-based solution for video classification
Nuwa Pytorch
⭐
466
Implementation of NÜWA, state of the art attention network for text to video synthesis, in Pytorch
Neural_sp
⭐
466
End-to-end ASR/LM implementation with PyTorch
Overlappredator
⭐
460
[CVPR 2021, Oral] PREDATOR: Registration of 3D Point Clouds with Low Overlap.
Megabyte Pytorch
⭐
458
Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch
Openstl
⭐
450
OpenSTL: A Comprehensive Benchmark of Spatio-Temporal Predictive Learning
Lamda Rlhf Pytorch
⭐
444
Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.
Palm Pytorch
⭐
439
Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways
Mirnet
⭐
436
[ECCV 2020] Learning Enriched Features for Real Image Restoration and Enhancement. SOTA results for image denoising, super-resolution, and image enhancement.
Geotransformer
⭐
422
[CVPR2022] Geometric Transformer for Fast and Robust Point Cloud Registration
Structured Self Attention
⭐
412
A Structured Self-attentive Sentence Embedding
Seam
⭐
408
Self-supervised Equivariant Attention Mechanism for Weakly Supervised Semantic Segmentation, CVPR 2020 (Oral)
Caranet
⭐
407
Context Axial Reverse Attention Network for Small Medical Objects Segmentation
Point Transformer Pytorch
⭐
402
Implementation of the Point Transformer layer, in Pytorch
Meshgpt Pytorch
⭐
394
Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch
Time Series Autoencoder
⭐
386
📈 PyTorch Dual-Attention LSTM-Autoencoder For Multivariate Time Series 📈
Paperrobot
⭐
384
Code for PaperRobot: Incremental Draft Generation of Scientific Ideas
Triplet Attention
⭐
383
Official PyTorch Implementation for "Rotate to Attend: Convolutional Triplet Attention Module." [WACV 2021]
Pytorch Original Transformer
⭐
376
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Swarms
⭐
376
Build, Deploy, and Scale Reliable Swarms of Autonomous Agents for Workflow Automation. Join our Community: https://discord.gg/DbjBMJTSWD
Enformer Pytorch
⭐
359
Implementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch
Action Recognition Visual Attention
⭐
338
Action recognition using soft attention based deep recurrent neural networks
Recurrent Memory Transformer Pytorch
⭐
335
Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
Changeformer
⭐
335
[IGARSS'22]: A Transformer-Based Siamese Network for Change Detection
Linformer Pytorch
⭐
323
My take on a practical implementation of Linformer for Pytorch.
Multimodalmamba
⭐
321
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
Robotic Transformer Pytorch
⭐
306
Implementation of RT1 (Robotic Transformer) in Pytorch
Keras Gat
⭐
301
Keras implementation of the graph attention networks (GAT) by Veličković et al. (2017; https://arxiv.org/abs/1710.10903)
Medical Chatgpt
⭐
300
Implementation of ChatGPT, but tailored towards primary care medicine, with the reward being able to collect patient histories in a thorough and efficient manner and come up with a reasonable differential diagnosis
Flash Pytorch
⭐
298
Implementation of the Transformer variant proposed in "Transformer Quality in Linear Time"
Stanet
⭐
296
official implementation of the spatial-temporal attention neural network (STANet) for remote sensing image change detection
Attention Mechanisms
⭐
294
Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Attention_is_all_you_need
⭐
293
Transformer of "Attention Is All You Need" (Vaswani et al. 2017) by Chainer.
Adaptiveattention
⭐
288
Implementation of "Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning"
Dodrio
⭐
287
Exploring attention weights in transformer-based models with linguistic knowledge.
Slot Attention
⭐
286
Implementation of Slot Attention from GoogleAI
Seq2seq_chatbot
⭐
286
基于seq2seq模型的简单对话系统的tf实现,具有embedding、attention、beam Movie Dialogs
Linear Attention Transformer
⭐
278
Transformer based on a variant of attention that is linear complexity in respect to sequence length
Tensorflow_end2end_speech_recognition
⭐
275
End-to-End speech recognition implementation base on TensorFlow (CTC, Attention, and MTL training)
Itransformer
⭐
275
Unofficial implementation of iTransformer - SOTA Time Series Forecasting using Attention networks, out of Tsinghua / Ant group
Related Searches
Python Attention Mechanism (517)
Deep Learning Attention Mechanism (370)
1-100 of 649 search results
Next >
Privacy
|
About
|
Terms
|
Follow Us On Twitter
Copyright 2018-2024 Awesome Open Source. All rights reserved.