Project Name | Stars | Downloads | Repos Using This | Packages Using This | Most Recent Commit | Total Releases | Latest Release | Open Issues | License | Language |
---|---|---|---|---|---|---|---|---|---|---|
Conv Emotion | 1,090 | 2 months ago | 7 | mit | Python | |||||
This repo contains implementation of different architectures for emotion recognition in conversations. | ||||||||||
Torchmoji | 678 | 3 years ago | 15 | mit | Python | |||||
😇A pyTorch implementation of the DeepMoji model: state-of-the-art deep learning model for analyzing sentiment, emotion, sarcasm etc | ||||||||||
Facepose_pytorch | 591 | a month ago | 120 | mit | Python | |||||
🔥🔥The pytorch implement of the head pose estimation(yaw,roll,pitch) and emotion detection with SOTA performance in real time.Easy to deploy, easy to use, and high accuracy.Solve all problems of face detection at one time.(极简,极快,高效是我们的宗旨) | ||||||||||
Deep Emotion | 149 | 10 months ago | 4 | Python | ||||||
Facial Expression Recognition Using Attentional Convolutional Network, Pytorch implementation | ||||||||||
Emotion_dataset | 140 | 5 months ago | ||||||||
:smile: Dataset for Emotion Classification | ||||||||||
Multimodal Speech Emotion Recognition | 133 | 2 years ago | 6 | mit | Jupyter Notebook | |||||
Lightweight and Interpretable ML Model for Speech Emotion Recognition and Ambiguity Resolution (trained on IEMOCAP dataset) | ||||||||||
Speech Emotion Classification With Pytorch | 100 | 7 months ago | 4 | Jupyter Notebook | ||||||
This repository contains PyTorch implementation of 4 different models for classification of emotions of the speech. | ||||||||||
Ned | 67 | 8 months ago | 2 | mit | Python | |||||
PyTorch implementation for NED (CVPR 2022). It can be used to manipulate the facial emotions of actors in videos based on emotion labels or reference styles. | ||||||||||
Group Level Emotion Recognition | 67 | 5 years ago | 3 | mit | Jupyter Notebook | |||||
Model submitted for the ICMI 2018 EmotiW Group-Level Emotion Recognition Challenge | ||||||||||
Emogen | 62 | 14 days ago | cc0-1.0 | Python | ||||||
PyTorch Implementation for Paper "Emotionally Enhanced Talking Face Generation" |
It's been a year since TorchMoji and DeepMoji were released. We're trying to understand how it's being used such that we can make improvements and design better models in the future.
You can help us achieve this by answering this 4-question Google Form. Thanks for your support!
Read our blog post about the implementation process here.
TorchMoji is a pyTorch implementation of the DeepMoji model developped by Bjarke Felbo, Alan Mislove, Anders Søgaard, Iyad Rahwan and Sune Lehmann.
This model trained on 1.2 billion tweets with emojis to understand how language is used to express emotions. Through transfer learning the model can obtain state-of-the-art performance on many emotion-related text modeling tasks.
Try the online demo of DeepMoji http://deepmoji.mit.edu! See the paper, blog post or FAQ for more details.
To start out with, have a look inside the examples/ directory. See score_texts_emojis.py for how to use DeepMoji to extract emoji predictions, encode_texts.py for how to convert text into 2304-dimensional emotional feature vectors or finetune_youtube_last.py for how to use the model for transfer learning on a new dataset.
Please consider citing the paper of DeepMoji if you use the model or code (see below for citation).
We assume that you're using Python 2.7-3.5 with pip installed.
First you need to install pyTorch (version 0.2+), currently by:
conda install pytorch -c pytorch
At the present stage the model can't make efficient use of CUDA. See details in the Hugging Face blog post.
When pyTorch is installed, run the following in the root directory to install the remaining dependencies:
pip install -e .
This will install the following dependencies:
Then, run the download script to downloads the pretrained torchMoji weights (~85MB) from here and put them in the model/ directory:
python scripts/download_weights.py
To run the tests, install nose. After installing, navigate to the tests/ directory and run:
cd tests
nosetests -v
By default, this will also run finetuning tests. These tests train the model for one epoch and then check the resulting accuracy, which may take several minutes to finish. If you'd prefer to exclude those, run the following instead:
cd tests
nosetests -v -a '!slow'
This code has been tested to work with Python 2.7 and 3.5 on Ubuntu 16.04 and macOS Sierra machines. It has not been optimized for efficiency, but should be fast enough for most purposes. We do not give any guarantees that there are no bugs - use the code on your own responsibility!
We welcome pull requests if you feel like something could be improved. You can also greatly help us by telling us how you felt when writing your most recent tweets. Just click here to contribute.
This code and the pretrained model is licensed under the MIT license.
The benchmark datasets are uploaded to this repository for convenience purposes only. They were not released by us and we do not claim any rights on them. Use the datasets at your responsibility and make sure you fulfill the licenses that they were released with. If you use any of the benchmark datasets please consider citing the original authors.
@inproceedings{felbo2017,
title={Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm},
author={Felbo, Bjarke and Mislove, Alan and S{\o}gaard, Anders and Rahwan, Iyad and Lehmann, Sune},
booktitle={Conference on Empirical Methods in Natural Language Processing (EMNLP)},
year={2017}
}