Project Name | Stars | Downloads | Repos Using This | Packages Using This | Most Recent Commit | Total Releases | Latest Release | Open Issues | License | Language |
---|---|---|---|---|---|---|---|---|---|---|
Tensorflow | 177,850 | 327 | 77 | 7 hours ago | 46 | October 23, 2019 | 2,047 | apache-2.0 | C++ | |
An Open Source Machine Learning Framework for Everyone | ||||||||||
Transformers | 112,568 | 64 | 1,869 | 4 hours ago | 114 | July 18, 2023 | 850 | apache-2.0 | Python | |
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. | ||||||||||
Keras | 59,451 | 578 | 8 hours ago | 80 | June 27, 2023 | 97 | apache-2.0 | Python | ||
Deep Learning for humans | ||||||||||
Tensorflow Examples | 42,312 | a year ago | 218 | other | Jupyter Notebook | |||||
TensorFlow Tutorial and Examples for Beginners (support TF v1 & v2) | ||||||||||
Photoprism | 28,751 | 4 | 5 days ago | 151 | April 25, 2021 | 388 | other | Go | ||
AI-Powered Photos App for the Decentralized Web 🌈💎✨ | ||||||||||
Ray | 27,947 | 80 | 298 | 4 hours ago | 87 | July 24, 2023 | 3,442 | apache-2.0 | Python | |
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads. | ||||||||||
Data Science Ipython Notebooks | 25,242 | 3 months ago | 34 | other | Python | |||||
Data science Python notebooks: Deep learning (TensorFlow, Theano, Caffe, Keras), scikit-learn, Kaggle, big data (Spark, Hadoop MapReduce, HDFS), matplotlib, pandas, NumPy, SciPy, Python essentials, AWS, and various command lines. | ||||||||||
Handson Ml | 25,030 | 3 months ago | 139 | apache-2.0 | Jupyter Notebook | |||||
⛔️ DEPRECATED – See https://github.com/ageron/handson-ml3 instead. | ||||||||||
Handson Ml2 | 24,694 | 5 months ago | 197 | apache-2.0 | Jupyter Notebook | |||||
A series of Jupyter notebooks that walk you through the fundamentals of Machine Learning and Deep Learning in Python using Scikit-Learn, Keras and TensorFlow 2. | ||||||||||
Netron | 24,086 | 4 | 69 | 14 hours ago | 587 | August 01, 2023 | 23 | mit | JavaScript | |
Visualizer for neural network, deep learning, and machine learning models |
This is our implementation for the paper:
Jun Xiao, Hao Ye, Xiangnan He, Hanwang Zhang, Fei Wu and Tat-Seng Chua (2017). Attentional Factorization Machines: Learning the Weight of Feature Interactions via Attention Networks IJCAI, Melbourne, Australia, August 19-25, 2017.
We have additionally released our TensorFlow implementation of Factorization Machines under our proposed neural network framework.
Please cite our IJCAI'17 paper if you use our codes. Thanks!
Author: Xiangnan He ([email protected]) and Hao Ye ([email protected])
We use the same input format as the LibFM toolkit (http://www.libfm.org/). In this instruction, we use MovieLens. The MovieLens data has been used for personalized tag recommendation, which contains 668,953 tag applications of users on movies. We convert each tag application (user ID, movie ID and tag) to a feature vector using one-hot encoding and obtain 90,445 binary features. The following examples are based on this dataset and it will be referred as ml-tag wherever in the files' name or inside the code. When the dataset is ready, the current directory should be like this:
Use the following command to train the model with the optimal parameters:
# step into the code folder
cd code
# train FM model and save as pretrain file
python FM.py --dataset ml-tag --epoch 100 --pretrain -1 --batch_size 4096 --hidden_factor 256 --lr 0.01 --keep 0.7
# train AFM model using the pretrained weights from FM
python AFM.py --dataset ml-tag --epoch 100 --pretrain 1 --batch_size 4096 --hidden_factor [8,256] --keep [1.0,0.5] --lamda_attention 2.0 --lr 0.1
The instruction of commands has been clearly stated in the codes (see the parse_args function).
The current implementation supports regression classification, which optimizes RMSE.
For the sake of a quick demonstration for the improvement of our AFM model compared to original FM, we set the dimension of the embedding factor to be 16 (instead of 256 in our paper), and epoch as 20.
Step into the code folder and train FM and AFM as follows. This will start to train our AFM model on the dataset frappe based on the pretrained model of FM. The parameters have been initialized optimally according to our experiments. It will loop 20 epochs and print the best epoch depending on the validation result.
# step into the code folder
cd code
# train FM model with optimal parameters
python FM.py --dataset ml-tag --epoch 20 --pretrain -1 --batch_size 4096 --hidden_factor 16 --lr 0.01 --keep 0.7
# train AFM model with optimal parameters
python AFM.py --dataset ml-tag --epoch 20 --pretrain 1 --batch_size 4096 --hidden_factor [16,16] --keep [1.0,0.5] --lamda_attention 100.0 --lr 0.1
After the trainning processes finish, the trained models will be saved into the pretrain folder, which should be like this:
Now it's time to evaluate the pretrained models with the test datasets, which can be done by running AFM.py and FM.py with --process evaluate as follows:
# evaluate the pretrained FM model
python FM.py --dataset ml-tag --epoch 20 --batch_size 4096 --lr 0.01 --keep 0.7 --process evaluate
# evaluate the pretrained AFM model
python AFM.py --dataset ml-tag --epoch 20 --pretrain 1 --batch_size 4096 --hidden_factor [16,16] --keep [1.0,0.5] --lamda_attention 100.0 --lr 0.1 --process evaluate
Last Update Date: Aug 2, 2017