Skip to content

Keras implementation of the graph attention networks (GAT) by Veličković et al. (2017; https://arxiv.org/abs/1710.10903)

License

Notifications You must be signed in to change notification settings

danielegrattarola/keras-gat

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Keras Graph Attention Network

DEPRECATED
This implementation of GAT is no longer actively maintained and may not work with modern versions of Tensorflow and Keras. Check out Spektral and its GAT example for a Tensorflow/Keras implementation of GAT.

This is a Keras implementation of the Graph Attention Network (GAT) model by Veličković et al. (2017, [arXiv link]).

Acknowledgements

I have no affiliation with the authors of the paper and I am implementing this code for non-commercial reasons.
The authors published their reference Tensorflow implementation here, so check it out for something that is guaranteed to work as intended. Their implementation is slightly different than mine, so that may be something to keep in mind. You should cite the paper if you use any of this code for your research:

@article{
  velickovic2018graph,
  title="{Graph Attention Networks}",
  author={Veli{\v{c}}kovi{\'{c}}, Petar and Cucurull, Guillem and Casanova, Arantxa and Romero, Adriana and Li{\`{o}}, Pietro and Bengio, Yoshua},
  journal={International Conference on Learning Representations},
  year={2018},
  url={https://openreview.net/forum?id=rJXMpikCZ},
  note={Accepted as poster},
}

If you would like to give me credit, feel free to link to my Github profile, blog, or Twitter.

I also copied the code in utils.py almost verbatim from this repo by Thomas Kipf, who I thank sincerely for sharing his work on GCNs and GAEs, and for giving me a few pointers on how to split the data into train/test/val sets.

Thanks to mawright, matthias-samwald, and vermaMachineLearning for helping me out with bugs, performance improvements, and running experiments.

Disclaimer

I do not own any rights to the datasets distributed with this code, but they are publicly available at the following links:

Installation

To install as a module:

$ git clone https://github.com/danielegrattarola/keras-gat.git
$ cd keras-gat
$ pip install .
$ python
>>> from keras_gat import GraphAttention

Or you can just copy and paste graph_attention_layer.py into your project.

Replicating experiments

To replicate the experimental results of the paper, simply run:

$ python examples/gat.py

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •  

Languages