Subgraph Based Learning of Contextual Embedding

Related tags

Deep Learning SLiCE
Overview

SLiCE

Self-Supervised Learning of Contextual Embeddings for Link Prediction in Heterogeneous Networks

Dataset details:

Install instructions:

  • Dependencies: Python 3.6, PyTorch 1.4.0 w/ CUDA 9.2, Pytorch Geometric
  • The specific Pytorch Geometric wheels we use are included in the repo for convenience in the 'wheels' directory
conda create -n slice python=3.6
conda activate slice
pip install -r requirements.txt

Training:

python main.py \
    --data_name 'amazon_s' \
    --data_path 'data' \
    --outdir 'output/amazon_s' \
    --pretrained_embeddings 'data/amazon_s/amazon_s.emd' \
    --n_epochs 10 \
    --n_layers 4 \
    --n_heads 4 \
    --gcn_option 'no_gcn' \
    --node_edge_composition_func 'mult' \
    --ft_input_option 'last4_cat' \
    --path_option 'shortest' \
    --ft_n_epochs 10 \
    --num_walks_per_node 1 \
    --max_length 6 \
    --walk_type 'dfs' \
    --is_pre_trained

Citation:

Please cite the following paper if you use this code in your work.

@inproceedings{wang2020self,
  title={Self-Supervised Learning of Contextual Embeddings for Link Prediction in Heterogeneous Networks},
  author={Wang, Ping and Agarwal, Khushbu and Ham, Colby and Choudhury, Sutanay and Reddy, Chandan K},
  booktitle={Proceedings of The Web Conference 2021},
  year={2021}
}

Notice

This material was prepared as an account of work sponsored by an agency of the United States Government. Neither the United States Government nor the United States Department of Energy, nor Battelle, nor any of their employees, nor any jurisdiction or organization that has cooperated in the development of these materials, makes any warranty, express or implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness or any information, apparatus, product, software, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States Government or any agency thereof, or Battelle Memorial Institute. The views and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof.

    PACIFIC NORTHWEST NATIONAL LABORATORY
    operated by
    BATTELLE
    for the
    UNITED STATES DEPARTMENT OF ENERGY
    under Contract DE-AC05-76RL01830
   

License

Released under the 3-Clause BSD license (see License.md)

You might also like...
Official Pytorch implementation for Deep Contextual Video Compression, NeurIPS 2021

Introduction Official Pytorch implementation for Deep Contextual Video Compression, NeurIPS 2021 Prerequisites Python 3.8 and conda, get Conda CUDA 11

Code and data for ImageCoDe, a contextual vison-and-language benchmark
Code and data for ImageCoDe, a contextual vison-and-language benchmark

ImageCoDe This repository contains code and data for ImageCoDe: Image Retrieval from Contextual Descriptions. Data All collected descriptions for the

the code used for the preprint Embedding-based Instance Segmentation of Microscopy Images.
the code used for the preprint Embedding-based Instance Segmentation of Microscopy Images.

EmbedSeg Introduction This repository hosts the version of the code used for the preprint Embedding-based Instance Segmentation of Microscopy Images.

Deep Text Search is an AI-powered multilingual text search and recommendation engine with state-of-the-art transformer-based multilingual text embedding (50+ languages).
Deep Text Search is an AI-powered multilingual text search and recommendation engine with state-of-the-art transformer-based multilingual text embedding (50+ languages).

Deep Text Search - AI Based Text Search & Recommendation System Deep Text Search is an AI-powered multilingual text search and recommendation engine w

Simple embedding based text classifier inspired by fastText, implemented in tensorflow

FastText in Tensorflow This project is based on the ideas in Facebook's FastText but implemented in Tensorflow. However, it is not an exact replica of

Repo for WWW 2022 paper: Progressively Optimized Bi-Granular Document Representation for Scalable Embedding Based Retrieval

BiDR Repo for WWW 2022 paper: Progressively Optimized Bi-Granular Document Representation for Scalable Embedding Based Retrieval. Requirements torch==

Code for
Code for "Learning the Best Pooling Strategy for Visual Semantic Embedding", CVPR 2021

Learning the Best Pooling Strategy for Visual Semantic Embedding Official PyTorch implementation of the paper Learning the Best Pooling Strategy for V

Paddle implementation for "Highly Efficient Knowledge Graph Embedding Learning with Closed-Form Orthogonal Procrustes Analysis" (NAACL 2021)

ProcrustEs-KGE Paddle implementation for Highly Efficient Knowledge Graph Embedding Learning with Orthogonal Procrustes Analysis 🙈 A more detailed re

Official PyTorch Implementation of Embedding Transfer with Label Relaxation for Improved Metric Learning, CVPR 2021
Official PyTorch Implementation of Embedding Transfer with Label Relaxation for Improved Metric Learning, CVPR 2021

Embedding Transfer with Label Relaxation for Improved Metric Learning Official PyTorch implementation of CVPR 2021 paper Embedding Transfer with Label

Comments
  • Installation Issues

    Installation Issues

    torch_scatter, torch_sparse, torch_spline_conv, torch_cluster are not getting installed even after explicitly running their wheel files. Kindly look into this and suggest something.

    opened by AnoushkaVyas 3
  • Issues abour eq(1)

    Issues abour eq(1)

    I am wondering whether the eq(1) equals message passing method? Like GCN, it's AXW. But it's WXA in paper. The dimension of W is d*d, X is |Vc|*d, A is |Vc| *|Vc|. The matirx multiplication WXA is ok? Or it should be AXW?

    opened by Theheavens 2
  • About node types in the datasets

    About node types in the datasets

    Hello, I'm trying to convert your dataset Healthcare into DGL Graph format, but I cannot find description about the node types. Where can I obtain the node types, or does it matter?

    opened by VoidHaruhi 1
  • Issues on the performance

    Issues on the performance

    I used the hyperparameters provided, but the performance is actually bad. Take DBLP and twitter as examples, each of the results is about 20% below your result in the paper. Here is the command lines I used and the results. DBLP: nohup python3 main.py --base_embedding_dim=128 --batch_size=128 --beam_width=4 --checkpoint=15 --d_ff=512 --d_k=64 --d_model=128 --d_v=64 --data_name='dblp' --false_edge_gen='double' --ft_batch_size=100 --ft_checkpoint=500 --ft_d_ff=512 --ft_drop_rate=0.1 --ft_input_option='last4_cat' --ft_layer='ffn' --ft_lr=5e-05 --ft_n_epochs=10 --gcn_option='no_gcn' --get_bert_encoder_embeddings=False --lr=0.0001 --max_length=12 --n_epochs=10 --n_heads=4 --n_layers=4 --node_edge_composition_func='mult' --num_gcn_layers=2 --num_walks_per_node=5 --outdir='./output/dblp/' --path_option='shortest' --pretrained_method='node2vec' --walk_type='dfs' 2>&1 & result Twitter: nohup python3 main.py --base_embedding_dim=128 --batch_size=128 --beam_width=4 --checkpoint=15 --d_ff=512 --d_k=64 --d_model=128 --d_v=64 --data_name='twitter' --emb_dir=None --false_edge_gen='double' --ft_batch_size=100 --ft_checkpoint=500 --ft_d_ff=512 --ft_drop_rate=0.1 --ft_input_option='last4_cat' --ft_layer='ffn' --ft_lr=5e-05 --ft_n_epochs=10 --gcn_option='no_gcn' --get_bert_encoder_embeddings=False --lr=0.0001 --max_length=12 --n_epochs=10 --n_heads=4 --n_layers=4 --node_edge_composition_func='mult' --num_gcn_layers=2 --num_walks_per_node=10 --path_option='shortest' --pretrained_embeddings='./embed/twitter/twitter.emd' --outdir='./output/twitter/' --pretrained_method='node2vec' --walk_type='dfs' 2>&1 & image

    opened by VoidHaruhi 23
Owner
Pacific Northwest National Laboratory
Pacific Northwest National Laboratory
Official PyTorch implementation of the paper: Improving Graph Neural Network Expressivity via Subgraph Isomorphism Counting.

Improving Graph Neural Network Expressivity via Subgraph Isomorphism Counting Official PyTorch implementation of the paper: Improving Graph Neural Net

Giorgos Bouritsas 58 Dec 31, 2022
Code for "Contextual Non-Local Alignment over Full-Scale Representation for Text-Based Person Search"

Contextual Non-Local Alignment over Full-Scale Representation for Text-Based Person Search This is an implementation for our paper Contextual Non-Loca

Tencent YouTu Research 50 Dec 3, 2022
Repo for CVPR2021 paper "QPIC: Query-Based Pairwise Human-Object Interaction Detection with Image-Wide Contextual Information"

QPIC: Query-Based Pairwise Human-Object Interaction Detection with Image-Wide Contextual Information by Masato Tamura, Hiroki Ohashi, and Tomoaki Yosh

null 105 Dec 23, 2022
banditml is a lightweight contextual bandit & reinforcement learning library designed to be used in production Python services.

banditml is a lightweight contextual bandit & reinforcement learning library designed to be used in production Python services. This library is developed by Bandit ML and ex-authors of Facebook's applied reinforcement learning platform, Reagent.

Bandit ML 51 Dec 22, 2022
Source code and data from the RecSys 2020 article "Carousel Personalization in Music Streaming Apps with Contextual Bandits" by W. Bendada, G. Salha and T. Bontempelli

Carousel Personalization in Music Streaming Apps with Contextual Bandits - RecSys 2020 This repository provides Python code and data to reproduce expe

Deezer 48 Jan 2, 2023
UmlsBERT: Clinical Domain Knowledge Augmentation of Contextual Embeddings Using the Unified Medical Language System Metathesaurus

UmlsBERT: Clinical Domain Knowledge Augmentation of Contextual Embeddings Using the Unified Medical Language System Metathesaurus General info This is

null 71 Oct 25, 2022
Generate Contextual Directory Wordlist For Target Org

PathPermutor Generate Contextual Directory Wordlist For Target Org This script generates contextual wordlist for any target org based on the set of UR

null 8 Jun 23, 2021
ICCV2021 - Mining Contextual Information Beyond Image for Semantic Segmentation

Introduction The official repository for "Mining Contextual Information Beyond Image for Semantic Segmentation". Our full code has been merged into ss

null 55 Nov 9, 2022
[2021 MultiMedia] CONQUER: Contextual Query-aware Ranking for Video Corpus Moment Retrieval

CONQUER: Contexutal Query-aware Ranking for Video Corpus Moment Retreival PyTorch implementation of CONQUER: Contexutal Query-aware Ranking for Video

Hou zhijian 23 Dec 26, 2022
Official implementation of NeurIPS 2021 paper "Contextual Similarity Aggregation with Self-attention for Visual Re-ranking"

CSA: Contextual Similarity Aggregation with Self-attention for Visual Re-ranking PyTorch training code for CSA (Contextual Similarity Aggregation). We

Hui Wu 19 Oct 21, 2022