prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the LinkDistMLP model in the Distilling Self-Knowledge From Contrastive Links to Classify Graph Nodes Without Passing Messages paper on the PubMed with Public Split: fixed 20 nodes per class dataset? | Accuracy, F1 |
What metrics were used to measure the ChebyNet model in the Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering paper on the PubMed with Public Split: fixed 20 nodes per class dataset? | Accuracy, F1 |
What metrics were used to measure the HeteroGraphSAGE model in the MuMiN: A Large-Scale Multilingual Multimodal Fact-Checked Misinformation Social Network Dataset paper on the MuMiN-medium dataset? | Claim Classification Macro-F1, Tweet Classification Macro-F1 |
What metrics were used to measure the LaBSE model in the MuMiN: A Large-Scale Multilingual Multimodal Fact-Checked Misinformation Social Network Dataset paper on the MuMiN-medium dataset? | Claim Classification Macro-F1, Tweet Classification Macro-F1 |
What metrics were used to measure the Majority class model in the MuMiN: A Large-Scale Multilingual Multimodal Fact-Checked Misinformation Social Network Dataset paper on the MuMiN-medium dataset? | Claim Classification Macro-F1, Tweet Classification Macro-F1 |
What metrics were used to measure the Random model in the MuMiN: A Large-Scale Multilingual Multimodal Fact-Checked Misinformation Social Network Dataset paper on the MuMiN-medium dataset? | Claim Classification Macro-F1, Tweet Classification Macro-F1 |
What metrics were used to measure the FAGCN model in the Beyond Low-frequency Information in Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-GCNII model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACMII-Snowball-2 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACMII-GCN+ model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GCNII* model in the Simple and Deep Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-GCN++ model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACMII-GCN model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACMII-GCN++ model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-GCNII* model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-GCN+ model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GCNII model in the Simple and Deep Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-Snowball-2 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACMII-Snowball-3 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the Snowball-2 model in the Break the Ceiling: Stronger Multi-scale Deep Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GCN model in the Semi-Supervised Classification with Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-Snowball-3 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-SGC-1 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the Snowball-3 model in the Break the Ceiling: Stronger Multi-scale Deep Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the ACM-SGC-2 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the SGC-2 model in the Simplifying Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the BernNet model in the BernNet: Learning Arbitrary Graph Spectral Filters via Bernstein Approximation paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the H2GCN model in the Beyond Low-frequency Information in Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the SGC-1 model in the Simplifying Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GraphSAGE model in the Inductive Representation Learning on Large Graphs paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the Geom-GCN* model in the Geom-GCN: Geometric Graph Convolutional Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the MLP-2 model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GAT+JK model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GCN+JK model in the Revisiting Heterophily For Graph Neural Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the APPNP model in the Predict then Propagate: Graph Neural Networks meet Personalized PageRank paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GPRGNN model in the Adaptive Universal Generalized PageRank Graph Neural Network paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the GAT model in the Graph Attention Networks paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the MixHop model in the MixHop: Higher-Order Graph Convolutional Architectures via Sparsified Neighborhood Mixing paper on the CiteSeer (60%/20%/20% random splits) dataset? | 1:1 Accuracy |
What metrics were used to measure the Exphormer model in the Exphormer: Sparse Transformers for Graphs paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the GPS model in the Recipe for a General, Powerful, Scalable Graph Transformer paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the GatedGCN model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the Transformer+LapPE model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the SAN+LapPE model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the GatedGCN+LapPE model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the SAN+RWSE model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the GCNII model in the Simple and Deep Graph Convolutional Networks paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the GINE model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the GCN model in the Long Range Graph Benchmark paper on the COCO-SP dataset? | macro F1 |
What metrics were used to measure the NLGAT model in the Non-Local Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GCNII model in the Simple and Deep Graph Convolutional Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GloGNN++ model in the Finding Global Homophily in Graph Neural Networks When Meeting Heterophily paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GloGNN model in the Finding Global Homophily in Graph Neural Networks When Meeting Heterophily paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACMII-GCN++ model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the WRGAT model in the Breaking the Limit of Graph Neural Networks by Improving the Assortativity of Graphs with Local Mixing Patterns paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACMII-GCN+ model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACM-GCN++ model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the NLGCN model in the Non-Local Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the FAGCN model in the Beyond Low-frequency Information in Graph Convolutional Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACM-GCN+ model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACMII-GCN model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GPRGCN model in the Adaptive Universal Generalized PageRank Graph Neural Network paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GGCN model in the Two Sides of the Same Coin: Heterophily and Oversmoothing in Graph Convolutional Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the H2GCN model in the Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACM-SGC-2 model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the MixHop model in the MixHop: Higher-Order Graph Convolutional Architectures via Sparsified Neighborhood Mixing paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the Gen-NSD model in the Neural Sheaf Diffusion: A Topological Perspective on Heterophily and Oversmoothing in GNNs paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the Diag-NSD model in the Neural Sheaf Diffusion: A Topological Perspective on Heterophily and Oversmoothing in GNNs paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the ACM-SGC-1 model in the Revisiting Heterophily For Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the O(d)-NSD model in the Neural Sheaf Diffusion: A Topological Perspective on Heterophily and Oversmoothing in GNNs paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GESN model in the Addressing Heterophily in Node Classification with Graph Echo State Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the Geom-GCN model in the Geom-GCN: Geometric Graph Convolutional Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the LINKX model in the Large Scale Learning on Non-Homophilous Graphs: New Benchmarks and Strong Simple Methods paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the NLMLP model in the Non-Local Graph Neural Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GREAD-BS model in the GREAD: Graph Neural Reaction-Diffusion Networks paper on the Cora (48%/32%/20% fixed splits) dataset? | 1:1 Accuracy, Accuracy |
What metrics were used to measure the GRIT model in the Graph Inductive Biases in Transformers without Message Passing paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the EGT model in the Global Self-Attention as a Replacement for Graph Convolution paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the Exphormer model in the Exphormer: Sparse Transformers for Graphs paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the GPTrans-Nano model in the Graph Propagation Transformer for Graph Representation Learning paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the GPS model in the Recipe for a General, Powerful, Scalable Graph Transformer paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the ARGNP model in the Automatic Relation-aware Graph Network Proliferation paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the GatedGCN-PE model in the Benchmarking Graph Neural Networks paper on the CLUSTER dataset? | Accuracy |
What metrics were used to measure the CoLinkDist model in the Distilling Self-Knowledge From Contrastive Links to Classify Graph Nodes Without Passing Messages paper on the Cora Full with Public Split dataset? | Accuracy |
What metrics were used to measure the LinkDist model in the Distilling Self-Knowledge From Contrastive Links to Classify Graph Nodes Without Passing Messages paper on the Cora Full with Public Split dataset? | Accuracy |
What metrics were used to measure the CoLinkDistMLP model in the Distilling Self-Knowledge From Contrastive Links to Classify Graph Nodes Without Passing Messages paper on the Cora Full with Public Split dataset? | Accuracy |
What metrics were used to measure the LinkDistMLP model in the Distilling Self-Knowledge From Contrastive Links to Classify Graph Nodes Without Passing Messages paper on the Cora Full with Public Split dataset? | Accuracy |
What metrics were used to measure the DJ-GNN model in the Diffusion-Jump GNNs: Homophiliation via Learnable Metric Filters paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the Dir-GNN model in the Edge Directionality Improves Learning on Heterophilic Graphs paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the FSGNN (8-hop) model in the Improving Graph Neural Networks with Simple Architecture Design paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the FSGNN (3-hop) model in the Improving Graph Neural Networks with Simple Architecture Design paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the HLP Concat model in the Simple Truncated SVD based Model for Node Classification on Heterophilic Graphs paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the Graph ESN model in the Beyond Homophily with Graph Echo State Networks paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the SADE-GCN model in the Self-attention Dual Embedding for Graphs with Heterophily paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the ACMII-GCN++ model in the Revisiting Heterophily For Graph Neural Networks paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the GCNII+DHGR model in the Make Heterophily Graphs Better Fit GNN: A Graph Rewiring Approach paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the ACMII-GCN+ model in the Revisiting Heterophily For Graph Neural Networks paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the UDGNN (GCN) model in the Universal Deep GNNs: Rethinking Residual Connection in GNNs from a Path Decomposition Perspective for Preventing the Over-smoothing paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the ACM-GCN+ model in the Revisiting Heterophily For Graph Neural Networks paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the ACM-GCN++ model in the Revisiting Heterophily For Graph Neural Networks paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the LW-GCN model in the Label-Wise Graph Convolutional Network for Heterophilic Graphs paper on the Chameleon dataset? | Accuracy |
What metrics were used to measure the SignGT model in the SignGT: Signed Attention-based Graph Transformer for Graph Representation Learning paper on the Chameleon dataset? | Accuracy |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.