paper_id string | claim_id string | claim string | label string | caption string | evi_type string | evi_path string | context string | domain string | use_context string | operation string | paper_path string | detail_others string | license_name string | license_url string | claim_id_pair string | evi_path_original string |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2403.19137 | val_tab_0001 | Table 1 shows that our probabilistic inference module consistently outperforms its deterministic counterpart in terms of Avg and Last accuracy. | Supported | Table 1 : Performance comparison of different methods averaged over three runs. Best scores are in bold . Second best scores are in blue . The results for L2P, DualPrompt, and PROOF are taken from [ 92 ] . See App. Table 14 for std. dev. scores. | table | tables_png/dev/val_tab_0001.png | To understand our probabilistic inference modules further, we examine their performance against the deterministic variant of ours (Ours w/o VI). | ml | yes | Change the cell values | papers/dev/ml_2403.19137.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0001 | tables/dev/val_tab_0001.tex | |
2403.19137 | val_tab_0002 | Table 1 shows that our probabilistic inference module consistently outperforms its deterministic counterpart in terms of Avg and Last accuracy. | Refuted | Table 1 : Performance comparison of different methods averaged over three runs. Best scores are in bold . Second best scores are in blue . The results for L2P, DualPrompt, and PROOF are taken from [ 92 ] . See App. Table 14 for std. dev. scores. | table | tables_png/dev/val_tab_0002.png | To understand our probabilistic inference modules further, we examine their performance against the deterministic variant of ours (Ours w/o VI). | ml | yes | Change the cell values | papers/dev/ml_2403.19137.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0001 | tables/dev/val_tab_0002.tex | |
2203.01212 | val_tab_0004 | If we compare the two-layer network results from GeoLIP and Sampling in Table 1 , which is a lower bound of true Lipschitz constant, the ratio is within 1.783 . | Supported | Table 1: \ell_{\infty} -FGL estimation of various methods: DGeoLIP and NGeoLIP induce the same values on two layer networks. DGeoLIP always produces tighter estimations than LiPopt and MP do. | table | tables_png/dev/val_tab_0004.png | We have also shown that the two-layer network \ell_{\infty} -FGL estimation from GeoLIP has a theoretical guarantee with the approximation factor K_{G}<1.783 ( Theorem 3.3 ). | ml | no | Change the cell values | papers/dev/ml_2203.01212.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0002 | tables/dev/val_tab_0004.tex | |
2209.12590 | val_tab_0005 | Table 2 shows that the adversary applies lower dropout probabilities to less informative words such as ‘unknown’ tokens that replace all out-of-dictionary words, and so offer little information about the next word. | Supported | Table 2 : Analysis of the Adversary. For selected SNLI (upper) and Yahoo (lower) sentences, word dropout scores are from a trained adversary and normalized per sentence. Darker colouring indicates a higher dropout probability. Boxed words are those selected to be dropped. <sos> = start of the sequence token; <eos> = en... | table | tables_png/dev/val_tab_0005.png | We obtain further insight into what the adversary learns by analyzing the word dropout scores for different sentences. | ml | no | Change the cell values | papers/dev/ml_2209.12590.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0003 | tables/dev/val_tab_0005.tex | |
2209.12590 | val_tab_0006 | Table 2 shows that the adversary applies lower dropout probabilities to less informative words such as ‘unknown’ tokens that replace all out-of-dictionary words, and so offer little information about the next word. | Refuted | Table 2 : Analysis of the Adversary. For selected SNLI (upper) and Yahoo (lower) sentences, word dropout scores are from a trained adversary and normalized per sentence. Darker colouring indicates a higher dropout probability. Boxed words are those selected to be dropped. <sos> = start of the sequence token; <eos> = en... | table | tables_png/dev/val_tab_0006.png | We obtain further insight into what the adversary learns by analyzing the word dropout scores for different sentences. | ml | no | Change the cell values | papers/dev/ml_2209.12590.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0003 | tables/dev/val_tab_0006.tex | |
2209.12590 | val_tab_0008 | We also show that adversarial training learns a useful generative model with meaningful latent space by interpolating between sentences (Table 3 ). | Supported | Table 3 : Sentence interpolation (Yelp dataset). Representations of two sentences (top, bottom) are obtained by feeding them through an adversarially trained VAE encoder. Three linearly interpolated representations are passed to the VAE decoder and sentences generated by greedy sampling (middle) . | table | tables_png/dev/val_tab_0008.png | We obtain further insight into what the adversary learns by analyzing the word dropout scores for different sentences. Table 2 shows that the adversary applies lower dropout probabilities to less informative words such as ‘unknown’ tokens that replace all out-of-dictionary words, and so offer little information about t... | ml | no | Change the cell values | papers/dev/ml_2209.12590.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0004 | tables/dev/val_tab_0008.tex | |
2205.11361 | val_tab_0009 | Lastly, we note that using larger values of \gamma naively does not guarantee better test performance: one has to fine tune the parameters \beta , \mu , \sigma , \eta appropriately to achieve favorable trade-off between training stability and test performance. | Supported | Table 1: Shallow neural nets trained on the airfoil data set. The results in parenthesis are achieved with the variant ( 11 ).
All the results are averaged over 5 models trained with different seed values. | table | tables_png/dev/val_tab_0009.png | For the training, we use a fully connected shallow neural network of width 16 with ReLU activation and train for 3000 epochs with the learning rate \eta=0.1 , using mean square error (MSE) as the loss and choosing \beta=0.5 . Table 1 reports the average root MSE (RMSE) and the RMSE gap (defined as test RMSE - train RMS... | ml | no | Change the cell values | papers/dev/ml_2205.11361.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0005 | tables/dev/val_tab_0009.tex | |
2205.11361 | val_tab_0010 | Table 2 shows that MPGD can lead to better test performance and lower accuracy gap when compared to the baseline (full batch GD) and Gaussian noise-perturbed GD. | Supported | Table 2: ResNet-18 trained on CIFAR-10 for 1000 epochs. Here, accuracy gap = training accuracy - validation accuracy. The results in parenthesis are achieved with the variant of MPGD ( 11 ). All the results are averaged over 5 models trained with different seed values. | table | tables_png/dev/val_tab_0010.png | ml | no | Change the cell values | papers/dev/ml_2205.11361.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0006 | tables/dev/val_tab_0010.tex | ||
2209.12362 | val_tab_0016 | We also show that our performance boost does not come from the additional training dataset of ActivityNet in Table 3 . | Supported | Table 3: Ablation experiments. We investigate the effectiveness of each component of our method as well as compare to vanilla multi-dataset training method.
“Vanilla” means using cross entropy (CE) loss in training.
“w/o informative los” means using CE and projection loss.
The numbers are top-1/top-5 accuracy, respecti... | table | tables_png/dev/val_tab_0016.png | We then compare our method with state-of-the-art on these datasets. We train a higher resolution model with larger spatial inputs (312p) and achieves better performance compared to recent multi-dataset training methods, CoVER [ 59 ] and PolyVit [ 38 ] , on Kinetics-400, and significantly better on MiT and SSv2, as show... | ml | no | Change the cell values | papers/dev/ml_2209.12362.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0007 | tables/dev/val_tab_0016.tex | |
2209.12362 | val_tab_0017 | Also, longer training schedule does not lead to better performance on some datasets, including SSv2, suggesting vanilla multi-dataset training is unstable. | Supported | Table 3: Ablation experiments. We investigate the effectiveness of each component of our method as well as compare to vanilla multi-dataset training method.
“Vanilla” means using cross entropy (CE) loss in training.
“w/o informative los” means using CE and projection loss.
The numbers are top-1/top-5 accuracy, respecti... | table | tables_png/dev/val_tab_0017.png | Does our proposed robust loss help? We compare our model training with vanilla multi-dataset training, where multiple classification heads are attached to the same backbone and the model is trained simply with cross-entropy loss. The vanilla model is trained from a K400 checkpoint as ours. As shown in Table 3 , we try ... | ml | no | Change the cell values | papers/dev/ml_2209.12362.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0008 | tables/dev/val_tab_0017.tex | |
2209.12362 | val_tab_0018 | Also, longer training schedule does not lead to better performance on some datasets, including SSv2, suggesting vanilla multi-dataset training is unstable. | Refuted | Table 3: Ablation experiments. We investigate the effectiveness of each component of our method as well as compare to vanilla multi-dataset training method.
“Vanilla” means using cross entropy (CE) loss in training.
“w/o informative los” means using CE and projection loss.
The numbers are top-1/top-5 accuracy, respecti... | table | tables_png/dev/val_tab_0018.png | Does our proposed robust loss help? We compare our model training with vanilla multi-dataset training, where multiple classification heads are attached to the same backbone and the model is trained simply with cross-entropy loss. The vanilla model is trained from a K400 checkpoint as ours. As shown in Table 3 , we try ... | ml | no | Change the cell values | papers/dev/ml_2209.12362.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0008 | tables/dev/val_tab_0018.tex | |
2209.12362 | val_tab_0019 | In terms of performance on ActivityNet, we observe that both training methods achieve good results, which might be because ActivityNet classes are highly overlapped with Kinetics-400 (65 out of 200). | Supported | Table 3: Ablation experiments. We investigate the effectiveness of each component of our method as well as compare to vanilla multi-dataset training method.
“Vanilla” means using cross entropy (CE) loss in training.
“w/o informative los” means using CE and projection loss.
The numbers are top-1/top-5 accuracy, respecti... | table | tables_png/dev/val_tab_0019.png | Does our proposed robust loss help? We compare our model training with vanilla multi-dataset training, where multiple classification heads are attached to the same backbone and the model is trained simply with cross-entropy loss. The vanilla model is trained from a K400 checkpoint as ours. As shown in Table 3 , we try ... | ml | no | Change the cell values | papers/dev/ml_2209.12362.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0009 | tables/dev/val_tab_0019.tex | |
2209.12362 | val_tab_0020 | As shown in Table 3 , the performance on MiT and SSv2 suffers by a large margin, indicating that the projection design helps boost training by better utilizing multi-dataset information. | Supported | Table 3: Ablation experiments. We investigate the effectiveness of each component of our method as well as compare to vanilla multi-dataset training method.
“Vanilla” means using cross entropy (CE) loss in training.
“w/o informative los” means using CE and projection loss.
The numbers are top-1/top-5 accuracy, respecti... | table | tables_png/dev/val_tab_0020.png | How important is the projection loss? We then experiment with removing the projection heads (Section 3.2 ) during multi-dataset training. The model is trained with the original cross-entropy loss and the informative loss. | ml | no | Change the cell values | papers/dev/ml_2209.12362.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0010 | tables/dev/val_tab_0020.tex | |
2211.01233 | val_tab_0022 | Finally, we show that ViTCA does not dramatically suffer when no explicit positioning is used—in contrast to typical Transformer-based models—as cells are still able to localize themselves by relying on their stored hidden information. | Supported | Quantitative ablation for denoising autoencoding with ViTCA (unless otherwise stated via prefix) on CelebA [ 47 ] . Boldface and underlining denote best and second best results. Italicized items denote baseline configuration settings. † Trained with gradient checkpointing [ 44 ] , which slightly alters round-off error ... | table | tables_png/dev/val_tab_0022.png | As shown in Tab. 2 , ViTCA benefits from an increase to most CA and Transformer-centric parameters, at the cost of computational complexity and/or an increase in parameter count. A noticeable decrease in performance is observed when embed size d\!=\!512 , most likely due to the vast increase in parameter count necessit... | ml | no | Change the cell values | papers/dev/ml_2211.01233.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0011 | tables/dev/val_tab_0022.tex | |
2211.01233 | val_tab_0023 | Finally, we show that ViTCA does not dramatically suffer when no explicit positioning is used—in contrast to typical Transformer-based models—as cells are still able to localize themselves by relying on their stored hidden information. | Refuted | Quantitative ablation for denoising autoencoding with ViTCA (unless otherwise stated via prefix) on CelebA [ 47 ] . Boldface and underlining denote best and second best results. Italicized items denote baseline configuration settings. † Trained with gradient checkpointing [ 44 ] , which slightly alters round-off error ... | table | tables_png/dev/val_tab_0023.png | As shown in Tab. 2 , ViTCA benefits from an increase to most CA and Transformer-centric parameters, at the cost of computational complexity and/or an increase in parameter count. A noticeable decrease in performance is observed when embed size d\!=\!512 , most likely due to the vast increase in parameter count necessit... | ml | no | Change the cell values | papers/dev/ml_2211.01233.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0011 | tables/dev/val_tab_0023.tex | |
2206.15241 | val_tab_0026 | In contrast, deep learning-based post-processing rather hinders performance for ‘heavy rain’ prediction. | Supported | Table 3: Evaluation metrics of KoMet and baseline models for precipitation while 12 variables are utilized for the training. Best performances are marked in bold. | table | tables_png/dev/val_tab_0026.png | Table 3 shows the results for various lead times ranging from 6 to 87 hours, according to the changes of the architectures. Compared to the statistics of GDAPS-KIM for predicting class ‘rain’, the qualities of post-processed predictions with the three networks are consistently improved while that of MetNet is the best. | ml | no | Change the cell values | papers/dev/ml_2206.15241.json | Public Domain | http://creativecommons.org/publicdomain/zero/1.0/ | 0012 | tables/dev/val_tab_0026.tex | |
2206.15241 | val_tab_0027 | In contrast, deep learning-based post-processing rather hinders performance for ‘heavy rain’ prediction. | Refuted | Table 3: Evaluation metrics of KoMet and baseline models for precipitation while 12 variables are utilized for the training. Best performances are marked in bold. | table | tables_png/dev/val_tab_0027.png | Table 3 shows the results for various lead times ranging from 6 to 87 hours, according to the changes of the architectures. Compared to the statistics of GDAPS-KIM for predicting class ‘rain’, the qualities of post-processed predictions with the three networks are consistently improved while that of MetNet is the best. | ml | no | Change the cell values | papers/dev/ml_2206.15241.json | Public Domain | http://creativecommons.org/publicdomain/zero/1.0/ | 0012 | tables/dev/val_tab_0027.tex | |
2210.05883 | val_tab_0029 | Results in Table 1 show that AD-Drop achieves consistent improvement, boosting the average scores of BERT base and RoBERTa base by 0.87 and 0.62, respectively. | Supported | Table 1: Overall results of fine-tuned models on the GLUE benchmark. The symbol \dagger denotes results directly taken from the original papers. The best average results are shown in bold. | table | tables_png/dev/val_tab_0029.png | We report the overall results of the fine-tuned models in Table 1 . We first compare AD-Drop with existing regularization methods on the development sets, including the original fine-tuning, SCAL [ 17 ] , SuperT [ 48 ] , R-Drop [ 18 ] , and HiddenCut [ 15 ] . We observe that AD-Drop surpasses the baselines on most of t... | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0013 | tables/dev/val_tab_0029.tex | |
2210.05883 | val_tab_0030 | Results in Table 1 show that AD-Drop achieves consistent improvement, boosting the average scores of BERT base and RoBERTa base by 0.87 and 0.62, respectively. | Refuted | Table 1: Overall results of fine-tuned models on the GLUE benchmark. The symbol \dagger denotes results directly taken from the original papers. The best average results are shown in bold. | table | tables_png/dev/val_tab_0030.png | We report the overall results of the fine-tuned models in Table 1 . We first compare AD-Drop with existing regularization methods on the development sets, including the original fine-tuning, SCAL [ 17 ] , SuperT [ 48 ] , R-Drop [ 18 ] , and HiddenCut [ 15 ] . We observe that AD-Drop surpasses the baselines on most of t... | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0013 | tables/dev/val_tab_0030.tex | |
2210.05883 | val_tab_0031 | Second, IGA outperforms GA in some cases. | Supported | Table 2: Results of ablation studies, in which r/w means “replace with” and w/o means “without”. | table | tables_png/dev/val_tab_0031.png | AD-Drop can be implemented with different attribution methods to generate the mask matrix in Eq. ( 1 ), such as integrated gradient attribution (IGA) introduced Eq. ( 3 ), attention weights for attribution (AA), and randomly generating the discard region (RD) in Eq. ( 6 ). We replace the gradient attribution (GA) in Eq... | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0014 | tables/dev/val_tab_0031.tex | |
2210.05883 | val_tab_0032 | Second, IGA outperforms GA in some cases. | Refuted | Table 2: Results of ablation studies, in which r/w means “replace with” and w/o means “without”. | table | tables_png/dev/val_tab_0032.png | AD-Drop can be implemented with different attribution methods to generate the mask matrix in Eq. ( 1 ), such as integrated gradient attribution (IGA) introduced Eq. ( 3 ), attention weights for attribution (AA), and randomly generating the discard region (RD) in Eq. ( 6 ). We replace the gradient attribution (GA) in Eq... | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0014 | tables/dev/val_tab_0032.tex | |
2210.05883 | val_tab_0033 | Third, AD-Drop improves the original BERT base and RoBERTa base with any of the masking strategies, demonstrating the robustness of AD-Drop to overfitting when fine-tuning these models. | Supported | Table 2: Results of ablation studies, in which r/w means “replace with” and w/o means “without”. | table | tables_png/dev/val_tab_0033.png | AD-Drop can be implemented with different attribution methods to generate the mask matrix in Eq. ( 1 ), such as integrated gradient attribution (IGA) introduced Eq. ( 3 ), attention weights for attribution (AA), and randomly generating the discard region (RD) in Eq. ( 6 ). We replace the gradient attribution (GA) in Eq... | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0015 | tables/dev/val_tab_0033.tex | |
2210.05883 | val_tab_0034 | As shown in Table 2 , removing cross-tuning causes noticeable performance degradation on most of the datasets. | Supported | Table 2: Results of ablation studies, in which r/w means “replace with” and w/o means “without”. | table | tables_png/dev/val_tab_0034.png | To verify the effectiveness of the cross-tuning strategy, we ablate it and apply only AD-Drop in all training epochs. | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0016 | tables/dev/val_tab_0034.tex | |
2210.05883 | val_tab_0035 | As shown in Table 2 , removing cross-tuning causes noticeable performance degradation on most of the datasets. | Refuted | Table 2: Results of ablation studies, in which r/w means “replace with” and w/o means “without”. | table | tables_png/dev/val_tab_0035.png | To verify the effectiveness of the cross-tuning strategy, we ablate it and apply only AD-Drop in all training epochs. | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0016 | tables/dev/val_tab_0035.tex | |
2210.05883 | val_tab_0036 | From Table 6 , we can see that RoBERTa with AD-Drop achieves better generalization, where AD-Drop boosts the performance by 0.66 on HANS and 3.35 on PAWS-X, illustrating that the model trained with AD-Drop generalizes better to OOD data. | Supported | Table 6: Testing AD-Drop on OOD datasets. | table | tables_png/dev/val_tab_0036.png | To further demonstrate AD-Drop is beneficial to reducing overfitting, we test AD-Drop with RoBERTa base on two out-of-distribution (OOD) datasets, i.e., HANS and PAWS-X. For HANS, we use the checkpoints trained on MNLI and test their performance on the validation set (the test set is not supplied). For PAWS-X, we use t... | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0017 | tables/dev/val_tab_0036.tex | |
2210.05883 | val_tab_0037 | As shown in Table 7 , although IGA achieves more favorable performance on one of the datasets, it requires higher computational costs than its counterparts, especially when applied in all the layers. | Supported | Table 7: Results of performance and computational cost of AD-Drop with different masking strategies (GA, IGA, AA, and RD) relative to the original fine-tuning. The symbol \ddagger means AD-Drop is only applied in the first layer. BERT is chosen as the base model. | table | tables_png/dev/val_tab_0037.png | To analyze the computational efficiency, we quantitatively study the computational cost of AD-Drop with different dropping strategies (GA, IGA, AA, and RD) relative to the original fine-tuning on CoLA, STS-B, MRPC, and RTE. BERT is chosen as the base model for this experiment. | ml | no | Change the cell values | papers/dev/ml_2210.05883.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0018 | tables/dev/val_tab_0037.tex | |
2209.15246 | val_tab_0038 | Note that resistance against the attack to both in- and out-sets is much harder than other cases since the perturbation budget has effectively been doubled. | Supported | Table 1: OOD detection AUROC under attack with \epsilon=\frac{8}{255} for various methods trained with CIFAR-10 or CIFAR-100 as the closed set. A clean evaluation is one where no attack is made on the data, whereas an in/out evaluation means that the corresponding data is attacked. The best and second-best results are ... | table | tables_png/dev/val_tab_0038.png | OOD detection under adversarial attack: To perform a comprehensive study, AUROC is computed in four different settings for each method. First, the standard OOD detection without any attack is conducted (Clean). Next, either in- or out-datasets are attacked (In/Out). Finally, both the in- and out-sets are attacked (In a... | ml | no | Change the cell values | papers/dev/ml_2209.15246.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0019 | tables/dev/val_tab_0038.tex | |
2209.15246 | val_tab_0040 | Moreover, we check the effect of attacking the generated images, which leads to a lower AUROC score. | Supported | Table 3: Ablation study on our method. Other choices for the feature extractor, training method, and generated data attacking are tested on CIFAR-10, but none of them is as effective as the setting used in the ATD. | table | tables_png/dev/val_tab_0040.png | Ablation study: ATD uses HAT as the feature extractor and performs adversarial training on open and closed sets to robustify the discriminator. As an ablation study using CIFAR-10 as the in-distribution data, we replace HAT with a standard trained model to check its effectiveness. Also, in another experiment, we replac... | ml | no | Change the cell values | papers/dev/ml_2209.15246.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0020 | tables/dev/val_tab_0040.tex | |
2209.15246 | val_tab_0041 | Moreover, we check the effect of attacking the generated images, which leads to a lower AUROC score. | Refuted | Table 3: Ablation study on our method. Other choices for the feature extractor, training method, and generated data attacking are tested on CIFAR-10, but none of them is as effective as the setting used in the ATD. | table | tables_png/dev/val_tab_0041.png | Ablation study: ATD uses HAT as the feature extractor and performs adversarial training on open and closed sets to robustify the discriminator. As an ablation study using CIFAR-10 as the in-distribution data, we replace HAT with a standard trained model to check its effectiveness. Also, in another experiment, we replac... | ml | no | Change the cell values | papers/dev/ml_2209.15246.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0020 | tables/dev/val_tab_0041.tex | |
2207.00461 | val_tab_0042 | As shown in Table 1 , ELIRL requires little extra training time versus MaxEnt IRL, even with the optional \mathbf{s}^{(t)} re-optimization, and runs significantly faster than GPIRL. | Supported | Table 1 : The average learning time per task. The standard error is reported after the \pm . | table | tables_png/dev/val_tab_0042.png | ml | no | Supported_claim_only | papers/dev/ml_2207.00461.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | no pair | tables/dev/val_tab_0042.tex | ||
2211.12551 | val_tab_0044 | As shown in Table 2 , the proposed method outperforms all three baselines. | Supported | Table 2 : Character-level language modeling results on Penn Tree Bank in test set bpd. | table | tables_png/dev/val_tab_0044.png | We use the Penn Tree Bank dataset with standard processing from Mikolov et al. [ 29 ] , which contains around 5M characters and a character-level vocabulary size of 50 . The data is split into sentences with a maximum sequence length of 288 . We compare with three competitive normalizing-flow-based models: Bipartite fl... | ml | no | Supported_claim_only | papers/dev/ml_2211.12551.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | no pair | tables/dev/val_tab_0044.tex | |
2405.17991 | val_tab_0054 | VeLoRA lowers the memory requirement of SSF [ 26 ] by 16% with only a minor degradation (0.1 pp ) in accuracy. | Supported | Results on a subset of the VTAB-1k benchmark. All methods use a ViT-Base-224/16 model pre-trained on ImageNet-21k. The batch sizes and ranks are the same across all tasks. | table | tables_png/dev/val_tab_0054.png | We conduct experiments evaluating the performance of VeLoRA for full-tuning and how it complements other PEFT methods. In Tab. 1 we reproduce a large set of results for LoRA [ 17 ] , SSF [ 26 ] , and Hydra [ 20 ] on a subset of the VTAB-1K benchmark, where the sub-token size for each experiment is given in the Appendix... | ml | no | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0021 | tables/dev/val_tab_0054.tex | |
2405.17991 | val_tab_0055 | VeLoRA lowers the memory requirement of SSF [ 26 ] by 16% with only a minor degradation (0.1 pp ) in accuracy. | Refuted | Results on a subset of the VTAB-1k benchmark. All methods use a ViT-Base-224/16 model pre-trained on ImageNet-21k. The batch sizes and ranks are the same across all tasks. | table | tables_png/dev/val_tab_0055.png | We conduct experiments evaluating the performance of VeLoRA for full-tuning and how it complements other PEFT methods. In Tab. 1 we reproduce a large set of results for LoRA [ 17 ] , SSF [ 26 ] , and Hydra [ 20 ] on a subset of the VTAB-1K benchmark, where the sub-token size for each experiment is given in the Appendix... | ml | no | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0021 | tables/dev/val_tab_0055.tex | |
2405.17991 | val_tab_0056 | It lowers the memory requirement of Hydra [ 20 ] by 7% while improving the accuracy by 0.1 pp . | Supported | Results on a subset of the VTAB-1k benchmark. All methods use a ViT-Base-224/16 model pre-trained on ImageNet-21k. The batch sizes and ranks are the same across all tasks. | table | tables_png/dev/val_tab_0056.png | We conduct experiments evaluating the performance of VeLoRA for full-tuning and how it complements other PEFT methods. In Tab. 1 we reproduce a large set of results for LoRA [ 17 ] , SSF [ 26 ] , and Hydra [ 20 ] on a subset of the VTAB-1K benchmark, where the sub-token size for each experiment is given in the Appendix... | ml | yes | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0022 | tables/dev/val_tab_0056.tex | |
2405.17991 | val_tab_0057 | Finally, it lowers the memory requirements of LoRA [ 17 ] by 4% while improving the accuracy by 0.6 pp | Supported | Results on a subset of the VTAB-1k benchmark. All methods use a ViT-Base-224/16 model pre-trained on ImageNet-21k. The batch sizes and ranks are the same across all tasks. | table | tables_png/dev/val_tab_0057.png | We conduct experiments evaluating the performance of VeLoRA for full-tuning and how it complements other PEFT methods. In Tab. 1 we reproduce a large set of results for LoRA [ 17 ] , SSF [ 26 ] , and Hydra [ 20 ] on a subset of the VTAB-1K benchmark, where the sub-token size for each experiment is given in the Appendix... | ml | no | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0023 | tables/dev/val_tab_0057.tex | |
2405.17991 | val_tab_0058 | Finally, it lowers the memory requirements of LoRA [ 17 ] by 4% while improving the accuracy by 0.6 pp | Refuted | Results on a subset of the VTAB-1k benchmark. All methods use a ViT-Base-224/16 model pre-trained on ImageNet-21k. The batch sizes and ranks are the same across all tasks. | table | tables_png/dev/val_tab_0058.png | We conduct experiments evaluating the performance of VeLoRA for full-tuning and how it complements other PEFT methods. In Tab. 1 we reproduce a large set of results for LoRA [ 17 ] , SSF [ 26 ] , and Hydra [ 20 ] on a subset of the VTAB-1K benchmark, where the sub-token size for each experiment is given in the Appendix... | ml | no | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0023 | tables/dev/val_tab_0058.tex | |
2405.17991 | val_tab_0059 | Our method further reduces the memory needed for training to 2.23GB, an improvement of 18% compared to LoRA, and 45% compared to GaLore, while still reaching higher results than either of them. | Supported | Comparison of our method with full fine-tuning, GaLore and LORA on GLUE benchmark using pre-trained RoBERTa-Base. Our method reaches the best overall results while showing significant memory improvements, especially compared to GaLore. We bold the best results from the considered PEFT methods. The GPU memory is measure... | table | tables_png/dev/val_tab_0059.png | We now evaluate our method with M=16 on various language tasks, using RoBERTa-Base [ 28 ] in the GLUE benchmark, and compare it with full fine-tuning, LoRA [ 17 ] and GaLore [ 53 ] , presenting the results in Tab. 2 .
We observe that both GaLore and LoRA lower the memory requirements compared to fine-tuning from 4.64GB... | ml | no | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0024 | tables/dev/val_tab_0059.tex | |
2405.17991 | val_tab_0060 | Our method further reduces the memory needed for training to 2.23GB, an improvement of 18% compared to LoRA, and 45% compared to GaLore, while still reaching higher results than either of them. | Refuted | Comparison of our method with full fine-tuning, GaLore and LORA on GLUE benchmark using pre-trained RoBERTa-Base. Our method reaches the best overall results while showing significant memory improvements, especially compared to GaLore. We bold the best results from the considered PEFT methods. The GPU memory is measure... | table | tables_png/dev/val_tab_0060.png | We now evaluate our method with M=16 on various language tasks, using RoBERTa-Base [ 28 ] in the GLUE benchmark, and compare it with full fine-tuning, LoRA [ 17 ] and GaLore [ 53 ] , presenting the results in Tab. 2 .
We observe that both GaLore and LoRA lower the memory requirements compared to fine-tuning from 4.64GB... | ml | no | Change the cell values | papers/dev/ml_2405.17991.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0024 | tables/dev/val_tab_0060.tex | |
2405.14903 | val_tab_0062 | Our method enhances the speed of both forward and backward computations compared to the previous state-of-the-art approach, benefiting both gradient-based and gradient-free optimization techniques. | Supported | Table 2 : Time Performance. Our method achieves one order of magnitude speedup across all resolutions compared to PhiFlow in both forward simulation and backward gradient propagation. | table | tables_png/dev/val_tab_0062.png | In this experiment, we demonstrate the performance efficiency of our framework through a comparison with PhiFlow. While PhiFlow operates with a TensorFlow-GPU backend, our framework is implemented in CUDA C++ and features a high-performance Geometric-Multigrid-Preconditioned-Conjugate-Gradient Poisson solver [ 17 ] . W... | ml | no | Supported_claim_only | papers/dev/ml_2405.14903.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | tables/dev/val_tab_0062.tex | |
2405.14507 | val_tab_0064 | The experimental results, as shown in Table 2 , reveal that enhancing the strong activation of SCMoE can further boost MoE models’ performance. | Supported | Table 2: Experimental results of different strong activations.
We set the weak activation with rank- 2 routing.
For each benchmark,
we select the top- k routing yielding the best performance in Figure 1 as the ideal strong activation.
The specific hyperparameter settings can be found in Table 9 . | table | tables_png/dev/val_tab_0064.png | As revealed by Figure 1 , using default top-2 routing is not optimal for all tasks. For instance, top-3 routing yields best results on GSM8K, while top-4 routing achieves the highest accuracy on HumanEval and StrategyQA. This leads us to consider whether enhancing the strong activation in SCMoE can further unlock MoE m... | ml | no | Change the cell values | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0025 | tables/dev/val_tab_0064.tex | |
2405.14507 | val_tab_0065 | The experimental results, as shown in Table 2 , reveal that enhancing the strong activation of SCMoE can further boost MoE models’ performance. | Refuted | Table 2: Experimental results of different strong activations.
We set the weak activation with rank- 2 routing.
For each benchmark,
we select the top- k routing yielding the best performance in Figure 1 as the ideal strong activation.
The specific hyperparameter settings can be found in Table 9 . | table | tables_png/dev/val_tab_0065.png | As revealed by Figure 1 , using default top-2 routing is not optimal for all tasks. For instance, top-3 routing yields best results on GSM8K, while top-4 routing achieves the highest accuracy on HumanEval and StrategyQA. This leads us to consider whether enhancing the strong activation in SCMoE can further unlock MoE m... | ml | no | Change the cell values | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0025 | tables/dev/val_tab_0065.tex | |
2405.14507 | val_tab_0066 | The results in Table 3 show that SCMoE increases the decoding time by a factor of 1.30x compared to greedy. | Supported | Table 3: Averaged decoding latency for each method. CS is short for contrastive search and CD is short for contrastive decoding. We set k = 3 for ensemble routing, while for dynamic routing we set threshold = 0.5. The speeds are tested on 4 A100 40G with batch size = 1. | table | tables_png/dev/val_tab_0066.png | We further evaluate the impact of SCMoE on decoding latency and compare it with other methods on Mixtral 8x7B. Specifically, we first input 32 tokens to each method and then force them to generate a sequence of 512 tokens to calculate the latency. | ml | no | Supported_claim_only | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | tables/dev/val_tab_0066.tex | |
2405.14507 | val_tab_0067 | Specifically, compared to greedy baseline, SCMoE demonstrates improvements across all tasks: it enhances mathematical reasoning by 1.82 on GSM8K, commonsense reasoning by 2.58 on StrategyQA, code generation by 2.00 on MBPP, and 1.22 on HumanEval. | Supported | Table 4: Experimental results on GSM8K, StrategyQA, MBPP and HumanEval with DeepSeekMoE-16B. We report the best results for each method here. The performance of each method with different hyperparameters can be found in the Appendix Table 10 . | table | tables_png/dev/val_tab_0067.png | We further explore the adaptability of SCMoE to other MoE models. We conduct experiments on DeepSeekMoE-16B [ 5 ] . DeepSeekMoE-16B employs fine-grained expert segmentation and shared expert isolation routing strategies, which is different from Mixtral 8x7B [ 17 ] . We detail the hyperparameters settings of experiments... | ml | yes | Supported_claim_only | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | tables/dev/val_tab_0067.tex | |
2405.14507 | val_tab_0068 | These results demonstrate that SCMoE can be successfully applied to other MoE models. | Supported | Table 4: Experimental results on GSM8K, StrategyQA, MBPP and HumanEval with DeepSeekMoE-16B. We report the best results for each method here. The performance of each method with different hyperparameters can be found in the Appendix Table 10 . | table | tables_png/dev/val_tab_0068.png | We further explore the adaptability of SCMoE to other MoE models. We conduct experiments on DeepSeekMoE-16B [ 5 ] . DeepSeekMoE-16B employs fine-grained expert segmentation and shared expert isolation routing strategies, which is different from Mixtral 8x7B [ 17 ] . We detail the hyperparameters settings of experiments... | ml | yes | Supported_claim_only | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | tables/dev/val_tab_0068.tex | |
2405.14800 | val_tab_0069 | We observe that under this over-training scenario, both of our methods nearly achieve ideal binary classification effectiveness. | Supported | Results under Over-training setting. We mark the best and second-best results for each metric in bold and underline , respectively. Additionally, the best results from baselines are marked in blue for comparison. | table | tables_png/dev/val_tab_0069.png | Over-training scenario. In Tab. 1 , models are trained for excessive steps on all three datasets, resulting in significant overfitting. | ml | no | Supported_claim_only | papers/dev/ml_2405.14800.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | no pair | tables/dev/val_tab_0069.tex | |
2405.14800 | val_tab_0070 | The best value of ASR and AUC of baseline methods decreases to around 65%, and the best value of TPR@1%FPR decreases to around 5%, indicating insufficient effectiveness of previous member inference methods in real-world training scenarios of text-to-image diffusion models. | Supported | Results under Real-world training setting. We also highlight key results according to Tab. 1 . | table | tables_png/dev/val_tab_0070.png | Real-world training scenario. In Tab. 2 , we adjust the training steps simulating real-word training scenario [ 19 ] and utilize default data augmentation [ 20 ] . | ml | other sources | Change the cell values | papers/dev/ml_2405.14800.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0026 | tables/dev/val_tab_0070.tex | |
2206.00050 | val_tab_0072 | Thanks to this design, our method requires only a very small number of additional parameters on top of the base network, e.g., converting a single ResNet-18 model to an ensemble of 16 models increases the parameter count by 1.3%, compared to an increase by 1500% when setting up a naïve ensemble, see Table 1 . | Supported | Table 1: Memory and inference complexity comparison (CIFAR-10/100 datasets): Number of additional trainable parameters to have 16 ensemble members for different backbones. The inference time (mult-adds) shown corresponds to the mean GPU time (number of multiply-add operations) required to run a forward pass for a batch... | table | tables_png/dev/val_tab_0072.png | Inspired by that line of work, we propose a new, efficient ensemble method, FiLM-Ensemble. Our method adapts
feature-wise linear modulation as an alternative way to construct an ensemble for (epistemic) uncertainty estimation. FiLM-Ensemble greatly reduces the computational overhead compared to the naïve ensemble appro... | ml | yes | Supported_claim_only | papers/dev/ml_2206.00050.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | tables/dev/val_tab_0072.tex | |
2205.13790 | val_tab_0073 | As shown in Table 4 , when we directly evaluate detectors trained without robustness augmentation, BEVFusion shows higher accuracy than the LiDAR-only stream and vanilla LiDAR-camera fusion approach in TransFusion. | Supported | Table 4: Results on robustness setting of object failure cases. Here, we report the results of baseline and our method that trained on the nuScenes dataset with and without the proposed robustness augmentation (Aug.). All settings are the same as in Table 3 . | table | tables_png/dev/val_tab_0073.png | LiDAR fails to receive object reflection points. Here exist common scenarios when LiDAR fails to receive points from the object. For example, on rainy days, the reflection rate of some common objects is below the threshold of LiDAR hence causing the issue of object failure AnonymousBenchmark . To simulate such a scenar... | ml | no | Change the cell values | papers/dev/ml_2205.13790.json | Public Domain | http://creativecommons.org/publicdomain/zero/1.0/ | 0027 | tables/dev/val_tab_0073.tex | |
2205.13790 | val_tab_0074 | When we finetune detectors on the robust augmented training set, BEVFusion largely improves PointPillars, CenterPoint, and TransFusion-L by 28.9%, 22.7%, and 15.7% mAP. | Supported | Table 4: Results on robustness setting of object failure cases. Here, we report the results of baseline and our method that trained on the nuScenes dataset with and without the proposed robustness augmentation (Aug.). All settings are the same as in Table 3 . | table | tables_png/dev/val_tab_0074.png | LiDAR fails to receive object reflection points. Here exist common scenarios when LiDAR fails to receive points from the object. For example, on rainy days, the reflection rate of some common objects is below the threshold of LiDAR hence causing the issue of object failure AnonymousBenchmark . To simulate such a scenar... | ml | no | Supported_claim_only | papers/dev/ml_2205.13790.json | Public Domain | http://creativecommons.org/publicdomain/zero/1.0/ | no pair | tables/dev/val_tab_0074.tex | |
2205.14612 | val_tab_0075 | A first observation is that one can transfer these weights to deeper ResNets without significantly affecting the test accuracy of the model: it remains above 94.5\% on CIFAR-10 and 72\% on ImageNet. | Supported | Table 2: Test accuracy (ResNet) | table | tables_png/dev/val_tab_0075.png | In addition, we also want our pretrained model to verify assumption 2 so we consider the following setup. On CIFAR (resp. ImageNet) we train a ResNet with 4 (resp. 8) blocks in each layer, where weights are tied within each layer. | ml | yes | Change the cell values | papers/dev/ml_2205.14612.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0028 | tables/dev/val_tab_0075.tex | |
2503.07300 | val_tab_0076 | Despite being executed on a high-performance server, the CMAES method mosleh2020hardware requires 144 seconds to process a 4K input image, which is considerably slow. | Supported | Experimental results demonstrating efficiency across varying input resolutions. Our method significantly outperforms other methods, achieving a speed enhancement of 260 times relative to the cascaded proxy method tseng2022neural at 720P resolution, and 117 times faster than the CMAES approach hansen2006cma ; mosleh2020... | table | tables_png/dev/val_tab_0076.png | As indicated in Tab. 2 , our method demonstrated superior efficiency, requiring only 1.23 seconds per execution with 4K input. While the monolithic proxy-based method outperformed the search-based method in terms of speed, it faced limitations as input resolutions increased, leading to out-of-memory (OOM) errors once G... | ml | no | Supported_claim_only | papers/dev/ml_2503.07300.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | no pair | tables/dev/val_tab_0076.tex | |
2203.01212 | val_tab_0077 | In the experiments of LiPopt, we only used LiPopt-2. | Supported | Table 1: \ell_{\infty} -FGL estimation of various methods: DGeoLIP and NGeoLIP induce the same values on two layer networks. DGeoLIP always produces tighter estimations than LiPopt and MP do. | table | tables_png/dev/val_tab_0077.png | ml | other sources | Change the cell values | papers/dev/ml_2203.01212.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0029 | tables/dev/val_tab_0077.tex | ||
2203.01212 | val_tab_0078 | In the experiments of LiPopt, we only used LiPopt-2. | Refuted | Table 1: \ell_{\infty} -FGL estimation of various methods: DGeoLIP and NGeoLIP induce the same values on two layer networks. DGeoLIP always produces tighter estimations than LiPopt and MP do. | table | tables_png/dev/val_tab_0078.png | ml | other sources | Change the cell values | papers/dev/ml_2203.01212.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0029 | tables/dev/val_tab_0078.tex | ||
2408.16862 | val_tab_0079 | We see that p-dLDS broadly outperforms existing methods in all metrics and significantly improves inference for decomposed models. | Supported | Table 1: Metrics for synthetic dynamical systems. Bold means best performance. (\uparrow) indicates higher score is better while (\downarrow) indicates that lower is better. ✗ indicates that value diverged towards -\infty . Switch events for decomposed models are defined as times where the active set of DOs change from... | table | tables_png/dev/val_tab_0079.png | Table 1 summarizes our quantitative evaluations on three metrics: 1) the mean squared error (MSE) between the learned and ground truth latent dynamics, 2) the MSE between the inferred and true switch rate to determine agreement of the discrete switching behavior,
and 3) the 100-step inference R^{2} to demonstrate that ... | ml | other sources | Change the cell values | papers/dev/ml_2408.16862.json | CC BY-SA 4.0 | http://creativecommons.org/licenses/by-sa/4.0/ | 0030 | tables/dev/val_tab_0079.tex | |
2408.16862 | val_tab_0080 | We see that p-dLDS broadly outperforms existing methods in all metrics and significantly improves inference for decomposed models. | Refuted | Table 1: Metrics for synthetic dynamical systems. Bold means best performance. (\uparrow) indicates higher score is better while (\downarrow) indicates that lower is better. ✗ indicates that value diverged towards -\infty . Switch events for decomposed models are defined as times where the active set of DOs change from... | table | tables_png/dev/val_tab_0080.png | Table 1 summarizes our quantitative evaluations on three metrics: 1) the mean squared error (MSE) between the learned and ground truth latent dynamics, 2) the MSE between the inferred and true switch rate to determine agreement of the discrete switching behavior,
and 3) the 100-step inference R^{2} to demonstrate that ... | ml | other sources | Change the cell values | papers/dev/ml_2408.16862.json | CC BY-SA 4.0 | http://creativecommons.org/licenses/by-sa/4.0/ | 0030 | tables/dev/val_tab_0080.tex | |
2408.16862 | val_tab_0081 | Moreover, this leads to improved estimation of latent dynamics, a switching rate that agrees with the true system, and improved multistep inference performance as shown in Table 1 . | Supported | Table 1: Metrics for synthetic dynamical systems. Bold means best performance. (\uparrow) indicates higher score is better while (\downarrow) indicates that lower is better. ✗ indicates that value diverged towards -\infty . Switch events for decomposed models are defined as times where the active set of DOs change from... | table | tables_png/dev/val_tab_0081.png | In figure 2 D, we see that rSLDS does not distinguish between the different speeds along the outer and inner sections of the attractor. Instead, the discrete states obscure the continuum of speeds by incorrectly grouping all activity in each lobe into a single regime. Furthermore, we observe that dLDS is limited withou... | ml | other sources | Change the cell values | papers/dev/ml_2408.16862.json | CC BY-SA 4.0 | http://creativecommons.org/licenses/by-sa/4.0/ | 0031 | tables/dev/val_tab_0081.tex | |
2501.08508 | val_tab_0082 | We see that FuncMol slightly improves VoxMol and both models perform worse compared to the equivariant point-cloud based baselines. | Supported | Table 1: QM9 results w.r.t. test set for 10000 samples per model. \uparrow / \downarrow indicate that higher/lower numbers are better. The row data are randomly sampled molecules from the validation set. We report 1-sigma error bars over 3 sampling runs. | table | tables_png/dev/val_tab_0082.png | Table 1 report the metrics described in Section ˜ 5.1 . | ml | other sources | Change the cell values | papers/dev/ml_2501.08508.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0032 | tables/dev/val_tab_0082.tex | |
2501.08508 | val_tab_0083 | We see that FuncMol slightly improves VoxMol and both models perform worse compared to the equivariant point-cloud based baselines. | Refuted | Table 1: QM9 results w.r.t. test set for 10000 samples per model. \uparrow / \downarrow indicate that higher/lower numbers are better. The row data are randomly sampled molecules from the validation set. We report 1-sigma error bars over 3 sampling runs. | table | tables_png/dev/val_tab_0083.png | Table 1 report the metrics described in Section ˜ 5.1 . | ml | other sources | Change the cell values | papers/dev/ml_2501.08508.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0032 | tables/dev/val_tab_0083.tex | |
2501.08508 | val_tab_0084 | We note that sampling time of FuncMol is an order of magnitude better than baselines. | Supported | Table 1: QM9 results w.r.t. test set for 10000 samples per model. \uparrow / \downarrow indicate that higher/lower numbers are better. The row data are randomly sampled molecules from the validation set. We report 1-sigma error bars over 3 sampling runs. | table | tables_png/dev/val_tab_0084.png | Table 1 report the metrics described in Section ˜ 5.1 . We see that FuncMol slightly improves VoxMol and both models perform worse compared to the equivariant point-cloud based baselines. | ml | other sources | Change the cell values | papers/dev/ml_2501.08508.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0033 | tables/dev/val_tab_0084.tex | |
2211.12551 | val_fig_0001 | It shows that both eParam and eFlow are reasonable pruning strategy, however, as we increase the percentage of pruned parameters, eFlow has less log-likelihoods drop compared with eParam . | Supported | (a) Comparison of heuristics eRand , eParam , and eFlow . Heuristic eFlow can prune up to 80% of the parameters without much loglikelihoods decrease.; (b) Histogram of parameters before (the same as in Figure 1 ) and after pruning. The parameter values take higher significance after pruning.; Empirical evaluation of th... | figure | figures/dev/val_fig_0001.png | Figure 5(a) compares the effect of pruning heuristics eParam , eFlow , as well as an uninformed strategy, prune randomly, which we denote as eRand . | ml | yes | Legend Swap | papers/dev/ml_2211.12551.json | CC BY-NC-SA 4.0 | http://creativecommons.org/licenses/by-nc-sa/4.0/ | 0034 | null | |
2211.16499 | val_fig_0002 | Thereafter, Swin becomes comparatively much more robust as occlusion becomes severe. | Supported | Figure 3: Random patch drop occlusion study on ConvNext and Swin networks on the ImageNet-1k validation set. Swin Transformers are slightly more robust to this type of artificial occlusion than ConvNext networks when the information loss is small, although they become comparatively much stronger as the information loss... | figure | figures/dev/val_fig_0002.png | For (1), we provide evidence that this difference between ResNets and DeiTs is not due to the nature of transformer vs. conv layers. In order to do so, we re-run the random patch drop experiments in [ 47 ] on all pairs of ConvNext and Swin networks. We present results for different levels of information loss in Fig. 3 ... | ml | yes | Legend Swap | papers/dev/ml_2211.16499.json | same image | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0035 | null |
2211.16499 | val_fig_0003 | Thereafter, Swin becomes comparatively much more robust as occlusion becomes severe. | Refuted | Figure 3: Random patch drop occlusion study on ConvNext and Swin networks on the ImageNet-1k validation set. Swin Transformers are slightly more robust to this type of artificial occlusion than ConvNext networks when the information loss is small, although they become comparatively much stronger as the information loss... | figure | figures/dev/val_fig_0003.png | For (1), we provide evidence that this difference between ResNets and DeiTs is not due to the nature of transformer vs. conv layers. In order to do so, we re-run the random patch drop experiments in [ 47 ] on all pairs of ConvNext and Swin networks. We present results for different levels of information loss in Fig. 3 ... | ml | yes | Legend Swap | papers/dev/ml_2211.16499.json | same image | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0035 | null |
2211.16499 | val_fig_0004 | For Base-22k and Large, the difference becomes large after the scale multiplier is under 0.6. | Supported | Figure 5: Counterfactual study of all sizes of ConvNext and Swin networks for object scale. | figure | figures/dev/val_fig_0004.png | In order to compare the robustness of ConvNext and Swin networks to scale, we first generate images of all main object models with unit object scale. Although objects have different volume in this state (e.g. microwaves are larger than spatulas), they take a fair amount of space in the frame and don’t present much of a... | ml | yes | Graph Swap | papers/dev/ml_2211.16499.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0036 | null | |
2403.19137 | val_fig_0005 | With task-specific encoders (right), the cross-task class centroids are further apart than those with one global encoder (left). | Supported | Figure 3 : Need for task-specific distribution encoders: Cosine distance between the centroids of class-specific latent variables produced without ( left ) and with ( right ) the use of task-specific encoders on CIFAR100 (10 tasks, 10 classes per task). Centroids of class-specific latent variables are more separable on... | figure | figures/dev/val_fig_0005.png | where \{z^{t}_{i}\}_{m=1}^{M} are the M task-specific MC samples. Task-specific encoders help us sample latent variables that are more discriminative across tasks. This is depicted in Fig. 3 using the cosine distance between the embeddings of class-specific samples. | ml | yes | Graph Swap | papers/dev/ml_2403.19137.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0037 | null | |
2403.19137 | val_fig_0006 | With task-specific encoders (right), the cross-task class centroids are further apart than those with one global encoder (left). | Refuted | Figure 3 : Need for task-specific distribution encoders: Cosine distance between the centroids of class-specific latent variables produced without ( left ) and with ( right ) the use of task-specific encoders on CIFAR100 (10 tasks, 10 classes per task). Centroids of class-specific latent variables are more separable on... | figure | figures/dev/val_fig_0006.png | where \{z^{t}_{i}\}_{m=1}^{M} are the M task-specific MC samples. Task-specific encoders help us sample latent variables that are more discriminative across tasks. This is depicted in Fig. 3 using the cosine distance between the embeddings of class-specific samples. | ml | yes | Graph Swap | papers/dev/ml_2403.19137.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0037 | null | |
2402.12416 | val_fig_0007 | As depicted in the figure, AgA with sign selection is aligned towards the fastest update direction, resulting in AgA progressing approximately 6 steps ahead of AgA without Sign at the end of the trajectory. | Supported | (a) Cleanup: AgA with different \lambda and AgA/Sign; (b) Cleanup: comparison between AgA and baselines; (c) Harvest: AgA with different \lambda and AgA/Sign; (d) Cleanup: comparison between AgA and baselines; Comparison of Episodic Social Welfare: our proposed AgA methods along with baselines on Cleanup (figure a and ... | figure | figures/dev/val_fig_0007.png | In addition, Fig. 3 presents a critical comparison between the trajectories of AgA (in red) and AgA without sign alignment (AgA w/o Sign, depicted in purple), as outlined in Corollary 4.3 . This side-by-side comparison covers 40 steps and features highlighted points every tenth step. It provides a visual illustration o... | ml | yes | Legend Swap | papers/dev/ml_2402.12416.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0038 | null | |
2402.12416 | val_fig_0008 | As depicted in the figure, AgA with sign selection is aligned towards the fastest update direction, resulting in AgA progressing approximately 6 steps ahead of AgA without Sign at the end of the trajectory. | Refuted | (a) Cleanup: AgA with different \lambda and AgA/Sign; (b) Cleanup: comparison between AgA and baselines; (c) Harvest: AgA with different \lambda and AgA/Sign; (d) Cleanup: comparison between AgA and baselines; Comparison of Episodic Social Welfare: our proposed AgA methods along with baselines on Cleanup (figure a and ... | figure | figures/dev/val_fig_0008.png | In addition, Fig. 3 presents a critical comparison between the trajectories of AgA (in red) and AgA without sign alignment (AgA w/o Sign, depicted in purple), as outlined in Corollary 4.3 . This side-by-side comparison covers 40 steps and features highlighted points every tenth step. It provides a visual illustration o... | ml | yes | Legend Swap | papers/dev/ml_2402.12416.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0038 | null | |
2205.14612 | val_fig_0010 | Approximated gradients, however, lead to a large test error at small depth, but give the same performance at large depth, hence confirming our results in Prop. 4 and 6 . | Supported | Figure 4: Comparison of the best test errors as a function of depth when using Euler or Heun’s discretization method with or without the adjoint method. | figure | figures/dev/val_fig_0010.png | We then apply a batch norm, a ReLU and iterate relation ( 1 ) where f is a pre-activation basic block (He et al., 2016b ) . We consider the zero residual initialisation: the last batch norm of each basic block is initialized to zero. We consider different values for the depth N and notice that in this setup, the deeper... | ml | yes | Supported_claim_only | papers/dev/ml_2205.14612.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | null | |
2410.17599 | val_fig_0011 | The underfitting delta model had a win rate of 54 when \alpha was 1.0, which increased to 64 when \alpha was 2.0. | Supported | (a) Impact on instruction tuning; (b) Impact on unlearning; Impact of strength coefficient \alpha on performance | figure | figures/dev/val_fig_0011.png | For instruction tuning, we tested values within the range of [0.5, 2] and evaluated them on the first 50 data points of AlpacaEval. As shown in Figure 4(a) , we found that the performance was optimal when the \alpha value was 1.0, and increasing or decreasing \alpha resulted in decreased performance. This is similar to... | ml | yes | Legend Swap | papers/dev/ml_2410.17599.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0039 | null | |
2201.12414 | val_fig_0014 | We also see that acquiring based on the lookahead posteriors incurs only a minimal increase in error compared to the sampling-based method, despite being far more efficient. | Supported | Figure 7: Average RMSE of reconstructions for greedy active feature acquisition. | figure | figures/dev/val_fig_0014.png | Figure 7 presents the root-mean-square error, averaged across the test instances, when imputing \mathbf{x}_{u} with different numbers of acquired features. We see that our models are able to achieve lower error than EDDI. | ml | yes | Graph Flip | papers/dev/ml_2201.12414.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0040 | null | |
2201.12414 | val_fig_0015 | We also see that acquiring based on the lookahead posteriors incurs only a minimal increase in error compared to the sampling-based method, despite being far more efficient. | Refuted | Figure 7: Average RMSE of reconstructions for greedy active feature acquisition. | figure | figures/dev/val_fig_0015.png | Figure 7 presents the root-mean-square error, averaged across the test instances, when imputing \mathbf{x}_{u} with different numbers of acquired features. We see that our models are able to achieve lower error than EDDI. | ml | yes | Graph Flip | papers/dev/ml_2201.12414.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0040 | null | |
2502.17771 | val_fig_0016 | Better sample selection due to jittering subsequently leads to significantly better performance in regression (Fig. 4 (c)). | Supported | Figure 4: Jittering analysis. (a) When trained without jittering, feature extractors easily overfit the noisy training data (yellow-shaded region), while jittering-regularized feature extractors robustly learn from the noisy training data. (b) Overfitted feature extractors (yellow-shaded region) on noisy samples increa... | figure | figures/dev/val_fig_0016.png | Fig. 4 (a) shows that with jittering, the feature extractor exhibits higher accuracy on the clean test data due to its regularization effect. In the sample selection stage (Fig. 4 (b)),
the feature extractor trained without jittering easily overfits the noise, resulting in over-selection and higher ERR (§ 4.2 ). In con... | ml | yes | Legend Swap | papers/dev/ml_2502.17771.json | Additionally, color changed. | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0041 | null |
2502.17771 | val_fig_0017 | Better sample selection due to jittering subsequently leads to significantly better performance in regression (Fig. 4 (c)). | Refuted | Figure 4: Jittering analysis. (a) When trained without jittering, feature extractors easily overfit the noisy training data (yellow-shaded region), while jittering-regularized feature extractors robustly learn from the noisy training data. (b) Overfitted feature extractors (yellow-shaded region) on noisy samples increa... | figure | figures/dev/val_fig_0017.png | Fig. 4 (a) shows that with jittering, the feature extractor exhibits higher accuracy on the clean test data due to its regularization effect. In the sample selection stage (Fig. 4 (b)),
the feature extractor trained without jittering easily overfits the noise, resulting in over-selection and higher ERR (§ 4.2 ). In con... | ml | yes | Legend Swap | papers/dev/ml_2502.17771.json | Additionally, color changed. | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0041 | null |
2502.17771 | val_fig_0018 | ConFrag achieves the lowest ERR while maintaining above-average selection rates, resulting in the best MRAE. | Supported | Figure 5: Selection/ERR/MRAE comparison between ConFrag and strong baselines of CNLCU-H, BMM, DY-S, AUX and Selfie on IMDB-Clean-B. We exclude the performance during the warm-up. | figure | figures/dev/val_fig_0018.png | Selection/ERR/MRAE comparison. Fig. 5 compares ConFrag to five selection and refurbishment baselines of CNLCU-H, BMM, DY-S, AUX, Selfie on IMDB-Clean-B using the selection rate, ERR, and MRAE. Ideally, a model should attain a high selection rate and a low ERR. It is worth noting that the relative importance of ERR and ... | ml | yes | Graph Flip | papers/dev/ml_2502.17771.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0042 | null | |
2502.17771 | val_fig_0019 | ConFrag achieves the lowest ERR while maintaining above-average selection rates, resulting in the best MRAE. | Refuted | Figure 5: Selection/ERR/MRAE comparison between ConFrag and strong baselines of CNLCU-H, BMM, DY-S, AUX and Selfie on IMDB-Clean-B. We exclude the performance during the warm-up. | figure | figures/dev/val_fig_0019.png | Selection/ERR/MRAE comparison. Fig. 5 compares ConFrag to five selection and refurbishment baselines of CNLCU-H, BMM, DY-S, AUX, Selfie on IMDB-Clean-B using the selection rate, ERR, and MRAE. Ideally, a model should attain a high selection rate and a low ERR. It is worth noting that the relative importance of ERR and ... | ml | yes | Graph Flip | papers/dev/ml_2502.17771.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0042 | null | |
2502.17771 | val_fig_0020 | The contrastive fragment pairing demonstrates superior performance to other pairing methods. | Supported | Figure 6: Analysis with 40% symmetric noise. (a) Comparison between the proposed contrastive pairing and other pairings on IMDB-Clean-B.
(b) Comparison between fragment numbers on SHIFT15M-B and IMDB-Clean-B. | figure | figures/dev/val_fig_0020.png | Fragment pairing. Fig. 6 (a) compares contrastive pairing to alternative pairings
using MRAE as a metric. | ml | yes | Legend Swap | papers/dev/ml_2502.17771.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0043 | null | |
2502.17771 | val_fig_0021 | The contrastive fragment pairing demonstrates superior performance to other pairing methods. | Refuted | Figure 6: Analysis with 40% symmetric noise. (a) Comparison between the proposed contrastive pairing and other pairings on IMDB-Clean-B.
(b) Comparison between fragment numbers on SHIFT15M-B and IMDB-Clean-B. | figure | figures/dev/val_fig_0021.png | Fragment pairing. Fig. 6 (a) compares contrastive pairing to alternative pairings
using MRAE as a metric. | ml | yes | Legend Swap | papers/dev/ml_2502.17771.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0043 | null | |
2403.19137 | val_fig_0029 | In Fig. 5(a) , the accuracy is poorer in range [1,10] , grows in range [10,20] , and saturates thereafter. | Supported | (a) Accuracy-runtime tradeoff; (b) Parameter count comparison; Ablations on CIFAR100 showing: (a) performance trade-off with the number of MC samples M , (b) the number of trainable parameters in different finetuning methods. | figure | figures/dev/val_fig_0029.png | We vary the number of MC samples M from 1 to 50. | ml | no | Graph Flip | papers/dev/ml_2403.19137.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0044 | null | |
2403.07937 | val_fig_0032 | On the other hand, increasing training data appears to have only a minor influence on robustness. | Supported | (a); (b); (c); NWER on non-adversarial and adversarial perturbations plotted against the number of parameters (a & c) and hours of training data used to train the DNN (b). Figures (a) and (b) plot all the models, while (c) plots three families of models (indicated by the color). The models within the family share simil... | figure | figures/dev/val_fig_0032.png | To determine if the prevailing practice of training DNNs with more parameters on larger datasets is yielding improvements in robustness, we plot NWER against the number of model parameters and the size of the training data of all the candidate models. These plots are shown in Figures 4(a) and 4(b) . We note that increa... | ml | no | Supported_claim_only | papers/dev/ml_2403.07937.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | null | |
2403.19863 | val_fig_0033 | An intriguing observation is that the decodability of both attributes decreases as the depth of the neural network increases, which is also observed in [ 10 ] . | Supported | (a); (b); Exploring the Effect of depth modulation: (a) illustrates how the linear decodability of features decreases as neural network depth increases, while (b) dives into the training dynamics of MLPs with varying depths under ERM. | figure | figures/dev/val_fig_0033.png | We employ the concept of feature decodability to assess the extent to which the specific features of a given dataset can be reliably decoded from the models with varying depths. Hermann et al . [ 10 ] demonstrated that the visual features can be decoded from the higher layers of untrained models. Additionally, they obs... | ml | no | Graph Flip | papers/dev/ml_2403.19863.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0045 | null | |
2403.19863 | val_fig_0035 | Hence, the disparity in decodability between color and digit attributes becomes significantly more pronounced in a 5-layer MLP in comparison to a 3-layer MLP. | Supported | (a); (b); Exploring the Effect of depth modulation: (a) illustrates how the linear decodability of features decreases as neural network depth increases, while (b) dives into the training dynamics of MLPs with varying depths under ERM. | figure | figures/dev/val_fig_0035.png | As observed in Fig. 2(b) , the initial phases of training for both networks emphasize color attribute (since bias is easy to learn), resulting in a notable enhancement in the decodability of color in both models. Also, as the training progresses, the decodability of the digit is higher in a 3-layer model when compared ... | ml | no | Graph Flip | papers/dev/ml_2403.19863.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0046 | null | |
2403.19863 | val_fig_0036 | As depicted in Fig. 4 , prior to the training, the deep branch demonstrates lower linear decodability for both the digit (core attribute) and color (bias attribute) in comparison to the linear decodability observed in the shallow branch. | Supported | Figure 4 : Training dynamics of DeNetDM during initial stages of training monitored in terms of linear decodability of bias attribute (color) and core attribute (digit). | figure | figures/dev/val_fig_0036.png | In Sec. 3.1 , we discussed the variability in the linear decodability at various depths and how the disparity in linear decodability between biased attributes and target attributes forms a significant motivation for debiasing. To further validate this intuition and identify the elements contributing to its effectivenes... | ml | no | Supported_claim_only | papers/dev/ml_2403.19863.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | null | |
2403.17919 | val_fig_0042 | In particular, LISA requires much less activation memory consumption than LoRA since it does not introduce additional parameters brought by the adaptor. | Supported | Figure 3: GPU memory consumption of LLaMA-2-7B with different methods and batch size 1. | figure | figures/dev/val_fig_0042.png | In Figure 3 , it is worth noticing that the memory reduction in LISA allows LLaMA-2-7B to be trained on a single RTX4090 (24GB) GPU, which makes high-quality fine-tuning affordable even on a laptop computer. | ml | no | Category Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0047 | null | |
2403.17919 | val_fig_0043 | In particular, LISA requires much less activation memory consumption than LoRA since it does not introduce additional parameters brought by the adaptor. | Refuted | Figure 3: GPU memory consumption of LLaMA-2-7B with different methods and batch size 1. | figure | figures/dev/val_fig_0043.png | In Figure 3 , it is worth noticing that the memory reduction in LISA allows LLaMA-2-7B to be trained on a single RTX4090 (24GB) GPU, which makes high-quality fine-tuning affordable even on a laptop computer. | ml | no | Category Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0047 | null | |
2403.17919 | val_fig_0045 | As shown in Figure 4 , LISA provides almost 2.9\times speedup when compared with full-parameter tuning, and \sim 1.5\times speedup against LoRA, partially due to the removal of adaptor structures. | Supported | Figure 4: Single-iteration time cost of LLaMA-2-7B with different methods and batch size 1. | figure | figures/dev/val_fig_0045.png | On top of that, a reduction in memory footprint from LISA also leads to an acceleration in speed. | ml | no | Category Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0048 | null | |
2403.17919 | val_fig_0046 | As shown in Figure 4 , LISA provides almost 2.9\times speedup when compared with full-parameter tuning, and \sim 1.5\times speedup against LoRA, partially due to the removal of adaptor structures. | Refuted | Figure 4: Single-iteration time cost of LLaMA-2-7B with different methods and batch size 1. | figure | figures/dev/val_fig_0046.png | On top of that, a reduction in memory footprint from LISA also leads to an acceleration in speed. | ml | no | Category Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0048 | null | |
2403.17919 | val_fig_0047 | Specifically, Figure 5 highlights the model’s performance in various aspects, particularly LISA’s superiority over all methods in aspects of Writing, Roleplay, and STEM. | Supported | Figure 5: Different aspects of LLaMA-2-70B model on MT-Bench. | figure | figures/dev/val_fig_0047.png | As shown in Table 5 , LISA consistently produces better or on-par performance when compared with LoRA. Furthermore, in instruction-tuning tasks, LISA again surpasses full-parameter training, rendering it a competitive method for this setting. | ml | no | Legend Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0049 | null | |
2403.17919 | val_fig_0048 | Specifically, Figure 5 highlights the model’s performance in various aspects, particularly LISA’s superiority over all methods in aspects of Writing, Roleplay, and STEM. | Refuted | Figure 5: Different aspects of LLaMA-2-70B model on MT-Bench. | figure | figures/dev/val_fig_0048.png | As shown in Table 5 , LISA consistently produces better or on-par performance when compared with LoRA. Furthermore, in instruction-tuning tasks, LISA again surpasses full-parameter training, rendering it a competitive method for this setting. | ml | no | Legend Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0049 | null | |
2403.17919 | val_fig_0049 | On top of that, LISA displays consistently higher performance than LoRA on all subtasks, underscoring LISA’s effectiveness across diverse tasks. | Supported | Figure 5: Different aspects of LLaMA-2-70B model on MT-Bench. | figure | figures/dev/val_fig_0049.png | As shown in Table 5 , LISA consistently produces better or on-par performance when compared with LoRA. Furthermore, in instruction-tuning tasks, LISA again surpasses full-parameter training, rendering it a competitive method for this setting. Specifically, Figure 5 highlights the model’s performance in various aspects,... | ml | no | Legend Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0050 | null | |
2403.17919 | val_fig_0050 | On top of that, LISA displays consistently higher performance than LoRA on all subtasks, underscoring LISA’s effectiveness across diverse tasks. | Refuted | Figure 5: Different aspects of LLaMA-2-70B model on MT-Bench. | figure | figures/dev/val_fig_0050.png | As shown in Table 5 , LISA consistently produces better or on-par performance when compared with LoRA. Furthermore, in instruction-tuning tasks, LISA again surpasses full-parameter training, rendering it a competitive method for this setting. Specifically, Figure 5 highlights the model’s performance in various aspects,... | ml | no | Legend Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0050 | null | |
2403.17919 | val_fig_0051 | The chart also contrasts the yellow LoRA line with the purple Vanilla line, revealing that in large models like the 70B, LoRA does not perform as well as expected, showing only marginal improvements on specific aspects. | Supported | Figure 5: Different aspects of LLaMA-2-70B model on MT-Bench. | figure | figures/dev/val_fig_0051.png | As shown in Table 5 , LISA consistently produces better or on-par performance when compared with LoRA. Furthermore, in instruction-tuning tasks, LISA again surpasses full-parameter training, rendering it a competitive method for this setting. Specifically, Figure 5 highlights the model’s performance in various aspects,... | ml | no | Supported_claim_only | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | null | |
2403.17919 | val_fig_0052 | It is also intriguing to observe from Figure 5 that Vanilla LLaMA-2-70B excelled in Writing, but full-parameter fine-tuning led to a decline in these areas, a phenomenon known as the “Alignment Tax” (Ouyang et al., 2022 ) . | Supported | Figure 5: Different aspects of LLaMA-2-70B model on MT-Bench. | figure | figures/dev/val_fig_0052.png | ml | no | Legend Swap | papers/dev/ml_2403.17919.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0051 | null | ||
2402.12416 | val_fig_0053 | Notably, AgA with \lambda=0.1 underperforms, even when contrasted with AgA’s ablation study. | Supported | (a) MMM2: AgA with various \lambda and AgA/Sign; (b) MMM2: comparison between AgA and baselines; The comparison of win rates on the mixed-motive MMM2 map of SMAC.
The left figure shows results of varying alignment parameter \lambda values in AgA, as well as AgA sans sign selection (AgA/Sign),
The right figure compare t... | figure | figures/dev/val_fig_0053.png | Results. Fig. 5(a) compares average win rates for various \lambda parameters in AgA, also encompassing AgA without sign selection, as stated in Corollary 4.3 . The results suggest comparable performance for \lambda=1 and 100 , with a lower value of 0.1 trailing noticeably. | ml | no | Legend Swap | papers/dev/ml_2402.12416.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0052 | null | |
2402.12416 | val_fig_0054 | Notably, AgA with \lambda=0.1 underperforms, even when contrasted with AgA’s ablation study. | Refuted | (a) MMM2: AgA with various \lambda and AgA/Sign; (b) MMM2: comparison between AgA and baselines; The comparison of win rates on the mixed-motive MMM2 map of SMAC.
The left figure shows results of varying alignment parameter \lambda values in AgA, as well as AgA sans sign selection (AgA/Sign),
The right figure compare t... | figure | figures/dev/val_fig_0054.png | Results. Fig. 5(a) compares average win rates for various \lambda parameters in AgA, also encompassing AgA without sign selection, as stated in Corollary 4.3 . The results suggest comparable performance for \lambda=1 and 100 , with a lower value of 0.1 trailing noticeably. | ml | no | Legend Swap | papers/dev/ml_2402.12416.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0052 | null | |
2405.14507 | val_fig_0056 | However, in our exploratory experiment on Mixtral 8x7B [ 17 ] , we find simply raising the number of activated experts (blue lines in Figure 1 ) does not lead to stable improvements and may even hurt performance on different tasks. | Supported | Figure 1 : Performance comparison between increasing the value of top- k ( i.e. , ensemble routing) and SCMoE. SCMoE surpasses the performance of ensemble routing across various benchmarks. | figure | figures/dev/val_fig_0056.png | In this paper, we investigate the impact of unchosen experts 1 1 1 Unchosen experts refer to the experts not selected by default routing ( e.g. , top-2 routing in Mixtral 8x7B). on the performance of MoE models and explore their suitable usage. A direct hypothesis is that incorporating more experts improves MoE models ... | ml | no | Graph Flip | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0053 | null | |
2405.14507 | val_fig_0057 | For instance, top-3 routing yields best results on GSM8K, while top-4 routing achieves the highest accuracy on HumanEval and StrategyQA. | Supported | Figure 1 : Performance comparison between increasing the value of top- k ( i.e. , ensemble routing) and SCMoE. SCMoE surpasses the performance of ensemble routing across various benchmarks. | figure | figures/dev/val_fig_0057.png | As revealed by Figure 1 , using default top-2 routing is not optimal for all tasks. | ml | no | Graph Swap | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0054 | null | |
2405.14507 | val_fig_0058 | For instance, top-3 routing yields best results on GSM8K, while top-4 routing achieves the highest accuracy on HumanEval and StrategyQA. | Refuted | Figure 1 : Performance comparison between increasing the value of top- k ( i.e. , ensemble routing) and SCMoE. SCMoE surpasses the performance of ensemble routing across various benchmarks. | figure | figures/dev/val_fig_0058.png | As revealed by Figure 1 , using default top-2 routing is not optimal for all tasks. | ml | no | Graph Swap | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0054 | null | |
2405.14507 | val_fig_0059 | Furthermore, SCMoE can enhance the major@20 accuracy from 75.59 to 78.31 (+2.72) on GSM8K. | Supported | Figure 5 : Experimental results on combining SCMoE with self-consistency on GSM8K using Mixtral 8x7B. | figure | figures/dev/val_fig_0059.png | Using self-consistency [ 35 ] for multiple sampling and taking a majority vote to determine the final answer is a common method to improve LLMs’ performance. Therefore, we explore whether SCMoE can combined with self-consistency. For vanilla self-consistency, we use temperature sampling with temperature \tau=0.7 to rea... | ml | no | Legend Swap | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0055 | null | |
2405.14507 | val_fig_0060 | Furthermore, SCMoE can enhance the major@20 accuracy from 75.59 to 78.31 (+2.72) on GSM8K. | Refuted | Figure 5 : Experimental results on combining SCMoE with self-consistency on GSM8K using Mixtral 8x7B. | figure | figures/dev/val_fig_0060.png | Using self-consistency [ 35 ] for multiple sampling and taking a majority vote to determine the final answer is a common method to improve LLMs’ performance. Therefore, we explore whether SCMoE can combined with self-consistency. For vanilla self-consistency, we use temperature sampling with temperature \tau=0.7 to rea... | ml | no | Legend Swap | papers/dev/ml_2405.14507.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | 0055 | null | |
2205.11361 | val_fig_0062 | Figure 2 demonstrates that MPGD can lead to successful optimization of the widening loss whereas the baseline GD and GD with Gaussian perturbations lead to poor solutions. | Supported | Figure 2: Evolution of GD with Gaussian perturbations (left plot) vs. that of MPGD with \gamma=0.7 , \beta=0.5 (middle plot), using \mu=0.02 , \sigma=0.05 , and \eta=0.01 ( m=100 ). Here we see that MPGD leads to successful optimization of the widening valley loss whereas that with Gaussian perturbations fails to conve... | figure | figures/dev/val_fig_0062.png | For the experiments, we start optimizing from the point (u_{0},0) , where u_{0}\sim 5\cdot\mathcal{U}(d) with d=10 , and use the learning rate \eta=0.01 . We study and compare the behavior of the following schemes: (i) baseline (vanilla GD), (ii) GD with uncorrelated Gaussian noise injection instead, and (iii) MPGD. | ml | no | Supported_claim_only | papers/dev/ml_2205.11361.json | CC BY 4.0 | http://creativecommons.org/licenses/by/4.0/ | no pair | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.