Upload ArXiv-Tweets-from-AK.csv
Browse files- ArXiv-Tweets-from-AK.csv +620 -0
ArXiv-Tweets-from-AK.csv
ADDED
|
@@ -0,0 +1,620 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
,id,tweet_text,paper_reference,like_count
|
| 2 |
+
0,1546707909748342784,"High-resource Language-specific Training for Multilingual Neural Machine Translation
|
| 3 |
+
abs: https://t.co/fYrwIPVpV2 https://t.co/b23EVZ6J5O",High-resource Language-specific Training for Multilingual Neural Machine Translation,11
|
| 4 |
+
1,1546669556789387264,"Exploring Length Generalization in Large Language Models
|
| 5 |
+
abs: https://t.co/7Gphb7Q8jJ https://t.co/cCpLTSrXfR",Exploring Length Generalization in Large Language Models,17
|
| 6 |
+
2,1546667351885729792,"LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action
|
| 7 |
+
abs:… https://t.co/lCk3P8KIwM","LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action",32
|
| 8 |
+
3,1546665636734140417,"Scaling the Number of Tasks in Continual Learning
|
| 9 |
+
abs: https://t.co/F4HxAxGUpI https://t.co/cyvXSBKthk",Scaling the Number of Tasks in Continual Learning,47
|
| 10 |
+
4,1546707909748342784,"High-resource Language-specific Training for Multilingual Neural Machine Translation
|
| 11 |
+
abs: https://t.co/fYrwIPVpV2 https://t.co/b23EVZ6J5O",High-resource Language-specific Training for Multilingual Neural Machine Translation,11
|
| 12 |
+
5,1546669556789387264,"Exploring Length Generalization in Large Language Models
|
| 13 |
+
abs: https://t.co/7Gphb7Q8jJ https://t.co/cCpLTSrXfR",Exploring Length Generalization in Large Language Models,17
|
| 14 |
+
6,1546667351885729792,"LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action
|
| 15 |
+
abs:… https://t.co/lCk3P8KIwM","LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action",32
|
| 16 |
+
7,1546665636734140417,"Scaling the Number of Tasks in Continual Learning
|
| 17 |
+
abs: https://t.co/F4HxAxGUpI https://t.co/cyvXSBKthk",Scaling the Number of Tasks in Continual Learning,47
|
| 18 |
+
8,1546379163803721729,"CausalAgents: A Robustness Benchmark for Motion Forecasting using Causal Relationships
|
| 19 |
+
abs: https://t.co/ozIrQ7gx68 https://t.co/gSGfnsZbji",CausalAgents: A Robustness Benchmark for Motion Forecasting using Causal Relationships,53
|
| 20 |
+
9,1546376106122567681,"The Harvard USPTO Patent Dataset: A Large-Scale, Well-Structured, and Multi-Purpose Corpus of Patent Applications
|
| 21 |
+
a… https://t.co/TOPpVPQbM8","The Harvard USPTO Patent Dataset: A Large-Scale, Well-Structured, and Multi-Purpose Corpus of Patent Applications",11
|
| 22 |
+
10,1546375104262725632,"Code Translation with Compiler Representations
|
| 23 |
+
abs: https://t.co/nTT3dmXH4c
|
| 24 |
+
|
| 25 |
+
method improves upon the state of the… https://t.co/wD4SozbilN",Code Translation with Compiler Representations,127
|
| 26 |
+
11,1546363822121820162,"End-to-End Binaural Speech Synthesis
|
| 27 |
+
abs: https://t.co/tR86cSAjQO
|
| 28 |
+
project page: https://t.co/nB1iSV68U2
|
| 29 |
+
|
| 30 |
+
end-to-end… https://t.co/OTzfVZTFqb",End-to-End Binaural Speech Synthesis,58
|
| 31 |
+
12,1545243820496936960,"Cross-Scale Vector Quantization for Scalable Neural Speech Coding
|
| 32 |
+
abs: https://t.co/AbE9rP0ApQ https://t.co/pZXUTNipgs",Cross-Scale Vector Quantization for Scalable Neural Speech Coding,25
|
| 33 |
+
13,1545240373328592897,"Finding Fallen Objects Via Asynchronous Audio-Visual Integration
|
| 34 |
+
abs: https://t.co/mv9Rvl0hFA
|
| 35 |
+
project page:… https://t.co/N8l4zaP9bH",Finding Fallen Objects Via Asynchronous Audio-Visual Integration,33
|
| 36 |
+
14,1545228848391938048,"Back to the Source: Diffusion-Driven Test-Time Adaptation
|
| 37 |
+
abs: https://t.co/5jmESOLQxG https://t.co/cI5UFyQI0B",Back to the Source: Diffusion-Driven Test-Time Adaptation,82
|
| 38 |
+
15,1544897525664169986,"When does Bias Transfer in Transfer Learning?
|
| 39 |
+
abs: https://t.co/tf8FWyf8Ge https://t.co/0l6vy8RHXI",When does Bias Transfer in Transfer Learning?,135
|
| 40 |
+
16,1544865587343630342,"Transformers are Adaptable Task Planners
|
| 41 |
+
abs: https://t.co/6lgFJD2Olt
|
| 42 |
+
|
| 43 |
+
TTP can be pre-trained on multiple preferenc… https://t.co/XrolcxlV22",Transformers are Adaptable Task Planners,82
|
| 44 |
+
17,1544853650316599299,"Ultra-Low-Bitrate Speech Coding with Pretrained Transformers
|
| 45 |
+
abs: https://t.co/rYRe5N7Bqu https://t.co/zOsCY53r2s",Ultra-Low-Bitrate Speech Coding with Pretrained Transformers,34
|
| 46 |
+
18,1544721641049145345,"CLEAR: Improving Vision-Language Navigation with Cross-Lingual, Environment-Agnostic Representations
|
| 47 |
+
|
| 48 |
+
abs:… https://t.co/6ng3UArKdE","CLEAR: Improving Vision-Language Navigation with Cross-Lingual, Environment-Agnostic Representations",52
|
| 49 |
+
19,1544521037274046464,"An Empirical Study of Implicit Regularization in Deep Offline RL
|
| 50 |
+
abs: https://t.co/rCjHkQ2jwL https://t.co/8hJOsVA6D0",An Empirical Study of Implicit Regularization in Deep Offline RL,45
|
| 51 |
+
20,1544519268234153984,"Offline RL Policies Should be Trained to be Adaptive
|
| 52 |
+
abs: https://t.co/kC7TPSOTt2 https://t.co/Ox2D028P33",Offline RL Policies Should be Trained to be Adaptive,34
|
| 53 |
+
21,1544491557293854721,"Efficient Representation Learning via Adaptive Context Pooling
|
| 54 |
+
abs: https://t.co/zZzezhvbN7 https://t.co/xJoStGBSqp",Efficient Representation Learning via Adaptive Context Pooling,163
|
| 55 |
+
22,1544488616734429185,"CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning
|
| 56 |
+
abs:… https://t.co/HqXmDpaUEh",CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning,102
|
| 57 |
+
23,1544485593991811072,"How Much More Data Do I Need? Estimating Requirements for Downstream Tasks
|
| 58 |
+
abs: https://t.co/RNXT4IRIaL https://t.co/uJGrEfgaAv",How Much More Data Do I Need? Estimating Requirements for Downstream Tasks,230
|
| 59 |
+
24,1544483235542990856,"Neural Networks and the Chomsky Hierarchy
|
| 60 |
+
abs: https://t.co/u6Jl2WvKMr
|
| 61 |
+
|
| 62 |
+
sota architectures, such as LSTMs and Trans… https://t.co/DyHnH8Q8z7",Neural Networks and the Chomsky Hierarchy,209
|
| 63 |
+
25,1544207617102331906,"GlowVC: Mel-spectrogram space disentangling model for language-independent text-free voice conversion
|
| 64 |
+
abs:… https://t.co/kFYdKhrhSA",GlowVC: Mel-spectrogram space disentangling model for language-independent text-free voice conversion,19
|
| 65 |
+
26,1544201186739458049,"Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation
|
| 66 |
+
abs:… https://t.co/yL9kWlUYfs",Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation,112
|
| 67 |
+
27,1544193877053161480,"WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents
|
| 68 |
+
abs: https://t.co/8hZyMt90Rv
|
| 69 |
+
pro… https://t.co/eHzGN2GHqj",WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents,52
|
| 70 |
+
28,1544127293660037120,"UserLibri: A Dataset for ASR Personalization Using Only Text
|
| 71 |
+
abs: https://t.co/0bug7OWU42 https://t.co/OMqJSGlqDx",UserLibri: A Dataset for ASR Personalization Using Only Text,9
|
| 72 |
+
29,1543981460964708352,"LaserMix for Semi-Supervised LiDAR Semantic Segmentation
|
| 73 |
+
abs: https://t.co/SvqHy1y7LI
|
| 74 |
+
project page:… https://t.co/jbQtQiDbDy",LaserMix for Semi-Supervised LiDAR Semantic Segmentation,74
|
| 75 |
+
30,1543766808309669889,"Rethinking Optimization with Differentiable Simulation from a Global Perspective
|
| 76 |
+
abs: https://t.co/trEcw4VZb2
|
| 77 |
+
proje… https://t.co/1UsI0q03IL",Rethinking Optimization with Differentiable Simulation from a Global Perspective,94
|
| 78 |
+
31,1543763117515182082,"Visual Pre-training for Navigation: What Can We Learn from Noise?
|
| 79 |
+
abs: https://t.co/Rn5UGvvMMz
|
| 80 |
+
github:… https://t.co/eKeMSlBxVx",Visual Pre-training for Navigation: What Can We Learn from Noise?,134
|
| 81 |
+
32,1543759817449390080,"DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale
|
| 82 |
+
abs:… https://t.co/IbF6IdUDj7",DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale,120
|
| 83 |
+
33,1543757524356272134,"When Does Differentially Private Learning Not Suffer in High Dimensions?
|
| 84 |
+
abs: https://t.co/yws7BhoBaP https://t.co/bD2Gz6B3GU",When Does Differentially Private Learning Not Suffer in High Dimensions?,28
|
| 85 |
+
34,1542740430084792320,"Implicit Neural Spatial Filtering for Multichannel Source Separation in the Waveform Domain
|
| 86 |
+
abs:… https://t.co/3cNoOlr5SD",Implicit Neural Spatial Filtering for Multichannel Source Separation in the Waveform Domain,31
|
| 87 |
+
35,1542713456268304384,"Denoised MDPs: Learning World Models Better Than the World Itself
|
| 88 |
+
abs: https://t.co/CPwlF0soWZ
|
| 89 |
+
project page:… https://t.co/5BBwGXYZ2l",Denoised MDPs: Learning World Models Better Than the World Itself,98
|
| 90 |
+
36,1542712192746782720,"Forecasting Future World Events with Neural Networks
|
| 91 |
+
abs: https://t.co/tD8F0ZC1rC
|
| 92 |
+
github: https://t.co/v8HZgye0ZH… https://t.co/eJaakYSUSw",Forecasting Future World Events with Neural Networks,77
|
| 93 |
+
37,1542709853516431361,"Learning Iterative Reasoning through Energy Minimization
|
| 94 |
+
abs: https://t.co/WDLx1hKPqG
|
| 95 |
+
project page:… https://t.co/oDEClr0ho1",Learning Iterative Reasoning through Energy Minimization,125
|
| 96 |
+
38,1542709029964849154,"Improving the Generalization of Supervised Models
|
| 97 |
+
abs: https://t.co/3CzEuuxvHt
|
| 98 |
+
project page: https://t.co/uSjiKvSMN8 https://t.co/ffUkpTL7Ng",Improving the Generalization of Supervised Models,189
|
| 99 |
+
39,1542325850036752394,"RegMixup: Mixup as a Regularizer Can Surprisingly Improve Accuracy and Out Distribution Robustness
|
| 100 |
+
abs:… https://t.co/iFAou98U0X",RegMixup: Mixup as a Regularizer Can Surprisingly Improve Accuracy and Out Distribution Robustness,172
|
| 101 |
+
40,1542316111743664133,"Masked World Models for Visual Control
|
| 102 |
+
abs: https://t.co/eZx53zuqnm
|
| 103 |
+
project page: https://t.co/hgZwrV3zO5
|
| 104 |
+
|
| 105 |
+
Can MAE… https://t.co/UfybFx81uj",Masked World Models for Visual Control,83
|
| 106 |
+
41,1542313347835731970,"Beyond neural scaling laws: beating power law scaling via data pruning
|
| 107 |
+
abs: https://t.co/OFYkTt5b2d https://t.co/7SKXMClaR8",Beyond neural scaling laws: beating power law scaling via data pruning,164
|
| 108 |
+
42,1542312585768435712,"3D-Aware Video Generation
|
| 109 |
+
abs: https://t.co/N64ARXFKMJ
|
| 110 |
+
project page: https://t.co/5MoGVKqItn https://t.co/uZdLIXWc1P",3D-Aware Video Generation,122
|
| 111 |
+
43,1541957148070010881,"DayDreamer: World Models for Physical Robot Learning
|
| 112 |
+
abs: https://t.co/quyTQGcjEA
|
| 113 |
+
project page:… https://t.co/DD67NUzgJy",DayDreamer: World Models for Physical Robot Learning,182
|
| 114 |
+
44,1541948699559006210,"Long Range Language Modeling via Gated State Spaces
|
| 115 |
+
abs: https://t.co/HEd2lwlGan https://t.co/tPOHv7dP0T",Long Range Language Modeling via Gated State Spaces,124
|
| 116 |
+
45,1541945827035332610,"ProGen2: Exploring the Boundaries of Protein Language Models
|
| 117 |
+
abs: https://t.co/kelWMlhH8r
|
| 118 |
+
github:… https://t.co/nzvei5pMJR",ProGen2: Exploring the Boundaries of Protein Language Models,64
|
| 119 |
+
46,1541626617490837504,"Multitask vocal burst modeling with ResNets and pre-trained paralinguistic Conformers
|
| 120 |
+
abs: https://t.co/QZLcoFOeSz https://t.co/315WfiVVRr",Multitask vocal burst modeling with ResNets and pre-trained paralinguistic Conformers,11
|
| 121 |
+
47,1541599748624351233,"Programmatic Concept Learning for Human Motion Description and Synthesis
|
| 122 |
+
abs: https://t.co/uIoxGozwhD
|
| 123 |
+
project page:… https://t.co/MmCMQouLF7",Programmatic Concept Learning for Human Motion Description and Synthesis,83
|
| 124 |
+
48,1541592312094101506,"Prompting Decision Transformer for Few-Shot Policy Generalization
|
| 125 |
+
abs: https://t.co/bD2f4SjRP6
|
| 126 |
+
project page:… https://t.co/ZfAxxx6zCu",Prompting Decision Transformer for Few-Shot Policy Generalization,48
|
| 127 |
+
49,1541590513241006080,"Repository-Level Prompt Generation for Large Language Models of Code
|
| 128 |
+
abs: https://t.co/GG1YHoCQdf
|
| 129 |
+
github:… https://t.co/Z9fUO4r8sU",Repository-Level Prompt Generation for Large Language Models of Code,56
|
| 130 |
+
50,1541588372631818241,"Your Autoregressive Generative Model Can be Better If You Treat It as an Energy-Based One
|
| 131 |
+
abs:… https://t.co/uJuKxO7XJC",Your Autoregressive Generative Model Can be Better If You Treat It as an Energy-Based One,121
|
| 132 |
+
51,1541226747533922308,"PSP: Million-level Protein Sequence Dataset for Protein Structure Prediction
|
| 133 |
+
abs: https://t.co/yXdFTqRWF3
|
| 134 |
+
|
| 135 |
+
dataset… https://t.co/ZDNMPI2NVR",PSP: Million-level Protein Sequence Dataset for Protein Structure Prediction,94
|
| 136 |
+
52,1541219433259175937,"Megapixel Image Generation with Step-Unrolled Denoising Autoencoders
|
| 137 |
+
abs: https://t.co/6fX9PseXBT
|
| 138 |
+
|
| 139 |
+
obtain FID score… https://t.co/HPodJ8xzPx",Megapixel Image Generation with Step-Unrolled Denoising Autoencoders,147
|
| 140 |
+
53,1540184734390706176,"Walk the Random Walk: Learning to Discover and Reach Goals Without Supervision
|
| 141 |
+
abs: https://t.co/NO2vzfdYdS https://t.co/WoN73BzgeQ",Walk the Random Walk: Learning to Discover and Reach Goals Without Supervision,66
|
| 142 |
+
54,1540176838017916933,"Offline RL for Natural Language Generation with Implicit Language Q Learning
|
| 143 |
+
abs: https://t.co/wYTtUgdryZ
|
| 144 |
+
project p… https://t.co/xS8JCODxwP",Offline RL for Natural Language Generation with Implicit Language Q Learning,43
|
| 145 |
+
55,1540161095930880001,"MaskViT: Masked Visual Pre-Training for Video Prediction
|
| 146 |
+
abs: https://t.co/uhMEB6ashb
|
| 147 |
+
project page:… https://t.co/gbnxrCxUrc",MaskViT: Masked Visual Pre-Training for Video Prediction,147
|
| 148 |
+
56,1540156319923060736,"The ArtBench Dataset: Benchmarking Generative Models with Artworks
|
| 149 |
+
abs: https://t.co/Zzq0A2i5ob
|
| 150 |
+
github:… https://t.co/SfQlvTLrk3",The ArtBench Dataset: Benchmarking Generative Models with Artworks,200
|
| 151 |
+
57,1539811680359796739,"TiCo: Transformation Invariance and Covariance Contrast for Self-Supervised Visual Representation Learning
|
| 152 |
+
abs:… https://t.co/UArbr7zhRE",TiCo: Transformation Invariance and Covariance Contrast for Self-Supervised Visual Representation Learning,85
|
| 153 |
+
58,1539794210190155778,"Jointist: Joint Learning for Multi-instrument Transcription and Its Applications
|
| 154 |
+
abs: https://t.co/xeuPUBcr01
|
| 155 |
+
proje… https://t.co/QmyCioKviJ",Jointist: Joint Learning for Multi-instrument Transcription and Its Applications,18
|
| 156 |
+
59,1539780412297330689,"GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
|
| 157 |
+
abs: https://t.co/pKS5mgoDkG
|
| 158 |
+
|
| 159 |
+
GEMv2 supports 40 docum… https://t.co/qMitHzTlO0",GEMv2: Multilingual NLG Benchmarking in a Single Line of Code,18
|
| 160 |
+
60,1539777865688010753,"reStructured Pre-training
|
| 161 |
+
abs: https://t.co/mYm7qbt59N https://t.co/O5T3tSY4PL",reStructured Pre-training,32
|
| 162 |
+
61,1539672920456298498,"Scaling Autoregressive Models for Content-Rich Text-to-Image Generation
|
| 163 |
+
paper: https://t.co/NKkTeHttLd
|
| 164 |
+
project page… https://t.co/CcKxsWPmjR",Scaling Autoregressive Models for Content-Rich Text-to-Image Generation,137
|
| 165 |
+
62,1539480179151712256,"Intra-Instance VICReg: Bag of Self-Supervised Image Patch Embedding
|
| 166 |
+
abs: https://t.co/Bq3GUQywPV https://t.co/iLTaoXm0yC",Intra-Instance VICReg: Bag of Self-Supervised Image Patch Embedding,66
|
| 167 |
+
63,1539460213211910150,"EnvPool: A Highly Parallel Reinforcement Learning Environment Execution Engine
|
| 168 |
+
abs: https://t.co/F4XkHLRxPi
|
| 169 |
+
github:… https://t.co/JiwSuMdkZH",EnvPool: A Highly Parallel Reinforcement Learning Environment Execution Engine,34
|
| 170 |
+
64,1539459120667021312,"EpiGRAF: Rethinking training of 3D GANs
|
| 171 |
+
abs: https://t.co/RcY2vQr0NH
|
| 172 |
+
project page: https://t.co/kuXPKA00bZ https://t.co/CVCsseAS21",EpiGRAF: Rethinking training of 3D GANs,145
|
| 173 |
+
65,1539453554578055168,"Unbiased Teacher v2: Semi-supervised Object Detection for Anchor-free and Anchor-based Detectors
|
| 174 |
+
abs:… https://t.co/noluSxtqzu",Unbiased Teacher v2: Semi-supervised Object Detection for Anchor-free and Anchor-based Detectors,72
|
| 175 |
+
66,1539435374103220226,"Global Context Vision Transformers
|
| 176 |
+
abs: https://t.co/d6go0yv7fu
|
| 177 |
+
github: https://t.co/rUYFs09ReC
|
| 178 |
+
|
| 179 |
+
On ImageNet-1K dat… https://t.co/HJnw5wclQV",Global Context Vision Transformers,89
|
| 180 |
+
67,1539421251076247554,"(Certified!!) Adversarial Robustness for Free!
|
| 181 |
+
abs: https://t.co/NTU6lioyII
|
| 182 |
+
|
| 183 |
+
show how to achieve sota certified adv… https://t.co/2VW1CDARya",(Certified!!) Adversarial Robustness for Free!,42
|
| 184 |
+
68,1539076449788997632,"A Closer Look at Smoothness in Domain Adversarial Training
|
| 185 |
+
abs: https://t.co/GgKE9695vj
|
| 186 |
+
github:… https://t.co/33MX6TZhjt",A Closer Look at Smoothness in Domain Adversarial Training,97
|
| 187 |
+
69,1538710356444471296,"Fast Finite Width Neural Tangent Kernel
|
| 188 |
+
abs: https://t.co/iY1lFoYMjA https://t.co/hWzzcCd5OZ",Fast Finite Width Neural Tangent Kernel,23
|
| 189 |
+
70,1538706936211951617,"What do navigation agents learn about their environment?
|
| 190 |
+
abs: https://t.co/eXelV0REgZ
|
| 191 |
+
github:… https://t.co/TGSzEQ1v1c",What do navigation agents learn about their environment?,37
|
| 192 |
+
71,1538698653493338114,"Bootstrapped Transformer for Offline Reinforcement Learning
|
| 193 |
+
abs: https://t.co/YiEY3uiTgL https://t.co/yle4hPgMmf",Bootstrapped Transformer for Offline Reinforcement Learning,137
|
| 194 |
+
72,1538695457550921728,"Bridge-Tower: Building Bridges Between Encoders in Vision-Language Representation Learning
|
| 195 |
+
abs:… https://t.co/uLQLmf4l3M",Bridge-Tower: Building Bridges Between Encoders in Vision-Language Representation Learning,42
|
| 196 |
+
73,1538692524830769152,"MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge
|
| 197 |
+
abs: https://t.co/etfGL1xnum
|
| 198 |
+
project pa… https://t.co/Fv1aLuEJSV",MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge,265
|
| 199 |
+
74,1538687423722541056,"Lossy Compression with Gaussian Diffusion
|
| 200 |
+
abs: https://t.co/tw5YiZAN3B
|
| 201 |
+
|
| 202 |
+
implement a proof of concept and find that… https://t.co/4nvLjhIX4e",Lossy Compression with Gaussian Diffusion,102
|
| 203 |
+
75,1538686489491648514,"NU-Wave 2: A General Neural Audio Upsampling Model for Various Sampling Rates
|
| 204 |
+
abs: https://t.co/4S8sBXq6Ko
|
| 205 |
+
|
| 206 |
+
a diffu… https://t.co/xd3eQ0ApQJ",NU-Wave 2: A General Neural Audio Upsampling Model for Various Sampling Rates,87
|
| 207 |
+
76,1538006265363738625,"iBoot: Image-bootstrapped Self-Supervised Video Representation Learning
|
| 208 |
+
abs: https://t.co/dkZUd4QC81 https://t.co/pJFpxd7ckU",iBoot: Image-bootstrapped Self-Supervised Video Representation Learning,73
|
| 209 |
+
77,1538000649933115393,"Neural Scene Representation for Locomotion on Structured Terrain
|
| 210 |
+
abs: https://t.co/68xY622f4w https://t.co/W3wTYp31f6",Neural Scene Representation for Locomotion on Structured Terrain,83
|
| 211 |
+
78,1537924151389736961,"Programmatic Concept Learning for Human Motion Description and Synthesis
|
| 212 |
+
paper: https://t.co/Qemk23gUHX
|
| 213 |
+
project pag… https://t.co/ImHeYQC5vj",Programmatic Concept Learning for Human Motion Description and Synthesis,60
|
| 214 |
+
79,1537640654968324099,"Spatially-Adaptive Multilayer Selection for GAN Inversion and Editing
|
| 215 |
+
abs: https://t.co/9tpvhXuaRw
|
| 216 |
+
project page:… https://t.co/XxpZg5PGke",Spatially-Adaptive Multilayer Selection for GAN Inversion and Editing,73
|
| 217 |
+
80,1537637590274277376,"MoDi: Unconditional Motion Synthesis from Diverse Data
|
| 218 |
+
abs: https://t.co/YBV9jSUemo https://t.co/o1uvG18RSk",MoDi: Unconditional Motion Synthesis from Diverse Data,70
|
| 219 |
+
81,1537630146244517889,"OmniMAE: Single Model Masked Pretraining on Images and Videos
|
| 220 |
+
abs: https://t.co/j9a3imUEJ6
|
| 221 |
+
|
| 222 |
+
single pretrained model… https://t.co/OiR2pY5emm",OmniMAE: Single Model Masked Pretraining on Images and Videos,146
|
| 223 |
+
82,1537622879386456064,"SAVi++: Towards End-to-End Object-Centric Learning from Real-World Videos
|
| 224 |
+
abs: https://t.co/0MkpFJiUzM
|
| 225 |
+
|
| 226 |
+
using spars… https://t.co/x1Hvgf13qE",SAVi++: Towards End-to-End Object-Centric Learning from Real-World Videos,54
|
| 227 |
+
83,1537621348339572736,"BYOL-Explore: Exploration by Bootstrapped Prediction
|
| 228 |
+
abs: https://t.co/xXQtolzjlP
|
| 229 |
+
|
| 230 |
+
BYOL-Explore achieves superhuman… https://t.co/uZvAbVd1Bb",BYOL-Explore: Exploration by Bootstrapped Prediction,79
|
| 231 |
+
84,1537618457365303296,"Know your audience: specializing grounded language models with the game of Dixit
|
| 232 |
+
abs: https://t.co/T8d5ir8LDQ https://t.co/zSk5oR2F9D",Know your audience: specializing grounded language models with the game of Dixit,39
|
| 233 |
+
85,1537323042380124160,"VCT: A Video Compression Transformer
|
| 234 |
+
abs: https://t.co/llH1L1ooKa
|
| 235 |
+
|
| 236 |
+
presented an elegantly simple transformer-based… https://t.co/ErovCWVDg3",VCT: A Video Compression Transformer,68
|
| 237 |
+
86,1537314480056672258,"Contrastive Learning as Goal-Conditioned Reinforcement Learning
|
| 238 |
+
abs: https://t.co/6dv7PNn0qq
|
| 239 |
+
project page:… https://t.co/vRSdekL9If",Contrastive Learning as Goal-Conditioned Reinforcement Learning,77
|
| 240 |
+
87,1537288570880368640,"Masked Siamese ConvNets
|
| 241 |
+
abs: https://t.co/YMG1O1ZZ5N https://t.co/LCVqVvFNfR",Masked Siamese ConvNets,83
|
| 242 |
+
88,1537265816609116161,"Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
|
| 243 |
+
abs: https://t.co/UgdYW9Cf1g
|
| 244 |
+
project page:… https://t.co/v2sTfFBq5r",Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone,89
|
| 245 |
+
89,1537257011657814016,"Variable Bitrate Neural Fields
|
| 246 |
+
abs: https://t.co/Rp1t2LaQaW
|
| 247 |
+
project page: https://t.co/e2t8OrznxI https://t.co/6hw7OwbjZN",Variable Bitrate Neural Fields,162
|
| 248 |
+
90,1537254679188488195,"A Unified Sequence Interface for Vision Tasks
|
| 249 |
+
abs: https://t.co/hXbVXdqHh1
|
| 250 |
+
|
| 251 |
+
explore a unified sequence interface fo… https://t.co/QG5UxvIgS4",A Unified Sequence Interface for Vision Tasks,50
|
| 252 |
+
91,1537252952666087424,"Prefix Language Models are Unified Modal Learners
|
| 253 |
+
abs: https://t.co/BD4b3rQnKg https://t.co/2ofScnMIKN",Prefix Language Models are Unified Modal Learners,66
|
| 254 |
+
92,1537248480074293251,"Diffusion Models for Video Prediction and Infilling
|
| 255 |
+
abs: https://t.co/MwfxwKXG4z
|
| 256 |
+
project page:… https://t.co/rnwB8eGFAs",Diffusion Models for Video Prediction and Infilling,103
|
| 257 |
+
93,1536879515883945984,"ReCo: Retrieve and Co-segment for Zero-shot Transfer
|
| 258 |
+
abs: https://t.co/YwxkCGGyG1
|
| 259 |
+
project page:… https://t.co/WzVhmfhWCz",ReCo: Retrieve and Co-segment for Zero-shot Transfer,58
|
| 260 |
+
94,1536872875885580288,"Object Scene Representation Transformer
|
| 261 |
+
abs: https://t.co/SUfNIBGAxt
|
| 262 |
+
project page: https://t.co/j8ebSAeM8v
|
| 263 |
+
|
| 264 |
+
scales… https://t.co/wa4vo3RJAK",Object Scene Representation Transformer,97
|
| 265 |
+
95,1536871347372052480,"Adversarial Audio Synthesis with Complex-valued Polynomial Networks
|
| 266 |
+
abs: https://t.co/ekeC0nKIhR
|
| 267 |
+
|
| 268 |
+
APOLLO results in… https://t.co/sDcl2nydkt",Adversarial Audio Synthesis with Complex-valued Polynomial Networks,23
|
| 269 |
+
96,1536526888289574915,"Large-Scale Retrieval for Reinforcement Learning
|
| 270 |
+
abs: https://t.co/fjzGvI3ZXB https://t.co/eFRHt8yXoq",Large-Scale Retrieval for Reinforcement Learning,86
|
| 271 |
+
97,1536522198785183744,"GLIPv2: Unifying Localization and Vision-Language Understanding
|
| 272 |
+
abs: https://t.co/3GomrHG8xq
|
| 273 |
+
github:… https://t.co/bD68NZk4Lp",GLIPv2: Unifying Localization and Vision-Language Understanding,73
|
| 274 |
+
98,1536521362898145280,"Self-critiquing models for assisting human evaluators
|
| 275 |
+
abs: https://t.co/8Zy2xfA5Qz https://t.co/qndZMS9zXa",Self-critiquing models for assisting human evaluators,19
|
| 276 |
+
99,1536515535202136064,"Multi-instrument Music Synthesis with Spectrogram Diffusion
|
| 277 |
+
abs: https://t.co/UNDV4e7A6R
|
| 278 |
+
|
| 279 |
+
use a simple two-stage pr… https://t.co/AebIraqLF2",Multi-instrument Music Synthesis with Spectrogram Diffusion,87
|
| 280 |
+
100,1536493418305703938,"How Much is Enough? A Study on Diffusion Times in Score-based Generative Models
|
| 281 |
+
abs: https://t.co/qFEZBDrdrq https://t.co/iBlNs4iNE2",How Much is Enough? A Study on Diffusion Times in Score-based Generative Models,60
|
| 282 |
+
101,1536491133513129990,"Meta Optimal Transport
|
| 283 |
+
abs: https://t.co/UKdYXKA8Vd
|
| 284 |
+
github: https://t.co/xb9FVcim7g
|
| 285 |
+
|
| 286 |
+
Meta OT models surpass the sta… https://t.co/OlfwZIC52r",Meta Optimal Transport,67
|
| 287 |
+
102,1535656084488192005,"Neural Prompt Search
|
| 288 |
+
abs: https://t.co/wZTUHIcqdv
|
| 289 |
+
github: https://t.co/vnYEMBrKzt
|
| 290 |
+
|
| 291 |
+
view existing parameter-efficien… https://t.co/pLvxNt84gV",Neural Prompt Search,174
|
| 292 |
+
103,1535521674233319424,"Deep Surrogate Assisted Generation of Environments
|
| 293 |
+
abs: https://t.co/1RYhxJ71tt
|
| 294 |
+
project page:… https://t.co/5MuAOKIePA",Deep Surrogate Assisted Generation of Environments,58
|
| 295 |
+
104,1535521046257975297,"Deep Hierarchical Planning from Pixels
|
| 296 |
+
abs: https://t.co/xXBDevsRnK
|
| 297 |
+
project page: https://t.co/LoNsGVecaR https://t.co/K7RKIq2hBT",Deep Hierarchical Planning from Pixels,101
|
| 298 |
+
105,1535506620624642048,"VN-Transformer: Rotation-Equivariant Attention for Vector Neurons
|
| 299 |
+
abs: https://t.co/OkS58YpYq8 https://t.co/ailLjhzsqa",VN-Transformer: Rotation-Equivariant Attention for Vector Neurons,144
|
| 300 |
+
106,1535469100436271105,"Factuality Enhanced Language Models for Open-Ended Text Generation
|
| 301 |
+
abs: https://t.co/YX83NnfpMU
|
| 302 |
+
|
| 303 |
+
factual-nucleus sa… https://t.co/suFGgO8Ajv",Factuality Enhanced Language Models for Open-Ended Text Generation,31
|
| 304 |
+
107,1535449832332177408,"Unveiling Transformers with LEGO: a synthetic reasoning task
|
| 305 |
+
abs: https://t.co/FCnAD9AjMY https://t.co/LsUblvE3Ig",Unveiling Transformers with LEGO: a synthetic reasoning task,77
|
| 306 |
+
108,1535392356068892674,"BigVGAN: A Universal Neural Vocoder with Large-Scale Training
|
| 307 |
+
abs: https://t.co/4NRS1WBePa
|
| 308 |
+
project page:… https://t.co/rpuKyOEGMH",BigVGAN: A Universal Neural Vocoder with Large-Scale Training,170
|
| 309 |
+
109,1535069067052195862,"Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
|
| 310 |
+
abs:… https://t.co/v2aIh9B5H2",Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models,158
|
| 311 |
+
110,1535067850435600403,"Draft-and-Revise: Effective Image Generation with Contextual RQ-Transformer
|
| 312 |
+
abs: https://t.co/0s94Tbwh3q
|
| 313 |
+
|
| 314 |
+
propose i… https://t.co/lQZWEHXeRI",Draft-and-Revise: Effective Image Generation with Contextual RQ-Transformer,52
|
| 315 |
+
111,1535066703075352601,"VideoINR: Learning Video Implicit Neural Representation for Continuous Space-Time Super-Resolution
|
| 316 |
+
abs:… https://t.co/UKXo53aomf",VideoINR: Learning Video Implicit Neural Representation for Continuous Space-Time Super-Resolution,146
|
| 317 |
+
112,1535061799975919633,"Diffusion probabilistic modeling of protein backbones in 3D for the motif-scaffolding problem
|
| 318 |
+
abs:… https://t.co/fUyM4hz22a",Diffusion probabilistic modeling of protein backbones in 3D for the motif-scaffolding problem,48
|
| 319 |
+
113,1535026713100537871,"Sparse Fusion Mixture-of-Experts are Domain Generalizable Learners
|
| 320 |
+
abs: https://t.co/koYO5SuiDQ
|
| 321 |
+
github:… https://t.co/1xMmVzboCC",Sparse Fusion Mixture-of-Experts are Domain Generalizable Learners,70
|
| 322 |
+
114,1534712305790894081,"STable: Table Generation Framework for Encoder-Decoder Models
|
| 323 |
+
abs: https://t.co/P8GcsztVFp https://t.co/lJnhODKXyn",STable: Table Generation Framework for Encoder-Decoder Models,32
|
| 324 |
+
115,1534702470202630144,"Neural Diffusion Processes
|
| 325 |
+
abs: https://t.co/do2pFgpRWY
|
| 326 |
+
|
| 327 |
+
empirically show that NDPs are able to capture functional… https://t.co/Fx5BFrA9qQ",Neural Diffusion Processes,229
|
| 328 |
+
116,1534701793183252485,"Patch-based Object-centric Transformers for Efficient Video Generation
|
| 329 |
+
abs: https://t.co/oeAa0hiBqZ
|
| 330 |
+
project page:… https://t.co/qCoaulnDfS",Patch-based Object-centric Transformers for Efficient Video Generation,30
|
| 331 |
+
117,1534700653628764160,"Accelerating Score-based Generative Models for High-Resolution Image Synthesis
|
| 332 |
+
abs: https://t.co/rC90ydANVJ
|
| 333 |
+
project… https://t.co/5reyDDPyBN",Accelerating Score-based Generative Models for High-Resolution Image Synthesis,69
|
| 334 |
+
118,1534476660355043329,"On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning
|
| 335 |
+
abs: https://t.co/1gEuTB7Sf1
|
| 336 |
+
|
| 337 |
+
multi-task pre… https://t.co/zx8QDoxq2l",On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning,39
|
| 338 |
+
119,1534465882512146432,"Few-Shot Learning by Dimensionality Reduction in Gradient Space
|
| 339 |
+
abs: https://t.co/IMwlsW0r5V
|
| 340 |
+
|
| 341 |
+
introduce SubGD, a no… https://t.co/YltxH8mUtF",Few-Shot Learning by Dimensionality Reduction in Gradient Space,204
|
| 342 |
+
120,1534376291453083648,"DETR++: Taming Your Multi-Scale Detection Transformer
|
| 343 |
+
abs: https://t.co/kOQ5V4vC3C
|
| 344 |
+
|
| 345 |
+
DETR++, a new architecture that… https://t.co/i7qtSX9eA3",DETR++: Taming Your Multi-Scale Detection Transformer,85
|
| 346 |
+
121,1534347375128547328,"Intra-agent speech permits zero-shot task acquisition
|
| 347 |
+
abs: https://t.co/2yVGA91kSA
|
| 348 |
+
|
| 349 |
+
with ~ 150 additional image cap… https://t.co/DtBczvw7lh",Intra-agent speech permits zero-shot task acquisition,60
|
| 350 |
+
122,1534343347334176770,"Universal Speech Enhancement with Score-based Diffusion
|
| 351 |
+
abs: https://t.co/jv1rQ14Do4
|
| 352 |
+
project page:… https://t.co/UMEE3irGWN",Universal Speech Enhancement with Score-based Diffusion,125
|
| 353 |
+
123,1534341405920870400,"Generating Long Videos of Dynamic Scenes
|
| 354 |
+
abs: https://t.co/SjMCJub1RO
|
| 355 |
+
project page: https://t.co/c97Jcf3lcC
|
| 356 |
+
|
| 357 |
+
presen… https://t.co/jgcfMwGMo6",Generating Long Videos of Dynamic Scenes,336
|
| 358 |
+
124,1533997063951765506,"Zero-Shot Voice Conditioning for Denoising Diffusion TTS Models
|
| 359 |
+
abs: https://t.co/iTfFppABzr
|
| 360 |
+
|
| 361 |
+
method requires a sho… https://t.co/GALvAsiQ0J",Zero-Shot Voice Conditioning for Denoising Diffusion TTS Models,89
|
| 362 |
+
125,1533996337557020672,"Drawing out of Distribution with Neuro-Symbolic Generative Models
|
| 363 |
+
abs: https://t.co/PcRRRLIVyV
|
| 364 |
+
|
| 365 |
+
DooD trained on MNI… https://t.co/h28KgM3m3k",Drawing out of Distribution with Neuro-Symbolic Generative Models,39
|
| 366 |
+
126,1533993050627776512,"Separable Self-attention for Mobile Vision Transformers
|
| 367 |
+
abs: https://t.co/Xj1aZMucFe
|
| 368 |
+
|
| 369 |
+
With ~ 3M parameters, MobileV… https://t.co/LTag2ck7Ew",Separable Self-attention for Mobile Vision Transformers,89
|
| 370 |
+
127,1533989659017199617,"Extreme Compression for Pre-trained Transformers Made Simple and Efficient
|
| 371 |
+
abs: https://t.co/7epbwDmV31 https://t.co/n9nppcTgGJ",Extreme Compression for Pre-trained Transformers Made Simple and Efficient,84
|
| 372 |
+
128,1533988146102288386,"On the duality between contrastive and non-contrastive self-supervised learning
|
| 373 |
+
abs: https://t.co/O2GdHjqiTz https://t.co/nUibodNE9M",On the duality between contrastive and non-contrastive self-supervised learning,83
|
| 374 |
+
129,1533982101653098503,"ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
|
| 375 |
+
abs:… https://t.co/tQuBWS3uaH",ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers,25
|
| 376 |
+
130,1533980842867015681,"Torsional Diffusion for Molecular Conformer Generation
|
| 377 |
+
abs: https://t.co/VfhEdlJLd7
|
| 378 |
+
github: https://t.co/DYpXh7NbKe https://t.co/khz3yO5FFZ",Torsional Diffusion for Molecular Conformer Generation,24
|
| 379 |
+
131,1533980437114232832,"Blended Latent Diffusion
|
| 380 |
+
abs: https://t.co/5K8QQnlQfz
|
| 381 |
+
project page: https://t.co/ztlJtR4Sio
|
| 382 |
+
|
| 383 |
+
present an accelerated… https://t.co/qzrdUJc4i9",Blended Latent Diffusion,55
|
| 384 |
+
132,1533979552761913344,"Diffusion-GAN: Training GANs with Diffusion
|
| 385 |
+
abs: https://t.co/rxRpORfP5U
|
| 386 |
+
|
| 387 |
+
DiffusionGAN can provide stable and data-… https://t.co/ScQTvm3XaA",Diffusion-GAN: Training GANs with Diffusion,237
|
| 388 |
+
133,1533676404063232000,"Beyond Tabula Rasa: Reincarnating Reinforcement Learning
|
| 389 |
+
abs: https://t.co/r8TcfqPyIs https://t.co/qSO5K11vYB",Beyond Tabula Rasa: Reincarnating Reinforcement Learning,34
|
| 390 |
+
134,1533649732345778177,"Improving Fairness in Large-Scale Object Recognition by CrowdSourced Demographic Information
|
| 391 |
+
abs:… https://t.co/3mGwmSsO6M",Improving Fairness in Large-Scale Object Recognition by CrowdSourced Demographic Information,17
|
| 392 |
+
135,1533634419986153472,"Positive Unlabeled Contrastive Learning
|
| 393 |
+
abs: https://t.co/LC33ii48Q6 https://t.co/eWLoasRamS",Positive Unlabeled Contrastive Learning,67
|
| 394 |
+
136,1533633258545610754,"Reinforcement Learning with Neural Radiance Fields
|
| 395 |
+
abs: https://t.co/8ESw75I2N9
|
| 396 |
+
project page:… https://t.co/DQrpZ5dyrb",Reinforcement Learning with Neural Radiance Fields,131
|
| 397 |
+
137,1533619945996697600,"Compositional Visual Generation with Composable Diffusion Models
|
| 398 |
+
abs: https://t.co/FEKYaDOlwf
|
| 399 |
+
project page:… https://t.co/qvaTyuj3un",Compositional Visual Generation with Composable Diffusion Models,122
|
| 400 |
+
138,1533611409069711368,"Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules
|
| 401 |
+
abs:… https://t.co/rQTNT4yfcB",Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules,40
|
| 402 |
+
139,1532729442321170433,"Deep Learning on Implicit Neural Datasets
|
| 403 |
+
abs: https://t.co/nPGleDBRSq
|
| 404 |
+
|
| 405 |
+
introduce the INR-Net, the first general fr… https://t.co/i1xT7bLhSN",Deep Learning on Implicit Neural Datasets,81
|
| 406 |
+
140,1532726423697465344,"SupMAE: Supervised Masked Autoencoders Are Efficient Vision Learners
|
| 407 |
+
abs: https://t.co/SIR2ufE89J
|
| 408 |
+
github:… https://t.co/tZoNFvtDFQ",SupMAE: Supervised Masked Autoencoders Are Efficient Vision Learners,178
|
| 409 |
+
141,1532558380119752705,"DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks
|
| 410 |
+
abs:… https://t.co/dHBUdpmqm9",DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks,31
|
| 411 |
+
142,1532554016072376323,"Cascaded Video Generation for Videos In-the-Wild
|
| 412 |
+
abs: https://t.co/wDkiRCEWXN https://t.co/GJSVK80qC0",Cascaded Video Generation for Videos In-the-Wild,57
|
| 413 |
+
143,1532547568567300096,"Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation
|
| 414 |
+
abs:… https://t.co/FAEEhSyQpY",Finding the Right Recipe for Low Resource Domain Adaptation in Neural Machine Translation,12
|
| 415 |
+
144,1532540853071265799,"BayesFormer: Transformer with Uncertainty Estimation
|
| 416 |
+
abs: https://t.co/0OqGgau2D2
|
| 417 |
+
|
| 418 |
+
introduce BayesFormer, a Transfo… https://t.co/znYfXmUPpJ",BayesFormer: Transformer with Uncertainty Estimation,188
|
| 419 |
+
145,1532539121662574605,"Improving Diffusion Models for Inverse Problems using Manifold Constraints
|
| 420 |
+
abs: https://t.co/Mt78QlNgZZ https://t.co/d6T7XFkqf1",Improving Diffusion Models for Inverse Problems using Manifold Constraints,115
|
| 421 |
+
146,1532538212438130697,"DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps
|
| 422 |
+
abs:… https://t.co/PBn2cEeEle",DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps,93
|
| 423 |
+
147,1532201565167267840,"Hopular: Modern Hopfield Networks for Tabular Data
|
| 424 |
+
abs: https://t.co/O5h6GYoGZd
|
| 425 |
+
github: https://t.co/kztLUsmzMY
|
| 426 |
+
pro… https://t.co/xqlUFoil7K",Hopular: Modern Hopfield Networks for Tabular Data,485
|
| 427 |
+
148,1532173830428442627,"PandA: Unsupervised Learning of Parts and Appearances in the Feature Maps of GANs
|
| 428 |
+
abs: https://t.co/MdoshW31xe
|
| 429 |
+
gith… https://t.co/d0PWKpIufP",PandA: Unsupervised Learning of Parts and Appearances in the Feature Maps of GANs,121
|
| 430 |
+
149,1532162242715721728,"Elucidating the Design Space of Diffusion-Based Generative Models
|
| 431 |
+
abs: https://t.co/WtodJSq1wa
|
| 432 |
+
|
| 433 |
+
improve efficiency… https://t.co/Fp84kzysBZ",Elucidating the Design Space of Diffusion-Based Generative Models,257
|
| 434 |
+
150,1531810146178957312,"Chefs' Random Tables: Non-Trigonometric Random Features
|
| 435 |
+
abs: https://t.co/qrt5BnhG2g https://t.co/AuWq9HKnl5",Chefs' Random Tables: Non-Trigonometric Random Features,19
|
| 436 |
+
151,1531802121280147457,"Few-Shot Diffusion Models
|
| 437 |
+
abs: https://t.co/Oz75eOx0Ue
|
| 438 |
+
|
| 439 |
+
At test time, the model is able to generate samples from pr… https://t.co/qw3Wdivfks",Few-Shot Diffusion Models,114
|
| 440 |
+
152,1531798720550952961,"SAMURAI: Shape And Material from Unconstrained Real-world Arbitrary Image collections
|
| 441 |
+
abs: https://t.co/eviBoaJ1Zw… https://t.co/XsdD2CSafR",SAMURAI: Shape And Material from Unconstrained Real-world Arbitrary Image collections,148
|
| 442 |
+
153,1531484127177936896,"Play it by Ear: Learning Skills amidst Occlusion through Audio-Visual Imitation Learning
|
| 443 |
+
abs:… https://t.co/yafGze7shH",Play it by Ear: Learning Skills amidst Occlusion through Audio-Visual Imitation Learning,36
|
| 444 |
+
154,1531466054492364800,"Dataset Condensation via Efficient Synthetic-Data Parameterization
|
| 445 |
+
abs: https://t.co/IA66WHQQCH
|
| 446 |
+
github:… https://t.co/PuBEVyx5EK",Dataset Condensation via Efficient Synthetic-Data Parameterization,110
|
| 447 |
+
155,1531465172262567937,"Neural Shape Mating: Self-Supervised Object Assembly with Adversarial Shape Priors
|
| 448 |
+
abs: https://t.co/25EYR1yE1A
|
| 449 |
+
pro… https://t.co/qdqxXZtyYx",Neural Shape Mating: Self-Supervised Object Assembly with Adversarial Shape Priors,56
|
| 450 |
+
156,1531460153152786432,"Teaching Models to Express Their Uncertainty in Words
|
| 451 |
+
abs: https://t.co/rKcZNhBLt5
|
| 452 |
+
|
| 453 |
+
GPT-3 model can learn to expres… https://t.co/Z3YCzXqaMX",Teaching Models to Express Their Uncertainty in Words,163
|
| 454 |
+
157,1531454478968406016,"Temporal Latent Bottleneck: Synthesis of Fast and Slow Processing Mechanisms in Sequence Learning
|
| 455 |
+
abs:… https://t.co/U47eMKEmf3",Temporal Latent Bottleneck: Synthesis of Fast and Slow Processing Mechanisms in Sequence Learning,36
|
| 456 |
+
158,1531451492120535041,"Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers
|
| 457 |
+
abs:… https://t.co/Ar0fNxMRi9",Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers,28
|
| 458 |
+
159,1531445364217237509,"Prompting ELECTRA: Few-Shot Learning with Discriminative Pre-Trained Models
|
| 459 |
+
abs: https://t.co/myWID3paI2 https://t.co/S0WUP71wz8",Prompting ELECTRA: Few-Shot Learning with Discriminative Pre-Trained Models,66
|
| 460 |
+
160,1531444059780308996,"Neural Volumetric Object Selection
|
| 461 |
+
abs: https://t.co/ZLiJ5iBZzQ
|
| 462 |
+
project page: https://t.co/YGsNO14XK7 https://t.co/4twrRcyExx",Neural Volumetric Object Selection,97
|
| 463 |
+
161,1531442002814025728,"Multi-Game Decision Transformers
|
| 464 |
+
abs: https://t.co/5JtgTx3B49
|
| 465 |
+
project page: https://t.co/rKk7h7wLga
|
| 466 |
+
|
| 467 |
+
a single trans… https://t.co/zcJXA5tDhR",Multi-Game Decision Transformers,105
|
| 468 |
+
162,1531440090161025024,"Diffusion-LM Improves Controllable Text Generation
|
| 469 |
+
abs: https://t.co/YYVX2fuWrM
|
| 470 |
+
|
| 471 |
+
Diffusion-LM iteratively denoises… https://t.co/1pJ5djHV9T",Diffusion-LM Improves Controllable Text Generation,145
|
| 472 |
+
163,1531176037400338432,"MyoSuite -- A contact-rich simulation suite for musculoskeletal motor control
|
| 473 |
+
abs: https://t.co/HpRvGT2UDz
|
| 474 |
+
project… https://t.co/6noxiVtz85",MyoSuite -- A contact-rich simulation suite for musculoskeletal motor control,47
|
| 475 |
+
164,1531174102572191744,"Neural Basis Models for Interpretability
|
| 476 |
+
abs: https://t.co/u0G7oK87X4 https://t.co/ML7UCNPDkP",Neural Basis Models for Interpretability,55
|
| 477 |
+
165,1531173694214656005,"Scalable Interpretability via Polynomials
|
| 478 |
+
abs: https://t.co/EKZDra09oM https://t.co/XyIoQHWftG",Scalable Interpretability via Polynomials,32
|
| 479 |
+
166,1531173081393336320,"Sharpness-Aware Training for Free
|
| 480 |
+
abs: https://t.co/R6SSrWAjL2 https://t.co/alHDGt3zQo",Sharpness-Aware Training for Free,155
|
| 481 |
+
167,1531165352037691392,"Global Normalization for Streaming Speech Recognition in a Modular Framework
|
| 482 |
+
abs: https://t.co/OfIb7wiVkx
|
| 483 |
+
|
| 484 |
+
demonstr… https://t.co/0iVBVXVBBs",Global Normalization for Streaming Speech Recognition in a Modular Framework,21
|
| 485 |
+
168,1531104909927628806,"Bongard-HOI: Benchmarking Few-Shot Visual Reasoning for Human-Object Interactions
|
| 486 |
+
abs: https://t.co/gVXiOx5Df3 https://t.co/eufEJbHHRr",Bongard-HOI: Benchmarking Few-Shot Visual Reasoning for Human-Object Interactions,47
|
| 487 |
+
169,1531100741166833664,"FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
|
| 488 |
+
abs: https://t.co/3aHeecihur
|
| 489 |
+
|
| 490 |
+
an IO-awa… https://t.co/GoJsOKYEgt",FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness,233
|
| 491 |
+
170,1531098962932944896,"Contrastive Siamese Network for Semi-supervised Speech Recognition
|
| 492 |
+
abs: https://t.co/SL374ByjZO
|
| 493 |
+
|
| 494 |
+
experiments show t… https://t.co/efVonWBQC5",Contrastive Siamese Network for Semi-supervised Speech Recognition,71
|
| 495 |
+
171,1531096569365282816,"X-ViT: High Performance Linear Vision Transformer without Softmax
|
| 496 |
+
abs: https://t.co/A6HZ2vXKDB https://t.co/kArY0Tm4VE",X-ViT: High Performance Linear Vision Transformer without Softmax,120
|
| 497 |
+
172,1531093245308059650,"Tranception: protein fitness prediction with autoregressive transformers and inference-time retrieval
|
| 498 |
+
|
| 499 |
+
transformer… https://t.co/OSLGlyUNqb",Tranception: protein fitness prediction with autoregressive transformers and inference-time retrieval,12
|
| 500 |
+
173,1531092289090736129,"Quark: Controllable Text Generation with Reinforced Unlearning
|
| 501 |
+
abs: https://t.co/OmS9AqhC7d
|
| 502 |
+
|
| 503 |
+
introduce Quantized Re… https://t.co/M4DHSUpwF3",Quark: Controllable Text Generation with Reinforced Unlearning,144
|
| 504 |
+
174,1531091654567919616,"Training and Inference on Any-Order Autoregressive Models the Right Way
|
| 505 |
+
abs: https://t.co/G8DNeKtoJK
|
| 506 |
+
|
| 507 |
+
leads to impr… https://t.co/JjXafy7iAu",Training and Inference on Any-Order Autoregressive Models the Right Way,22
|
| 508 |
+
175,1531090584231890947,"Contrastive Learning Rivals Masked Image Modeling in Fine-tuning via Feature Distillation
|
| 509 |
+
abs:… https://t.co/binMlc2scV",Contrastive Learning Rivals Masked Image Modeling in Fine-tuning via Feature Distillation,52
|
| 510 |
+
176,1531089687263293442,"Maximum Likelihood Training of Implicit Nonlinear Diffusion Models
|
| 511 |
+
abs: https://t.co/U2YtYUURqH https://t.co/lw7hcspT7o",Maximum Likelihood Training of Implicit Nonlinear Diffusion Models,110
|
| 512 |
+
177,1531088458839740416,"Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters
|
| 513 |
+
a… https://t.co/e1H5ZyvcQg","Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters",20
|
| 514 |
+
178,1531086920461307906,"Learning to Reason with Neural Networks: Generalization, Unseen Data and Boolean Measures
|
| 515 |
+
abs:… https://t.co/7DWwix1kP1","Learning to Reason with Neural Networks: Generalization, Unseen Data and Boolean Measures",81
|
| 516 |
+
179,1531017163284393987,"CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers
|
| 517 |
+
github: https://t.co/1JuOHU7puc https://t.co/Wilcq2Xxb9",CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers,1498
|
| 518 |
+
180,1530278551676657665,"Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
|
| 519 |
+
abs: https://t.co/swtjYLryr5 https://t.co/Ny4wTtkaAI",Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality,31
|
| 520 |
+
181,1530029153101168645,"Towards Learning Universal Hyperparameter Optimizers with Transformers
|
| 521 |
+
abs: https://t.co/yON7zKZCRy
|
| 522 |
+
|
| 523 |
+
extensive expe… https://t.co/UWv7nrCmhF",Towards Learning Universal Hyperparameter Optimizers with Transformers,129
|
| 524 |
+
182,1530028097692647449,"BiT: Robustly Binarized Multi-distilled Transformer
|
| 525 |
+
abs: https://t.co/buQ40Vo9ee https://t.co/Q8iyC2Auql",BiT: Robustly Binarized Multi-distilled Transformer,37
|
| 526 |
+
183,1530018008667660300,"Evaluating Multimodal Interactive Agents
|
| 527 |
+
abs: https://t.co/CtrOihrZBZ https://t.co/sThFVydSUZ",Evaluating Multimodal Interactive Agents,23
|
| 528 |
+
184,1530013711645253632,"Matryoshka Representations for Adaptive Deployment
|
| 529 |
+
abs: https://t.co/KkqN7sxmnN
|
| 530 |
+
|
| 531 |
+
flexibility within the learned Mat… https://t.co/RYra48uEKN",Matryoshka Representations for Adaptive Deployment,69
|
| 532 |
+
185,1530010193836244992,"Green Hierarchical Vision Transformer for Masked Image Modeling
|
| 533 |
+
abs: https://t.co/r4Y9LfE4HC
|
| 534 |
+
github:… https://t.co/o7ZihujhkM",Green Hierarchical Vision Transformer for Masked Image Modeling,26
|
| 535 |
+
186,1529673576835698691,"Inception Transformer
|
| 536 |
+
abs: https://t.co/EoPDBOafSS
|
| 537 |
+
|
| 538 |
+
iFormer-S hits the top-1 accuracy of 83.4% on ImageNet-1K, much… https://t.co/24J3SnTBdm",Inception Transformer,117
|
| 539 |
+
187,1529640184081534977,"FLEURS: Few-shot Learning Evaluation of Universal Representations of Speech
|
| 540 |
+
abs: https://t.co/IABvUreqHv https://t.co/iUUzNPaPFp",FLEURS: Few-shot Learning Evaluation of Universal Representations of Speech,30
|
| 541 |
+
188,1529637573462831105,"Autoformalization with Large Language Models
|
| 542 |
+
abs: https://t.co/SoGYXkMGhV
|
| 543 |
+
|
| 544 |
+
methodology results in a new state-of-th… https://t.co/pTxpC00QFC",Autoformalization with Large Language Models,24
|
| 545 |
+
189,1529630110885851136,"AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
|
| 546 |
+
abs: https://t.co/aD0daO7HEa
|
| 547 |
+
|
| 548 |
+
By… https://t.co/NW3DbOJdwH",AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models,64
|
| 549 |
+
190,1529625016471633920,"An Evolutionary Approach to Dynamic Introduction of Tasks in Large-scale Multitask Learning Systems
|
| 550 |
+
abs:… https://t.co/gks4xeDd22",An Evolutionary Approach to Dynamic Introduction of Tasks in Large-scale Multitask Learning Systems,10
|
| 551 |
+
191,1529341790335246336,"Policy Compliance Detection via Expression Tree Inference
|
| 552 |
+
abs: https://t.co/Ic7Wm852Qz https://t.co/4RtEnug1RD",Policy Compliance Detection via Expression Tree Inference,8
|
| 553 |
+
192,1529309686318653441,"History Compression via Language Models in Reinforcement Learning
|
| 554 |
+
abs: https://t.co/N1smkJUAW9 https://t.co/4v1an4CkTU",History Compression via Language Models in Reinforcement Learning,85
|
| 555 |
+
193,1529303237572034560,"On the Role of Bidirectionality in Language Model Pre-Training
|
| 556 |
+
abs: https://t.co/fG2SbUhB1W
|
| 557 |
+
|
| 558 |
+
propose a new framewor… https://t.co/Gc40i0zyeV",On the Role of Bidirectionality in Language Model Pre-Training,26
|
| 559 |
+
194,1529301315221917699,"Large Language Models are Zero-Shot Reasoners
|
| 560 |
+
abs: https://t.co/GgdLms77wF
|
| 561 |
+
|
| 562 |
+
LLMs are decent zero-shot reasoners by… https://t.co/PTH6QpdSo2",Large Language Models are Zero-Shot Reasoners,85
|
| 563 |
+
195,1529278657856000000,"Naive Few-Shot Learning: Sequence Consistency Evaluation
|
| 564 |
+
abs: https://t.co/ySAzuujz2O https://t.co/aVVLHJdBUC",Naive Few-Shot Learning: Sequence Consistency Evaluation,19
|
| 565 |
+
196,1529075001256824834,"All Birds with One Stone: Multi-task Text Classification for Efficient Inference with One Forward Pass
|
| 566 |
+
abs:… https://t.co/fcPGWaFEk5",All Birds with One Stone: Multi-task Text Classification for Efficient Inference with One Forward Pass,12
|
| 567 |
+
197,1529071850860453888,"StreamingQA: A Benchmark for Adaptation to New Knowledge over Time in Question Answering Models
|
| 568 |
+
abs:… https://t.co/MDT1Bxw9by",StreamingQA: A Benchmark for Adaptation to New Knowledge over Time in Question Answering Models,20
|
| 569 |
+
198,1528909940324192256,"Contrastive and Non-Contrastive Self-Supervised Learning Recover Global and Local Spectral Embedding Methods
|
| 570 |
+
abs:… https://t.co/B65LGrnCLg",Contrastive and Non-Contrastive Self-Supervised Learning Recover Global and Local Spectral Embedding Methods,38
|
| 571 |
+
199,1528907841335066625,"Flexible Diffusion Modeling of Long Videos
|
| 572 |
+
abs: https://t.co/Cx1BUqA7zM
|
| 573 |
+
|
| 574 |
+
demonstrate improved video modeling over p… https://t.co/Y15RoaMAFg",Flexible Diffusion Modeling of Long Videos,84
|
| 575 |
+
200,1528904484553900034,"Scaling Laws and Interpretability of Learning from Repeated Data
|
| 576 |
+
abs: https://t.co/UbSQazzMwa
|
| 577 |
+
|
| 578 |
+
performance of 800M… https://t.co/4HHdSCe8ZT",Scaling Laws and Interpretability of Learning from Repeated Data,46
|
| 579 |
+
201,1528851863306752000,"Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding
|
| 580 |
+
project page:… https://t.co/5yJZQIqMdn",Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding,2724
|
| 581 |
+
202,1528584642407841792,"Self-Supervised Depth Estimation with Isometric-Self-Sample-Based Learning
|
| 582 |
+
abs: https://t.co/rE7gjT0COx https://t.co/EtbaT2jTle",Self-Supervised Depth Estimation with Isometric-Self-Sample-Based Learning,116
|
| 583 |
+
203,1528576691152494592,"Pre-Train Your Loss: Easy Bayesian Transfer Learning with Informative Priors
|
| 584 |
+
abs: https://t.co/bXqr4sP1V4
|
| 585 |
+
github:… https://t.co/efonU0Az1m",Pre-Train Your Loss: Easy Bayesian Transfer Learning with Informative Priors,59
|
| 586 |
+
204,1528569396456828929,"Why GANs are overkill for NLP
|
| 587 |
+
abs: https://t.co/zwjCFxh22z https://t.co/tuM1ufFC7x",Why GANs are overkill for NLP,139
|
| 588 |
+
205,1528555058916429828,"Lossless Acceleration for Seq2seq Generation with Aggressive Decoding
|
| 589 |
+
abs: https://t.co/7bmGFXe47E
|
| 590 |
+
github:… https://t.co/bXfTVfP56t",Lossless Acceleration for Seq2seq Generation with Aggressive Decoding,48
|
| 591 |
+
206,1528541664561844229,"Planning with Diffusion for Flexible Behavior Synthesis
|
| 592 |
+
abs: https://t.co/HSoQhC6WBV
|
| 593 |
+
project page:… https://t.co/PA69vLOYmb",Planning with Diffusion for Flexible Behavior Synthesis,119
|
| 594 |
+
207,1527765335528591361,"Disentangling Visual Embeddings for Attributes and Objects
|
| 595 |
+
abs: https://t.co/QlDsekM1rH https://t.co/YfsJGNzjlX",Disentangling Visual Embeddings for Attributes and Objects,253
|
| 596 |
+
208,1527452603264733184,"RankGen: Improving Text Generation with Large Ranking Models
|
| 597 |
+
abs: https://t.co/uVVfXNnZeR
|
| 598 |
+
github:… https://t.co/GfRxgf4hKe",RankGen: Improving Text Generation with Large Ranking Models,37
|
| 599 |
+
209,1527450826343604233,"Robust and Efficient Medical Imaging with Self-Supervision
|
| 600 |
+
abs: https://t.co/oBqk2TTp73
|
| 601 |
+
|
| 602 |
+
strategy leads to strong d… https://t.co/ptwGGG2NkL",Robust and Efficient Medical Imaging with Self-Supervision,45
|
| 603 |
+
210,1527097137825202177,"Masked Autoencoders As Spatiotemporal Learners
|
| 604 |
+
abs: https://t.co/MWlK2uV6qF
|
| 605 |
+
|
| 606 |
+
MAE method can learn strong representa… https://t.co/KX2kb7Zf0m",Masked Autoencoders As Spatiotemporal Learners,288
|
| 607 |
+
211,1527092033374113795,"Meta-Learning Sparse Compression Networks
|
| 608 |
+
abs: https://t.co/pDKyAXyGmg https://t.co/ExJQyGQefn",Meta-Learning Sparse Compression Networks,34
|
| 609 |
+
212,1526825242026512385,"An Empirical Investigation of Representation Learning for Imitation
|
| 610 |
+
abs: https://t.co/P6C15OJ0ft https://t.co/C0PcBJ72kH",An Empirical Investigation of Representation Learning for Imitation,40
|
| 611 |
+
213,1526798231191048192,"Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space
|
| 612 |
+
abs: https://t.co/kKjj0oKSyE https://t.co/bk1DPxlwZ9",Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space,49
|
| 613 |
+
214,1526738002621472768,"SKILL: Structured Knowledge Infusion for Large Language Models
|
| 614 |
+
abs: https://t.co/vbExGmg4hx https://t.co/3hVTWxLVE1",SKILL: Structured Knowledge Infusion for Large Language Models,46
|
| 615 |
+
215,1526435187093123072,"FactPEGASUS: Factuality-Aware Pre-training and Fine-tuning for Abstractive Summarization
|
| 616 |
+
abs:… https://t.co/qG4s6MlGmd",FactPEGASUS: Factuality-Aware Pre-training and Fine-tuning for Abstractive Summarization,46
|
| 617 |
+
216,1526428632868167682,"PrefixRL: Optimization of Parallel Prefix Circuits using Deep Reinforcement Learning
|
| 618 |
+
abs: https://t.co/JTAU1vnmst https://t.co/fyonWO1rKa",PrefixRL: Optimization of Parallel Prefix Circuits using Deep Reinforcement Learning,45
|
| 619 |
+
217,1526373200183033857,"Diffusion Models for Adversarial Purification
|
| 620 |
+
abs: https://t.co/VdSXsTahOY https://t.co/lFxumNcuIj",Diffusion Models for Adversarial Purification,135
|