Ananthusajeev190 commited on
Commit
490bf5e
·
verified ·
1 Parent(s): 6cee5e8

Upload index.json

Browse files
Files changed (1) hide show
  1. index.json +1017 -0
index.json ADDED
@@ -0,0 +1,1017 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "created_at": "2025-10-14T09:52:41Z",
3
+ "repos": [
4
+ {
5
+ "repo_url": "https://github.com/ageron/handson-ml2.git",
6
+ "repo_name": "handson-ml2",
7
+ "archived_at": "2025-10-14T09:52:45Z",
8
+ "num_files_copied": 36,
9
+ "total_bytes": 33551736,
10
+ "readme_snippet": "Machine Learning Notebooks\n==========================\n\n# \u26a0 The 3rd edition of my book will be released in October 2022. The notebooks are available at [ageron/handson-ml3](https://github.com/ageron/handson-ml3) and contain more up-to-date code.\n\nThis project aims at teaching you the fundamentals of Machine Learning in\npython. It contains the example code and solutions to the exercises in the second edition of my O'Reilly book [Hands-on Machine Learning with Scikit-Learn, Keras and TensorFlow](https://www.oreilly.com/library/view/hands-on-machine-learning/9781492032632/):\n\n<img src=\"https://images-na.ssl-images-amazon.com/images/I/51aqYc1QyrL._SX379_BO1,204,203,200_.jpg\" title=\"book\" width=\"150\" />\n\n**Note**: If you are looking for the first edition notebooks, check out [ageron/handson-ml](https://github.com/ageron/handson-ml). For the third edition, check out [ageron/handson-ml3](https://github.com/ageron/handson-ml3).\n\n## Quick Start\n\n### Want to play with these notebooks online without having to install anything?\nUse any of the following services (I recommended Colab or Kaggle, since they offer free GPUs and TPUs).\n\n**WARNING**: _Please be aware that these services provide temporary environments: anything you do will be deleted after a while, so make sure you download any data you care about._\n\n* <a href=\"https://colab.research.google.com/github/ageron/handson-ml2/blob/master/\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n\n* <a href=\"https://homl.info/kaggle/\"><img src=\"https://kaggle.com/static/images/open-in-kaggle.svg\" alt=\"Open in Kaggle\" /></a>\n\n* <a href=\"https://mybinder.org/v2/gh/ageron/handson-ml2/HEAD?filepath=%2Findex.ipynb\"><img src=\"https://mybinder.org/badge_logo.svg\" alt=\"Launch binder\" /></a>\n\n* <a href=\"https://homl.info/deepnote/\"><img src=\"https://deepnote.com/buttons/launch-in-deepnote-small.svg\" alt=\"Launch in Deepnote\" /></a>\n\n### Just want to quickly look at some notebooks, wi",
11
+ "license_snippet": " Apache License\n Version 2.0, January 2004\n http://www.apache.org/licenses/\n\n TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION\n\n 1. Definitions.\n\n \"License\" shall mean the terms and conditions for use, reproduction,\n and distribution as defined by Sections 1 through 9 of this document.\n\n \"Licensor\" shall mean the copyright owner or entity authorized by\n the copyright owner that is granting the License.\n\n \"Legal Entity\" shall mean the union of the acting entity and all\n other entities that control, are controlled by, or are under common\n control with that entity. For the purposes of this definition,\n \"control\" means (i) the power, direct or indirect, to cause the\n direction or management of such entity, whether by contract or\n otherwise, or (ii) ownership of fifty percent (50%) or more of the\n outstanding shares, or (iii) beneficial ownership of such entity.\n\n \"You\" (or \"Your\") shall mean an individual or Legal Entity\n exercising permissions granted by this License.\n\n \"Source\" form shall mean the preferred form for making modifications,\n including but not limited to software source code, documentation\n source, and configuration files.\n\n \"Object\" form shall mean any form resulting from mechanical\n transformation or translation of a Source form, including but\n not limited to compiled object code, generated documentation,\n and conversions to other media types.\n\n \"Work\" shall mean the work of authorship, whether in Source or\n Object form, made available under the License, as indicated by a\n copyright notice that is included in or attached to the work\n (an example is provided in the Appendix below).\n\n \"Derivative Works\" shall mean any work, whether in Source or Object\n form, that is based on (or derived from) the Work and for which the\n editori",
12
+ "files": [
13
+ {
14
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/09_unsupervised_learning.ipynb",
15
+ "archive_path": "handson-ml2/09_unsupervised_learning.ipynb",
16
+ "extension": ".ipynb",
17
+ "size_bytes": 3937187
18
+ },
19
+ {
20
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/extra_gradient_descent_comparison.ipynb",
21
+ "archive_path": "handson-ml2/extra_gradient_descent_comparison.ipynb",
22
+ "extension": ".ipynb",
23
+ "size_bytes": 310911
24
+ },
25
+ {
26
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/12_custom_models_and_training_with_tensorflow.ipynb",
27
+ "archive_path": "handson-ml2/12_custom_models_and_training_with_tensorflow.ipynb",
28
+ "extension": ".ipynb",
29
+ "size_bytes": 188683
30
+ },
31
+ {
32
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/10_neural_nets_with_keras.ipynb",
33
+ "archive_path": "handson-ml2/10_neural_nets_with_keras.ipynb",
34
+ "extension": ".ipynb",
35
+ "size_bytes": 454450
36
+ },
37
+ {
38
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/changes_in_2nd_edition.md",
39
+ "archive_path": "handson-ml2/changes_in_2nd_edition.md",
40
+ "extension": ".md",
41
+ "size_bytes": 9604
42
+ },
43
+ {
44
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/13_loading_and_preprocessing_data.ipynb",
45
+ "archive_path": "handson-ml2/13_loading_and_preprocessing_data.ipynb",
46
+ "extension": ".ipynb",
47
+ "size_bytes": 561951
48
+ },
49
+ {
50
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/environment.yml",
51
+ "archive_path": "handson-ml2/environment.yml",
52
+ "extension": ".yml",
53
+ "size_bytes": 2898
54
+ },
55
+ {
56
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/18_reinforcement_learning.ipynb",
57
+ "archive_path": "handson-ml2/18_reinforcement_learning.ipynb",
58
+ "extension": ".ipynb",
59
+ "size_bytes": 388304
60
+ },
61
+ {
62
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/02_end_to_end_machine_learning_project.ipynb",
63
+ "archive_path": "handson-ml2/02_end_to_end_machine_learning_project.ipynb",
64
+ "extension": ".ipynb",
65
+ "size_bytes": 1333744
66
+ },
67
+ {
68
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/03_classification.ipynb",
69
+ "archive_path": "handson-ml2/03_classification.ipynb",
70
+ "extension": ".ipynb",
71
+ "size_bytes": 819860
72
+ },
73
+ {
74
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/04_training_linear_models.ipynb",
75
+ "archive_path": "handson-ml2/04_training_linear_models.ipynb",
76
+ "extension": ".ipynb",
77
+ "size_bytes": 763839
78
+ },
79
+ {
80
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/math_differential_calculus.ipynb",
81
+ "archive_path": "handson-ml2/math_differential_calculus.ipynb",
82
+ "extension": ".ipynb",
83
+ "size_bytes": 607435
84
+ },
85
+ {
86
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/08_dimensionality_reduction.ipynb",
87
+ "archive_path": "handson-ml2/08_dimensionality_reduction.ipynb",
88
+ "extension": ".ipynb",
89
+ "size_bytes": 3800716
90
+ },
91
+ {
92
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/16_nlp_with_rnns_and_attention.ipynb",
93
+ "archive_path": "handson-ml2/16_nlp_with_rnns_and_attention.ipynb",
94
+ "extension": ".ipynb",
95
+ "size_bytes": 259287
96
+ },
97
+ {
98
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/17_autoencoders_and_gans.ipynb",
99
+ "archive_path": "handson-ml2/17_autoencoders_and_gans.ipynb",
100
+ "extension": ".ipynb",
101
+ "size_bytes": 6576798
102
+ },
103
+ {
104
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/11_training_deep_neural_networks.ipynb",
105
+ "archive_path": "handson-ml2/11_training_deep_neural_networks.ipynb",
106
+ "extension": ".ipynb",
107
+ "size_bytes": 391237
108
+ },
109
+ {
110
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/tools_pandas.ipynb",
111
+ "archive_path": "handson-ml2/tools_pandas.ipynb",
112
+ "extension": ".ipynb",
113
+ "size_bytes": 454974
114
+ },
115
+ {
116
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/07_ensemble_learning_and_random_forests.ipynb",
117
+ "archive_path": "handson-ml2/07_ensemble_learning_and_random_forests.ipynb",
118
+ "extension": ".ipynb",
119
+ "size_bytes": 534446
120
+ },
121
+ {
122
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/19_training_and_deploying_at_scale.ipynb",
123
+ "archive_path": "handson-ml2/19_training_and_deploying_at_scale.ipynb",
124
+ "extension": ".ipynb",
125
+ "size_bytes": 72608
126
+ },
127
+ {
128
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/05_support_vector_machines.ipynb",
129
+ "archive_path": "handson-ml2/05_support_vector_machines.ipynb",
130
+ "extension": ".ipynb",
131
+ "size_bytes": 868577
132
+ },
133
+ {
134
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/06_decision_trees.ipynb",
135
+ "archive_path": "handson-ml2/06_decision_trees.ipynb",
136
+ "extension": ".ipynb",
137
+ "size_bytes": 216835
138
+ },
139
+ {
140
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/math_linear_algebra.ipynb",
141
+ "archive_path": "handson-ml2/math_linear_algebra.ipynb",
142
+ "extension": ".ipynb",
143
+ "size_bytes": 673875
144
+ },
145
+ {
146
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/tools_matplotlib.ipynb",
147
+ "archive_path": "handson-ml2/tools_matplotlib.ipynb",
148
+ "extension": ".ipynb",
149
+ "size_bytes": 3596357
150
+ },
151
+ {
152
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/INSTALL.md",
153
+ "archive_path": "handson-ml2/INSTALL.md",
154
+ "extension": ".md",
155
+ "size_bytes": 6228
156
+ },
157
+ {
158
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/ml-project-checklist.md",
159
+ "archive_path": "handson-ml2/ml-project-checklist.md",
160
+ "extension": ".md",
161
+ "size_bytes": 7689
162
+ },
163
+ {
164
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/15_processing_sequences_using_rnns_and_cnns.ipynb",
165
+ "archive_path": "handson-ml2/15_processing_sequences_using_rnns_and_cnns.ipynb",
166
+ "extension": ".ipynb",
167
+ "size_bytes": 1261718
168
+ },
169
+ {
170
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/tools_numpy.ipynb",
171
+ "archive_path": "handson-ml2/tools_numpy.ipynb",
172
+ "extension": ".ipynb",
173
+ "size_bytes": 630631
174
+ },
175
+ {
176
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/14_deep_computer_vision_with_cnns.ipynb",
177
+ "archive_path": "handson-ml2/14_deep_computer_vision_with_cnns.ipynb",
178
+ "extension": ".ipynb",
179
+ "size_bytes": 4452678
180
+ },
181
+ {
182
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/extra_autodiff.ipynb",
183
+ "archive_path": "handson-ml2/extra_autodiff.ipynb",
184
+ "extension": ".ipynb",
185
+ "size_bytes": 32998
186
+ },
187
+ {
188
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/01_the_machine_learning_landscape.ipynb",
189
+ "archive_path": "handson-ml2/01_the_machine_learning_landscape.ipynb",
190
+ "extension": ".ipynb",
191
+ "size_bytes": 323557
192
+ },
193
+ {
194
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/index.ipynb",
195
+ "archive_path": "handson-ml2/index.ipynb",
196
+ "extension": ".ipynb",
197
+ "size_bytes": 5598
198
+ },
199
+ {
200
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/docker/jupyter_notebook_config.py",
201
+ "archive_path": "handson-ml2/docker/jupyter_notebook_config.py",
202
+ "extension": ".py",
203
+ "size_bytes": 691
204
+ },
205
+ {
206
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/docker/docker-compose.yml",
207
+ "archive_path": "handson-ml2/docker/docker-compose.yml",
208
+ "extension": ".yml",
209
+ "size_bytes": 664
210
+ },
211
+ {
212
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/.github/ISSUE_TEMPLATE/feature-or-improvement-request.md",
213
+ "archive_path": "handson-ml2/.github/ISSUE_TEMPLATE/feature-or-improvement-request.md",
214
+ "extension": ".md",
215
+ "size_bytes": 1270
216
+ },
217
+ {
218
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/.github/ISSUE_TEMPLATE/bug_report.md",
219
+ "archive_path": "handson-ml2/.github/ISSUE_TEMPLATE/bug_report.md",
220
+ "extension": ".md",
221
+ "size_bytes": 1666
222
+ },
223
+ {
224
+ "source_path": "/tmp/gh_clone_cw5x0lnz/handson-ml2/.github/ISSUE_TEMPLATE/clarification-request.md",
225
+ "archive_path": "handson-ml2/.github/ISSUE_TEMPLATE/clarification-request.md",
226
+ "extension": ".md",
227
+ "size_bytes": 1772
228
+ }
229
+ ]
230
+ },
231
+ {
232
+ "repo_url": "https://github.com/pytorch/examples.git",
233
+ "repo_name": "examples",
234
+ "archived_at": "2025-10-14T09:52:50Z",
235
+ "num_files_copied": 129,
236
+ "total_bytes": 514568,
237
+ "readme_snippet": "# PyTorch Examples\n\nhttps://pytorch.org/examples/\n\n`pytorch/examples` is a repository showcasing examples of using [PyTorch](https://github.com/pytorch/pytorch). The goal is to have curated, short, few/no dependencies _high quality_ examples that are substantially different from each other that can be emulated in your existing work.\n\n- For tutorials: https://github.com/pytorch/tutorials\n- For changes to pytorch.org: https://github.com/pytorch/pytorch.github.io\n- For a general model hub: https://pytorch.org/hub/ or https://huggingface.co/models\n- For recipes on how to run PyTorch in production: https://github.com/facebookresearch/recipes\n- For general Q&A and support: https://discuss.pytorch.org/\n\n## Available models\n\n- [Image classification (MNIST) using Convnets](./mnist/README.md)\n- [Word-level Language Modeling using RNN and Transformer](./word_language_model/README.md)\n- [Training Imagenet Classifiers with Popular Networks](./imagenet/README.md)\n- [Generative Adversarial Networks (DCGAN)](./dcgan/README.md)\n- [Variational Auto-Encoders](./vae/README.md)\n- [Superresolution using an efficient sub-pixel convolutional neural network](./super_resolution/README.md)\n- [Hogwild training of shared ConvNets across multiple processes on MNIST](mnist_hogwild)\n- [Training a CartPole to balance with actor-critic](./reinforcement_learning/README.md)\n- [Natural Language Inference (SNLI) with GloVe vectors, LSTMs, and torchtext](snli)\n- [Time sequence prediction - use an LSTM to learn Sine waves](./time_sequence_prediction/README.md)\n- [Implement the Neural Style Transfer algorithm on images](./fast_neural_style/README.md)\n- [Reinforcement Learning with Actor Critic and REINFORCE algorithms on OpenAI gym](./reinforcement_learning/README.md)\n- [PyTorch Module Transformations using fx](./fx/README.md)\n- Distributed PyTorch examples with [Distributed Data Parallel](./distributed/ddp/README.md) and [RPC](./distributed/rpc)\n- [Several examples illustrating the C++ Frontend](cpp)\n- [I",
238
+ "license_snippet": "BSD 3-Clause License\n\nCopyright (c) 2017, Pytorch contributors\nAll rights reserved.\n\nRedistribution and use in source and binary forms, with or without\nmodification, are permitted provided that the following conditions are met:\n\n* Redistributions of source code must retain the above copyright notice, this\n list of conditions and the following disclaimer.\n\n* Redistributions in binary form must reproduce the above copyright notice,\n this list of conditions and the following disclaimer in the documentation\n and/or other materials provided with the distribution.\n\n* Neither the name of the copyright holder nor the names of its\n contributors may be used to endorse or promote products derived from\n this software without specific prior written permission.\n\nTHIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\nAND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\nIMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\nDISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\nFOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\nDAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\nSERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\nCAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\nOR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\nOF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n",
239
+ "files": [
240
+ {
241
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/run_cpp_examples.sh",
242
+ "archive_path": "examples/run_cpp_examples.sh",
243
+ "extension": ".sh",
244
+ "size_bytes": 4116
245
+ },
246
+ {
247
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/run_python_examples.sh",
248
+ "archive_path": "examples/run_python_examples.sh",
249
+ "extension": ".sh",
250
+ "size_bytes": 8648
251
+ },
252
+ {
253
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/CONTRIBUTING.md",
254
+ "archive_path": "examples/CONTRIBUTING.md",
255
+ "extension": ".md",
256
+ "size_bytes": 2395
257
+ },
258
+ {
259
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/run_distributed_examples.sh",
260
+ "archive_path": "examples/run_distributed_examples.sh",
261
+ "extension": ".sh",
262
+ "size_bytes": 3029
263
+ },
264
+ {
265
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/utils.sh",
266
+ "archive_path": "examples/utils.sh",
267
+ "extension": ".sh",
268
+ "size_bytes": 1597
269
+ },
270
+ {
271
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/CODE_OF_CONDUCT.md",
272
+ "archive_path": "examples/CODE_OF_CONDUCT.md",
273
+ "extension": ".md",
274
+ "size_bytes": 3353
275
+ },
276
+ {
277
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/mnist/mnist.cpp",
278
+ "archive_path": "examples/cpp/mnist/mnist.cpp",
279
+ "extension": ".cpp",
280
+ "size_bytes": 4746
281
+ },
282
+ {
283
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/distributed/dist-mnist.cpp",
284
+ "archive_path": "examples/cpp/distributed/dist-mnist.cpp",
285
+ "extension": ".cpp",
286
+ "size_bytes": 5596
287
+ },
288
+ {
289
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/dcgan/dcgan.cpp",
290
+ "archive_path": "examples/cpp/dcgan/dcgan.cpp",
291
+ "extension": ".cpp",
292
+ "size_bytes": 8000
293
+ },
294
+ {
295
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/dcgan/display_samples.py",
296
+ "archive_path": "examples/cpp/dcgan/display_samples.py",
297
+ "extension": ".py",
298
+ "size_bytes": 891
299
+ },
300
+ {
301
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/transfer-learning/convert.py",
302
+ "archive_path": "examples/cpp/transfer-learning/convert.py",
303
+ "extension": ".py",
304
+ "size_bytes": 586
305
+ },
306
+ {
307
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/transfer-learning/main.h",
308
+ "archive_path": "examples/cpp/transfer-learning/main.h",
309
+ "extension": ".h",
310
+ "size_bytes": 2412
311
+ },
312
+ {
313
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/transfer-learning/classify.cpp",
314
+ "archive_path": "examples/cpp/transfer-learning/classify.cpp",
315
+ "extension": ".cpp",
316
+ "size_bytes": 2676
317
+ },
318
+ {
319
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/transfer-learning/main.cpp",
320
+ "archive_path": "examples/cpp/transfer-learning/main.cpp",
321
+ "extension": ".cpp",
322
+ "size_bytes": 9714
323
+ },
324
+ {
325
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/autograd/autograd.cpp",
326
+ "archive_path": "examples/cpp/autograd/autograd.cpp",
327
+ "extension": ".cpp",
328
+ "size_bytes": 6003
329
+ },
330
+ {
331
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/custom-dataset/custom-dataset.cpp",
332
+ "archive_path": "examples/cpp/custom-dataset/custom-dataset.cpp",
333
+ "extension": ".cpp",
334
+ "size_bytes": 6949
335
+ },
336
+ {
337
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/regression/regression.cpp",
338
+ "archive_path": "examples/cpp/regression/regression.cpp",
339
+ "extension": ".cpp",
340
+ "size_bytes": 2321
341
+ },
342
+ {
343
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/tools/download_mnist.py",
344
+ "archive_path": "examples/cpp/tools/download_mnist.py",
345
+ "extension": ".py",
346
+ "size_bytes": 2723
347
+ },
348
+ {
349
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/cpp/tools/InstallingOpenCV.md",
350
+ "archive_path": "examples/cpp/tools/InstallingOpenCV.md",
351
+ "extension": ".md",
352
+ "size_bytes": 958
353
+ },
354
+ {
355
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/vae/main.py",
356
+ "archive_path": "examples/vae/main.py",
357
+ "extension": ".py",
358
+ "size_bytes": 4899
359
+ },
360
+ {
361
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/mnist/main.py",
362
+ "archive_path": "examples/mnist/main.py",
363
+ "extension": ".py",
364
+ "size_bytes": 5392
365
+ },
366
+ {
367
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fast_neural_style/download_saved_models.py",
368
+ "archive_path": "examples/fast_neural_style/download_saved_models.py",
369
+ "extension": ".py",
370
+ "size_bytes": 931
371
+ },
372
+ {
373
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fast_neural_style/neural_style/transformer_net.py",
374
+ "archive_path": "examples/fast_neural_style/neural_style/transformer_net.py",
375
+ "extension": ".py",
376
+ "size_bytes": 3782
377
+ },
378
+ {
379
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fast_neural_style/neural_style/vgg.py",
380
+ "archive_path": "examples/fast_neural_style/neural_style/vgg.py",
381
+ "extension": ".py",
382
+ "size_bytes": 1398
383
+ },
384
+ {
385
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fast_neural_style/neural_style/utils.py",
386
+ "archive_path": "examples/fast_neural_style/neural_style/utils.py",
387
+ "extension": ".py",
388
+ "size_bytes": 997
389
+ },
390
+ {
391
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fast_neural_style/neural_style/neural_style.py",
392
+ "archive_path": "examples/fast_neural_style/neural_style/neural_style.py",
393
+ "extension": ".py",
394
+ "size_bytes": 10889
395
+ },
396
+ {
397
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fast_neural_style/neural_style/__init__.py",
398
+ "archive_path": "examples/fast_neural_style/neural_style/__init__.py",
399
+ "extension": ".py",
400
+ "size_bytes": 0
401
+ },
402
+ {
403
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/mnist_forward_forward/main.py",
404
+ "archive_path": "examples/mnist_forward_forward/main.py",
405
+ "extension": ".py",
406
+ "size_bytes": 5833
407
+ },
408
+ {
409
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/imagenet/extract_ILSVRC.sh",
410
+ "archive_path": "examples/imagenet/extract_ILSVRC.sh",
411
+ "extension": ".sh",
412
+ "size_bytes": 2654
413
+ },
414
+ {
415
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/imagenet/main.py",
416
+ "archive_path": "examples/imagenet/main.py",
417
+ "extension": ".py",
418
+ "size_bytes": 20959
419
+ },
420
+ {
421
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/time_sequence_prediction/generate_sine_wave.py",
422
+ "archive_path": "examples/time_sequence_prediction/generate_sine_wave.py",
423
+ "extension": ".py",
424
+ "size_bytes": 275
425
+ },
426
+ {
427
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/time_sequence_prediction/train.py",
428
+ "archive_path": "examples/time_sequence_prediction/train.py",
429
+ "extension": ".py",
430
+ "size_bytes": 3333
431
+ },
432
+ {
433
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/legacy/snli/model.py",
434
+ "archive_path": "examples/legacy/snli/model.py",
435
+ "extension": ".py",
436
+ "size_bytes": 2655
437
+ },
438
+ {
439
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/legacy/snli/train.py",
440
+ "archive_path": "examples/legacy/snli/train.py",
441
+ "extension": ".py",
442
+ "size_bytes": 5354
443
+ },
444
+ {
445
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/legacy/snli/util.py",
446
+ "archive_path": "examples/legacy/snli/util.py",
447
+ "extension": ".py",
448
+ "size_bytes": 3885
449
+ },
450
+ {
451
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/reinforcement_learning/actor_critic.py",
452
+ "archive_path": "examples/reinforcement_learning/actor_critic.py",
453
+ "extension": ".py",
454
+ "size_bytes": 5370
455
+ },
456
+ {
457
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/reinforcement_learning/reinforce.py",
458
+ "archive_path": "examples/reinforcement_learning/reinforce.py",
459
+ "extension": ".py",
460
+ "size_bytes": 3384
461
+ },
462
+ {
463
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/super_resolution/super_resolve.py",
464
+ "archive_path": "examples/super_resolution/super_resolve.py",
465
+ "extension": ".py",
466
+ "size_bytes": 1720
467
+ },
468
+ {
469
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/super_resolution/dataset.py",
470
+ "archive_path": "examples/super_resolution/dataset.py",
471
+ "extension": ".py",
472
+ "size_bytes": 1083
473
+ },
474
+ {
475
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/super_resolution/model.py",
476
+ "archive_path": "examples/super_resolution/model.py",
477
+ "extension": ".py",
478
+ "size_bytes": 1060
479
+ },
480
+ {
481
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/super_resolution/data.py",
482
+ "archive_path": "examples/super_resolution/data.py",
483
+ "extension": ".py",
484
+ "size_bytes": 2044
485
+ },
486
+ {
487
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/super_resolution/main.py",
488
+ "archive_path": "examples/super_resolution/main.py",
489
+ "extension": ".py",
490
+ "size_bytes": 3094
491
+ },
492
+ {
493
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/tensor_parallelism/log_utils.py",
494
+ "archive_path": "examples/distributed/tensor_parallelism/log_utils.py",
495
+ "extension": ".py",
496
+ "size_bytes": 611
497
+ },
498
+ {
499
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/tensor_parallelism/llama2_model.py",
500
+ "archive_path": "examples/distributed/tensor_parallelism/llama2_model.py",
501
+ "extension": ".py",
502
+ "size_bytes": 16585
503
+ },
504
+ {
505
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/tensor_parallelism/tensor_parallel_example.py",
506
+ "archive_path": "examples/distributed/tensor_parallelism/tensor_parallel_example.py",
507
+ "extension": ".py",
508
+ "size_bytes": 3975
509
+ },
510
+ {
511
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/tensor_parallelism/sequence_parallel_example.py",
512
+ "archive_path": "examples/distributed/tensor_parallelism/sequence_parallel_example.py",
513
+ "extension": ".py",
514
+ "size_bytes": 3324
515
+ },
516
+ {
517
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/tensor_parallelism/fsdp_tp_example.py",
518
+ "archive_path": "examples/distributed/tensor_parallelism/fsdp_tp_example.py",
519
+ "extension": ".py",
520
+ "size_bytes": 6114
521
+ },
522
+ {
523
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/tensor_parallelism/run_example.sh",
524
+ "archive_path": "examples/distributed/tensor_parallelism/run_example.sh",
525
+ "extension": ".sh",
526
+ "size_bytes": 497
527
+ },
528
+ {
529
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/download_dataset.sh",
530
+ "archive_path": "examples/distributed/FSDP/download_dataset.sh",
531
+ "extension": ".sh",
532
+ "size_bytes": 285
533
+ },
534
+ {
535
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/T5_training.py",
536
+ "archive_path": "examples/distributed/FSDP/T5_training.py",
537
+ "extension": ".py",
538
+ "size_bytes": 7857
539
+ },
540
+ {
541
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/summarization_dataset.py",
542
+ "archive_path": "examples/distributed/FSDP/summarization_dataset.py",
543
+ "extension": ".py",
544
+ "size_bytes": 2835
545
+ },
546
+ {
547
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/utils/environment.py",
548
+ "archive_path": "examples/distributed/FSDP/utils/environment.py",
549
+ "extension": ".py",
550
+ "size_bytes": 982
551
+ },
552
+ {
553
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/utils/train_utils.py",
554
+ "archive_path": "examples/distributed/FSDP/utils/train_utils.py",
555
+ "extension": ".py",
556
+ "size_bytes": 3187
557
+ },
558
+ {
559
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/utils/__init__.py",
560
+ "archive_path": "examples/distributed/FSDP/utils/__init__.py",
561
+ "extension": ".py",
562
+ "size_bytes": 203
563
+ },
564
+ {
565
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/model_checkpointing/checkpoint_handler.py",
566
+ "archive_path": "examples/distributed/FSDP/model_checkpointing/checkpoint_handler.py",
567
+ "extension": ".py",
568
+ "size_bytes": 8863
569
+ },
570
+ {
571
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/model_checkpointing/__init__.py",
572
+ "archive_path": "examples/distributed/FSDP/model_checkpointing/__init__.py",
573
+ "extension": ".py",
574
+ "size_bytes": 292
575
+ },
576
+ {
577
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/configs/training.py",
578
+ "archive_path": "examples/distributed/FSDP/configs/training.py",
579
+ "extension": ".py",
580
+ "size_bytes": 388
581
+ },
582
+ {
583
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/configs/__init__.py",
584
+ "archive_path": "examples/distributed/FSDP/configs/__init__.py",
585
+ "extension": ".py",
586
+ "size_bytes": 65
587
+ },
588
+ {
589
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/configs/fsdp.py",
590
+ "archive_path": "examples/distributed/FSDP/configs/fsdp.py",
591
+ "extension": ".py",
592
+ "size_bytes": 654
593
+ },
594
+ {
595
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/policies/mixed_precision.py",
596
+ "archive_path": "examples/distributed/FSDP/policies/mixed_precision.py",
597
+ "extension": ".py",
598
+ "size_bytes": 862
599
+ },
600
+ {
601
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/policies/wrapping.py",
602
+ "archive_path": "examples/distributed/FSDP/policies/wrapping.py",
603
+ "extension": ".py",
604
+ "size_bytes": 1152
605
+ },
606
+ {
607
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/policies/activation_checkpointing_functions.py",
608
+ "archive_path": "examples/distributed/FSDP/policies/activation_checkpointing_functions.py",
609
+ "extension": ".py",
610
+ "size_bytes": 816
611
+ },
612
+ {
613
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP/policies/__init__.py",
614
+ "archive_path": "examples/distributed/FSDP/policies/__init__.py",
615
+ "extension": ".py",
616
+ "size_bytes": 128
617
+ },
618
+ {
619
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP2/example.py",
620
+ "archive_path": "examples/distributed/FSDP2/example.py",
621
+ "extension": ".py",
622
+ "size_bytes": 4136
623
+ },
624
+ {
625
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP2/model.py",
626
+ "archive_path": "examples/distributed/FSDP2/model.py",
627
+ "extension": ".py",
628
+ "size_bytes": 4604
629
+ },
630
+ {
631
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP2/checkpoint.py",
632
+ "archive_path": "examples/distributed/FSDP2/checkpoint.py",
633
+ "extension": ".py",
634
+ "size_bytes": 7707
635
+ },
636
+ {
637
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP2/utils.py",
638
+ "archive_path": "examples/distributed/FSDP2/utils.py",
639
+ "extension": ".py",
640
+ "size_bytes": 675
641
+ },
642
+ {
643
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/FSDP2/run_example.sh",
644
+ "archive_path": "examples/distributed/FSDP2/run_example.sh",
645
+ "extension": ".sh",
646
+ "size_bytes": 348
647
+ },
648
+ {
649
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp/example.py",
650
+ "archive_path": "examples/distributed/ddp/example.py",
651
+ "extension": ".py",
652
+ "size_bytes": 3481
653
+ },
654
+ {
655
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp/run_example.sh",
656
+ "archive_path": "examples/distributed/ddp/run_example.sh",
657
+ "extension": ".sh",
658
+ "size_bytes": 394
659
+ },
660
+ {
661
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/pipeline/main.py",
662
+ "archive_path": "examples/distributed/rpc/pipeline/main.py",
663
+ "extension": ".py",
664
+ "size_bytes": 9675
665
+ },
666
+ {
667
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/rnn/rnn.py",
668
+ "archive_path": "examples/distributed/rpc/rnn/rnn.py",
669
+ "extension": ".py",
670
+ "size_bytes": 3692
671
+ },
672
+ {
673
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/rnn/main.py",
674
+ "archive_path": "examples/distributed/rpc/rnn/main.py",
675
+ "extension": ".py",
676
+ "size_bytes": 2485
677
+ },
678
+ {
679
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/batch/reinforce.py",
680
+ "archive_path": "examples/distributed/rpc/batch/reinforce.py",
681
+ "extension": ".py",
682
+ "size_bytes": 9892
683
+ },
684
+ {
685
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/batch/parameter_server.py",
686
+ "archive_path": "examples/distributed/rpc/batch/parameter_server.py",
687
+ "extension": ".py",
688
+ "size_bytes": 4705
689
+ },
690
+ {
691
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/parameter_server/rpc_parameter_server.py",
692
+ "archive_path": "examples/distributed/rpc/parameter_server/rpc_parameter_server.py",
693
+ "extension": ".py",
694
+ "size_bytes": 11628
695
+ },
696
+ {
697
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/ddp_rpc/main.py",
698
+ "archive_path": "examples/distributed/rpc/ddp_rpc/main.py",
699
+ "extension": ".py",
700
+ "size_bytes": 6922
701
+ },
702
+ {
703
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/rpc/rl/main.py",
704
+ "archive_path": "examples/distributed/rpc/rl/main.py",
705
+ "extension": ".py",
706
+ "size_bytes": 9133
707
+ },
708
+ {
709
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/multigpu.py",
710
+ "archive_path": "examples/distributed/ddp-tutorial-series/multigpu.py",
711
+ "extension": ".py",
712
+ "size_bytes": 3674
713
+ },
714
+ {
715
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/datautils.py",
716
+ "archive_path": "examples/distributed/ddp-tutorial-series/datautils.py",
717
+ "extension": ".py",
718
+ "size_bytes": 331
719
+ },
720
+ {
721
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/single_gpu.py",
722
+ "archive_path": "examples/distributed/ddp-tutorial-series/single_gpu.py",
723
+ "extension": ".py",
724
+ "size_bytes": 2784
725
+ },
726
+ {
727
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/multinode.py",
728
+ "archive_path": "examples/distributed/ddp-tutorial-series/multinode.py",
729
+ "extension": ".py",
730
+ "size_bytes": 4106
731
+ },
732
+ {
733
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/multigpu_torchrun.py",
734
+ "archive_path": "examples/distributed/ddp-tutorial-series/multigpu_torchrun.py",
735
+ "extension": ".py",
736
+ "size_bytes": 4021
737
+ },
738
+ {
739
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/slurm/sbatch_run.sh",
740
+ "archive_path": "examples/distributed/ddp-tutorial-series/slurm/sbatch_run.sh",
741
+ "extension": ".sh",
742
+ "size_bytes": 553
743
+ },
744
+ {
745
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/ddp-tutorial-series/slurm/setup_pcluster_slurm.md",
746
+ "archive_path": "examples/distributed/ddp-tutorial-series/slurm/setup_pcluster_slurm.md",
747
+ "extension": ".md",
748
+ "size_bytes": 1114
749
+ },
750
+ {
751
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/run_example.sh",
752
+ "archive_path": "examples/distributed/minGPT-ddp/run_example.sh",
753
+ "extension": ".sh",
754
+ "size_bytes": 326
755
+ },
756
+ {
757
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/model.py",
758
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/model.py",
759
+ "extension": ".py",
760
+ "size_bytes": 11128
761
+ },
762
+ {
763
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/gpt2_train_cfg.yaml",
764
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/gpt2_train_cfg.yaml",
765
+ "extension": ".yaml",
766
+ "size_bytes": 468
767
+ },
768
+ {
769
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/trainer.py",
770
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/trainer.py",
771
+ "extension": ".py",
772
+ "size_bytes": 5993
773
+ },
774
+ {
775
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/char_dataset.py",
776
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/char_dataset.py",
777
+ "extension": ".py",
778
+ "size_bytes": 1410
779
+ },
780
+ {
781
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/main.py",
782
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/main.py",
783
+ "extension": ".py",
784
+ "size_bytes": 2139
785
+ },
786
+ {
787
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/slurm/sbatch_run.sh",
788
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/slurm/sbatch_run.sh",
789
+ "extension": ".sh",
790
+ "size_bytes": 543
791
+ },
792
+ {
793
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/distributed/minGPT-ddp/mingpt/slurm/setup_pcluster_slurm.md",
794
+ "archive_path": "examples/distributed/minGPT-ddp/mingpt/slurm/setup_pcluster_slurm.md",
795
+ "extension": ".md",
796
+ "size_bytes": 1205
797
+ },
798
+ {
799
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/dcgan/main.py",
800
+ "archive_path": "examples/dcgan/main.py",
801
+ "extension": ".py",
802
+ "size_bytes": 10653
803
+ },
804
+ {
805
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/mnist_hogwild/train.py",
806
+ "archive_path": "examples/mnist_hogwild/train.py",
807
+ "extension": ".py",
808
+ "size_bytes": 2038
809
+ },
810
+ {
811
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/mnist_hogwild/main.py",
812
+ "archive_path": "examples/mnist_hogwild/main.py",
813
+ "extension": ".py",
814
+ "size_bytes": 4205
815
+ },
816
+ {
817
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/gat/main.py",
818
+ "archive_path": "examples/gat/main.py",
819
+ "extension": ".py",
820
+ "size_bytes": 16116
821
+ },
822
+ {
823
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/word_language_model/model.py",
824
+ "archive_path": "examples/word_language_model/model.py",
825
+ "extension": ".py",
826
+ "size_bytes": 5951
827
+ },
828
+ {
829
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/word_language_model/data.py",
830
+ "archive_path": "examples/word_language_model/data.py",
831
+ "extension": ".py",
832
+ "size_bytes": 1482
833
+ },
834
+ {
835
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/word_language_model/generate.py",
836
+ "archive_path": "examples/word_language_model/generate.py",
837
+ "extension": ".py",
838
+ "size_bytes": 3886
839
+ },
840
+ {
841
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/word_language_model/main.py",
842
+ "archive_path": "examples/word_language_model/main.py",
843
+ "extension": ".py",
844
+ "size_bytes": 11669
845
+ },
846
+ {
847
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/mnist_rnn/main.py",
848
+ "archive_path": "examples/mnist_rnn/main.py",
849
+ "extension": ".py",
850
+ "size_bytes": 5658
851
+ },
852
+ {
853
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/regression/main.py",
854
+ "archive_path": "examples/regression/main.py",
855
+ "extension": ".py",
856
+ "size_bytes": 1595
857
+ },
858
+ {
859
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/language_translation/main.py",
860
+ "archive_path": "examples/language_translation/main.py",
861
+ "extension": ".py",
862
+ "size_bytes": 10809
863
+ },
864
+ {
865
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/language_translation/src/model.py",
866
+ "archive_path": "examples/language_translation/src/model.py",
867
+ "extension": ".py",
868
+ "size_bytes": 2859
869
+ },
870
+ {
871
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/language_translation/src/data.py",
872
+ "archive_path": "examples/language_translation/src/data.py",
873
+ "extension": ".py",
874
+ "size_bytes": 5083
875
+ },
876
+ {
877
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/siamese_network/main.py",
878
+ "archive_path": "examples/siamese_network/main.py",
879
+ "extension": ".py",
880
+ "size_bytes": 12938
881
+ },
882
+ {
883
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/primitive_library.py",
884
+ "archive_path": "examples/fx/primitive_library.py",
885
+ "extension": ".py",
886
+ "size_bytes": 6365
887
+ },
888
+ {
889
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/custom_tracer.py",
890
+ "archive_path": "examples/fx/custom_tracer.py",
891
+ "extension": ".py",
892
+ "size_bytes": 4178
893
+ },
894
+ {
895
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/module_tracer.py",
896
+ "archive_path": "examples/fx/module_tracer.py",
897
+ "extension": ".py",
898
+ "size_bytes": 6648
899
+ },
900
+ {
901
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/wrap_output_dynamically.py",
902
+ "archive_path": "examples/fx/wrap_output_dynamically.py",
903
+ "extension": ".py",
904
+ "size_bytes": 2653
905
+ },
906
+ {
907
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/inline_function.py",
908
+ "archive_path": "examples/fx/inline_function.py",
909
+ "extension": ".py",
910
+ "size_bytes": 3195
911
+ },
912
+ {
913
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/profiling_tracer.py",
914
+ "archive_path": "examples/fx/profiling_tracer.py",
915
+ "extension": ".py",
916
+ "size_bytes": 5966
917
+ },
918
+ {
919
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/proxy_based_graph_creation.py",
920
+ "archive_path": "examples/fx/proxy_based_graph_creation.py",
921
+ "extension": ".py",
922
+ "size_bytes": 2115
923
+ },
924
+ {
925
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/invert.py",
926
+ "archive_path": "examples/fx/invert.py",
927
+ "extension": ".py",
928
+ "size_bytes": 2140
929
+ },
930
+ {
931
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/replace_op.py",
932
+ "archive_path": "examples/fx/replace_op.py",
933
+ "extension": ".py",
934
+ "size_bytes": 2434
935
+ },
936
+ {
937
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/subgraph_rewriter_basic_use.py",
938
+ "archive_path": "examples/fx/subgraph_rewriter_basic_use.py",
939
+ "extension": ".py",
940
+ "size_bytes": 2486
941
+ },
942
+ {
943
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/native_interpreter/interpreter.cpp",
944
+ "archive_path": "examples/fx/native_interpreter/interpreter.cpp",
945
+ "extension": ".cpp",
946
+ "size_bytes": 7343
947
+ },
948
+ {
949
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/fx/native_interpreter/use_interpreter.py",
950
+ "archive_path": "examples/fx/native_interpreter/use_interpreter.py",
951
+ "extension": ".py",
952
+ "size_bytes": 5134
953
+ },
954
+ {
955
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/docs/source/conf.py",
956
+ "archive_path": "examples/docs/source/conf.py",
957
+ "extension": ".py",
958
+ "size_bytes": 2631
959
+ },
960
+ {
961
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/gcn/main.py",
962
+ "archive_path": "examples/gcn/main.py",
963
+ "extension": ".py",
964
+ "size_bytes": 11760
965
+ },
966
+ {
967
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/ISSUE_TEMPLATE/feature_template.md",
968
+ "archive_path": "examples/.github/ISSUE_TEMPLATE/feature_template.md",
969
+ "extension": ".md",
970
+ "size_bytes": 553
971
+ },
972
+ {
973
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/ISSUE_TEMPLATE/doc_template.md",
974
+ "archive_path": "examples/.github/ISSUE_TEMPLATE/doc_template.md",
975
+ "extension": ".md",
976
+ "size_bytes": 206
977
+ },
978
+ {
979
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/ISSUE_TEMPLATE/bug_template.md",
980
+ "archive_path": "examples/.github/ISSUE_TEMPLATE/bug_template.md",
981
+ "extension": ".md",
982
+ "size_bytes": 1327
983
+ },
984
+ {
985
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/workflows/main_cpp.yml",
986
+ "archive_path": "examples/.github/workflows/main_cpp.yml",
987
+ "extension": ".yml",
988
+ "size_bytes": 1401
989
+ },
990
+ {
991
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/workflows/main_distributed.yaml",
992
+ "archive_path": "examples/.github/workflows/main_distributed.yaml",
993
+ "extension": ".yaml",
994
+ "size_bytes": 1095
995
+ },
996
+ {
997
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/workflows/main_python.yml",
998
+ "archive_path": "examples/.github/workflows/main_python.yml",
999
+ "extension": ".yml",
1000
+ "size_bytes": 1045
1001
+ },
1002
+ {
1003
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/workflows/doc-build.yml",
1004
+ "archive_path": "examples/.github/workflows/doc-build.yml",
1005
+ "extension": ".yml",
1006
+ "size_bytes": 895
1007
+ },
1008
+ {
1009
+ "source_path": "/tmp/gh_clone_cw5x0lnz/examples/.github/PULL_REQUEST_TEMPLATE/pull_request_template.md",
1010
+ "archive_path": "examples/.github/PULL_REQUEST_TEMPLATE/pull_request_template.md",
1011
+ "extension": ".md",
1012
+ "size_bytes": 1405
1013
+ }
1014
+ ]
1015
+ }
1016
+ ]
1017
+ }