You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- Check out the [YOLOv5 CLI example](ultralytics-yolov5/tutorials/sparse-transfer-learning.md) for more details on the YOLOv5 training pipeline
167
+
- Check out the [YOLOv5 CLI example](integrations/ultralytics-yolov5/tutorials/sparse-transfer-learning.md) for more details on the YOLOv5 training pipeline
168
168
- Check out the [Hugging Face CLI example](integrations/huggingface-transformers/tutorials/sparse-transfer-learning-bert.md) for more details on the available NLP training pipelines
169
169
- Check out the [Torchvision CLI example](integrations/torchvision/tutorials/sparse-transfer-learning.md) for more details on the image classification training pipelines
Copy file name to clipboardExpand all lines: integrations/huggingface-transformers/tutorials/sparse-transfer-learning-bert-python.md
+8-8
Original file line number
Diff line number
Diff line change
@@ -302,13 +302,13 @@ The command creates a `./deployment` folder in your local directory, which conta
302
302
303
303
Take a look at the tutorials for more examples in other use cases:
304
304
305
-
- [Sparse Transfer with GLUE Datasets (SST2) for sentiment analysis](tutorials/sentiment-analysis/docs-sentiment-analysis-python-sst2.ipynb)
306
-
- [Sparse Transfer with Custom Datasets (RottenTomatoes) and Custom Teacher from HF Hub for sentiment analysis](tutorials/sentiment-analysis/docs-sentiment-analysis-python-custom-teacher-rottentomatoes)
307
-
- [Sparse Transfer with GLUE Datasets (QQP) for multi-input text classification](tutorials/text-classification/docs-text-classification-python-qqp.ipynb)
308
-
- [Sparse Transfer with Custom Datasets (SICK) for multi-input text classification](tutorials/text-classification/docs-text-classification-python-sick.ipynb)
309
-
- [Sparse Transfer with Custom Datasets (TweetEval) and Custom Teacher for single input text classificaiton](tutorials/text-classification/docs-text-classification-python-custom-teacher-tweeteval.ipynb)
310
-
- [Sparse Transfer with Custom Datasets (GoEmotions) for multi-label text classification](tutorials/text-classification/docs-text-classification-python-multi-label-go_emotions.ipynb)
311
-
- [Sparse Transfer with Conll2003 for named entity recognition](tutorials/token-classification/docs-token-classification-python-conll2003.ipynb)
312
-
- [Sparse Transfer with Custom Datasets (WNUT) and Custom Teacher for named entity recognition](tutorials/token-classification/docs-token-classification-custom-teacher-wnut.ipynb)
305
+
- [Sparse Transfer with GLUE Datasets (SST2) for sentiment analysis](sentiment-analysis/docs-sentiment-analysis-python-sst2.ipynb)
306
+
- [Sparse Transfer with Custom Datasets (RottenTomatoes) and Custom Teacher from HF Hub for sentiment analysis](sentiment-analysis/docs-sentiment-analysis-python-custom-teacher-rottentomatoes.ipynb)
307
+
- [Sparse Transfer with GLUE Datasets (QQP) for multi-input text classification](text-classification/docs-text-classification-python-qqp.ipynb)
308
+
- [Sparse Transfer with Custom Datasets (SICK) for multi-input text classification](text-classification/docs-text-classification-python-sick.ipynb)
309
+
- [Sparse Transfer with Custom Datasets (TweetEval) and Custom Teacher for single input text classification](text-classification/docs-text-classification-python-custom-teacher-tweeteval.ipynb)
310
+
- [Sparse Transfer with Custom Datasets (GoEmotions) for multi-label text classification](text-classification/docs-text-classification-python-multi-label-go_emotions.ipynb)
311
+
- [Sparse Transfer with Conll2003 for named entity recognition](token-classification/docs-token-classification-python-conll2003.ipynb)
312
+
- [Sparse Transfer with Custom Datasets (WNUT) and Custom Teacher for named entity recognition](token-classification/docs-token-classification-python-custom-teacher-wnut.ipynb)
313
313
- Sparse Transfer with SQuAD (example coming soon!)
314
314
- Sparse Transfer with Squadshifts Amazon (example coming soon!)
Copy file name to clipboardExpand all lines: integrations/huggingface-transformers/tutorials/token-classification/token-classification-cli.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -175,7 +175,7 @@ A `deployment` folder is created in your local directory, which has all of the f
175
175
176
176
## Sparse Transfer Learning with a Custom Dataset (WNUT_17)
177
177
178
-
Beyond the Conll2003 dataset, we can also use a dataset from the Hugging Face Hub or from local files. Let's try an example of each for the sentiment analysis using [WNUT 17](wnut_17), which is also a NER task.
178
+
Beyond the Conll2003 dataset, we can also use a dataset from the Hugging Face Hub or from local files. Let's try an example of each for the sentiment analysis using WNUT_17, which is also a NER task.
179
179
180
180
For simplicity, we will perform the fine-tuning without distillation. Although the transfer learning recipe contains distillation
181
181
modifiers, by setting `--distill_teacher disable` we instruct SparseML to skip distillation.
Copy file name to clipboardExpand all lines: integrations/old-examples/pytorch/tutorials/classification_sparse_transfer_learning_tutorial.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -74,7 +74,7 @@ Note: The models above were originally trained and sparsified on the [ImageNet](
74
74
75
75
- After noting respective SparseZoo model stub, [train.py](../train.py) script can be used to download checkpoint and [Imagenette](https://github.com/fastai/imagenette) and kick-start transfer learning.
76
76
The transfer learning process itself is guided using recipes; We include example [recipes](../recipes) for classification along with others in the SparseML [GitHub repository](https://github.com/neuralmagic/sparseml).
77
-
[Learn more about recipes and modifiers](../../../docs/source/recipes.md).
77
+
[Learn more about recipes and modifiers](https://github.com/neuralmagic/sparseml/tree/main/docs/source/recipes.md).
78
78
79
79
- Run the following example command to kick off transfer learning for [ResNet-50](https://arxiv.org/abs/1512.03385) starting from a moderately pruned checkpoint from [SparseZoo](https://sparsezoo.neuralmagic.com/):
The latest commit hash that `train.py` is based on is included in the docstring.
86
86
87
87
@@ -112,5 +112,5 @@ python export.py \
112
112
--config ./path/to/checkpoint/args.yaml
113
113
```
114
114
115
-
The DeepSparse Engine [accepts ONNX formats](https://docs.neuralmagic.com/sparseml/source/onnx_export.html) and is engineered to significantly speed up inference on CPUs for the sparsified models from this integration.
115
+
The DeepSparse Engine [accepts ONNX formats](https://docs.neuralmagic.com/archive/sparseml/source/onnx_export.html) and is engineered to significantly speed up inference on CPUs for the sparsified models from this integration.
116
116
Examples for loading, benchmarking, and deploying can be found in the [DeepSparse repository here](https://github.com/neuralmagic/deepsparse).
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/README.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -35,8 +35,8 @@ The techniques include, but are not limited to:
35
35
36
36
## Tutorials
37
37
38
-
-[Sparsifying YOLOv3 Using Recipes](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/tutorials/sparsifying_yolov3_using_recipes.md)
39
-
-[Sparse Transfer Learning With YOLOv3](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/tutorials/yolov3_sparse_transfer_learning.md)
38
+
-[Sparsifying YOLOv3 Using Recipes](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/tutorials/sparsifying_yolov3_using_recipes.md)
39
+
-[Sparse Transfer Learning With YOLOv3](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/tutorials/yolov3_sparse_transfer_learning.md)
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -149,7 +149,7 @@ pruning_modifiers:
149
149
150
150
This recipe creates a sparse, [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model that achieves 97% recovery of its baseline accuracy on the COCO detection dataset.
151
151
Training was done using 4 GPUs at half precision using a total training batch size of 256 with the
152
-
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3).
152
+
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3).
153
153
154
154
When running, adjust hyperparameters based on training environment and dataset.
155
155
@@ -159,7 +159,7 @@ When running, adjust hyperparameters based on training environment and dataset.
159
159
160
160
## Training
161
161
162
-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
162
+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
163
163
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
164
164
The contents of the `hyp.pruned.yaml` hyperparameters file is given below.
165
165
Adjust the script command for your GPU device setup.
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned.short.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -148,7 +148,7 @@ pruning_modifiers:
148
148
149
149
This recipe creates a sparse [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model in a shortened schedule as compared to the original pruned recipe.
150
150
It will train faster, but will recover slightly worse.
151
-
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3) to run.
151
+
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3) to run.
152
152
153
153
When running, adjust hyperparameters based on training environment and dataset.
154
154
@@ -158,7 +158,7 @@ When running, adjust hyperparameters based on training environment and dataset.
158
158
159
159
## Training
160
160
161
-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
161
+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
162
162
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
163
163
The contents of the `hyp.pruned.yaml` hyperparameters file is given below.
164
164
Adjust the script command for your GPU device setup.
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned_quantized.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -223,7 +223,7 @@ quantization_modifiers:
223
223
224
224
This recipe creates a sparse-quantized, [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model that achieves 94% recovery of its baseline accuracy on the COCO detection dataset.
225
225
Training was done using 4 GPUs at half precision using a total training batch size of 256 with the
226
-
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3).
226
+
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3).
227
227
228
228
When running, adjust hyperparameters based on training environment and dataset.
229
229
@@ -237,7 +237,7 @@ This additionally means that the checkpoints are saved using state_dicts rather
237
237
238
238
## Training
239
239
240
-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
240
+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
241
241
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
242
242
The contents of the `hyp.pruned_quantized.yaml` hyperparameters file is given below.
243
243
Adjust the script command for your GPU device setup.
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned_quantized.short.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -223,7 +223,7 @@ quantization_modifiers:
223
223
224
224
This recipe creates a sparse-quantized [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model in a shortened shcedule as compared to the original pruned recipe.
225
225
It will train faster, but will recover slightly worse.
226
-
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3) to run.
226
+
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3) to run.
227
227
228
228
When running, adjust hyperparameters based on training environment and dataset.
229
229
@@ -237,7 +237,7 @@ This additionally means that the checkpoints are saved using state_dicts rather
237
237
238
238
## Training
239
239
240
-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
240
+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
241
241
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
242
242
The contents of the `hyp.pruned_quantized.yaml` hyperparameters file is given below.
243
243
Adjust the script command for your GPU device setup.
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.test.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -224,7 +224,7 @@ quantization_modifiers:
224
224
This is a test recipe useful for quickly evaluating the time and resources needed for pruning and quantizing a model.
225
225
In addition, it offers a quick integration tests pathway.
226
226
This recipe creates a sparse-quantized [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model that will not be accurate.
227
-
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3) to run.
227
+
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3) to run.
228
228
229
229
Note that half-precision, EMA, and pickling are not supported for quantization.
230
230
Therefore, once quantization is run, all three will be disabled for the training pipeline.
@@ -236,7 +236,7 @@ This additionally means that the checkpoints are saved using state_dicts rather
236
236
237
237
## Training
238
238
239
-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
239
+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
240
240
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
241
241
The contents of the `hyp.pruned_quantized.yaml` hyperparameters file is given below.
242
242
Adjust the script command for your GPU device setup.
Copy file name to clipboardExpand all lines: integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.transfer_learn_pruned.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -130,7 +130,7 @@ pruning_modifiers:
130
130
This recipe transfer learns from a sparse, [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model.
131
131
It was originally tested on the VOC dataset and achieved 0.84 [email protected].
132
132
133
-
Training was done using 4 GPUs at half precision with the [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3).
133
+
Training was done using 4 GPUs at half precision with the [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3).
134
134
135
135
When running, adjust hyperparameters based on training environment and dataset.
136
136
@@ -142,7 +142,7 @@ The training results for this recipe are made available through Weights and Bias
142
142
143
143
## Training
144
144
145
-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
145
+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
146
146
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
147
147
Adjust the script command for your GPU device setup.
0 commit comments