Skip to content

Commit 90b664a

Browse files
authored
Create linkcheck.yml (#1525)
* Create linkcheck.yml * Fix links * Missed saving state * . * Update
1 parent df570d1 commit 90b664a

23 files changed

+80
-50
lines changed

.github/workflows/linkcheck.yml

+22
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,22 @@
1+
name: Check Markdown links
2+
3+
on:
4+
push:
5+
branches:
6+
- main
7+
pull_request:
8+
branches:
9+
- main
10+
11+
# Allows you to run this workflow manually from the Actions tab
12+
workflow_dispatch:
13+
14+
jobs:
15+
markdown-link-check:
16+
runs-on: ubuntu-latest
17+
steps:
18+
- uses: actions/checkout@v2
19+
- uses: gaurav-nelson/github-action-markdown-link-check@v1
20+
with:
21+
use-quiet-mode: 'yes'
22+
config-file: '.github/workflows/mlc_config.json'

.github/workflows/mlc-config.json

+13
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
{
2+
"ignorePatterns": [
3+
{
4+
"pattern": ".*localhost.*"
5+
},
6+
{
7+
"pattern": ".*127\\.0\\.0\\.1.*"
8+
},
9+
{
10+
"pattern": ".*0\\.0\\.0\\.0.*"
11+
}
12+
]
13+
}

README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -164,7 +164,7 @@ sparseml.yolov5.train \
164164
--hyp hyps/hyp.finetune.yaml --cfg yolov5s.yaml --patience 0
165165
```
166166

167-
- Check out the [YOLOv5 CLI example](ultralytics-yolov5/tutorials/sparse-transfer-learning.md) for more details on the YOLOv5 training pipeline
167+
- Check out the [YOLOv5 CLI example](integrations/ultralytics-yolov5/tutorials/sparse-transfer-learning.md) for more details on the YOLOv5 training pipeline
168168
- Check out the [Hugging Face CLI example](integrations/huggingface-transformers/tutorials/sparse-transfer-learning-bert.md) for more details on the available NLP training pipelines
169169
- Check out the [Torchvision CLI example](integrations/torchvision/tutorials/sparse-transfer-learning.md) for more details on the image classification training pipelines
170170

integrations/huggingface-transformers/tutorials/sparse-transfer-learning-bert-python.md

+8-8
Original file line numberDiff line numberDiff line change
@@ -302,13 +302,13 @@ The command creates a `./deployment` folder in your local directory, which conta
302302

303303
Take a look at the tutorials for more examples in other use cases:
304304

305-
- [Sparse Transfer with GLUE Datasets (SST2) for sentiment analysis](tutorials/sentiment-analysis/docs-sentiment-analysis-python-sst2.ipynb)
306-
- [Sparse Transfer with Custom Datasets (RottenTomatoes) and Custom Teacher from HF Hub for sentiment analysis](tutorials/sentiment-analysis/docs-sentiment-analysis-python-custom-teacher-rottentomatoes)
307-
- [Sparse Transfer with GLUE Datasets (QQP) for multi-input text classification](tutorials/text-classification/docs-text-classification-python-qqp.ipynb)
308-
- [Sparse Transfer with Custom Datasets (SICK) for multi-input text classification](tutorials/text-classification/docs-text-classification-python-sick.ipynb)
309-
- [Sparse Transfer with Custom Datasets (TweetEval) and Custom Teacher for single input text classificaiton](tutorials/text-classification/docs-text-classification-python-custom-teacher-tweeteval.ipynb)
310-
- [Sparse Transfer with Custom Datasets (GoEmotions) for multi-label text classification](tutorials/text-classification/docs-text-classification-python-multi-label-go_emotions.ipynb)
311-
- [Sparse Transfer with Conll2003 for named entity recognition](tutorials/token-classification/docs-token-classification-python-conll2003.ipynb)
312-
- [Sparse Transfer with Custom Datasets (WNUT) and Custom Teacher for named entity recognition](tutorials/token-classification/docs-token-classification-custom-teacher-wnut.ipynb)
305+
- [Sparse Transfer with GLUE Datasets (SST2) for sentiment analysis](sentiment-analysis/docs-sentiment-analysis-python-sst2.ipynb)
306+
- [Sparse Transfer with Custom Datasets (RottenTomatoes) and Custom Teacher from HF Hub for sentiment analysis](sentiment-analysis/docs-sentiment-analysis-python-custom-teacher-rottentomatoes.ipynb)
307+
- [Sparse Transfer with GLUE Datasets (QQP) for multi-input text classification](text-classification/docs-text-classification-python-qqp.ipynb)
308+
- [Sparse Transfer with Custom Datasets (SICK) for multi-input text classification](text-classification/docs-text-classification-python-sick.ipynb)
309+
- [Sparse Transfer with Custom Datasets (TweetEval) and Custom Teacher for single input text classification](text-classification/docs-text-classification-python-custom-teacher-tweeteval.ipynb)
310+
- [Sparse Transfer with Custom Datasets (GoEmotions) for multi-label text classification](text-classification/docs-text-classification-python-multi-label-go_emotions.ipynb)
311+
- [Sparse Transfer with Conll2003 for named entity recognition](token-classification/docs-token-classification-python-conll2003.ipynb)
312+
- [Sparse Transfer with Custom Datasets (WNUT) and Custom Teacher for named entity recognition](token-classification/docs-token-classification-python-custom-teacher-wnut.ipynb)
313313
- Sparse Transfer with SQuAD (example coming soon!)
314314
- Sparse Transfer with Squadshifts Amazon (example coming soon!)

integrations/huggingface-transformers/tutorials/token-classification/token-classification-cli.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,7 @@ A `deployment` folder is created in your local directory, which has all of the f
175175

176176
## Sparse Transfer Learning with a Custom Dataset (WNUT_17)
177177

178-
Beyond the Conll2003 dataset, we can also use a dataset from the Hugging Face Hub or from local files. Let's try an example of each for the sentiment analysis using [WNUT 17](wnut_17), which is also a NER task.
178+
Beyond the Conll2003 dataset, we can also use a dataset from the Hugging Face Hub or from local files. Let's try an example of each for the sentiment analysis using WNUT_17, which is also a NER task.
179179

180180
For simplicity, we will perform the fine-tuning without distillation. Although the transfer learning recipe contains distillation
181181
modifiers, by setting `--distill_teacher disable` we instruct SparseML to skip distillation.

integrations/old-examples/dbolya-yolact/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -106,7 +106,7 @@ The following table lays out the root-level files and folders along with a descr
106106
| Folder/File Name | Description |
107107
|-------------------------------------------------|-----------------------------------------------------------------------------------------------------------------------|
108108
| [recipes](./recipes) | Typical recipes for sparsifying YOLACT models along with any downloaded recipes from the SparseZoo. |
109-
| [yolact](./yolact) | Integration repository folder used to train and sparsify YOLACT models (`setup_integration.sh` must run first). |
109+
| yolact | Integration repository folder used to train and sparsify YOLACT models (`setup_integration.sh` must run first). |
110110
| [README.md](./README.md) | Readme file. |
111111
| [tutorials](./tutorials) | Easy to follow sparsification tutorials for YOLACT models. |
112112

integrations/old-examples/keras/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ The techniques include, but are not limited to:
3131

3232
## Tutorials
3333

34-
- [Classification](https://github.com/neuralmagic/sparseml/blob/main/integrations/keras/notebooks/classification.ipynb)
34+
- [Classification](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/keras/notebooks/classification.ipynb)
3535

3636
## Installation
3737

integrations/old-examples/pytorch/tutorials/classification_sparse_transfer_learning_tutorial.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ Note: The models above were originally trained and sparsified on the [ImageNet](
7474

7575
- After noting respective SparseZoo model stub, [train.py](../train.py) script can be used to download checkpoint and [Imagenette](https://github.com/fastai/imagenette) and kick-start transfer learning.
7676
The transfer learning process itself is guided using recipes; We include example [recipes](../recipes) for classification along with others in the SparseML [GitHub repository](https://github.com/neuralmagic/sparseml).
77-
[Learn more about recipes and modifiers](../../../docs/source/recipes.md).
77+
[Learn more about recipes and modifiers](https://github.com/neuralmagic/sparseml/tree/main/docs/source/recipes.md).
7878

7979
- Run the following example command to kick off transfer learning for [ResNet-50](https://arxiv.org/abs/1512.03385) starting from a moderately pruned checkpoint from [SparseZoo](https://sparsezoo.neuralmagic.com/):
8080
```

integrations/old-examples/pytorch/tutorials/sparsifying_pytorch_models_using_recipes.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ pip install "sparseml[torchvision, dev]"
4949
Recipes are YAML or Markdown files that SparseML uses to easily define and control the sparsification of a model.
5050
Recipes consist of a series of `Modifiers` that can influence the training process in different ways. A list of
5151
common modifiers and their uses is provided
52-
[here](../../../docs/source/recipes.md#modifiers-intro).
52+
[here](https://github.com/neuralmagic/sparseml/tree/main/docs/source/recipes.md#modifiers-intro).
5353

5454
SparseML provides a recipe for sparsifying a ResNet-50 model trained on the tiny Imagenette dataset. The recipe can
5555
be viewed in the browser
@@ -71,7 +71,7 @@ of the parameters list for a single `GMPruningModifier`.
7171

7272
Recipes can integrated into training flows with a couple of lines of code by using a `ScheduledModifierManager`
7373
that wraps the PyTorch `Optimizer` step. An example of how this is done can be found
74-
[here](../../../docs/source/code.md#pytorch-sparsification).
74+
[here](https://github.com/neuralmagic/sparseml/tree/main/docs/source/code.md#pytorch-sparsification).
7575

7676
For this example, we can use the `sparseml.image_classification.train` utility. This utility runs a
7777
PyTorch training flow that is modified by a `ScheduledModifierManager` and takes a recipe as an input.

integrations/old-examples/rwightman-timm/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ python train.py \
8181
```
8282

8383
Documentation on the original script can be found
84-
[here](https://rwightman.github.io/pytorch-image-models/scripts/).
84+
[here](https://huggingface.co/docs/timm/training_script).
8585
The latest commit hash that `train.py` is based on is included in the docstring.
8686

8787

@@ -112,5 +112,5 @@ python export.py \
112112
--config ./path/to/checkpoint/args.yaml
113113
```
114114

115-
The DeepSparse Engine [accepts ONNX formats](https://docs.neuralmagic.com/sparseml/source/onnx_export.html) and is engineered to significantly speed up inference on CPUs for the sparsified models from this integration.
115+
The DeepSparse Engine [accepts ONNX formats](https://docs.neuralmagic.com/archive/sparseml/source/onnx_export.html) and is engineered to significantly speed up inference on CPUs for the sparsified models from this integration.
116116
Examples for loading, benchmarking, and deploying can be found in the [DeepSparse repository here](https://github.com/neuralmagic/deepsparse).

integrations/old-examples/tensorflow_v1/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ The techniques include, but are not limited to:
3131

3232
## Tutorials
3333

34-
- [Classification](https://github.com/neuralmagic/sparseml/blob/main/integrations/tensorflow_v1/notebooks/classification.ipynb)
34+
- [Classification](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/tensorflow_v1/notebooks/classification.ipynb)
3535

3636
## Installation
3737

integrations/old-examples/ultralytics-yolov3/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -35,8 +35,8 @@ The techniques include, but are not limited to:
3535

3636
## Tutorials
3737

38-
- [Sparsifying YOLOv3 Using Recipes](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/tutorials/sparsifying_yolov3_using_recipes.md)
39-
- [Sparse Transfer Learning With YOLOv3](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/tutorials/yolov3_sparse_transfer_learning.md)
38+
- [Sparsifying YOLOv3 Using Recipes](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/tutorials/sparsifying_yolov3_using_recipes.md)
39+
- [Sparse Transfer Learning With YOLOv3](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/tutorials/yolov3_sparse_transfer_learning.md)
4040

4141
## Installation
4242

integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -149,7 +149,7 @@ pruning_modifiers:
149149

150150
This recipe creates a sparse, [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model that achieves 97% recovery of its baseline accuracy on the COCO detection dataset.
151151
Training was done using 4 GPUs at half precision using a total training batch size of 256 with the
152-
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3).
152+
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3).
153153

154154
When running, adjust hyperparameters based on training environment and dataset.
155155

@@ -159,7 +159,7 @@ When running, adjust hyperparameters based on training environment and dataset.
159159

160160
## Training
161161

162-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
162+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
163163
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
164164
The contents of the `hyp.pruned.yaml` hyperparameters file is given below.
165165
Adjust the script command for your GPU device setup.

integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned.short.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -148,7 +148,7 @@ pruning_modifiers:
148148

149149
This recipe creates a sparse [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model in a shortened schedule as compared to the original pruned recipe.
150150
It will train faster, but will recover slightly worse.
151-
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3) to run.
151+
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3) to run.
152152

153153
When running, adjust hyperparameters based on training environment and dataset.
154154

@@ -158,7 +158,7 @@ When running, adjust hyperparameters based on training environment and dataset.
158158

159159
## Training
160160

161-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
161+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
162162
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
163163
The contents of the `hyp.pruned.yaml` hyperparameters file is given below.
164164
Adjust the script command for your GPU device setup.

integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned_quantized.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -223,7 +223,7 @@ quantization_modifiers:
223223

224224
This recipe creates a sparse-quantized, [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model that achieves 94% recovery of its baseline accuracy on the COCO detection dataset.
225225
Training was done using 4 GPUs at half precision using a total training batch size of 256 with the
226-
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3).
226+
[SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3).
227227

228228
When running, adjust hyperparameters based on training environment and dataset.
229229

@@ -237,7 +237,7 @@ This additionally means that the checkpoints are saved using state_dicts rather
237237

238238
## Training
239239

240-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
240+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
241241
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
242242
The contents of the `hyp.pruned_quantized.yaml` hyperparameters file is given below.
243243
Adjust the script command for your GPU device setup.

integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.pruned_quantized.short.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -223,7 +223,7 @@ quantization_modifiers:
223223

224224
This recipe creates a sparse-quantized [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model in a shortened shcedule as compared to the original pruned recipe.
225225
It will train faster, but will recover slightly worse.
226-
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3) to run.
226+
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3) to run.
227227

228228
When running, adjust hyperparameters based on training environment and dataset.
229229

@@ -237,7 +237,7 @@ This additionally means that the checkpoints are saved using state_dicts rather
237237

238238
## Training
239239

240-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
240+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
241241
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
242242
The contents of the `hyp.pruned_quantized.yaml` hyperparameters file is given below.
243243
Adjust the script command for your GPU device setup.

integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.test.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -224,7 +224,7 @@ quantization_modifiers:
224224
This is a test recipe useful for quickly evaluating the time and resources needed for pruning and quantizing a model.
225225
In addition, it offers a quick integration tests pathway.
226226
This recipe creates a sparse-quantized [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model that will not be accurate.
227-
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3) to run.
227+
Use the following [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3) to run.
228228

229229
Note that half-precision, EMA, and pickling are not supported for quantization.
230230
Therefore, once quantization is run, all three will be disabled for the training pipeline.
@@ -236,7 +236,7 @@ This additionally means that the checkpoints are saved using state_dicts rather
236236

237237
## Training
238238

239-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
239+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
240240
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
241241
The contents of the `hyp.pruned_quantized.yaml` hyperparameters file is given below.
242242
Adjust the script command for your GPU device setup.

integrations/old-examples/ultralytics-yolov3/recipes/yolov3-spp.transfer_learn_pruned.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -130,7 +130,7 @@ pruning_modifiers:
130130
This recipe transfer learns from a sparse, [YOLOv3-SPP](https://arxiv.org/abs/1804.02767) model.
131131
It was originally tested on the VOC dataset and achieved 0.84 [email protected].
132132

133-
Training was done using 4 GPUs at half precision with the [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov3).
133+
Training was done using 4 GPUs at half precision with the [SparseML integration with ultralytics/yolov3](https://github.com/neuralmagic/sparseml/tree/main/integrations/old-examples/ultralytics-yolov3).
134134

135135
When running, adjust hyperparameters based on training environment and dataset.
136136

@@ -142,7 +142,7 @@ The training results for this recipe are made available through Weights and Bias
142142

143143
## Training
144144

145-
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/ultralytics-yolov3/README.md).
145+
To set up the training environment, follow the instructions on the [integration README](https://github.com/neuralmagic/sparseml/blob/main/integrations/old-examples/ultralytics-yolov3/README.md).
146146
Using the given training script from the `yolov3` directory the following command can be used to launch this recipe.
147147
Adjust the script command for your GPU device setup.
148148
Ultralytics supports both DataParallel and DDP.

0 commit comments

Comments
 (0)