You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
will load an SSD model pretrained on COCO dataset from Torch Hub.
27
-
28
-
Setting precision='fp16' will load a checkpoint trained with [mixed precision](https://arxiv.org/abs/1710.03740) into architecture enabling execution on [Tensor Cores](https://developer.nvidia.com/tensor-cores).
29
-
Handling mixed precision data requires [Apex](https://github.com/NVIDIA/apex) library.
30
-
31
-
32
-
33
20
### Model Description
34
21
35
22
This SSD300 model is based on the
@@ -56,17 +43,17 @@ they are enhanced by additional BatchNorm layers after each convolution.
56
43
57
44
### Example
58
45
59
-
In the example below we will use the pretrained SSD model loaded from Torch Hub to detect objects in sample images and visualize the result.
60
-
61
-
To run the example you need some extra python packages installed.
62
-
These are needed for preprocessing images and visualization.
46
+
In the example below we will use the pretrained SSD model to detect objects in sample images and visualize the result.
63
47
48
+
To run the example you need some extra python packages installed. These are needed for preprocessing images and visualization.
64
49
```bash
65
50
pip install numpy scipy scikit-image matplotlib
66
51
```
67
52
68
-
For convenient and comprehensive formatting of input and output of the model, load a set of utility methods.
53
+
Load an SSD model pretrained on COCO dataset, as well as a set of utility methods for convenient and comprehensive formatting of input and output of the model.
Alternatively, play it right away in a notebook with IPython widgets
97
-
98
85
```python
99
86
from IPython.display import Audio
100
87
Audio(audio_numpy, rate=rate)
101
88
```
102
89
103
90
### Details
104
-
For detailed information on model input and output, training recipies, inference and performance visit: [github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2) and/or [NGC](https://ngc.nvidia.com/catalog/model-scripts/nvidia:tacotron_2_and_waveglow_for_pytorch)
91
+
For detailed information on model input and output, training recipies, inference and performance visit: [github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2) and/or [NGC](https://ngc.nvidia.com/catalog/resources/nvidia:tacotron_2_and_waveglow_for_pytorch)
105
92
106
93
### References
107
94
108
95
-[Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions](https://arxiv.org/abs/1712.05884)
109
96
-[WaveGlow: A Flow-based Generative Network for Speech Synthesis](https://arxiv.org/abs/1811.00002)
110
-
-[Tacotron2 and WaveGlow on NGC](https://ngc.nvidia.com/catalog/model-scripts/nvidia:tacotron_2_and_waveglow_for_pytorch)
111
-
-[Tacotron2 and Waveglow on github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2)
97
+
-[Tacotron2 and WaveGlow on NGC](https://ngc.nvidia.com/catalog/resources/nvidia:tacotron_2_and_waveglow_for_pytorch)
98
+
-[Tacotron2 and Waveglow on github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2)
Alternatively, play it right away in a notebook with IPython widgets
95
-
96
87
```python
97
88
from IPython.display import Audio
98
89
Audio(audio_numpy, rate=rate)
99
90
```
100
91
101
92
### Details
102
-
For detailed information on model input and output, training recipies, inference and performance visit: [github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2) and/or [NGC](https://ngc.nvidia.com/catalog/model-scripts/nvidia:tacotron_2_and_waveglow_for_pytorch)
93
+
For detailed information on model input and output, training recipies, inference and performance visit: [github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2) and/or [NGC](https://ngc.nvidia.com/catalog/resources/nvidia:tacotron_2_and_waveglow_for_pytorch)
103
94
104
95
### References
105
96
106
97
-[Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions](https://arxiv.org/abs/1712.05884)
107
98
-[WaveGlow: A Flow-based Generative Network for Speech Synthesis](https://arxiv.org/abs/1811.00002)
108
-
-[Tacotron2 and WaveGlow on NGC](https://ngc.nvidia.com/catalog/model-scripts/nvidia:tacotron_2_and_waveglow_for_pytorch)
109
-
-[Tacotron2 and Waveglow on github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2)
99
+
-[Tacotron2 and WaveGlow on NGC](https://ngc.nvidia.com/catalog/resources/nvidia:tacotron_2_and_waveglow_for_pytorch)
100
+
-[Tacotron2 and Waveglow on github](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechSynthesis/Tacotron2)
0 commit comments