PyTorch 1.6 Released, Microsoft To Take Care Of The Windows Version of PyTorch

Recently, Facebook introduced the provision of the most recent model of PyTorch, PyTorch 1.6. The social media big additionally made a large announcement that Microsoft has expanded its participation within the PyTorch group and is taking possession of the event and upkeep of the PyTorch to construct for Windows.

PyTorch is among the hottest machine studying libraries in Python. The model 1.6 launch contains a number of new APIs, instruments for efficiency enchancment and profiling, in addition to important updates to each distributed data-parallel (DDP) and distant process name (RPC) primarily based distributed coaching.

According to the blog post, from this launch onward, options will probably be categorised as Stable, Beta and Prototype, the place Prototype options usually are not included as a part of the binary distribution and are as a substitute obtainable by way of both constructing from supply, utilizing nightlies or through a compiler flag. 

New Features

The important updates on this model of PyTorch are as below-

Automatic Mixed Precision (AMP) Training

Automatic blended precision (AMP) coaching is now natively supported and is a secure characteristic. AMP permits customers to simply allow automated mixed-precision coaching permitting increased efficiency and reminiscence financial savings of as much as 50 per cent on Tensor Core GPUs.  


TensorPipe Backend for RPC

PyTorch 1.6 introduces a brand new backend for the RPC module which leverages the TensorPipe library. TensorPipe library is a tensor-aware point-to-point communication primitive focused at machine studying, which is meant to enrich the present primitives for distributed coaching in PyTorch.

Fork Parallelism

PyTorch 1.6 provides help for a language-level assemble together with runtime help for coarse-grained parallelism in TorchScript code. This characteristic is beneficial for working fashions in an ensemble in parallel, or working bidirectional parts of recurrent nets in parallel, and permits the flexibility to unlock the computational energy of parallel architectures for task-level parallelism.

Memory Profiler

The torch.autograd.profiler API now features a reminiscence profiler that permits you to examine the tensor reminiscence value of various operators inside your CPU and GPU fashions.


DDP is used for full sync data-parallel coaching of fashions, and the RPC framework permits distributed mannequin parallelism. PyTorch 1.6 has mixed these two options to attain each information parallelism and mannequin parallelism on the identical time.   

Torchvision 0.7

Torchvision 0.7 introduces two new pre-trained semantic segmentation fashions, FCN ResNet50 and DeepLabV3 ResNet50, which is each skilled on COCO and utilizing smaller reminiscence footprints than the ResNet101 spine.  

Besides these newly up to date options, there are additionally quite a few enhancements and new options in distributed coaching & RPC, area libraries in addition to frontend APIs.  

PyTorch For Windows

Researchers from Microsoft have been engaged on including help for PyTorch on Windows. However, as a consequence of some restricted assets, together with lack of check protection, TorchAudio area library, distributed coaching help, amongst others, Windows help for PyTorch has lagged behind different platforms. 

See Also

With the discharge of PyTorch, the tech big improved the core high quality of the Windows construct by bringing check protection at par with Linux for core PyTorch and its area libraries, and by automating tutorial testing. 

In a weblog put up, the builders at Microsoft said, “Thanks to the broader PyTorch community, which contributed TorchAudio support to Windows, we were able to add test coverage to all three domain libraries: TorchVision, TorchText and TorchAudio.”

They added, “In subsequent releases of PyTorch, we will continue improving the Windows experience based on community feedback and requests. So far, the feedback we received from the community points to distributed training support and a better installation experience using pip as the next areas of improvement.” 

Installing On Windows

To set up PyTorch utilizing Anaconda with the most recent GPU help, run the command below-

conda set up pytorch torchvision cudatoolkit=10.2 -c pytorch

Provide your feedback beneath


If you really liked this story, do be a part of our Telegram Community.

Also, you possibly can write for us and be one of many 500+ consultants who’ve contributed tales at AIM. Share your nominations here.

Ambika Choudhury

Ambika Choudhury

A Technical Journalist who loves writing about Machine Learning and Artificial Intelligence. A lover of music, writing and studying one thing out of the field. Contact:


Please enter your comment!
Please enter your name here