Skip to content

Latest commit

 

History

History
837 lines (570 loc) · 25.2 KB

README.md

File metadata and controls

837 lines (570 loc) · 25.2 KB

FastAI XLA Extensions Library

The FastAI XLA Extensions library package allows your fastai/Pytorch models to run on TPUs using the Pytorch-XLA library.

Open In Colab

Documentation Site

You can view the documentation here: https://butchland.github.io/fastai_xla_extensions

Install

pip install fastai_xla_extensions

How to use

Configure TPU Environment Access

The Pytorch XLA package requires an environment supporting TPUs (Kaggle kernels, GCP or Colab environments required).

Nominally, Pytorch XLA also supports GPUs so please see the Pytorch XLA site for more instructions.

If running on Colab, make sure the Runtime Type is set to TPU.

Install pytorch 1.7.1

Use pytorch 1.7.1 as this is latest version of pytorch supported by the fastai package

#hide_output
#colab
# install pytorch 1.7.1 b/c fastai doesn't support pytorch 1.8 just yet
!pip install -Uqq --no-cache-dir torch==1.7.1+cu101 torchvision==0.8.2+cu101 torchtext==0.8.0 -f https://download.pytorch.org/whl/torch_stable.html

Install fastai

Use the latest fastai and fastcore versions (note: due to a compatibility issue, we are using fastai==2.3.0 for now)

#hide_output
#colab
!pip install -Uqq fastcore --upgrade
!pip install -Uqq fastai==2.3.0 

Install Pytorch XLA

This is the official way to install Pytorch-XLA 1.7 as per the instructions here

#hide_output
#colab
!pip install -Uqq cloud-tpu-client==0.10 https://storage.googleapis.com/tpu-pytorch/wheels/torch_xla-1.7-cp37-cp37m-linux_x86_64.whl

Check if XLA is available

#colab
import warnings
try:
    import torch_xla
except ImportError as e:
    if DEBUG: warnings.warn('TPU environment not available')
    
WARNING:root:Waiting for TPU to be start up with version pytorch-1.7...
WARNING:root:Waiting for TPU to be start up with version pytorch-1.7...
WARNING:root:TPU has started up successfully with version pytorch-1.7

Import the libraries

Import the fastai and fastai_xla_extensions libraries

from fastai.vision.all import *
#colab
#hide_output
from fastai_xla_extensions.all import *

Single and Multi Core TPU Modes

The fastai_xla_extensions package allows you to train your models using either a single TPU core or multiple TPU cores.

Using multi TPU cores is usually faster than single TPU core, but due to some limitations, not all fastai features are supported in multi TPU core mode.

  • Multi Core TPU mode

    In order to run in multi TPU core mode, the fastai_xla_extensions package patched some additional methods to the Learner class.

    These additional methods are all prefixed by xla. For example, to run the Learner.fit method in multi core tpu mode, you can invoke the equivalent Learner.xla_fit method, with almost the same parameters.

    Note that in multi core TPU mode, when calling normal learner or data loader methods (not prefixed by xla), the process does not use the multi core TPUs, but instead runs them using the CPU.

    • List of xla prefixed methods

      • xla_fit
      • xla_fine_tune
      • xla_fit_one_cycle
      • xla_fit_sgdr
      • xla_fit_flat_cos
      • xla_lr_find
      • xla_get_preds
    • Additional xla Parameters (for Multi Core TPU mode)

      All xla prefixed methods support the following additional parameters:

      • num_cores: this is defaulted to 8 (representing the number of TPU cores per TPU device) - valid values are 8 or 1 on Google Colab. For running GCP TPUs, other values maybe possible. See the Pytorch XLA page for more details.

      • start_method: this is defaulted to fork which the only value supported on Google Colab. On GCP, other values supported include spawn. Again, see the Pytorch XLA page for more details.

      • master_cbs: these are callbacks that are run only on the master ordinal process (rank = 0). This allows the addition of callbacks that may interfere with each other if run simultaneously on more than one process by restricting them to run on only one process - the master ordinal process.

  • Single Core TPU mode

    In order to run in single TPU core mode, the fastai_xla_extensions package added a method to_xla to the Learner class.

    After invoking this method, the learner will now use an associated TPU core device to the learner process which will be used when training the model.

    This means we can use the same fit methods in Learner that we normally use to train the model when using a GPU or a CPU.

    Also note that the single TPU core mode is incompatible with the Multi TPU Core mode, and that the kernel will have to restarted once the single TPU core mode is set. This is because once a TPU core is associated with the main process, it cannot be unset without restarting the kernel.

Multi TPU Core Example

Build a MNIST classifier -- adapted from fastai course Lesson 4 notebook

Load MNIST dataset

path = untar_data(URLs.MNIST_TINY)

Create Fastai DataBlock

datablock = DataBlock(
    blocks=(ImageBlock,CategoryBlock),
    get_items=get_image_files,
    splitter=GrandparentSplitter(),
    get_y=parent_label,
    item_tfms=Resize(28),
    # batch transforms run on the CPU and are slow, so only using the Normalize batch transform
    batch_tfms=[Normalize.from_stats(*imagenet_stats)]
)
#colab
datablock.summary(path)
Setting-up type transforms pipelines
Collecting items from /root/.fastai/data/mnist_tiny
Found 1428 items
2 datasets of sizes 709,699
Setting up Pipeline: PILBase.create
Setting up Pipeline: parent_label -> Categorize -- {'vocab': None, 'sort': True, 'add_na': False}

Building one sample
  Pipeline: PILBase.create
    starting from
      /root/.fastai/data/mnist_tiny/train/7/8791.png
    applying PILBase.create gives
      PILImage mode=RGB size=28x28
  Pipeline: parent_label -> Categorize -- {'vocab': None, 'sort': True, 'add_na': False}
    starting from
      /root/.fastai/data/mnist_tiny/train/7/8791.png
    applying parent_label gives
      7
    applying Categorize -- {'vocab': None, 'sort': True, 'add_na': False} gives
      TensorCategory(1)

Final sample: (PILImage mode=RGB size=28x28, TensorCategory(1))


Collecting items from /root/.fastai/data/mnist_tiny
Found 1428 items
2 datasets of sizes 709,699
Setting up Pipeline: PILBase.create
Setting up Pipeline: parent_label -> Categorize -- {'vocab': None, 'sort': True, 'add_na': False}
Setting up after_item: Pipeline: Resize -- {'size': (28, 28), 'method': 'crop', 'pad_mode': 'reflection', 'resamples': (2, 0), 'p': 1.0} -> ToTensor
Setting up before_batch: Pipeline: 
Setting up after_batch: Pipeline: IntToFloatTensor -- {'div': 255.0, 'div_mask': 1} -> Normalize -- {'mean': tensor([[[[0.4850]],

         [[0.4560]],

         [[0.4060]]]]), 'std': tensor([[[[0.2290]],

         [[0.2240]],

         [[0.2250]]]]), 'axes': (0, 2, 3)}

Building one batch
Applying item_tfms to the first sample:
  Pipeline: Resize -- {'size': (28, 28), 'method': 'crop', 'pad_mode': 'reflection', 'resamples': (2, 0), 'p': 1.0} -> ToTensor
    starting from
      (PILImage mode=RGB size=28x28, TensorCategory(1))
    applying Resize -- {'size': (28, 28), 'method': 'crop', 'pad_mode': 'reflection', 'resamples': (2, 0), 'p': 1.0} gives
      (PILImage mode=RGB size=28x28, TensorCategory(1))
    applying ToTensor gives
      (TensorImage of size 3x28x28, TensorCategory(1))

Adding the next 3 samples

No before_batch transform to apply

Collating items in a batch

Applying batch_tfms to the batch built
  Pipeline: IntToFloatTensor -- {'div': 255.0, 'div_mask': 1} -> Normalize -- {'mean': tensor([[[[0.4850]],

         [[0.4560]],

         [[0.4060]]]]), 'std': tensor([[[[0.2290]],

         [[0.2240]],

         [[0.2250]]]]), 'axes': (0, 2, 3)}
    starting from
      (TensorImage of size 4x3x28x28, TensorCategory([1, 1, 1, 1]))
    applying IntToFloatTensor -- {'div': 255.0, 'div_mask': 1} gives
      (TensorImage of size 4x3x28x28, TensorCategory([1, 1, 1, 1]))
    applying Normalize -- {'mean': tensor([[[[0.4850]],

         [[0.4560]],

         [[0.4060]]]]), 'std': tensor([[[[0.2290]],

         [[0.2240]],

         [[0.2250]]]]), 'axes': (0, 2, 3)} gives
      (TensorImage of size 4x3x28x28, TensorCategory([1, 1, 1, 1]))

Create the dataloader

dls = datablock.dataloaders(path, bs=8) # use a small batch size as mnist_tiny dataset is very small
#colab
dls.show_batch()

png

Create a Fastai CNN Learner

learner = cnn_learner(dls, resnet18, metrics=accuracy, concat_pool=False) # fastai AdaptivePooling causes a lowering which slows down the training so using nn.AvgPool2D
                      
Downloading: "https://download.pytorch.org/models/resnet18-5c106cde.pth" to /root/.cache/torch/hub/checkpoints/resnet18-5c106cde.pth
#colab
learner.summary()

Using the lr_find works

#colab
learner.xla_lr_find()
start fit









SuggestedLRs(lr_min=7.585775847473997e-08, lr_steep=6.309573450380412e-07)

png

Run one cycle training.

#colab
learner.xla_fit_one_cycle(5,lr_max=slice(1e-4,0.02))
start fit
epoch train_loss valid_loss accuracy time
0 0.190064 0.391755 0.848011 00:17
1 0.273210 0.288629 0.882102 00:06
2 0.322163 0.430995 0.875000 00:06
3 0.354422 0.534526 0.830966 00:06
4 0.372328 0.517053 0.833807 00:06

Further fine-tuning

#colab
learner.xla_fit_one_cycle(5,lr_max=slice(7e-4, 1e-3))
start fit
epoch train_loss valid_loss accuracy time
0 0.046700 0.451159 0.867898 00:14
1 0.056459 0.340751 0.896307 00:07
2 0.061811 0.376011 0.892045 00:06
3 0.071214 0.394541 0.879261 00:06
4 0.137522 0.407396 0.899148 00:06

Model params are using CPU (since it is the spawned process models that are moved to TPU)

#colab
one_param(learner.model).device
device(type='cpu')

Plot loss seems to be working fine.

#colab
learner.recorder.plot_loss()

png

Single TPU Core Example

Build a MNIST classifier -- adapted from fastai course Lesson 4 notebook

Load MNIST dataset

path = untar_data(URLs.MNIST_TINY)

Create Fastai DataBlock

datablock = DataBlock(
    blocks=(ImageBlock,CategoryBlock),
    get_items=get_image_files,
    splitter=GrandparentSplitter(),
    get_y=parent_label,
    item_tfms=Resize(28),
    # affine transforms are performed on the CPU, other batch transforms are done on the TPU
    batch_tfms=aug_transforms(do_flip=False,min_scale=0.8)
)
#colab
datablock.summary(path)
Setting-up type transforms pipelines
Collecting items from /root/.fastai/data/mnist_tiny
Found 1428 items
2 datasets of sizes 709,699
Setting up Pipeline: PILBase.create
Setting up Pipeline: parent_label -> Categorize -- {'vocab': None, 'sort': True, 'add_na': False}

Building one sample
  Pipeline: PILBase.create
    starting from
      /root/.fastai/data/mnist_tiny/train/7/8791.png
    applying PILBase.create gives
      PILImage mode=RGB size=28x28
  Pipeline: parent_label -> Categorize -- {'vocab': None, 'sort': True, 'add_na': False}
    starting from
      /root/.fastai/data/mnist_tiny/train/7/8791.png
    applying parent_label gives
      7
    applying Categorize -- {'vocab': None, 'sort': True, 'add_na': False} gives
      TensorCategory(1)

Final sample: (PILImage mode=RGB size=28x28, TensorCategory(1))


Collecting items from /root/.fastai/data/mnist_tiny
Found 1428 items
2 datasets of sizes 709,699
Setting up Pipeline: PILBase.create
Setting up Pipeline: parent_label -> Categorize -- {'vocab': None, 'sort': True, 'add_na': False}
Setting up after_item: Pipeline: Resize -- {'size': (28, 28), 'method': 'crop', 'pad_mode': 'reflection', 'resamples': (2, 0), 'p': 1.0} -> ToTensor
Setting up before_batch: Pipeline: 
Setting up after_batch: Pipeline: IntToFloatTensor -- {'div': 255.0, 'div_mask': 1} -> Warp -- {'magnitude': 0.2, 'p': 1.0, 'draw_x': None, 'draw_y': None, 'size': None, 'mode': 'bilinear', 'pad_mode': 'reflection', 'batch': False, 'align_corners': True, 'mode_mask': 'nearest'} -> RandomResizedCropGPU -- {'size': None, 'min_scale': 0.8, 'ratio': (1, 1), 'mode': 'bilinear', 'valid_scale': 1.0, 'p': 1.0} -> Brightness -- {'max_lighting': 0.2, 'p': 1.0, 'draw': None, 'batch': False}

Building one batch
Applying item_tfms to the first sample:
  Pipeline: Resize -- {'size': (28, 28), 'method': 'crop', 'pad_mode': 'reflection', 'resamples': (2, 0), 'p': 1.0} -> ToTensor
    starting from
      (PILImage mode=RGB size=28x28, TensorCategory(1))
    applying Resize -- {'size': (28, 28), 'method': 'crop', 'pad_mode': 'reflection', 'resamples': (2, 0), 'p': 1.0} gives
      (PILImage mode=RGB size=28x28, TensorCategory(1))
    applying ToTensor gives
      (TensorImage of size 3x28x28, TensorCategory(1))

Adding the next 3 samples

No before_batch transform to apply

Collating items in a batch

Applying batch_tfms to the batch built
  Pipeline: IntToFloatTensor -- {'div': 255.0, 'div_mask': 1} -> Warp -- {'magnitude': 0.2, 'p': 1.0, 'draw_x': None, 'draw_y': None, 'size': None, 'mode': 'bilinear', 'pad_mode': 'reflection', 'batch': False, 'align_corners': True, 'mode_mask': 'nearest'} -> RandomResizedCropGPU -- {'size': None, 'min_scale': 0.8, 'ratio': (1, 1), 'mode': 'bilinear', 'valid_scale': 1.0, 'p': 1.0} -> Brightness -- {'max_lighting': 0.2, 'p': 1.0, 'draw': None, 'batch': False}
    starting from
      (TensorImage of size 4x3x28x28, TensorCategory([1, 1, 1, 1]))
    applying IntToFloatTensor -- {'div': 255.0, 'div_mask': 1} gives
      (TensorImage of size 4x3x28x28, TensorCategory([1, 1, 1, 1]))
    applying Warp -- {'magnitude': 0.2, 'p': 1.0, 'draw_x': None, 'draw_y': None, 'size': None, 'mode': 'bilinear', 'pad_mode': 'reflection', 'batch': False, 'align_corners': True, 'mode_mask': 'nearest'} gives
      (TensorImage of size 4x3x28x28, TensorCategory([1, 1, 1, 1]))
    applying RandomResizedCropGPU -- {'size': None, 'min_scale': 0.8, 'ratio': (1, 1), 'mode': 'bilinear', 'valid_scale': 1.0, 'p': 1.0} gives
      (TensorImage of size 4x3x26x26, TensorCategory([1, 1, 1, 1]))
    applying Brightness -- {'max_lighting': 0.2, 'p': 1.0, 'draw': None, 'batch': False} gives
      (TensorImage of size 4x3x26x26, TensorCategory([1, 1, 1, 1]))

Create the dataloader

dls = datablock.dataloaders(path)
#colab
dls.show_batch()

png

Create a Fastai CNN Learner

learner = cnn_learner(dls, resnet18, metrics=accuracy)
                      
#colab
learner.summary()

Set Learner to XLA mode This will setup the learner to use the XLA Device

#colab
learner.to_xla()
<fastai.learner.Learner at 0x7fc75a5a8090>

Using the lr_find works

#colab
learner.lr_find()
SuggestedLRs(lr_min=0.02089296132326126, lr_steep=0.0030199517495930195)

png

Run one cycle training.

#colab
learner.fit_one_cycle(5,lr_max=slice(1e-4,0.02))
epoch train_loss valid_loss accuracy time
0 0.896238 1.876374 0.505007 00:51
1 0.725275 1.018629 0.532189 00:13
2 0.614300 3.292356 0.513591 00:03
3 0.549166 2.262487 0.580830 00:03
4 0.535161 1.760773 0.660944 00:03

Further fine-tuning

#colab
learner.fit_one_cycle(5,slice(7e-4, 1e-3))
epoch train_loss valid_loss accuracy time
0 0.280359 0.992468 0.688126 00:12
1 0.315132 1.013533 0.712446 00:03
2 0.306187 0.745202 0.711016 00:03
3 0.282278 0.701246 0.736767 00:03
4 0.288474 0.854370 0.711016 00:03

Model params are using TPU

#colab
one_param(learner.model).device
device(type='xla', index=1)

Dataloader device is set to None (as it is the device mover transform that moves the input to the TPU)

learner.dls.device is None
True

Plot loss seems to be working fine.

#colab
learner.recorder.plot_loss()

png

Performance troubleshooting

To check if your model is hiting an aten operation (an operation that is not handled by accelerator device and returned to CPU for default implementation) you can check it with ands then you can report to pytorch xla team.

#colab
from fastai_xla_extensions.utils import print_aten_ops

print_aten_ops()
needs lowering: Counter: aten::adaptive_max_pool2d
  Value: 311
needs lowering: Counter: aten::adaptive_max_pool2d_backward
  Value: 201

Samples

Other examples of fastai notebooks using the fastai_xla_extensions package are also available here:

More samples will be added in the future as we fix issues and implement more capabilities.

Status

The fastai XLA extensions library is still in early development phase but is getting close to a 1.0 release.

The Multi Core TPU mode has been well tested on vision tasks (Classification) but the other modules (Tabular, Collab Filtering, and Text) have not.

If you use the package and encounter issues, please file bugs and issues encountered in the project's issue tracker.

Feel free to suggest enhancements or feature requests. (Though they might not necessarily be addressed any time soon).

If you wish to contribute to the project, fork it and make pull request.

This project uses nbdev -- a jupyter notebook first development environment and is being developed on Colab.

Development Setup

see this note on how to setup your environment for development