blob: 26b80ac323ba0e50875719c0d098230ad64b1e07 [file] [log] [blame]
__version__ = '{{VERSION}}'
debug = False
cuda = '{{CUDA_VERSION}}'
hip = None
# This is a gross monkey-patch hack that depends on the order of imports
# in torch/__init__.py
# TODO: find a more elegant solution to set `USE_GLOBAL_DEPS` for the bazel build
import torch
torch.USE_GLOBAL_DEPS = False