Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

XLA_GPU and no GPU utilization #38

Open
bkj opened this issue Sep 21, 2020 · 2 comments
Open

XLA_GPU and no GPU utilization #38

bkj opened this issue Sep 21, 2020 · 2 comments

Comments

@bkj
Copy link

bkj commented Sep 21, 2020

When I run

CUDA_VISIBLE_DEVICES=0 python fixmatch.py --filters=32 --dataset=cifar10.0@40-1 --train_dir ./experiments/fixmatch

I see the process use GPU memory in nvidia-smi, but there is 0% GPU utilization and training is super slow. When I look at the devices returned by libml/utils.py:get_available_gpus, the local_device_protos are all XLA_GPU instead of GPU. Any ideas on what might be going on here and how to fix? Presumably this is some kind of version issue?

(Apologies that this is a more general TF question, but I wasn't able to find a working fix by Googling)

@david-berthelot
Copy link
Collaborator

Sorry I have no idea, looks like a TensorFlow question.

@khalilsarwari
Copy link

I had a similar issue and was able to resolve it by using conda to install tensorflow (and its dependencies/cudnn etc) instead of pip

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants