Slow NER training with GPU - SpaCy v3.2 #9876
Replies: 2 comments 5 replies
-
What makes you say your training is slow? Is it faster in another context? You say that memory usage is low and therefore it's slow but I don't understand the connection. Also, how big is your training set, your GPU, your RAM...? It's not clear if SLURM is relevant here or not either. |
Beta Was this translation helpful? Give feedback.
-
Thank you for your reply. My training dataset contains 574 files (short texts), I was able to retrieve an accounting info on the submitted job (on GPU v100) which ran ~ 11 minutes for 7 epochs (~ 24 minutes on CPU). Is this normal? I think that if we have a ratio of less than 4, we can talk about slowness. |
Beta Was this translation helpful? Give feedback.
-
Hi,
I am using GPU to train a NER model from Scratch in SpaCy v3.2 (with the --gpu-id option) and SLURM job scheduler:
sbatch -p gpu --gres = gpu: v100: 1 my_script.sh
Here is the "my_script.sh" submission script:
#! / bin / bash
python -m spacy train config.cfg --output ./output --paths.train ./train.spacy --paths.dev ./train.spacy --gpu-id 0
When I use nvidia-smi, I can clearly see GPU usage at 7% with memory usage at 0% (slow). That's why, I think that adjustments on my side are to be made at the level of the SpaCy to optimize its use.
Could you please tell me where this slow training with GPU comes from?
Thanks in advance,
FA
Beta Was this translation helpful? Give feedback.
All reactions