Skip to content
This repository has been archived by the owner on Jan 3, 2023. It is now read-only.

training process is killed because OOM #48

Open
fancyerii opened this issue Aug 29, 2017 · 4 comments
Open

training process is killed because OOM #48

fancyerii opened this issue Aug 29, 2017 · 4 comments
Assignees

Comments

@fancyerii
Copy link

fancyerii commented Aug 29, 2017

I have trained neon for librispeech data. But it's always killed because OOM. My machine has 24GB memory and GeForce GTX 1070 card of 8G memory.

I found this msg by dmsg
[3017506.733819] Out of memory: Kill process 25635 (python) score 974 or sacrifice child
[3017506.736861] Killed process 25635 (python) total-vm:55518724kB, anon-rss:23902876kB, file-rss:154436kB

is neon leaking memory or it require more memory to train?

The command I run is:
python train.py --manifest train:/bigdata/lili/deepspeech/librispeech/train-clean-100/train-manifest.csv --manifest val:/bigdata/lili/deepspeech/librispeech/train-clean-100/val-manifest.csv -e 20 -z 16 -s models -b gpu

@gardenia22
Copy link

Try reduce batch size.

@fancyerii
Copy link
Author

fancyerii commented Sep 4, 2017

I changed batch_size to 8 but it's still killed.
[3256824.391743] Killed process 9666 (python) total-vm:53893188kB, anon-rss:23892380kB, file-rss:152808kB

it use too much memory

@Neuroschemata
Copy link
Contributor

I suspect the source of the problem is unrelated to the model size. With the default parameters using the command you posted above, I get the following:

batch size GPU memory footprint
32 6949 GB
16 3915 GB
8 2415 GB

So your 8GB GPU has the capacity to handle a batch size of up to 32.

@fancyerii
Copy link
Author

so what's wrong? From the /var/log. it seems this python process used 23892380kB(23GB) cpu memory(not gpu memory).

[3256824.391743] Killed process 9666 (python) total-vm:53893188kB, anon-rss:23892380kB, file-rss:152808kB

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants