Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Import Model] High transient memory usage when installing large models with Docker #566

Open
davidkyle opened this issue Jul 19, 2023 · 0 comments
Labels
topic:NLP Issue or PR about NLP model support and eland_import_hub_model

Comments

@davidkyle
Copy link
Member

davidkyle commented Jul 19, 2023

Using the docker container to run eland_import_hub_model and install a large model imposes large memory requirements. For example using the command below to install xlm-roberta-base requires the container to have more than 8GB of memory.

docker run -it --rm --network host \
    elastic/eland \
    eland_import_hub_model \
      --url 'https://elastic:XXX@host:9200/' \
      --hub-model-id xlm-roberta-base \
      --task-type fill_mask

If the container does not have enough memory the process exits shortly after the download has completed prior to uploading the model to Elasticsearch. Investigate the what is causing the high memory usage.

  • What is causing the high memory usage?
    • Tracing the mode?
  • How can a failure due to insufficient memory be detected and reported?
@davidkyle davidkyle added the topic:NLP Issue or PR about NLP model support and eland_import_hub_model label Jul 19, 2023
@davidkyle davidkyle changed the title [Import Model] High transient memory usage when installing large models [Import Model] High transient memory usage when installing large models with Docker Aug 24, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
topic:NLP Issue or PR about NLP model support and eland_import_hub_model
Projects
None yet
Development

No branches or pull requests

1 participant