Skip to content

from-307-dockerize-the-best-run-from-HPO-study #309

@david-thrower

Description

@david-thrower

TLDR Package the settings from best run in Docker

Issue

We have found the optimal model on the HPO srudy done on branch 307. We need to package this run in a suitable GPU enabled container to scale it up and control the dependencies.

Task

Make this run in tensorflow/tensorflow:2.20.0-gpu

  • Dockerize the script
  • Set up a volume mount for artifacts
  • Clean up dependencies
  • Parameterize the script (Dataset, number of samples, sample expansion, ...)

To DO:

  • Replace the prompt samples with the ones from the data set.
  • Set the stage 1-a and 1-b model checkpoints to save in the artifacts folder.
  • Make sure model serialization still works on the mounted volume
  • Add conditional MlFlow logging of params (log if MLFLOW_PORT != 0).
  • Set MlFlow up on SQLite on the monted volume.
  • Add MlFlow system metrics.
  • make sure the experiment name is unique automatically. (Generate a unique name to avoid naming collisions.)
  • Rebuild and push the latest container.

Metadata

Metadata

Assignees

No one assigned

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions