WARNING: Jenkins X version 2.x is deprecated.
Please refer to the v3 documentation for the latest supported version.
In most scenarios, you will be expecting to remotely access data hosted elsewhere on your network and can manage this in code as part of your training scripts and service implementations. There are however a couple of situations in which Jenkins X can help you to manage certain types of data.
Handling data in Buckets
If you are working with data in the form of arbitrary files, you can transfer these to your training environment via a Storage Bucket in your Cloud project with the following helper step within your build pipeline:
> jx step unstash -u <URL of bucket> -o <output filename or directory>
The URL should be in the form: s3://mybucket/tests/myOrg/myData/trainingset.xml where the protocol is set as appropriate to the Cloud provider you are using.
You must ensure that the data has been uploaded to this bucket prior to starting the training build and should bear in mind that this command copies the specified file from the bucket to the working volume of the build container executing the current build step.
Working with Volumes
Under some circumstances, you may wish to create versioned collections of immutable training data that can be shared across multiple models and which are too large to easily copy from buckets in a timely manner.
Under these circumstances, it is straightforward to create a named, persistent Kubernetes Volume within your Cloud project, mount it in read/write mode and upload your training data files to it, then unmount it ready for use.
Within your training pipelines, you can then specify that this volume be mounted during the training build, in read-only mode, in more than one project in parallel.
To do this, you need to modify the
jenkins-x.yml file in your training projects to reference the
volumeMount config necessary to connect the build container instance to your training data volume.
buildPack: ml-python-gpu-training pipelineConfig: pipelines: overrides: - pipeline: release stage: training name: training volumes: - name: trainingset gcePersistentDisk: pdName: mytrainingvolume01 fsType: ext4 readonly: true containerOptions: resources: limits: cpu: 4 memory: 32Gi nvidia.com/gpu: 1 requests: cpu: 0.5 memory: 8Gi nvidia.com/gpu: 1 volumeMounts: - name: trainingset mountPath: /trainingset readOnly: true
Note that it is only possible to simultaneously share volumes that are mounted read-only.
Was this page helpful?
Glad to hear it! Please tell us how we can improve.
Sorry to hear that. Please tell us how we can improve.