Run the Pipeline Container (AWS)
This section will teach you:
Configure Disdat with an ECR prefix and the name of your AWS Batch job queue
How to execute your container on Batch
How to look at your resulting bundles
Have you setup your AWS credentials? You will also need to stand up AWS Batch. That means creating 1.) a compute environment and, 2.) a batch job submission queue.
You then need to add your AWS Batch job queue name to the Disdat configuration file.
We're still assuming your running in (switched into) your examples context that has a remote attached:
Push container to ECR
Let's push our container up to AWS ECR. We use the same dockerize command. We say --no-build
because you built it in the prior step.
You should see a bunch of transfer status updates as Docker moves the container to ECR.
Submit the job to batch
If you log in to your AWS account you should see something like:
Once the job moves to the SUCCEEDED state, then the container has run successfully. If it didn't you can click on the job and follow the clicks through to its CloudWatch logs. Those look like:
Let's grab our results from the remote context
Yay! We pushed a container, ran it up on AWS, and then grabbed all of our results!
Continuing on!
There's a lot more to know about how Disdat manages bundles when it runs pipelines remotely. And how to control the size of the container instance of your job.
I need to know more details about building pipelines (using dependencies, return types, etc.)
I need more information on running in AWS
I need to build containers with unix*, R, or other dependencies
Last updated