Installation

Installation to run locally

Follow the steps below to install the library, create a job folder and execute them in a docker environment:

$ pip install yaetos
$ cd /path/to/an/empty/folder/that/will/contain/pipeline/code
$ yaetos setup  # to create sub-folders and setup framework files.
$ yaetos launch_env # to launch the docker container
# From inside the docker container, try a test pipeline with
$ python jobs/examples/ex1_frameworked_job.py --dependencies

Installation to push jobs to the cloud

To get jobs executed and/or scheduled in AWS, You need to:
  • Get permissions setup in AWS to access: S3 (at least 1 bucket), EMR, AWS Secrets, AWS Data Pipeline, and AWS Airflow

  • fill AWS parameters in conf/config.cfg. More details below.

  • have ~/.aws/ folder setup to give access to AWS secret keys. If not, run pip install awscli and aws configure, adding –profile=your_profile if a profile is already registered in AWS CLI.

If not available, the “Key Pair” parameter from the conf/config.cfg can be created following the instructions at https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/create-key-pairs.html. It involves going to https://console.aws.amazon.com/ec2/.

_images/Screenshot_2022-05-02_at_12.57.55.png