-
Notifications
You must be signed in to change notification settings - Fork 2
Jenkins for Pipelines
pipelines playbook includes a role to install configure and use jenkins for pipelines data processing jobs.
Follow the role README to install some jenkins plugins dependencies.
Some of these preconfigured jobs will need some extra script approval during their first Run:
Navigate to jenkins > Manage jenkins > In-process Script Approval
Navigate to jenkins > Manage jenkins > Configure > Labels and verify that your jenkins has a correct Label that match your spark and pipelines node name, and master
.
Manage Jenkins > Manage credentials > Jenkins (global) > Global credantials > Add Credential
You will need several things to run this job. In your production cassandra:
- Create this directory
# mkdir /data/uuid-exports/
# chown jenkins:jenkins /data/uuid-exports/
- copy in
/data/uuid-exports/uuid-export.sh
this script and give execution permissions. - Be sure that you can connect passwordless from your jenkins to your cassandra as spark user.
Copy some of your datasets to test data/dwca-export
like this:
/data/dwca-export
├── dr289
│ └── dr289.zip
├── dr490
│ └── dr490.zip
├── dr603
│ └── dr603.zip
(...)
└── dr879
└── dr879.zip
to start with smoke tests, please try with small datasets (< 300,000).
Index
- Wiki home
- Community
- Getting Started
- Support
- Portals in production
- ALA modules
- Demonstration portal
- Data management in ALA Architecture
- DataHub
- Customization
- Internationalization (i18n)
- Administration system
- Contribution to main project
- Study case