-
Notifications
You must be signed in to change notification settings - Fork 2
Jenkins for Pipelines
pipelines playbook includes a role to install configure and use jenkins for pipelines data processing jobs.
Follow the role README to install some jenkins plugins dependencies.
Some of these preconfigured jobs will need some extra script approval during their first Run:
Navigate to jenkins > Manage jenkins > In-process Script Approval
Navigate to jenkins > Manage jenkins > Configure > Labels and verify that your jenkins has a correct Label that match your spark and pipelines node name, and master
.
Manage Jenkins > Manage credentials > Jenkins (global) > Global credantials > Add Credential
You will need several things to run this job. In your production cassandra:
- Create this directory
# mkdir /data/uuid-exports/
# chown someuser:someuser /data/uuid-exports/ # optionally
- copy in
/data/uuid-exports/uuid-export.sh
this script and give execution permissions.
# chmod +x /data/uuid-exports/uuid-export.sh
- Be sure that you can connect ssh passwordless from your pipelines jenkins to your cassandra, from the spark user to
someuser
in cassandra that can run the previous script. - Adapt the migration-uuid job to fit to your infrastructure and users
Copy some of your datasets to test data/dwca-export
like this:
/data/dwca-export
├── dr289
│ └── dr289.zip
├── dr490
│ └── dr490.zip
├── dr603
│ └── dr603.zip
(...)
└── dr879
└── dr879.zip
to start with smoke tests, please try with small datasets (< 300,000).
Index
- Wiki home
- Community
- Getting Started
- Support
- Portals in production
- ALA modules
- Demonstration portal
- Data management in ALA Architecture
- DataHub
- Customization
- Internationalization (i18n)
- Administration system
- Contribution to main project
- Study case