Skip to content

VaibhavBansal26/Realtime-Logs-Processing

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Realtime-Logs-Processing

Realtime Logs Processing With Apache Airflow, Kafka & Elastisearch

System Architecture

1

Demo Video

demo

Screenshots

2

3

4

5

6

airflow commands

airflow init db
airflow migrate
airflow db init
airflow migrate
airflow webserver -p 8180
airflow scheduler
export AIRFLOW_HOME=${pwd}
airflow users create \
--username admin \
--firstname Vaibhav \
--lastname Bansal \
--role Admin \
--email vaibhav.bansal2020@gmail.com
  1. Store Secrets in AWS Secrets Manager a. KAFKA_SASL_USERNAME b. KAFKA_SASL_PASSWORD c. KAFKA_BOOTSTRAP_SERVER d. ELASTICSEARCH_URL e. ELASTICSEARCH_API_KEY

  2. Create account in confluent kafka -> Create Environment -> Provision Cluster -> Create Topic

  3. Get bootstrap Server from Cluster Settings in Kafka

  4. Get bootstrap server url, username and password from API keys (Generate if not available)

  5. Create environment and cluster in confluent kafka

  6. Create account in elasticsearch

  7. Create index in elasticserarch with partitions with a particular name in dags and get the elasticsearch url and api key to be stored in aws secrets manager

  8. CREATE IAM USER

  9. CREATE S3 Bucket

  10. Go to AWS MANAGER APACHE AIRFLOW -> Create New Environment -> Link it to your dags -> Run Dag On Airflow

  11. Configure the path for dag folder, S3 bucket, & requirements file

  12. Create VPC and Add Secret Manager Read-Write policy to role in MWAA

About

Realtime Logs Processing With Apache Airflow, Kafka & Elastisearch

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages