First and foremost, welcome! 🎉 Willkommen! 🎊 Bienvenue! 🎈
Thank you for visiting the Churn Insights
project repository. This README file provides essential information about our project. Feel free to jump straight to one of the sections below, or simply scroll down to learn more.
Our Churn Insights
team, comprised of adept developers and data scientists, is driven by a shared passion for technology and data analytics. Our expertise in data analysis underpins our commitment to this project.
Churn Insights
is an analytical platform dedicated to understanding and reducing customer attrition. Aimed at stakeholders such as business managers, data analysts, and customer relationship teams, it provides actionable insights through the careful analysis of customer data to identify patterns and reasons behind customer departures.
As a data analytics team, we strive to provide our target audience with a comprehensive tool that elucidates the various determinants of customer churn.
The business environment is intricate, often characterized by competitive market dynamics. There's a pronounced need for understanding customer behavior, especially to retain valuable customers and reduce churn rates.
With Churn Insights
, you can explore customer trends, analyze attributes influencing churn, and view real-time churn predictions, all through an interactive and intuitive interface.
Our Churn Insights
dashboard is an empowering tool that enables:
- The exploration of customer data with customizable features and filters.
- Community-driven contributions to our database, fostering growth and inclusivity.
- The synthesis of valuable insights to facilitate well-informed business decisions.
Churn Insights is a comprehensive project that involved several key steps, starting with the ETL process in PostgreSQL via pgAdmin to load and transform the data. We then conducted data cleaning within PostgreSQL to ensure the dataset was accurate and consistent. Following this, we applied data transformations and created enhanced visualizations in Tableau to effectively present our findings. The final phase involved building and evaluating multiple machine learning models, including KNN, Decision Tree, Random Forest, RBF SVM, and Logistic Regression, within Jupyter Notebook. We conducted thorough Exploratory Data Analysis (EDA) and selected the best model to make predictions and present the results.
The current iteration of Churn Insights
boasts:
- Interactive filters for localized customer data analysis.
- Correlative bar plots that elucidate the influence of customer features on churn.
- An interactive map to visualize geographical data and churn distributions.
- A table with an implemented churn probability slider for interactive filtering.
We are poised to refine Churn Insights
further, enhancing its interactivity and user interface. Our ambitions include the integration of advanced predictive analytics and the expansion of our dataset to encompass global markets.
- ETL Process in PostgreSQL from pgAdmin
- Data Cleaning in PostgreSQL
- Tableau Transformations
- Tableau Visualization & Enhancing Visuals
- Build Machine Learning Model - EDA and build KNN, Decision tree, Random Forest, RBF SVM, Logistic regression models in Jupyter Notebook and finally pick the best model to predict and show the results.
Encountered an issue or have a question? Feel free to open an issue in this repository, and our team will be happy to assist you.
Our project is a collaborative endeavor, with each team member playing a pivotal role:
Our analysis is anchored in robust customer datasets, obtained from various sources. Details are provided within the project repository.
The Churn Insights
codebase is MIT licensed, as found in the LICENSE in this repository.
To run Churn Insights
locally:
-
Clone the Repository
Clone the repository using Git:
git clone https://github.com/your-repo/churn-insights.git cd churn-insights
-
Set Up PostgreSQL Database
- Open pgAdmin and create a new database.
- Run the SQL scripts provided in the
sql
folder to set up the ETL process and clean the data.
-
Tableau Visualization
- Open Tableau and connect to your PostgreSQL database.
- Use the transformations provided in the
tableau
folder to create and enhance visuals.
-
Run Machine Learning Models
- Navigate to the
notebooks
folder. - Open the Jupyter Notebooks provided and run the machine learning models.
- Navigate to the
-
Create a Conda Environment
Create a Conda environment named
churn_insights
using theenvironment.yml
file. This file contains all necessary dependencies:conda env create -f environment.yml
-
Activate the Conda Environment
Activate the newly created environment:
conda activate churn_insights
-
Run Jupyter Notebook
jupyter notebook
-
Open and Execute Notebooks
- Open the notebooks from the Jupyter interface and run the cells to execute the machine learning models.
Dive into Churn Insights
and join us in forging a more transparent and inclusive business environment. Your contributions are what make the open-source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated. Please read our Contribution Guidelines for details on our code of conduct and the process for submitting pull requests to us.
Thank you so much (Danke schön! Merci beaucoup!) for visiting our project. We hope you'll join us on this exciting journey to empower businesses with Churn Insights.
- README file: A document introducing an open project to the public and potential contributors, typically including installation, usage, and contribution information.
- Repository (or Repo): A collection of documents related to your project, where you create and save new code or content, often hosted on platforms like GitHub or Bitbucket.
- Roadmap: A document outlining the planned schedule of work for a project.
- Milestone: An event or state marking a significant stage in development within the project.
- Issue: The GitHub term for tasks, enhancements, bugs, discussions, and questions related to your projects.