Data Engineer with over 2 years of experience in designing and optimizing data pipelines, focusing on big data processing and efficient workflow automation. Skilled in leveraging Python and PySpark for large-scale data processing, advanced SQL for querying and optimization, and data modeling for building scalable solutions. Proven ability to automate workflows, ensure data quality, and deliver high-performance data solutions that support business insights and decision-making.
- 🔭 I’m currently working on Developing scalable data pipelines, optimizing big data workflows, and building real-time data processing systems
- 🌱 I’m currently learning Advanced data warehousing concepts, optimizing distributed systems, and integrating machine learning with data pipelines
- 👯 I’m looking to collaborate on Big Data Engineering, Data Warehousing, Data Infrastructure and Distributed Data Processing
- 💬 Ask me about Python, PySpark, SQL, and Data Pipeline Automation
- 📫 How to reach me lokeshreddy.lingala18@gmail.com
- ⚡ Fun fact I’m passionate about building efficient, scalable data solutions that drive business insights
Python, Pyspark, SQL, HiveQL, Java, HTML, CSS, R, C Programming, Redshift, Snowflake, Iceberg, BigQuery, PostgreSQL, MySQL, DynamoDB, MongoDB, AWS Services: S3, Glue, Lambda, EC2, Athena, RDS, EMR, Kinesis, DMS, Pandas, Trino, Power BI, Kafka, Spark, Airflow, Flink, Git, dbt, Docker, VS Code