Designing, building, and maintaining scalable data pipelines and distributed systems. Passionate about reliable data systems, automation, and performance engineering.
I'm a Data Engineer with 2+ years of experience designing and maintaining high-performance data systems. My expertise lies in building streaming and batch pipelines using tools like Kafka, Python, and AWS.
I specialize in creating reliable, scalable systems that handle large volumes of data efficiently. With a strong foundation in distributed systems and data engineering principles, I focus on delivering solutions that are both performant and maintainable.
Building robust data pipelines
AWS, Docker, and DevOps
Scalable and efficient systems
Python (Pandas, BeautifulSoup, Selenium, Django, Flask), SQL
Apache Kafka, RabbitMQ, ETL/ELT, Cron, Distributed Systems
AWS (S3, EC2), Docker, Git, CI/CD
MongoDB, PostgreSQL, Elasticsearch
Built end-to-end streaming pipeline with Python, Apache Kafka, and PostgreSQL for real-time stock data simulation. Containerized ecosystem using Docker and created live dashboards with Grafana for data visualization.
Developed Django web application enabling users to upload datasets and generate interactive visualizations. Implemented data processing with Pandas and visualization with Matplotlib/Seaborn. Designed responsive UI with Bootstrap.
Deogiri Institute of Engineering and Management Studies
2018 – 2022
I'm open to new data engineering roles, collaborations, or freelance projects. Reach out if you'd like to discuss opportunities or projects.