Hi, I'm Jayakumar

Personalize Theme

*Theme settings will be saved for
your next visit

What I Do

I work as a Data Engineer building end-to-end data pipelines using cloud platforms, Databricks, Talend, and Apache Airflow. I specialize in scalable data ingestion, transformation, and optimization with distributed processing frameworks, and integrate AI and machine learning components to enable advanced analytics and data-driven decision-making.

More about me

I am passionate about learning new skills and exploring emerging technologies.

Currently, I’m focused on combining AI with data engineering to tackle real-world problems and create intelligent solutions. I’m enthusiastic about collaboration and always excited about new opportunities to learn, innovate, and contribute.



My skills

Data Engineer | AI & Gen AI Enthusiast

Building end-to-end data pipelines, integrating AI and Generative AI solutions, and exploring innovative data-driven applications : Download Resume

  • Databricks
  • Pyspark
  • Python
  • SQL
  • Talend
  • Apache Airflow
  • Gen AI
  • Git / Github
  • CI / CD Pipelines
  • AWS Cloud Services
  • Java
  • -
  • -
  • -
  • -
  • -
  • -
  • -
  • -
  • -
  • -
  • -
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Proficient
  • Intermediate

💼 Experience

Business Analyst, Genpact Bengaluru, Karnataka | Oct 2024 - Present
  • Architected and delivered enterprise-grade Lakehouse solutions on Databricks at scale.
  • Designed and built high-performance ETL/ELT pipelines using PySpark and Spark SQL following the Medallion Architecture.
  • Engineered complex, large-scale data transformations handling millions of records efficiently.
  • Led Delta Lake and Unity Catalog implementation for governed data access, lineage, and cross-team collaboration.
  • Developed and migrated custom EDL jobs from Talend to Databricks, modernizing legacy workflows.
  • Optimized Spark workloads using advanced partitioning, caching, and query tuning techniques.
  • Built reusable, production-grade Python frameworks with robust logging and fault tolerance.
  • Owned end-to-end integration of Databricks pipelines with Apache Airflow, managing over 200 DAGs.
  • Enabled reliable, fully automated data platforms powering analytics, BI, and ML use cases.
Data/AI Intern, Genpact Bengaluru, Karnataka | Feb 2024 - Aug 2024
  • Completed the NextGen 24 Analytics program with hands-on experience in Power BI, Python, and Excel.
  • Worked in the GenAI COE on an AI-powered financial report summarization project, processing large documents using chunking and embeddings, and generating concise DOCX summaries for financial advisors.

Some of my college projects

Loan prediction using supervised machine learning models

Prediction of loan by verifying all the satisfactory constraints to be done for loan approval by using supervised machine learning algorithms.

Github Link
Pneumonia classifier

Deep learning based Pneumonia classifier which classifies the respective pneumonia class with the help of chest X-Ray images by using 7 different CNN models.

Github Link
Flower identification using tensorflow

Identification of flowers by using popular flower image dataset which classifies the respective flower's name with the help of tensorflow.

Github Link
Bitcoin price prediction

Prediction of the decentralized Bitcoin's price based on the previous data obtained from the yfinance website.

Github Link
Smart Stick For Blind People

Helping aid for blind people which helps them via object detection, GPS tracking and messaging the location using GSM to the guardian.

Github Link
Personal portfolio website

Simple portfolio website using HTML and CSS which showcase my skills.

Github Link

Contact Me

Let's connect and collaborate to bring your ideas to life!

Call Me

+91 86084 13629

Location

Bengaluru - Karnataka - India

Made with 🗿 by Jayakumar S