About Me

I am an AWS & Azure certified, data professional with 3+ years of experience in transforming complex data into actionable insights, spanning roles in data engineering, analysis, and management. I specialize in making data accessible and impactful, bridging gaps between raw data and business intelligence to guide strategic decisions.


Currently, I am enhancing data efficiency and governance at Texas A&M University, where I develop and implement data-driven features in Slate CRM. By optimizing database structures and managing workflows, I ensure seamless, targeted communication and precise reporting. This role has strengthened my expertise in data quality, segmentation, and accessibility, supporting informed decision-making through reliable data systems.


In my previous role at L&T Infotech, I contributed to major projects, including integrating supply chain data for Keurig Dr. Pepper, where I transformed diverse data sources into meaningful analytics. My leadership of the Rotaract Club of Hyderabad Corporates also taught me the importance of empathy, adaptability, and teamwork in achieving shared goals and making a positive community impact.


Now, as I pursue my Master’s in Management Information Systems, I am expanding my analytical and technical expertise, actively seeking roles as a Data Engineer, Data Analyst, or Business Analyst. I aim to leverage my skills in data transformation, big data, and cloud platforms to drive impactful, data-informed solutions.

Key Competencies

  • Accountable: proven experience in delivering high-impact results
  • Team player: great collaboration skills
  • Agile and adaptable: can quickly adapt to changing requirements
  • Positive influence: optimistic driver of new initiatives

Work Experience

Texas A&M University College Station, Texas

Graduate Student Assistant - CRM Database AdministratorAug 2023 - Present

  • Developed and implemented data-driven features and functions in Slate CRM, including the creation of forms, portals, event templates, and scheduler tools, while managing data imports, populations, rules, and campaigns. Optimized database structures to ensure accurate data segmentation, integrity, and accessibility for targeted communication and reporting.
  • Streamlined and maintained Slate CRM data workflows, building and managing queries, reports, and databases. Enhanced data governance by ensuring clean, reliable datasets, troubleshooting application issues, and proposing optimization strategies to improve data efficiency and support business decision-making.

Keurig Dr. Pepper (via LTIMindtree) Hyderabad, India

Senior Data EngineerJan 2022 - July 2023

  • Designed a custom database architecture and automated end-to-end data pipelines using Qlik, SQL, and MicroStrategy allowing an American beverage conglomerate to expedite its supply chain and marketing analytics turnaround time to 2 days.
  • Collaborated with business teams to gather requirements and built ETL pipelines using Snowflake, Python & Informatica, resulting in 20% quicker access to the most recent data; managed delivery timelines using JIRA.
  • Created analytical dashboards and forecasting reports using Tableau & Power BI ensuring accurate logistics data tracking and achieved a 12% efficiency boost through automation of manual tasks.

Data EngineerJuly 2021 - Jan 2022

  • Engaged with the client marketing teams to understand the DTC objectives; developed Python scripts to enable accurate and effective visualization of the as-is, and to-be trends & strategies.
  • Extracted data from various sources and transformed them into unified, actionable datasets, further generated analytical reports enabling the supply chain management team to decrease the warehouse-to-rack delivery timeline by 3 days.

Research Centre, ImarathHyderabad, India

Data Engineer InternOct 2019 - Sep 2020

  • Designed, implemented, and supported maintenance of data capturing and ETL processes for the 'UDP Data Stream Capture & Analyzer' project, facilitating agile adaptation to receiving a live stream of data.
  • Utilized Python scripting to capture, transform, and analyze the live data for defense engineering R&D, complemented by a UI dashboard for customizable analytics and informed decision-making.

Education

Texas A&M University College Station, Texas

Master of Science in Management Information SystemsAugust 2023 - May 2025

Relevant Coursework: Advanced Database Management, Data Warehousing, Engineering Data Analytics and Machine Learning, System Analysis and Design


Vasavi College of Engineering Hyderabad, India

Bachelor of Engineering in Information TechnologyJuly 2017 - May 2021

Relevant Coursework: Big Data Analysis, Data Mining & Visualization, Database Management Systems, Design and Analysis of Algorithms, Operating Systems, Software Engineering

Projects

Explore my data engineering projects in this section.

Stock Market Data Streaming  Code  Demo

Tech Stack: Docker, Apache kafka, Spark, InfluxDB, Grafana

Built a real-time stock market data pipeline using Kafka, Spark, InfluxDB, and Grafana. Deployed an autonomous dockerized system for handling historical, incremental, and live data. Optimized Kafka and Spark jobs for low-latency processing of stock price updates. Configured InfluxDB retention policies and deployed Grafana dashboards for real-time market insights to support financial decision-making.


SQL Server to Azure Cloud - Data Migration Engineering Code

Tech Stack: Azure, SQL Server, Python, Databricks

Migrated an on-premises SQL Server Point-of-Sales database to Azure Cloud. Developed an automated data pipeline using Azure Data Factory for data transfer, Azure Databricks for data transformation, and Azure Synapse Analytics for efficient data querying and reporting needs. Built Tableau dashboards to visualize key metrics and display product and sales insights.

Stock Market Data Streaming Code

Tech Stack: Docker, Apache Kafka, Spark, InfluxDB, Grafana

Built a real-time stock market data pipeline using Kafka, Spark, InfluxDB, and Grafana. Deployed an autonomous dockerized system for handling historical, incremental, and live data. Optimized Kafka and Spark jobs for low-latency processing of stock price updates. Configured InfluxDB retention policies and deployed Grafana dashboards for real-time market insights to support financial decision-making.


SQL Server to Azure Cloud - Data Migration Engineering Code

Tech Stack: Azure, SQL Server, Python, Databricks

Migrated an on-premises SQL Server Point-of-Sales database to Azure Cloud. Developed an automated data pipeline using Azure Data Factory for data transfer, Azure Databricks for data transformation, and Azure Synapse Analytics for efficient data querying and reporting needs. Built Tableau dashboards to visualize key metrics and display product and sales insights.


Dominicks Data Warehouse Code

Tech Stack: SSMS, SSIS, SSRS, SQL, PowerBI, Tableau, OLAP Data Warehousing

Designed and implemented an OLAP data warehouse for Dominick’s Fine Food using Kimball methodology, creating a star schema with fact and dimension tables. Built ETL processes with SSIS to ensure data consistency and calculated key metrics. Developed Tableau and Power BI dashboards, SSRS reports, and SSAS cubes for multi-dimensional analysis. The project optimized sales, pricing, inventory, and customer insights, improving operational efficiency and business strategy.


Concrete Strength Analysis Code

Tech Stack: ML: Exploratory Data Analysis, Linear Regression, Decision Trees, and SVM's

Built a machine learning model to predict the compressive strength of concrete based on its composition. Collected and cleaned data on concrete mixtures, ensuring accuracy by removing invalid values. Identified key relationships between ingredients and strength through exploratory data analysis. Tested models like linear regression, decision trees, and support vector machines, selecting the most accurate one. The best model highlighted crucial factors influencing strength.


WhatsStats using Python Code

Tech Stack: Python, Libraries : Numpy, Pandas

Developed a Python script to parse and analyze WhatsApp group chat data. Automated the processing of chat logs to uncover key insights such as message frequency, peak activity periods, emoji usage, and the most active participants. Leveraged libraries like Pandas and Matplotlib to visualize trends and provide data-driven insights into group dynamics, member engagement and conversation patterns, offering a deeper understanding of social interactions within the group.

Scholarly Works

Explore my publications and articles here.

IoT-Based Smart Parking System Publication

Tech Stack: IOT, Python, HTML, CSS

Designed a smart parking system utilizing IoT technology to alleviate parking challenges in urban areas. The system integrates sensors and microcontrollers to detect available parking slots, updating real-time data to the cloud. Users can access this information through a web application, significantly reducing the time and effort required to find parking.

Skills

Here is a snapshot of skills that I bring to the table.

Programming Languages

  • Python
  • SQL
  • Shell-scripting
  • Java
  • C Programming

Big Data

  • Snowflake
  • Hadoop
  • Spark

Databases

  • MySQL
  • PostgreSQL
  • MS SQL
  • NoSQL (MariaDB)

Visualization and Other Tools

  • Tableau
  • Power BI
  • Advanced MS Excel
  • Git
  • Jira

Cloud and Containers

  • Azure (Data Factory, Data Lake, Databricks, Synapse Analytics)
  • AWS (S3, EC2)
  • Docker

Certifications

  • AWS Certified Data Engineer Associate
  • Microsoft Certified Power BI Data Analyst Associate
  • Advanced SQL - Hacker Rank
  • Intermediate Python - Hacker Rank
  • Other Datacamp & LinkedIn Certifications

Contact

Feel free to reach out to me on the details mentioned below.

Write to me:    saikrishna.koppula@tamu.edu  

Follow: