Name: Punya Ira Anand

Interested Roles: Software Engineering/ Data Engineering/ Business Intelligence Analyst

Experience: 4 Years FullTime (India)

Address: Dallas, United States

Skills

PYTHON
OBJECT ORIENTED PROGRAMMING
SQL
Cloud

About

About Me

Hello! I'm Punya , a passionate Software Enthusiast based in Dallas, United States. I'm currently pursuing my Master of Science in Information Technology and Management at the University of Texas at Dallas. With over four years of industry experience as a Data Engineer and Business Intelligence Engineer, I've developed a strong foundation in Software and Data Engineering, Visualization, and interpretation. I thrive on transforming complex datasets into actionable insights that drive business growth and decision-making.Outside of data, I'm an avid learner and enjoy exploring new technologies and methodologies to stay ahead in the rapidly evolving field of data science.

  • Profile: Data Engineering & Analytics
  • Domain: Healthcare,Banking,Finance & IT
  • Education: MS Information Technology & Management
  • BI Tools: Tableau, Jupyter, Figma, SAS DI Studio, SAS Management Console, IBM- Flow Manager, Power-BI
  • Databases: PostgreSQL, MS SQL Server, MongoDB, Hadoop (HDFS, HBase, Hive)
  • Programing: Python, SQL, No SQL, Unix, Shell Scripting, Base SAS, HiveQL
  • Certifications:AWS Solutions Architect
  • CI/CD:Git, GitHub, Jenkins
  • Core Skills:Data Engineering, BusinessIntelligence, Data Analysis, ETL/ELT Automation, SDLC

Resume

Resume

Experienced Software Engineer with 4+ years of proven track record in leveraging data-driven insights to steer business strategies. Specialized in Business Intelligence, statistical analysis, Data Visualization, Automation, and Prediction Modeling.

Experience


Oct 2022- Jan 2024

Senior Software Engineer

Wipro Technologies

Wipro Technologies is a global technology consulting and digital solutions company, providing end-to-end IT services to clients across various industries.

  • Implemented a stored process to integrate, clean, and transform Claims data from four distinct sources. Enabled seamless data availability and automated reporting process, reduced computation time by 20 hours.
  • Designed and implemented jobs to maintain the integrity of incoming data streams, ensuring 75% of GDPR compliance.
  • Leveraged Proc SQL to automate processes and generate reports, optimizing data storage within SAS data warehouse, resulting in a 30% reduction in report generation time and a 15% improvement in storage efficiency.
  • Developed and optimized advanced SQL queries for healthcare data, achieving a 25% boost in processing efficiency and a 20% reduction in report turnaround time.
  • Managed and optimized analytics environments using SQL Server Management Studio (SSMS) and PostgreSQL. Ensured database performance through real-time monitoring, query tuning, and system reliability practices by 25%.

Jan 2020- Sept 2022

Software Engineer

Larsen & Toubro Infotech

Wipro Technologies is a global technology consulting and digital solutions company, providing end-to-end IT services to clients across various industries.

  • Developed data mining jobs using Proc SQL and Python Scripts to extract insights from policy data, enriching the data landscape. This initiated a result in a 5% improvement for the business. .
  • Enhanced a SQL server database with 50,000+ customer records for healthcare clients, enacting SQL queries (Stored Procedures, Joins, Indexing,Views).
  • Developed a Python-based Time Series Forecasting model to predict monthly ticket closure rates, leading to a substantial 45% improvement in operational efficiency.
  • Conducted information and requirement-gathering sessions with business stakeholders, and served as a subject-matter expert, identified, and addressed main points in the data delivery pipeline.
  • Constructed and managed shell scripts to automate batch processing tasks for data loading and transformation, improving workflow efficiency by 30% and optimizing performance across SAS environments.



Education


Jan 2024 - May 2025

Master of Science, Information Technology & Management

University of Texas at Dallas

GPA: 3.8

July 2015 - June 2019

Bachelor of Engineering, Information Technology

Rajiv Gandhi Proudyogiki Vishwavidyalaya

GPA: 3.7


Projects

Projects

Below are data analytics projects briefly describing the technology stack used to solve cases.

Data Engineering for Trending Youtube Video Analysis in AWS

Below is the data analytics project developed using AWS and its services.

Steps: Develop a data pipeline in AWS to manage, streamline, and transform the structured and unstructured data containing 200+ trending youtube video datasets to perform analysis based on categories and trending metrics.

Skills: Python, Spark, AWS(S3, IAM, Glue, Athena, Lambda, Quicksight, AWS CLI)

Data Engineering for Trending Youtube Video Analysis in AWS

• Project Learnings - Data Ingestion, Data Lake, ETL Operations, Cloud Reporting

• Description: Developed a data pipeline in AWS to manage, streamline, and transform the structured and unstructured data containing 200+ trending youtube video datasets to perform analysis based on categories and trending metrics.

Twitter Data Pipeline with Airflow

The project involves setting up an ETL (Extract, Transform, Load) pipeline to extract data from Twitter using Tweepy, store the data in AWS S3 storage, and then load formatted data into a Redshift data warehouse for sentimental analysis.

Steps: Setting up AWS infrastructure with S3 storage and Redshift, creating a Python script using Tweepy to extract 1 million tweets and store them securely in S3, formatting and loading the data into Redshift for sentiment analysis, and automating the entire process with Airflow DAGs to ensure orderly task execution.

Skills: S3, Redshift Datawarehouse, Python, DAGs

Twitter Data Pipeline with Airflow

• Project Learnings - By leveraging Airflow and AWS services, the project streamlines the process of data ingestion, processing, and analysis, making it easier and faster for analysts to access and analyze Twitter data.

• Description: To automate and run scheduled jobs, the project uses Airflow's DAG (Directed Acyclic Graph) concepts to orchestrate the pipeline.

Data Visualization

Below is the data analytics project developed using Tableau.

Steps: data loading, data cleaning and preprocessing, and EDA (exploratory data analysis) & building Tableau Dashboard.

Skills: Calculated fields, LOD operations

Covid 19 Analysis

• Analyzed Covid 19 data using Tableau to identify pandemic trends, geographical spread, and healthcare resource allocation.

• Description: The dataset contains records of Covid-19 cases, deaths, and vaccine records by country in 2020-2021.

Python

Below is the data analytics project developed using Python.

Skills: data cleaning, data analysis, data visualization,Prediction Analysis

Technology: Python,Jupyter Notebook,Pandas,Seaborn,Matplotlib

Data Analysis and Visualization Using Python

• Description: Developed a data visualization project using Python to analyze job application trends, providing actionable insights.

• Converted raw data from a pickle file into structured CSV format, leveraging object-oriented programming.

• Utilized pandas library for data manipulation and Matplotlib/Seaborn libraries for creating insightful visualizations, highlighting key metrics such as application volume, job types, and temporal trends

SQL

Below is the SQL project for data querying and analysis.

Skills: MySQL, MS SQL Server, Pgadmin4

Cafe Database Management System

• Description: Designed DB Schema in PostgreSQL, normalized data for storage, retrieval, and manipulation.

• Developed SQL queries on Oracle Database to provide insights for café performance and customer behavior.

Big Data

Below is the HDFS project for data querying and analysis.

Skills: Hive, Impala, HDFS, Tableau

Truck Fleet Data Analysis

• As the fleet manager of AZ National Trucking, our foremost challenge is to ensure adherence to company regulations, aiming to reduce insurance risks. This encompasses tackling issues such as speeding, unsafe following distances, lane departure incidents, and other hazardous driving practices among our fleet drivers.

• Description: The dataset contains 7 different data files such as Truck ,Truck mileage,Geolocation and more

System Analysis and Project Management

Below is the project for which UML diagrams were designed.

Skills: Figma, Jira, Confluence, Microsoft Visio

Park In – Renting Parking Space System

• Simplify urban parking with a seamless platform for discovering and booking spaces while boosting revenue for space owners.

• Park In transforms parking from a stressful daily chore into a convenient and profitable experience. Our solution not only enhances individual lives but also contributes to better urban living. We ensure full compliance with technological and regulatory standards, leveraging cutting-edge digital tools to keep our platform ahead of the competition.

Leadership & Awards

Teaching Assistant 0 Mentored Students AIMDB - Cooperate Relations

More projects on Github

I love to solve business problems & uncover hidden data stories


GitHub

Contact

Contact Me

Address

Dallas, TX, USA

Email Address

punyaianand01@gmail.com