Close this

Sriya Panda

Development
TX, United States
Skills
Data Engineering
About
SRIYA PANDA's skills align with System Developers and Analysts (Information and Communication Technology). SRIYA also has skills associated with Programmers (Information and Communication Technology). SRIYA PANDA has 8 years of work experience.
Work Experience

Bigdata Engineer

DreamBox Learning
October 2023 - Present
  • Project Description: Common Data Acquisition (DreamBox Learning) CDA is built for collecting data from various source systems to perform enrichment in the DataLake. It generates various KPIs to help businesses focus on product enhancements thereby driving sales revenue. Imported data from various sources into Spark data frames for processing. Involved in all the phases of SDLC including Requirements, Collection, Design & Analysis of the Customer Specifications from Business Analyst. Used Jenkins, Git Stash, Ansible like CI/CD tools to make daily builds and deploys. Used JSON for validation, documentation data required by a given application. Used AWS Infrastructure and features of AWS like S3, EC2, Glue, Lambda, CloudWatch. Worked on reading multiple data formats on HDFS using python. Implemented Spark using Python (PySpark) and Spark SQL for faster testing and processing of data. Load the data into Spark RDD and do in memory data Computation. Involved in converting Hive/SQL queries into Spark transformations using APIs like Spark SQL, Data Frames, and python. Worked with data in multiple file formats including Avro, Parquet, Sequence files, ORC and Text/ CSV.

Bigdata Engineer

February 2021 - February 2022
  • Project Description: DataLake implementation for Byju's Suite of products (Byju's) Developed various data loading strategies and performed various transformations for analyzing the datasets by using Hortonworks Distribution for Hadoop ecosystem. Experience in Developing Spark applications using Spark - SQL in Databricks for data extraction, transformation, and aggregation from multiple file formats for analyzing and transforming the data to uncover insights into the customer usage patterns Worked in Loading and transforming large sets of structured, semi structured, and unstructured data. Involved in collecting, aggregating, and moving data from servers to HDFS using Flume. Collecting data from various Flume agents that are imported on various servers using Multi-hop Flow. Knowledge on various flume sources, channels and sink by which data is ingested into HDFS the data onto HDFS.

Data Engineer

GlobalLogic India
April 2018 - March 2020
  • Project Description: Operation Processing Engine (GlobalLogic India) Engaged in OPE transformations built in Spark runs on the transactions that come in through various source systems and performs derivation/lookup for the various key attributes across domains. Part of an 8-member team to build various modules like: Customer Identification, Partner Identification, Sales Compensation in Spark. Built Reprocessing framework to reprocess a configurable Window period of data from source. Built a custom Spark class to read the data from the raw layer, process it and finally load it into the ref layer. Performance tuning of large data sets and thereby creating Partitions/Buckets to handle Terabytes of data. Analysis of the Spark summit jobs through Spark UI and take corrective actions accordingly in case of failures.

Sr Software Engineer

December 2016 - April 2018
  • Project Description: KAI - Bigdata Modernization (L&T Technology Services - Huawei) Played a key role in project planning, coordination, and implementation, encompassing coding based on design documents, conducting Unit Testing, and supporting the automation of system testing. Developed applications using Servlets, JSP, and Struts-based frameworks. Tuning large data sets on HIVE. Creating Hive Partitions in Hive to handle the incremental data from the MDM systems. Performed data enrichments using Spark Data frame API's. Analyzed Spark Application logs to debug and fix development issues. Implemented Performance tuning techniques in developing PySpark jobs.

Software Engineer

Pearson India Education Pvt Ltd
December 2015 - December 2016
  • Project Description: Pearson Datamart uses Talend for Big Data and Hive. (Pearson India Education Pvt Ltd) Pearson Datamart was created to redesign the Enterprise data lake using big data components. Responsible for building a framework that ingests data into Hadoop from a variety of data sources providing high storage efficiency. Tuning large data sets on HIVE. Creating Hive Partitions in Hive to handle the incremental data from the MDM systems. Creating the ETL framework using Talend for Big data to load data to the target.

Software Engineer

Pearson India Education Pvt Ltd
January 2014 - December 2015
  • Project Description: www.edurite.com (Pearson India Education Pvt Ltd) Edurite is a part of Pearson, the world's leading learning company. Pearson is an international media company with world leading businesses in education, business information and consumer publishing. Pearson programs help educate more than 100 million people worldwide. Developed and maintained complex user interface for web applications like www.edurite.com and eduriteonline.com using html5, css3, PHP, MySQL, jQuery, Ajax, Bootstrap etc. Integrated user interface designs with back-end services and databases, resulting in a 30% reduction in page load times. Collaborated with cross-functional teams to implement accessibility standards and guidelines, resulting in a vast increase in accessibility compliance and a more inclusive user experience. Internal
Education

Bachelor's in engineering