Amulya Ramasahayam
Development
Telangana, India
Skills
Data Engineering
About
Amulya Ramasahayam's skills align with Programmers (Information and Communication Technology). Amulya also has skills associated with System Developers and Analysts (Information and Communication Technology). Amulya Ramasahayam has 11 years of work experience.
Work Experience
SQL Developer
TCS (TTSL)
December 2012 - September 2014
- In this application we were indulged in activities extraction of raw call data transactions from the client database and processing the raw data. Once the processed files are stored onto UNIX servers, they are provided to the down streams for their further usage. Responsibilities: * Implementation of Bulk upload for Wireless/GSM Subscribers. * Handling and configuring changes for new requirements and saving the money for business. * Client interaction to gather business requirements related to enhancements. * Involved in Major and Minor development projects. Environment: UNIX, JDK1.6, Oracle11g, PL/SQL, Edit Plus, TOAD, SQL Developer
Sr. Big Data Engineer
INNOVA SOLUTIONS
May 2023 - Present
- (DATA.AI) PREVIOUS ORGANIZATION I: BA CONTINUUM PREVIOUS ORGANIZATION II: COGNIZANT PREVIOUS ORGANIZATION III: TCS Client: Innova Solutions(Data.AI) Sr. Big Data Engineer May 2023 - Till date Project: Store- Intelligence Description: Data.AI delivers data analytics for various mobile apps and provides insights into their business in order to Improve app performance with the leading app store data platform. Responsibilities: * Develop and maintain pipeline infrastructure that supports ML models * Generating analytics reports pertaining to mobile app data insights * Recommended improvements in data quality and performance. * Dealt with migrating the existing pipelines to unity catalog, to implement data governance and data quality check. * Developed airflow dags to orchestrate the pipelines. * Monitoring and maintenance of pipelines in production environment * Implemented CI/CD using Git and Jenkins * Communicated with stakeholders for requirement discussions and production issue handling. * Handled stakeholder communications and issue resolutions. Recommended enhancements in data quality and performance. * Supported weekly customer deliverables. * Enhancements made to the code as per business requirement and maintain the high level of code quality. Environment: Spark - Python API, AWS, EMR, Databricks, hive, JIRA, Git, Jenkins, Unix, Airflow, Pandas
Sr. Data Engineer
INNOVA SOLUTIONS
September 2021 - April 2023
- Project: Connected Consumer Health Description: The Connected Consumer Health is delivered in a software-as-a-service model, with speedy deployment, no hardware to procure, and minimal IT resources required for customers to access and use the platform. Change Healthcare's Connected Consumer Health platform transforms the administrative tasks to more of the retail-style experience consumers have come to expect in other industries such as banking, travel, and entertainment. As a result, patients are empowered to shop, book, and pay for healthcare online as they would for any other product or service, revolutionizing a provider's ability to engage with patients and capture revenue pre-service. Responsibilities: * Develop spark pipelines to process the data as per the business requirement using spark scala * Develop lambdas to implement the required functionality at low cost by making using of aws serverless architecture. * Schedule the jobs using Airflow and dag development * Deployment of the code to various environments. * Involved in planning process of iterations under the Agile Scrum methodology. Environment: Spark - Scala API, AWS, EMR, LAMBDA, SSM(and many other components in AWS), hive, JIRA, Git, Jenkins, Unix, Airflow, Redshift, Pandas
Spark Developer
Bank of America
August 2019 - September 2021
- Project: Global Information and Security - Identity Access Management Module 1: Revocation Repository Description: The Revocation Repository is designed to be a central location to record and track revocation requests throughout the enterprise. The development of this project has been taken up to establish a central platform to record and track the global status of the revocations processing. Furthermore this helped in eliminating the inconsistent and unpredictable processing of the revocations and hence reprovisioning them. Module 2: CSDB (Centralized Security Database) Description: Centralized Security Database is the entitlement aggregator for Mainframe, Midrange and WAN and Application level entitlements. It stores data from various applications hosted on platforms that fall in the above mentioned categories. Project aims to extract the complete datasets from CSDB and build entitlement data on cloudera in business required way to provide detailed report on the entitlements. This gives us in sights on the entitlements active and helps us in identifying anomalies. Responsibilities: * Extract data from CSDB on to cloudera environment using EWS. * Processing the data stored on the HDFS using spark scala and have hive tables on top of the hdfs processed data locations. * Publish the log events of the process to down streams. * Involved in planning process of iterations under the Agile Scrum methodology. * Generated detailed design documentation for the source-to-target transformations. Environment: Spark - Scala API, Cloudera 6.3.3, hadoop, hive, sqoop, JIRA, Git, Jenkins, Bitbucket, Unix, Autosys
Spark Developer
Cognizant (Nike)
December 2017 - July 2019
- Description: Nike sells its products under its own name as well as Nike Golf, Nike Pro, Nike+, Air Jordan, and Team Starter. Nike also owns the brands Bauer, Cole Haan, Hurley International, and Converse. Nike, Incorporated also funds certain sports events such as the FIFA World Cup 2014. Responsibilities: * Hive and Spark tuning with partitioning/bucketing of Parquet and executors/driver's memory * Performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning. * Used Spark API over AWS EMR to perform analytics on data in Hive. * Generated detailed design documentation for the source-to-target transformations. * Involved in planning process of iterations under the Agile Scrum methodology. * Processed S3 data and created external tables using Hive and developed scripts to ingest and repair tables that can be reused across the project. * Handled importing of data from various data sources, performed transformations using Hive, , loaded data into S3 * Developed Spark scripts by using python commands as per the requirement. Environment: Hadoop, HDFS, Hive, Spark, Python, AWS EMR, AWS S3, AWS Athena, AWS Snowflake and PARQUET data files.
Spark Developer
TCS (Humana)
May 2015 - September 2017
- Description: The purpose of the project is the migration of the existing logic into hive scripts to suffice the need to store the huge data and to perform required analytic s over the data, which helps in enhancing the business in turn. Responsibilities: * Responsible for building scalable distributed data solutions using Spark. * Operating the cluster on AWS by using EC2, EMR, S3. * Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive. * Build Spark 1.6.1 source code over yarn for matching of production Cloudera (CDH 5.7) Hadoop 2.7 version. * Involved in running Hadoop jobs for processing millions of records of text data. * Hands on experience in AWS Cloud in various AWS services such EC2 and S3 for small data sets. Environment: Hadoop, HDFS, Hive, Spark, Python, AWS EMR, AWS S3, AWS and PARQUET data files.
Hadoop Developer
TCS (TTSL)
September 2014 - April 2015
- Description: TTSL is an Indian broadband and telecommunications service provider based in Mumbai. In this application we were indulged in activities like injection of processed files into hive and generating journals, providing monthly reports and ad-hoc reports as per the customer requirement. We also provide revenue journals to the customers on the circle basis. Responsibilities: * Created Hive tables to store the processed results in a tabular format. * Developed the Sqoop scripts in order to make the interaction Oracle Database. * Involved in gathering the requirements, designing, development and testing. * Writing the script files for processing data and loading to HDFS. * Developed the UNIX shell scripts for creating the reports from Hive data. * Setting up crone job to delete Hadoop logs/local old job files/cluster temp files. Environment: Hadoop, Hive, SQOOP, UNIX, Shell
Education