Job Duties: Data Analysis & Requirements Interpretation
- Work with product owners and business stakeholders to understand analytical requirements, data dependencies, and expected application capabilities
- Translate functional needs into technical specifications for data processing, validation, and reporting
- Participate in Agile ceremonies and contribute to planning, estimation, and prioritization of data-related tasks
Data Processing and ETL Development
- Develop and optimize Snowflake JavaScript-based ETL procedures to process structured and semi-structured JSON data, applying performance-tuning techniques to support large-scale datasets
- Implement incremental, truncated, and soft-delete load strategies (SCD-I) to support accurate and efficient data refresh cycles
- Build snowflake tasks with scheduled execution to monitor and validate dataloads, ensuring timely and accurate data availability
Data Engineering Support for Analytics
- Design and implement ETL pipelines for ingesting upstream data into the Hadoop Data Lake, maintaining fact and dimension structures used for analytical reporting
- Create and optimize Hive queries to process and store large datasets in internal and external tables, including partitioning and bucketing for efficient access
- Translate existing HiveQL logic into Spark to enable parallel processing and improved performance for analytical workloads
- Develop CDC logic using Spark RDDs, Dataframes, and Hive to support SCD types 1, 2, and 3 for downstream analytical consumption
Data Modeling and Quality Assurance
- Design Snowflake data models using star and snowflake schemas aligned with analytical and reporting needs
- Establish testing environments using Snowflake Clone and leverage Time Travel features to validate data accuracy and support audit requirements
- Define and implement Snowflake roles and privileges to ensure secure and compliant access to analytical datasets
Cross-functional collaboration and SDLC participation
- Collaborate with full-stack development teams to integrate data solutions into broader application workflows
- Participate in all phases of the SDLC, including analysis, design, development, testing, CI/CD, release and post-production support
- Apply appropriate software engineering and data management patterns to build scalable, reliable analytical systems
Operational Support
- Support production deployments and data pipeline operations across Hadoop and Snowflake environments
- Monitor data workflows, troubleshoot issues, and assist in escalation support to maintain high availability of analytical datasets
Build partnerships with other technology teams to improve efficiency in the use of shared data platforms and services.
Location: Anywhere in USA
Please submit your resume and cover letter to: hr@praisetechsol.com
Job Category: Software Engineer
Job Type: Full Time
Job Location: USA