Snowflake Developer

Overview

Remote
Depends on Experience
Accepts corp to corp applications
Contract - Independent
Contract - W2
Contract - 8 Month(s)
No Travel Required

Skills

Snowflake Data Warehouse Developer
Oracle
PostgreSQL
HA
cloning
RBAC
SSO
SCIM
Informatica
ADF
ETL
ELT
Python
Shell scripting
Erwin
Visio
MS Project
agile
JIRA
Cognos
CI
CD
PII
PHI
HL7
Azure
Version control
Workflow
Scalability
Snow flake schema
Sprint
Microsoft Visio
Network
Microsoft Excel
Microsoft Exchange
Microsoft Office
Microsoft PowerPoint
IBM Cognos
Good Clinical Practice
Google Cloud Platform
Health care
Data structure
Data loading
Data masking
Data modeling
Data security
Access control
Amazon Web Services
Analytics
Streaming
Technical writing
Microsoft Azure
Modeling
Optimization
Real-time
SQL
Scripting
Extract
transform
load
High availability
Identity management
Mapping
Data integration
Blueprint
Cloud computing
Continuous delivery
Continuous integration
Data governance
Data storage
Data warehouse
Database
Encryption
Management

Job Details

Job ID: TX-529501270

Remote/Local Snowflake Data Warehouse Developer with Oracle, PostgreSQL, HA, cloning, RBAC, SSO, SCIM, Informatica, ADF, ETL/ELT, Python/Shell scripting, Erwin/Visio, MS Project, agile, JIRA, Cognos, CI/CD, PII/PHI/HL7, Azure experience

Location: Austin, TX (HHSC)
Duration: 8 Months
Position: 1
Any location within the State of Texas

Skills:
8 Required Experience with data modeling, data integration, data warehousing, data governance, and data security
8 Required Experience with Oracle and/or PostgreSQL in HA deployments and Expertise in data storage
8 Required Proficiency in Snowflake architecture and its components.
8 Required Hands-on experience with Snowflake objects such as Databases, Procedures, Tasks, and Streams.
8 Required Expertise in using Snowflake s cloning capabilities for databases and schemas.
8 Required Proven experience in managing Snowflake Warehouses and optimizing performance for efficient query execution.
8 Required Proficiency in Snowflake RBAC (Role-Based Access Control), including implementation of roles and privileges.
8 Required Experience with integrating Snowflake SSO (Single Sign-On) and SCIM (System for Cross-domain Identity Management) for secure access and identity management.
8 Required Experience working with data integration tools like Informatica and ADF for seamless ETL/ELT processes.
8 Required Ability to automate administrative tasks using Snowflake SQL and scripting languages like Python or Shell scripting.
8 Required Expertise in monitoring and troubleshooting Snowflake environments, including usage tracking and query profiling.
8 Required Strong understanding of Snowflake s security features such as data masking, encryption, and network policies.
8 Required Technical writing and diagramming skills, including proficiency with modeling and mapping tools (e.g., Visio, Erwin), and the Microsoft Office Suite (Word, Excel, and PowerPoint) and MS Project.
8 Required Experience on an agile sprint team
8 Required Experience with JIRA software
8 Required Experience working with multiple teams concurrently, being able to prioritize and complete work on time with high quality
8 Required Knowledge of Informatica 10.5
8 Required Developing reports in Cognos Analytics 11.1
5 Preferred Familiarity with CI/CD pipelines and version control for managing Snowflake code deployments.
5 Preferred Prior experience in the Healthcare Industry
5 Preferred Prior experience with an HHS agency
5 Preferred Prior experience working with PII or PHI data
5 Preferred Prior experience working with HL7 data
5 Preferred Prior experience with Azure
4 Preferred Bachelor s degree in computer science, Information Systems, or Business or equivalent experience.

Description:
The Department of Information Resources (DIR) requires the services of One Developer hereafter referred to as Worker, who meets the general qualification of Developer 3 Emerging and the specifications outlined in this document for Health & Human Services Commission Social Services Applications team.

All work products resulting from the project shall be considered works made for hire and are the property of the HHSC. HHSC may include pre-selection requirements that potential Vendors (and their Workers) submit to and satisfy criminal background checks as authorized by the Texas law. HHSC will pay no fees for interviews or discussions, which occur during the process of selecting a Worker(s).

Design overall data structure, ensuring that Snowflake s features (e.g., data sharing, scalability, secure data exchange, etc.) are fully utilized to meet the business requirements.
Create a blueprint for how data will be stored, processed, and accessed within the Snowflake platform.
Perform optimization of data pipelines and workflows for performance, scalability, and cost-efficiency.
Design ETL (Extract, Transform, Load) or ELT (Extract, Load, Transform) processes, and optimize queries and data storage strategies.
Integrate with other cloud services (e.g., AWS, Azure, Google Cloud Platform), third-party tools, and on-premises data systems.
Designs and implements strategies to control access to sensitive data, applying encryption, role-based access control, and data masking as necessary.
Work closely with data engineers, data scientists, business analysts, and other stakeholders to understand their requirements and ensure the Snowflake environment meets those needs.
Monitor the performance of the Snowflake environment, identifying bottlenecks, and ensuring optimal query performance.
Automate administrative tasks using Snowflake SQL and scripting languages like Python or Shell scripting.
Preform data loading methods (bulk loading using COPY INTO, Snowpipe for real-time ingestion, and External tables).
Perform Snowflake cloning capabilities for databases and schemas.
Perform configuration and management of Snowflake Virtual Warehouses including scaling, resizing and auto-suspend/resume settings.
Implement roles and privileges for managing secure access utilizing Snowflake RBAC (Role-Based Access Control)
Integrate Snowflake SSO (Single Sign-On) and SCIM (System for Cross-domain Identity Management) for secure access and identity management.
Configure alerts and monitor data pipeline failures, resource spikes, and cost thresholds.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.