Overview
Remote
Depends on Experience
Contract - W2
Contract - Independent
Contract - 6 Month(s)
No Travel Required
Skills
Dataiku
Python
AI/ML
Data Engineering
Job Details
Dataiku Developer
Location: Remote
Employment Type: Contract
Job Summary:
We are seeking a skilled Dataiku Developer to design, develop, and optimize data workflows and machine learning pipelines using Dataiku DSS. The ideal candidate will have a strong background in data engineering, analytics, and automation, along with experience in integrating Dataiku with cloud platforms and big data technologies.
Key Responsibilities:
- Develop, configure, and optimize Dataiku DSS workflows to streamline data processing and machine learning operations.
- Design and implement ETL pipelines and data transformations using Python, SQL, and Dataiku's visual recipes.
- Integrate Dataiku DSS with cloud platforms (AWS, Azure, Google Cloud Platform) and big data technologies such as Snowflake, Hadoop, and Spark.
- Implement and optimize machine learning models within Dataiku for predictive analytics and AI-driven solutions.
- Automate data workflows, monitor job performance, and ensure scalable execution.
- Develop and maintain custom Python/R scripts within Dataiku to enhance analytics capabilities.
- Collaborate with data scientists, engineers, and business teams to understand requirements and provide actionable insights.
- Ensure compliance with data governance, security, and best practices within the Dataiku environment.
- Troubleshoot and resolve performance bottlenecks in data pipelines.
- Train and support users on Dataiku DSS best practices and functionalities.
Required Qualifications:
- 3-5+ years of experience in data engineering, analytics, or machine learning development.
- Hands-on experience with Dataiku DSS, including visual recipes, automation nodes, and API integration.
- Strong proficiency in Python, SQL, and scripting languages for data processing.
- Experience working with relational and NoSQL databases (e.g., PostgreSQL, Snowflake, MongoDB).
- Knowledge of cloud computing platforms (AWS, Azure, Google Cloud Platform) and big data tools (Spark, Hadoop).
- Familiarity with CI/CD practices for data pipeline deployment.
- Strong problem-solving skills and the ability to optimize data workflows for performance and scalability.
- Experience in dashboarding and visualization tools (e.g., Tableau, Power BI) is a plus.
Preferred Qualifications:
- Experience with AI/ML model deployment and monitoring in Dataiku.
- Knowledge of MLOps and best practices in machine learning lifecycle management.
- Exposure to APIs and web services for data integration.
- Familiarity with Git, Docker, Kubernetes for model deployment.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.