https://bayt.page.link/FPUX6gEavgQuHpbU8
Create a job alert for similar positions

Job Description

Build the future of data. Join the Snowflake team.



We are seeking a skilled Data Platform Engineer to join our dynamic team at Snowflake. As a Data Platform Engineer, you will play a critical role in developing and maintaining our data platform used by data teams across the organization. The ideal candidate will have 6-8 years of experience in data engineering and a strong proficiency in SQL, Python, Snowflake, Airflow, Docker & Kubernetes. 


Key Responsibilities:


  • Design and Develop Data Platform Tools: Collaborate with cross-functional teams to design, develop, and maintain data platform tools and infrastructure components using cutting-edge technologies such as Kubernetes, Docker, and Airflow.
  • Optimize Data Pipelines: Implement efficient data pipelines to ingest, transform, and load data into Snowflake using SQL and Python, ensuring high performance and reliability.
  • Monitor and Maintain Data Platform: Monitor the health and performance of data platform tools and infrastructure, proactively identifying and resolving issues to ensure maximum uptime and reliability.
  • Automate Deployment Processes: Implement automation scripts and tools to streamline the deployment and configuration of data platform components using Kubernetes and Docker.
  • Collaborate with Data Teams: Work closely with data engineering, data science, and analytics teams to understand their requirements and provide support in leveraging data platform tools effectively.
  • Ensure Security and Compliance: Implement security best practices and ensure compliance with data governance policies and regulations across all data platform components.
  • Documentation and Knowledge Sharing: Document technical designs, procedures, and best practices, and actively participate in knowledge sharing sessions to facilitate cross-team collaboration and learning.
  • Product Mindset: A strong focus on delivering high-quality products that meet the needs of internal stakeholders and customers.
  • Fast-Paced Environment: Ability to thrive in a fast-paced, dynamic environment.

Qualifications:


  • Bachelor's degree in Computer Science, Engineering, or a related field.
  • 6-8 years of experience in data engineering or related roles.
  • Strong proficiency in SQL and Python for data manipulation and scripting.
  • Experience with containerization technologies such as Docker and container orchestration platforms like Kubernetes.
  • Experience with workflow management tools like Airflow for orchestrating complex data pipelines.
  • Hands-on experience with Snowflake or other cloud data warehouse platforms is highly desirable.
  • Strong problem-solving skills and the ability to troubleshoot complex issues in a distributed, cloud-based environment.
  • Excellent communication and collaboration skills, with the ability to work effectively in a cross-functional team environment.

Job Details

Job Location
Pune India
Company Industry
Other Business Support Services
Company Type
Unspecified
Employment Type
Unspecified
Monthly Salary Range
Unspecified
Number of Vacancies
Unspecified

Do you need help in adding the right mix of strong keywords to your CV?

Let our experts design a Professional CV for you.

You have reached your limit of 15 Job Alerts. To create a new Job Alert, delete one of your existing Job Alerts first.
Similar jobs alert created successfully. You can manage alerts in settings.
Similar jobs alert disabled successfully. You can manage alerts in settings.