Job Description
At FreePL, you will be an integral part of a small team that is currently on a mission to build something big from the ground up. You will need to be comfortable with ambiguity and fast pace, have an entrepreneurial mindset, a can-do attitude and will be expected to take ownership of your work and be comfortable making decisions and solving problems independently.
As part of a quickly growing team, you will have the opportunity to shape the future of FreePL and make a real impact. You will work closely with the founding team to develop and implement strategies for growth and success. If you are self-motivated, results-driven, and have a passion for delivering excellence, this is your chance to not only join us in our launch phase but to grow into a leader at FreePL.
What We Look For
We are looking for a talented Data Engineer to join our team and contribute to building efficient, scalable, and reliable data infrastructure to drive our data-driven decision-making processes.
Responsabilities
- Design, build, and maintain efficient and scalable data pipelines using Python and other modern tools.
- Collaborate with cross-functional teams to understand data requirements and deliver robust solutions.
- Develop and optimize ETL/ELT processes to extract, transform, and load data from multiple sources.
- Manage, monitor, and improve database systems to ensure high performance and availability.
- Implement and maintain data models and architecture to support analytics and machine learning workflows.
- Ensure data security, privacy, and compliance with relevant regulations.
- Perform root cause analysis for data-related issues and propose solutions.
- Stay updated on the latest technologies, frameworks, and industry
Qualifications & Work Experience
- Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
- 3+ years of experience as a Data Engineer or in a similar role.
- Proficiency in Python with a strong understanding of libraries like Pandas, NumPy, and PySpark.
- Hands-on experience with relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., MongoDB, Cassandra).
- Strong knowledge of data integration tools and frameworks (e.g., Apache Airflow, Apache Kafka, etc.).
- Experience with Azure cloud platforms and managing cloud-based data pipelines.
- Familiarity with big data technologies such as Hadoop, Spark, or Snowflake is a plus.
- Solid understanding of data warehousing concepts and tools (e.g., Redshift, BigQuery).
- Knowledge of containerization (e.g., Docker) and orchestration tools (e.g., Kubernetes) is an advantage.
- Excellent problem-solving and communication skills.
- Experience in the logistics or supply chain domain.
- Familiarity with real-time data streaming and processing.
- Knowledge of data governance, quality frameworks, and best practices.