Cloud Platform Engineer

  • Remote, On-site, Hybrid
  • Prague + 1 more
  • Part-time, Full-time

Are you passionate about building robust, scalable cloud infrastructure that powers the next generation of data-driven applications? 🤖 Do you thrive on automating everything through code and ensuring seamless delivery pipelines? 

We are looking for a Cloud Platform Engineer who bridges the gap between infrastructure and data engineering. If you have an ownership mindset, love solving complex platform puzzles, and want to work in a high-impact, cloud-native environment, we’d love to meet you.


⚒️ What You’ll Do

As a Cloud Platform Engineer at DataSentics, you will be the backbone of our technical delivery, ensuring our data and AI solutions run on world-class infrastructure. You will work closely with data engineers and scientists to build stable, secure, and automated environments. Your core responsibilities will include:

  • Architecting the Foundation: Design and implement cloud infrastructure (primarily AWS/Azure) using Infrastructure as Code (Terraform).

  • Automating Delivery: Build and maintain robust CI/CD pipelines (GitLab/GitHub) to ensure frequent and reliable deployments.

  • Platform Enablement: Own and solve platform-related topics around Databricks, including Unity Catalog integration, permissions, and service principal management.

  • Containerization: Manage and orchestrate services using Docker and Kubernetes (or Azure Container Apps) to host internal data applications.

  • Data Ops & Security: Set up monitoring, logging, and alerting systems while ensuring top-tier security through managed identities, secrets management, and access control.


⭐️ Why Join Us?

At DataSentics, we are a team of tech enthusiasts who believe that great data science is only possible with great engineering. We combine deep technical expertise with a practical, product-driven approach. You’ll be part of a culture that values technical excellence, continuous learning, and the freedom to suggest and implement the best tools for the job.


🫆 Our Projects & Technology Stack

We build enterprise-grade data platforms and internal applications across industries like finance, retail, and manufacturing.

  • Cloud Ecosystems: Deep dives into AWS and Azure.

  • Data Orchestration: Advanced environments involving Databricks job clusters, complex pipelines, and automated permissions.

  • Modern DevOps: Heavy use of Terraform, Docker, and Git-based automation.

  • How we work: Agile teams (2–5 people), remote-first culture, and a focus on "building it right" the first time.

  • Flexibility: Work from anywhere with optional offices in Prague and Brno; flexible roles starting at 30 hours/week.


🧩 What We’re Looking For

Must-haves:

  • Cloud Proficiency: Experience with cloud platforms, ideally AWS/Azure.

  • Automation Expert: Proven track record with CI/CD (GitLab/GitHub pipelines) and Terraform.

  • Container Skills: Hands-on experience with Docker (Kubernetes or Azure Container Apps is a plus).

  • Platform Troubleshooting: Ability to solve complex integrations involving Databricks and data applications.

  • Mindset: A strong sense of ownership, independence, and the "investigative" skills needed to fix broken systems.

Nice-to-haves:

  • Familiarity with Databricks administration (Unity Catalog, permissions, service principals).

  • Basic to intermediate Python skills for automation and scripting.

  • Understanding of Data Engineering lifecycles (pipelines, orchestration, storage).

  • Experience with cloud security (Secrets, IAM) and monitoring/alerting tools.


🎁 What You’ll Get

  • The opportunity to shape the infrastructure of cutting-edge AI and Data products.

  • A collaborative environment focused on innovation and technical excellence.

  • Flexibility, remote work options, and a supportive, curious team culture.


Curious to know more? If you're excited about infrastructure as code and want to empower data teams to build amazing things, let’s talk!