Job Title

DataOps / DevOps Engineer (Python | AWS | Snowflake)

Experience Range: 4 - 8 years

Employment Type: Full-time (On-site)

Location:

Vadodara, Gujarat

About the Role


We are looking for a DataOps/DevOps Engineer to streamline and automate data platform operations across AWS and Snowflake environments. The ideal candidate will have strong experience in Python scripting, CI/CD automation, infrastructure-as-code, and data pipeline reliability. You’ll collaborate with data engineers and analysts to ensure data products are deployed, monitored, and scaled efficiently.

Key Responsibilities

  • DataOps Automation
  • Build and maintain CI/CD pipelines for data ingestion, transformation, and analytics workloads
  • Automate ETL/ELT job deployments using tools like GitHub Actions, Jenkins, or AWS CodePipeline
  • Develop Python utilities for orchestration, metadata tracking, and operational workflows
  • Implement data quality and validation checks as part of deployment workflows
  • DevOps & Cloud Infrastructure
  • Manage and optimize AWS infrastructure for data workloads — EC2, Lambda, S3, Step Functions, Glue, ECS/EKS, and IAM
  • Design and manage infrastructure-as-code (IaC) using Terraform or AWS CloudFormation
  • Set up monitoring, alerting, and logging via CloudWatch, Prometheus, or Datadog
  • Ensure high availability, backup, and DR strategies for Snowflake and data pipelines
  • Snowflake & Data Platform
  • Automate Snowflake warehouse management, user provisioning, and role-based access using Python or Terraform
  • Manage Snowpipe, Streams, and Tasks for continuous data ingestion
  • Collaborate with data engineers on performance tuning and query optimizatio.
  • Build integrations between Snowflake, AWS (S3, Lambda, Kinesis), and BI tools (e.g., Superset, Tableau)
  • Collaboration & Best Practices
  • Work closely with data engineering and analytics teams to improve delivery efficiency and reliability
  • Define and enforce data versioning, testing, and release standards
  • Contribute to observability and incident response playbooks
  • Promote a culture of automation, testing, and continuous improvement


Required Skills

  • Programming: Python (must have); Bash scripting
  • Cloud: AWS (S3, Lambda, Glue, ECS, IAM, CloudFormation)
  • Data Platform: Snowflake (SQL, Snowpipe, Streams, Tasks, RBAC)
  • DevOps: CI/CD pipelines (GitHub Actions, Jenkins, or AWS CodePipeline)
  • IaC: Terraform or AWS CloudFormation
  • Version Control: Git, GitHub, GitLab
  • Monitoring: CloudWatch, Grafana, or Datadog
  • OS: Linux / Unix


Nice to Have

  • Experience with Airflow, dbt, or Flink/Spark for pipeline orchestration
  • Exposure to Kubernetes (EKS) for scalable data workloads
  • Knowledge of Cost Optimization for Snowflake and AWS resources
  • Familiarity with Data Governance and security frameworks (e.g., AWS KMS, Secrets Manager)

Working with Us

The Commerce Team Global (TCTG) is a leading eCommerce specialist, developing solutions on Salesforce Commerce Cloud (SFCC), BigCommerce, Shopify, and other platforms to drive client success.

We are a team of 60+ tech enthusiasts, expanding our global presence with happy customers worldwide. Our depth of experience and expertise across diverse development platforms has made us a partner of choice for many leading retail clients.

Send your CV to HR@thecommerceteam.com