DataOps / DevOps Engineer (Python | AWS | Snowflake)
Experience Range: 4 - 8 years
Employment Type: Full-time (On-site)
Location:
Vadodara, Gujarat
About the Role
We are looking for a DataOps/DevOps Engineer to streamline and automate data platform operations across AWS and Snowflake environments. The ideal candidate will have strong experience in Python scripting, CI/CD automation, infrastructure-as-code, and data pipeline reliability. You’ll collaborate with data engineers and analysts to ensure data products are deployed, monitored, and scaled efficiently.
Key Responsibilities
- DataOps Automation
- Build and maintain CI/CD pipelines for data ingestion, transformation, and analytics workloads
- Automate ETL/ELT job deployments using tools like GitHub Actions, Jenkins, or AWS CodePipeline
- Develop Python utilities for orchestration, metadata tracking, and operational workflows
- Implement data quality and validation checks as part of deployment workflows
- DevOps & Cloud Infrastructure
- Manage and optimize AWS infrastructure for data workloads — EC2, Lambda, S3, Step Functions, Glue, ECS/EKS, and IAM
- Design and manage infrastructure-as-code (IaC) using Terraform or AWS CloudFormation
- Set up monitoring, alerting, and logging via CloudWatch, Prometheus, or Datadog
- Ensure high availability, backup, and DR strategies for Snowflake and data pipelines
- Snowflake & Data Platform
- Automate Snowflake warehouse management, user provisioning, and role-based access using Python or Terraform
- Manage Snowpipe, Streams, and Tasks for continuous data ingestion
- Collaborate with data engineers on performance tuning and query optimizatio.
- Build integrations between Snowflake, AWS (S3, Lambda, Kinesis), and BI tools (e.g., Superset, Tableau)
- Collaboration & Best Practices
- Work closely with data engineering and analytics teams to improve delivery efficiency and reliability
- Define and enforce data versioning, testing, and release standards
- Contribute to observability and incident response playbooks
- Promote a culture of automation, testing, and continuous improvement
Required Skills
- Programming: Python (must have); Bash scripting
- Cloud: AWS (S3, Lambda, Glue, ECS, IAM, CloudFormation)
- Data Platform: Snowflake (SQL, Snowpipe, Streams, Tasks, RBAC)
- DevOps: CI/CD pipelines (GitHub Actions, Jenkins, or AWS CodePipeline)
- IaC: Terraform or AWS CloudFormation
- Version Control: Git, GitHub, GitLab
- Monitoring: CloudWatch, Grafana, or Datadog
- OS: Linux / Unix
Nice to Have
- Experience with Airflow, dbt, or Flink/Spark for pipeline orchestration
- Exposure to Kubernetes (EKS) for scalable data workloads
- Knowledge of Cost Optimization for Snowflake and AWS resources
- Familiarity with Data Governance and security frameworks (e.g., AWS KMS, Secrets Manager)
Working with Us
The Commerce Team Global (TCTG) is a leading eCommerce specialist, developing solutions on Salesforce Commerce Cloud (SFCC), BigCommerce, Shopify, and other platforms to drive client success.
We are a team of 60+ tech enthusiasts, expanding our global presence with happy customers worldwide. Our depth of experience and expertise across diverse development platforms has made us a partner of choice for many leading retail clients.
Send your CV to HR@thecommerceteam.com
