Overview
Salla, a leading e-commerce platform, is seeking a Senior Data Quality Engineer to ensure the accuracy, consistency, and reliability of our organization\'s data pipelines. As a Data Quality Engineer, you will be responsible for defining data quality standards, implementing quality frameworks, monitoring key metrics, and driving continuous improvements. This role requires both technical expertise and strong communication skills to collaborate across teams including data engineering, analytics, and business stakeholders.
Responsibilities
- Ensure availability and performance of database systems. As a DBRE you ensure applying engineering principles, operational discipline, and automation to ensure the long-term reliability and scalability of database systems
- Automating database deployments, upgrades, and routine maintenance
- Building self-service tools and platforms for developers to provision database environments
- Ensuring consistent configuration across all database instances
- Handling performance issues and maintaining database availability as environments grow
- Treating databases like code, using version control and automated pipelines
- Performance tuning and indexing. Expert in performance tuning, indexing strategies and architecture‑level improvements. Eagle eye on analyze slow queries, execution plans and system‑level performance (I / O, memory, CPU)
- Backup, disaster recovery and replication. Help design and execute robust backup and disaster recovery strategies and manage high‑availability features such as replication and failover. Automate backup scheduling and recovery procedures to meet RPO / RTO objectives
- Monitoring and observability. Proactively monitor database performance, system health and workload trends using Prometheus, Grafana and similar metric tools. Develop metrics and dashboards to identify issues before they impact customers and own database observability
- Automation and infrastructure‑as‑code. Develop and maintain automation for provisioning, patching, backups, schema changes and alerting using scripting (Python, Bash) or configuration tools. Embrace DevOps / SRE practices to minimize manual toil
- Collaborate with development and SRE teams. Work closely with developers to refine schema design, assist with query tuning, improve data architecture and share best practices
- Security and compliance. Enforce database security policies, perform audits and ensure data protection. Implement role‑based access, encryption, and compliance controls, and maintain audit logs
- Capacity planning and scaling. Participate in capacity planning, load testing and forecasting to ensure the database platform meets current and future needs, balancing optimization with resource utilization
- Incident response and root‑cause analysis. Lead incident management for database issues, conduct blameless post‑mortems, analyze recurring incidents to identify trends, and drive improvements to prevent future failures
- Documentation and mentorship. Maintain up‑to‑date documentation (runbooks, architecture diagrams, KB articles) and share lessons learned. Provide mentorship to colleagues and advocate for architectural and operational improvements
- Modern Approach : Operates within a DevOps model, working to break down database silos and empower developers with automated processes
Requirements
Manage and tune high-availability database and storage services including SQL & NoSQL databases (MySQL, PostgreSQL, MongoDB, Redis, ClickHouse) and object storage platforms (MinIO, S3-compatible stores)Plan and execute backups, disaster recovery (DR), replication, and failover strategies for both data related services including any object storage to meet RPO / RTO targetsOptimize storage performance and cost — implement multi-tier storage, cold / hot data separation, S3 offloading, bucket lifecycle policies, and data retention strategiesExpert in Kubernetes storage management — deep understanding of StorageClasses, dynamic provisioning, CSI drivers, and the ability to perform advanced PersistentVolumeClaim (PVC) troubleshooting and recovery during peak production incidents with confidence (especially for stateful sets & mission-critical volumes)Hands-on with scaling and securing object storage clusters such as MinIO (erasure coding, distributed mode, TLS, access control), and integrating them with Kubernetes workloads for high-throughput data pipelines and application storage needsExperience with block storage (EBS / io2 / gp3), shared file systems (EFS, NFS), and hybrid approaches to achieve performance, resiliency, and cost-efficiencyBenefits
Medical Health InsurancePerformance BonusOther BenefitsSeniority level
AssociateEmployment type
Full-timeJob function
OtherIndustries
IT Services and IT Consulting#J-18808-Ljbffr