About the Role
We are looking for a highly skilled DataOps Engineer to join our Data Platform team.
This role is responsible for ensuring the stability, reliability, automation, performance, and monitoring of our fully on-premise data infrastructure—including Kafka clusters, NiFi pipelines, Elasticsearch, SQL Server, MySQL, PostgreSQL, MongoDB, ClickHouse, Redis, Vector Databases, and Graph Databases.
Key Responsibilities
- Operate, monitor, and optimize real-time and batch data infrastructures (Kafka, NiFi, Flink, SQL Server, MySQL, PostgreSQL, MongoDB, ClickHouse).
- Administer and tune SQL Server (AlwaysOn AGs, performance tuning, backups, maintenance).
- Administer and tune MySQL servers: replication, performance tuning, indexing, backups, schema migrations.
- Manage Kafka clusters including replication, partitions, retention, consumer lag, throughput optimization.
- Operate and optimize Apache NiFi clusters, access control, performance tuning, and flow stability.
- Manage and maintain the Elastic Stack (Elasticsearch, Logstash, Kibana, Beats) for logs and observability.
- Automate infrastructure provisioning and configuration using Ansible (playbooks, roles, orchestration).
- Deploy and manage workloads in Kubernetes (resource limits, ingress, deployments, autoscaling).
- Implement CI/CD pipelines for data system deployments and configuration updates.
- Provide operational support for SQL Server, MySQL, PostgreSQL, MongoDB, Redis, and ClickHouse.
- Support new storage engines including Vector Databases (FAISS, Milvus, Vespa) and Graph Databases (Neo4j, JanusGraph).
- Build dashboards and alerting mechanisms in Elastic Stack for system reliability and anomaly detection.
- Troubleshoot real-time issues: network, storage, cluster imbalance, latency, CPU/memory bottlenecks.
- Automate operational tasks via Bash and Python scripts.
- Collaborate closely with Data Engineering, Backend, BI, and Product teams for seamless data delivery.
Required Skills & Qualifications
- 2–5+ years of experience in DataOps roles supporting distributed systems.
- Strong background in Linux administration, networking, and troubleshooting.
- Hands-on experience with SQL Server operations (AG, tuning, backups) and MySQL administration (replication, tuning).
- Strong understanding of Kafka internals (brokers, partitions, ISR, retention, quotas, consumer lag).
- Experience with Apache NiFi clustering and performance configuration.
- Experience with Elastic Stack for log analytics and monitoring.
- Proficiency in Ansible for automation and configuration management.
- Experience deploying and supporting services on Kubernetes.
- Familiarity with PostgreSQL, MongoDB, Redis, and ClickHouse operations.
- Understanding or willingness to learn Vector DBs and Graph DBs.
- Strong scripting skills using Bash and Python.
- Experience operating in on-premise environments (no cloud).
Nice to Have
- Kafka Connect, Schema Registry, MirrorMaker2 experience.
- Experience with MySQL replication (ROW/BINLOG), MySQL tuning, or Percona tools.
- Experience with HAProxy, Nginx, or load balancers in front of databases or Kafka.
- Knowledge of DR, backup/restore automation, and high-availability patterns.
Soft Skills
- Strong ownership and accountability for systems.
- Calm and structured problem-solving under incident pressure.
- Excellent communication and teamwork.
- Proactive mindset toward automation and reliability.
- High attention to detail and analytical thinking.
What We Offer
- The chance to work with mission-critical, large-scale, real-time data systems.
- A technically strong, collaborative Data Platform team.
- Competitive salary and benefits.
- Career growth path toward Senior DataOps or Platform Engineering.
- Exposure to advanced systems: Kafka, NiFi, Elasticsearch, SQL Server, MySQL, ClickHouse, Redis, Vector DBs, Graph DBs.