Data Engineer

Job Details

Industry Information Technology and Services Location: Remote
Function Data Analyst Employment Type: Full-time
Work Exp. Reqd: Mid-senior level Education Reqd: Master's Degree

Role Overview

We are seeking an exceptional Data Engineer to design, develop, and optimise intelligent, large-scale data systems that power advanced analytics, AI models, and cybersecurity intelligence within vQsystems’ innovation ecosystem.

This role demands a deep understanding of distributed data architectures, high-performance data pipelines, and scalable data storage strategies. You will work with cutting-edge technologies in a multi-cloud environment, collaborating with AI/ML, backend, and platform teams to deliver a real-time, insight-driven data infrastructure that fuels innovation across our enterprise.

If you are passionate about solving complex data problems, building resilient data architectures, and enabling AI-powered decision systems, this role offers a unique opportunity to shape the foundation of next-generation intelligent platforms.

Responsibilities

  • Design, develop, and maintain scalable, real-time and batch data pipelines using tools like Apache Spark, Kafka, and Airflow.
  • Build and optimise data lake and warehouse architectures on AWS, Azure, or GCP (Redshift, BigQuery, Snowflake, or Synapse).
  • Implement ETL/ELT workflows that ensure high-quality, consistent, and secure data ingestion from multiple structured and unstructured sources.
  • Collaborate with AI/ML engineers to design data pipelines optimised for machine learning models and continuous training.
  • Develop and enforce data governance, lineage, and quality frameworks for enterprise-grade compliance and traceability.
  • Implement monitoring, observability, and automation for all data flows to ensure reliability and minimal downtime.
  • Work closely with software engineers and product teams to integrate real-time analytics and predictive insights into production systems.
  • Continuously evaluate and integrate emerging data technologies to improve scalability, performance, and automation.

Requirements

  • 4+ years of professional experience as a Data Engineer, preferably in complex, data-intensive environments.
  • Strong proficiency in Python and SQL for data manipulation, transformation, and automation.
  • Hands-on experience with big data technologies such as Apache Spark, Kafka, Hadoop, and Airflow.
  • Proven expertise with cloud data platforms (AWS Glue, Redshift, GCP BigQuery, or Azure Synapse).
  • Deep understanding of data modelling, warehousing, and lakehouse architectures.
  • Experience in ETL/ELT design, data partitioning, and performance tuning.
  • Familiarity with containerised and microservice-based architectures (Docker, Kubernetes).
  • Exposure to AI/ML data workflows, including feature stores and model-serving pipelines, is a plus.
  • Strong knowledge of data governance, compliance (GDPR), and security best practices.
  • Excellent problem-solving skills, attention to detail, and ability to work collaboratively in a fast-paced, innovation-driven environment.


Benefits

  • Competitive compensation and equity package.
  • Hybrid or fully remote work flexibility.
  • Opportunity to work on AI-integrated data infrastructure projects at enterprise scale.
  • Access to modern data technologies and high-performance computing environments.
  • Continuous professional learning.
  • Inclusive, forward-thinking engineering culture that rewards innovation, precision, and long-term impact.