Published: Fri, 03 Jan 2025 16:50:05 GMT
Data Engineer – Equity-Only Role
Stealth-Mode Start-Up Client is seeking experienced Data Engineers to join our team. As a Data Engineer, you will be responsible for designing, building, and maintaining scalable data pipelines and infrastructure to support advanced analytics, AI, and machine learning initiatives. This is a unique opportunity to be part of a fast-growing start-up and play a critical role in shaping our data strategy.
Key Responsibilities:
– Design and implement ETL/ELT data pipelines to collect, process, and store large datasets from diverse sources.
– Develop and maintain a robust data architecture to support real-time and batch data processing workflows.
– Integrate data from multiple platforms, APIs, and third-party services into centralized data lakes and warehouses.
– Optimize and manage cloud-based data infrastructure (e.g., AWS Redshift, GCP BigQuery, Azure Synapse).
– Implement data validation and integrity checks to ensure accuracy, completeness, and consistency.
– Build and manage real-time data streaming pipelines using technologies like Apache Kafka, Spark Streaming, or similar.
– Collaborate with Data Scientists, Analysts, and Product Managers to understand data requirements and ensure data availability for analytical models.
– Continuously optimize data storage and retrieval processes for performance and cost efficiency.
– Ensure data security and compliance with regulatory standards (e.g., GDPR, CCPA).
– Set up monitoring systems and logging mechanisms to track data flow, detect anomalies, and resolve pipeline failures.
– Maintain detailed technical documentation for all data pipelines, workflows, and infrastructure components.
Requirements:
– Minimum 3+ years of experience as a Data Engineer or in a similar role.
– Excellent English language skills.
– Previous start-up experience is a plus.
– Proficiency with ETL/ELT tools (e.g., Apache Airflow, Talend, dbt).
– Strong experience in data processing and automation using languages like Python, Java, or Scala.
– Experience with big data technologies like Hadoop, Spark, and Kafka.
– Proficiency with SQL and NoSQL databases (e.g., PostgreSQL, MongoDB, Cassandra).
– Hands-on experience with cloud data platforms such as AWS Redshift, GCP BigQuery, or Azure Data Lake.
– Knowledge of data security best practices and compliance with global privacy regulations (e.g., GDPR).
– Experience with monitoring and logging platforms (e.g., Prometheus, Datadog, Splunk).
– Strong analytical and problem-solving skills.
– Ability to work cross-functionally with data scientists, analysts, product, and engineering teams.
Ideal Candidate Profile:
– Detail-oriented engineer with a passion for building scalable data systems.
– Thrives on solving complex data integration challenges across diverse datasets and platforms.
– Proactive in identifying bottlenecks and optimizing workflows for better efficiency.
– Collaborative with excellent communication skills for sharing technical details with non-technical stakeholders.
– Continuously curious about emerging data technologies and tools.
– Committed to maintaining data integrity, security, and compliance.
– Willing to transition to a salaried, full-time position when funding is secured.
Compensation & Benefits:
– Equity-only at present, to transition to a salaried, full-time permanent position when funding is secured.
– Remote and flexible working arrangements.
– Opportunity to be part of a potentially epic start-up with potential opportunities for global travel.
– Access to industry conferences and workshops in due course.
To apply, please provide a CV, your compensation requirements (including salary expectations for when funding is secured), and a cover letter explaining your interest and qualifications for the role. Note that submissions without all requested information will be automatically disqualified and rejected. Interviews will commence in February 2025. Apply link