Founding Data Engineer
Source: Arbeitnow
AI Summary Powered by Gemini
Cygrid GmbH is seeking a Founding Data Engineer in Berlin to design and build the core data foundations for a next-generation AI-powered risk management platform in the cybersecurity space. This hybrid role requires hands-on experience with scalable data pipelines, cloud-native systems, and a passion for building complex platforms from scratch in a fast-paced startup environment.
Job Description
Berlin • Cybersecurity & AI Startup • Recently FundedA newly funded cybersecurity startup in Berlin is looking for a Data Engineer to join as a founding team member and help design the core security, identity, and enforcement foundations of a next-generation AI-powered risk management platform.We are an early-stage startup founded by experienced cybersecurity leaders with backgrounds in Israeli intelligence units.We are looking for a hands-on Founding Data Engineer for a hybrid role based in central Berlin. If you are passionate about cybersecurity and AI, thrive in fast-paced startup environments, and enjoy building complex platforms from the ground up, this is an opportunity to make a real impact.We’re building a next-generation cyber risk platform that helps enterprises understand, measure, and reduce identity risk at scale. Our goal is to turn complex identity and security data into clear, actionable risk insights that CISOs and CTOs can trust. You will work from scratch on core platform components from day one, shaping how modern enterprises manage and reason about risk using cloud-native systems, AI-driven analysis, and automated enforcement through AI agents.Key ResponsibilitiesDesign, build, and operate scalable data pipelines ingesting large volumes of security, identity, and SaaS telemetry dataDevelop batch and streaming data processing workflows powering analytics, risk scoring, and AI-driven insightsBuild and operate data platforms across cloud-native environmentsOwn data ingestion from multiple sources including APIs, event streams, and cloud servicesDesign and maintain data models, schemas, and transformations for analytics and downstream AI workloadsEnsure data quality, consistency, reliability, and observability across pipelinesOptimize data processing performance and infrastructure cost across storage and compute layersCollaborate closely with backend, AI, and security teams to align data architecture with product needsImplement monitoring, logging, alerting, and failure handling for data pipelinesSupport experimentation, PoCs, and rapid iteration in an early-stage startup environmentContribute to data security, access control, and compliance readinessEstablish Data best practices as the company scalesContinuously improve platform cost efficiency, reliability, and operational maturityDecide on Data Tech Stack and build whole Data platform setup from scratch Must have skills10+ years of experience in data engineering or data platform roles5+ years experience with Data Tools on AWS in production environmentsProven experience with design of the Data Warehouse, Data Lake and LakehouseAdvanced Python skills for data processing and pipeline developmentExperience with at least one modern data platform: Databricks, Snowflake, or Amazon RedshiftStrong SQL skills and experience with analytical and operational databasesExperience with Building data topics around AI agents, LLMs, RAG, Embeddings and Vector DBsHands-on experience with Kafka for streaming data pipelinesExperience working with as a data lake or primary storage layerExperience building and operating ETL / ELT pipelines and orchestration at scale(Fivetran, Airflow, AWS Glue, dbt, etc.)Experience operating data workloads in cloud environments (AWS), AWS S3, AWS BedrockUnderstanding of data quality, validation, and observability practicesAbility to troubleshoot and optimize data pipelines end to endHigh ownership mentality, startup mindset, and ability to move quickly.Experience with PostgreSQL (RDS / Aurora), Neo4J(or similar)Experience with Coding CoPilots(Claude Code, GitHub Copilot, AWS Whisperer, Codex, etc.)Knowledge of observability standards (OpenTelemetry, Prometheus)Experience supporting production SaaS platformsAbility to build things from scratch at early stage startupExperience in startup or scale-up environmentsLow ego, high commitment, and a true tech startup mentalityExcellent English communication skillsNice-to-have requirementsExperience with Apache SparkExperience with real-time streaming architectures and event-driven systemsExposure to feature stores and ML data pipelinesKnowledge of data security, access control, and encryptionFamiliarity with compliance frameworks (ISO 27001, SOC 2, NIS2, DORA)Experience in cybersecurity, identity, or risk management domainsInterest in security or risk management productsKnowledge of RBAC/ABAC or fine-grained access modelsExperience designing permission models, identity graphs, or access control frameworksKnowledge of Zero Trust principles, least privilege, and continuous verificationExperience with Security related to AI/LLM systems, agent logic, or willingness to learn fastPrevious experience in cybersecurity, identity startupsYou are a fan of rock or metal musicGerman language knowledgeBenefitsFounding Team Member Equity — ESOPs in a well funded cybersecurity companyCompetitive Compensation — salary aligned with market, flexible early-stage arrangements possibleFounding Team Member Status — shape the company from day one and own critical data decisionsFresh Funding Secured — immediate runway to build, hire, and execute fastSecurity Contribution — contribute to data architecture across identity, AI agents, and enforcementHigh Autonomy — freedom to design and build without corporate overheadWork with a Technical CEO and CPTO — fast iteration, strong alignment, clear directionCutting-Edge Domain — identity security, SaaS-to-SaaS access, AI agent governanceMassive Learning Curve — quickly gain depth in cybersecurity, identity, and AI systemsFast Career Path — opportunity evolve into leadership as we scale the data teamBerlin Location — top-tier tech environment with strong cybersecurity and AI ecosystemsDirect Impact — your data architecture and code become the foundation of the entire platformBest Equipment and Tools - High quality equipment and modern tools to help you do your best work from day oneTeam Offsites - Regular team offsites designed to build strong connections, alignment, and shared momentumFind Jobs in Germany on Arbeitnow
Full Description
Berlin • Cybersecurity & AI Startup • Recently FundedA newly funded cybersecurity startup in Berlin is looking for a Data Engineer to join as a founding team member and help design the core security, identity, and enforcement foundations of a next-generation AI-powered risk management platform.We are an early-stage startup founded by experienced cybersecurity leaders with backgrounds in Israeli intelligence units.We are looking for a hands-on Founding Data Engineer for a hybrid role based in central Berlin. If you are passionate about cybersecurity and AI, thrive in fast-paced startup environments, and enjoy building complex platforms from the ground up, this is an opportunity to make a real impact.We’re building a next-generation cyber risk platform that helps enterprises understand, measure, and reduce identity risk at scale. Our goal is to turn complex identity and security data into clear, actionable risk insights that CISOs and CTOs can trust. You will work from scratch on core platform components from day one, shaping how modern enterprises manage and reason about risk using cloud-native systems, AI-driven analysis, and automated enforcement through AI agents.Key ResponsibilitiesDesign, build, and operate scalable data pipelines ingesting large volumes of security, identity, and SaaS telemetry dataDevelop batch and streaming data processing workflows powering analytics, risk scoring, and AI-driven insightsBuild and operate data platforms across cloud-native environmentsOwn data ingestion from multiple sources including APIs, event streams, and cloud servicesDesign and maintain data models, schemas, and transformations for analytics and downstream AI workloadsEnsure data quality, consistency, reliability, and observability across pipelinesOptimize data processing performance and infrastructure cost across storage and compute layersCollaborate closely with backend, AI, and security teams to align data architecture with product needsImplement monitoring, logging, alerting, and failure handling for data pipelinesSupport experimentation, PoCs, and rapid iteration in an early-stage startup environmentContribute to data security, access control, and compliance readinessEstablish Data best practices as the company scalesContinuously improve platform cost efficiency, reliability, and operational maturityDecide on Data Tech Stack and build whole Data platform setup from scratch Must have skills10+ years of experience in data engineering or data platform roles5+ years experience with Data Tools on AWS in production environmentsProven experience with design of the Data Warehouse, Data Lake and LakehouseAdvanced Python skills for data processing and pipeline developmentExperience with at least one modern data platform: Databricks, Snowflake, or Amazon RedshiftStrong SQL skills and experience with analytical and operational databasesExperience with Building data topics around AI agents, LLMs, RAG, Embeddings and Vector DBsHands-on experience with Kafka for streaming data pipelinesExperience working with as a data lake or primary storage layerExperience building and operating ETL / ELT pipelines and orchestration at scale(Fivetran, Airflow, AWS Glue, dbt, etc.)Experience operating data workloads in cloud environments (AWS), AWS S3, AWS BedrockUnderstanding of data quality, validation, and observability practicesAbility to troubleshoot and optimize data pipelines end to endHigh ownership mentality, startup mindset, and ability to move quickly.Experience with PostgreSQL (RDS / Aurora), Neo4J(or similar)Experience with Coding CoPilots(Claude Code, GitHub Copilot, AWS Whisperer, Codex, etc.)Knowledge of observability standards (OpenTelemetry, Prometheus)Experience supporting production SaaS platformsAbility to build things from scratch at early stage startupExperience in startup or scale-up environmentsLow ego, high commitment, and a true tech startup mentalityExcellent English communication skillsNice-to-have requirementsExperience with Apache SparkExperience with real-time streaming architectures and event-driven systemsExposure to feature stores and ML data pipelinesKnowledge of data security, access control, and encryptionFamiliarity with compliance frameworks (ISO 27001, SOC 2, NIS2, DORA)Experience in cybersecurity, identity, or risk management domainsInterest in security or risk management productsKnowledge of RBAC/ABAC or fine-grained access modelsExperience designing permission models, identity graphs, or access control frameworksKnowledge of Zero Trust principles, least privilege, and continuous verificationExperience with Security related to AI/LLM systems, agent logic, or willingness to learn fastPrevious experience in cybersecurity, identity startupsYou are a fan of rock or metal musicGerman language knowledgeBenefitsFounding Team Member Equity — ESOPs in a well funded cybersecurity companyCompetitive Compensation — salary aligned with market, flexible early-stage arrangements possibleFounding Team Member Status — shape the company from day one and own critical data decisionsFresh Funding Secured — immediate runway to build, hire, and execute fastSecurity Contribution — contribute to data architecture across identity, AI agents, and enforcementHigh Autonomy — freedom to design and build without corporate overheadWork with a Technical CEO and CPTO — fast iteration, strong alignment, clear directionCutting-Edge Domain — identity security, SaaS-to-SaaS access, AI agent governanceMassive Learning Curve — quickly gain depth in cybersecurity, identity, and AI systemsFast Career Path — opportunity evolve into leadership as we scale the data teamBerlin Location — top-tier tech environment with strong cybersecurity and AI ecosystemsDirect Impact — your data architecture and code become the foundation of the entire platformBest Equipment and Tools - High quality equipment and modern tools to help you do your best work from day oneTeam Offsites - Regular team offsites designed to build strong connections, alignment, and shared momentumFind Jobs in Germany on Arbeitnow