Architect batch + stream pipelines (Airflow, Kafka, dbt) for diverse structured and unstructured marked data. Provide reusable SDKs in Python and Go for internal data producers. Implement and tune S3, column‑oriented and time‑series data storage for petabyte‑scale analytics; own partitioning, compression, TTL, versioning and cost optimisation. Develop internal libraries for schema management, data contracts, validation and lineage.
Job listings
We’re looking for an experienced, hands-on Data Scientist to embed into and support analytics within RevOps and Finance teams. You will ensure data quality & integrity, develop data models, build intuitive dashboards for revenue insights, support ad hoc analysis and strategic questions, and document processes & methodologies.
In this job, you will join our efforts in the application of Network Science to our business, building and maintaining pipelines that transform vanilla tabular data into informative graphs and using your analytical skills to come up with useful graph architectures. As a member of a fully remote and distributed team, you are expected to complete tasks autonomously, being highly collaborative and self-driven, with the curiosity of a child and the responsibility of a grown-up.
Design and implement scalable, well-documented data pipelines and semantic models using Microsoft Fabric and Power BI to support enterprise-wide analytics. Translate complex business requirements into performant data models, calculated measures, and interactive dashboards using DAX and Power Query. Integrate and transform data from Dynamics 365 and other enterprise sources into analytics-ready formats for self-service BI and decision support.