Seargin

Careers At Seargin

Break into the IT industry without coding or tech skills and join teams working on international projects.

Data Engineer

Seargin is a dynamic multinational tech company operating in 50 countries. At Seargin, we drive innovation and create projects that shape the future and greatly enhance the quality of life. You will find our solutions in the space industry, supporting scientists in the development of cancer drugs, and implementing innovative technological solutions for industrial clients worldwide. These are just some of the areas in which we operate.

Position:

Data Engineer

Location:

Remote

Country:

UE

Form of employment:

B2B

Experience level:

Senior

Responsibilities:

  • Data Integration Across Snowflake Accounts

    Customize and configure SAP Financial Accounting (FI) and Controlling (CO) modules to align with business requirements. This includes setting up organizational structures, account settings, and cost elements

  • KPI Engine Development

    Develop a robust KPI engine using Snowflake and Python to calculate clinical trial metrics. This involves gathering relevant data, performing calculations, and ensuring data integrity for decision-making

  • Data Pipeline Design

    Design and implement end-to-end data pipelines that extract, transform, and load (ETL) data from multiple business areas into a central repository for further processing

  • Clinical Data Collection

    Gather clinical trial data from various business areas, ensuring that the data is correctly formatted and complete for integration into the KPI engine

  • Implementation of Data Visualization Interfaces

    Develop interfaces that deliver processed data and calculated metrics to the data visualization team. Ensure that the data is presented clearly and in a format that supports the visualization needs

  • Optimization of Snowflake Queries

    Optimize Snowflake SQL queries to enhance performance when handling large datasets from clinical trials, improving overall processing efficiency

  • Automation of Data Collection

    Build automation scripts in Python to streamline the data collection process, ensuring regular, consistent updates from different business areas

  • Data Validation and Quality Control

    Implement validation checks within the data pipelines to ensure the accuracy and quality of the gathered clinical trial data before it's integrated into the KPI engine

  • Collaboration with Visualization Teams

    Work closely with the data visualization team to ensure that the metrics delivered meet their technical requirements and are formatted properly for visualization tools like Power BI or Tableau

  • Data Security and Compliance

    Ensure all data handling processes comply with industry standards and regulations, particularly regarding the sensitive nature of clinical trial data, such as HIPAA or GDPR

  • Data Transformation and Enrichment

    Use Python and Snowflake to transform raw clinical data into enriched datasets that provide deeper insights and actionable metrics for decision-makers

  • Support for Metrics Reporting

    Provide support in generating reports from the calculated KPIs, ensuring that all necessary metrics are included and correctly reported to stakeholders

  • Troubleshooting and Debugging

    Identify and resolve issues in data pipelines, ensuring that data flows seamlessly and that any interruptions in the process are promptly fixed

  • Documentation of Processes and Pipelines

    Maintain comprehensive documentation for all data integration processes, KPI engine logic, and pipeline workflows to ensure easy handover and transparency for future development

What we offer

Requirements:

  • Proficiency in Snowflake

    Strong expertise in Snowflake, with the ability to manage data warehousing, write complex SQL queries, and handle data integrations across multiple Snowflake accounts

  • Advanced Python Skills

    Proven experience in Python, including the development of data pipelines, KPI engines, and automation scripts. Ability to write efficient, clean, and maintainable code

  • Cloud Platform Knowledge (AWS/Azure)

    Hands-on experience with cloud platforms like AWS or Azure, specifically in the context of data storage, computation, and integration services

  • Version Control with GitLab

    Strong knowledge of GitLab for version control, managing codebases, and collaborating with other developers on large-scale projects

  • CI/CD Pipeline Experience

    Familiarity with GitLab CI/CD basics to automate testing, deployment, and integration workflows in the development process.

  • Python Testing Basics

    Knowledge of Python testing frameworks (e.g., PyTest, unittest) to ensure high-quality code through unit tests and continuous integration

  • Data Pipeline Design Experience

    Proven ability to design and implement scalable, efficient ETL pipelines to move and transform data from multiple sources into Snowflake for KPI generation

  • Data Integration and Transformation Skills

    Expertise in data integration and transformation processes to gather, cleanse, and standardize clinical trial data from various business areas for further analysis

  • Experience with Cloud Data Solutions

    Experience in using cloud-based data solutions like AWS Redshift or Azure Synapse to handle large volumes of clinical trial data, ensuring scalability and performance

  • Data Visualization Collaboration

    Ability to work closely with data visualization teams, ensuring that the delivered metrics are properly structured and compatible with visualization tools such as Tableau or Power BI

  • Automation Skills

    Experience in automating data extraction and transformation processes using Python, streamlining data workflows, and ensuring timely updates to datasets

  • Troubleshooting and Debugging

    Proven ability to troubleshoot issues in data pipelines and Snowflake processes, with strong debugging skills in both Python and SQL

  • Data Security and Compliance

    Understanding of data security principles and compliance requirements, especially in handling sensitive clinical trial data. Experience with GDPR, HIPAA, or equivalent is a plus

  • Experience in Cloud Data Warehousing

    Expertise in managing cloud-based data warehouses like Snowflake to support large-scale data operations, ensuring secure, fast, and reliable data storage

  • Strong Communication and Collaboration Skills

    Ability to communicate effectively with cross-functional teams, including business stakeholders, data visualization experts, and other developers, to deliver end-to-end solutions

Apply & join the team




    Ready to elevate your business? Let’s start the conversation.

    Reach out to learn more