Seargin is a dynamic multinational tech company operating in 50 countries. At Seargin, we drive innovation and create projects that shape the future and greatly enhance the quality of life. You will find our solutions in the space industry, supporting scientists in the development of cancer drugs, and implementing innovative technological solutions for industrial clients worldwide. These are just some of the areas in which we operate.
Data Engineer
Remote
UE
B2B
Senior
Customize and configure SAP Financial Accounting (FI) and Controlling (CO) modules to align with business requirements. This includes setting up organizational structures, account settings, and cost elements
Develop a robust KPI engine using Snowflake and Python to calculate clinical trial metrics. This involves gathering relevant data, performing calculations, and ensuring data integrity for decision-making
Design and implement end-to-end data pipelines that extract, transform, and load (ETL) data from multiple business areas into a central repository for further processing
Gather clinical trial data from various business areas, ensuring that the data is correctly formatted and complete for integration into the KPI engine
Develop interfaces that deliver processed data and calculated metrics to the data visualization team. Ensure that the data is presented clearly and in a format that supports the visualization needs
Optimize Snowflake SQL queries to enhance performance when handling large datasets from clinical trials, improving overall processing efficiency
Build automation scripts in Python to streamline the data collection process, ensuring regular, consistent updates from different business areas
Implement validation checks within the data pipelines to ensure the accuracy and quality of the gathered clinical trial data before it's integrated into the KPI engine
Work closely with the data visualization team to ensure that the metrics delivered meet their technical requirements and are formatted properly for visualization tools like Power BI or Tableau
Ensure all data handling processes comply with industry standards and regulations, particularly regarding the sensitive nature of clinical trial data, such as HIPAA or GDPR
Use Python and Snowflake to transform raw clinical data into enriched datasets that provide deeper insights and actionable metrics for decision-makers
Provide support in generating reports from the calculated KPIs, ensuring that all necessary metrics are included and correctly reported to stakeholders
Identify and resolve issues in data pipelines, ensuring that data flows seamlessly and that any interruptions in the process are promptly fixed
Maintain comprehensive documentation for all data integration processes, KPI engine logic, and pipeline workflows to ensure easy handover and transparency for future development
Employment based on a B2B contract
Opportunity to work in a stable, dynamically developing international company
Chance to participate in interesting projects and work with the latest information technologies
Attractive remuneration rates offered
Involvement in the most prestigious international projects
Access to Multisport benefits and private healthcare services
Strong expertise in Snowflake, with the ability to manage data warehousing, write complex SQL queries, and handle data integrations across multiple Snowflake accounts
Proven experience in Python, including the development of data pipelines, KPI engines, and automation scripts. Ability to write efficient, clean, and maintainable code
Hands-on experience with cloud platforms like AWS or Azure, specifically in the context of data storage, computation, and integration services
Strong knowledge of GitLab for version control, managing codebases, and collaborating with other developers on large-scale projects
Familiarity with GitLab CI/CD basics to automate testing, deployment, and integration workflows in the development process.
Knowledge of Python testing frameworks (e.g., PyTest, unittest) to ensure high-quality code through unit tests and continuous integration
Proven ability to design and implement scalable, efficient ETL pipelines to move and transform data from multiple sources into Snowflake for KPI generation
Expertise in data integration and transformation processes to gather, cleanse, and standardize clinical trial data from various business areas for further analysis
Experience in using cloud-based data solutions like AWS Redshift or Azure Synapse to handle large volumes of clinical trial data, ensuring scalability and performance
Ability to work closely with data visualization teams, ensuring that the delivered metrics are properly structured and compatible with visualization tools such as Tableau or Power BI
Experience in automating data extraction and transformation processes using Python, streamlining data workflows, and ensuring timely updates to datasets
Proven ability to troubleshoot issues in data pipelines and Snowflake processes, with strong debugging skills in both Python and SQL
Understanding of data security principles and compliance requirements, especially in handling sensitive clinical trial data. Experience with GDPR, HIPAA, or equivalent is a plus
Expertise in managing cloud-based data warehouses like Snowflake to support large-scale data operations, ensuring secure, fast, and reliable data storage
Ability to communicate effectively with cross-functional teams, including business stakeholders, data visualization experts, and other developers, to deliver end-to-end solutions