Who We Are
TetraScience is the Scientific Data and AI Cloud company. We are catalyzing the Scientific AI revolution by designing and industrializing AI-native scientific data sets, which we bring to life in a growing suite of next gen lab data management solutions, scientific use cases, and AI-enabled outcomes.
TetraScience is the category leader in this vital new market, generating more revenue than all other companies in the aggregate. In the last year alone, the world’s dominant players in compute, cloud, data, and AI infrastructure have converged on TetraScience as the de facto standard, entering into co-innovation and go-to-market partnerships:
In connection with your candidacy, you will be asked to carefully review the Tetra Way letter, authored directly by Patrick Grady, our co-founder and CEO. This letter is designed to assist you in better understanding whether TetraScience is the right fit for you from a values and ethos perspective.
It is impossible to overstate the importance of this document and you are encouraged to take it literally and reflect on whether you are aligned with our unique approach to company and team building. If you join us, you will be expected to embody its contents each day.
The Role
As a Senior Platform and Data Lake Engineer at TetraScience, you will play a critical role in building and maintaining our data infrastructure. You will work closely with cross-functional teams to ensure the seamless ingestion, processing, and storage of significant volumes of scientific data. The role requires extensive experience building and supporting data pipeline infrastructure and hands-on experience in the Databricks ecosystem.
What You Will Do
- Design, develop, and optimize data lake solutions to support our scientific data pipelines and analytics capabilities.
- Design, develop, and optimize data pipelines and workflows within the Databricks platform.
- Design and architect services to meet customer data processing needs.
- Implement data quality and governance frameworks to ensure data integrity and compliance.
- 8+ years of experience in the software development industry, preferably in data engineering, data warehousing r data analytics companies and teams.
- Experienced in designing and implementing complex, scalable data pipelines/ETL services.
- Expert level of Python, Java, and Typescript.
- Extensive in cloud-based data storage and processing technologies, particularly AWS services such as S3, Step Functions, Lambda, and Airflow.
- Expert level of understanding and hands-on experience with Lake House architecture.
- Ability to articulate ideas clearly, present findings persuasively, and build rapport with clients and team members.
Nice to Have
- Knowledge of basic DevOps and MLOps principles
- 3+ year of experience with the DataBricks ecosystem.
- Expert level of experience with Spark/Glue and Delta tables/iseberg.
- Working knowledge of Snowflake
- Experience in working with Data Scientists and ML Developers
- Experience in management and lead developer roles from technology services companies
- Hands-on experience with data warehousing solutions and ETL tools.
- 100% employer-paid benefits for all eligible employees and immediate family members
- Unlimited paid time off (PTO)
- 401K
- Flexible working arrangements - Remote work
- Company paid Life Insurance, LTD/STD
- A culture of continuous improvement where you can grow your career and get coaching
We are not currently providing visa sponsorship for this position.
Top Skills
What We Do
TetraScience is the Scientific Data Cloud company with a mission to accelerate scientific discovery and improve and extend human life.
The Tetra Scientific Data Cloud(TM) is the only open, cloud-native platform purpose-built for science that connects lab instruments, informatics software, and data apps across the biopharma value chain and delivers the foundation of harmonized, actionable scientific data necessary to transform raw data into accelerated and improved scientific outcomes.
Through the Tetra Partner Network, market-leading vendors access the power of our cloud to help customers maximize the value of their data.