About TeraCyte
TeraCyte is seeking a Data Engineer with 2–5 years of experience to architect, build, and maintain robust, scalable data pipelines and infrastructure. You will be a key contributor to the team powering our AI and analytics efforts, working alongside DevOps, AI, and software engineering teams.
At TeraCyte, we are revolutionizing biological research and industrial processes with our cutting-edge Temporal Cytometry technology. Our end-to-end solution, from data capture to AI-driven predictions, provides unparalleled data integrity and supports groundbreaking research and applications.
If you are passionate about advancements in biotechnology and medicine, join us in shaping the future of cell biology by driving the success of our products in the market.
Key Responsibilities
- Integrate AI/ML model endpoints with production systems and ensure reproducibility.
- Lead continuous improvements in system observability, error handling, and testing strategies.
- Optimize frontend and backend codebases for performance and maintainability.
- Collaborate closely with AI teams to prepare datasets for training, validation, and inference workflows.
- Collaborate with product and AI teams to build intuitive UIs and efficient APIs using modern frameworks.
- Implement and maintain robust data storage and retrieval strategies using blob storage, SQL/NoSQL databases, and file systems.
- Design, develop, and maintain scalable full stack applications with integrated CI/CD pipelines using Azure DevOps.
- Ensure data integrity, lineage, and security across all stages of the pipeline.
- Develop internal dashboards and tools for deployment tracking, system health, and monitoring.
Qualifications
- 2–5 years of experience as a Data Engineer or in a similar backend/data infrastructure role.
- Proficiency in Python and SQL; strong Linux and scripting (Bash) knowledge.
- Experience with distributed systems, asynchronous tasks, and workflow orchestration tools.
- Proficiency with modern frontend technologies (e.g., React, Vue) and RESTful API development.
- Comfortable with Docker, Kubernetes, and Linux-based development environments (Azure preferred).
- Independent and self-motivated, with the ability to manage and improve core data systems.
Nice to Have
- Experience working on ML/AI pipelines and deploying model inference services.
- Understanding of secure data handling, compliance, and role-based access control.
- Exposure to cloud-native databases, serverless architecture, or edge computing.
- Familiarity with Prometheus, Grafana, or the ELK stack for monitoring and logging.
- Experience working with Celery, RabbitMQ, Redis, or equivalent messaging frameworks.
- Hands-on experience with automation tools and frameworks for infrastructure, testing, or workflows.
Why Join Us
- Competitive salary and comprehensive benefits package.
- Opportunity to work on innovative projects in a dynamic and collaborative environment.
- Professional development opportunities to enhance your skills and career growth.
- Flexible working hours and supportive work culture.
Join us and support groundbreaking innovation in biotechnology. If you are passionate about efficient data flow, clean architecture, and operational excellence, this is the role for you., we want to hear from you!