AI Engineer - Multi-Modal Microscopy Representation Alignment & Post-Training

Reposted 5 Days Ago
Campus, IL, USA
In-Office
96K-300K Annually
Mid level
Biotech
The Role
Develop and fine-tune multi-modal microscopy vision foundation models, implement supervised adaptation and cross-modality representation alignment, build reproducible pipelines for segmentation, tracking, and classification, produce uncertainty-aware predictions, lead rigorous model evaluations, collaborate with scientists, and deploy models and workflows in scalable GPU-based environments.
Summary Generated by Built In
Primary Work Address: 19700 Helix Drive, Ashburn, VA, 20147

Current HHMI Employees, click here to apply via your Workday account.

TLDR: Build the model backbone for the next era of AI-powered spatial biology.

Please include a cover letter with your application detailing your qualifications and experience for this position. Describe a deep learning project you have executed, ideally a creative use of supervised fine tuning of a pre-trained vision transformer, U-Net architecture, or related topic. Projects in computer vision for microscopy image analysis are especially relevant. Include a link to a code repository if possible. If you contributed to a joint project, please describe your specific contributions. Briefly discuss the project's results, limitations, and challenges you encountered. Finally, include a link to your GitHub profile, personal website, or similar and/ or any relevant projects at the bottom of your cover letter.

About the role:

AI@HHMI: HHMI is investing $500 million over the next 10 years to support AI-driven projects and to embed AI systems throughout every stage of the scientific process in labs across HHMI. The Foundational Microscopy Image Analysis (MIA) project sits at the heart of AI@HHMI. Our ambition is big: to create one of the world’s most comprehensive, multimodal 3D/4D microscopy datasets and use it to power a vision foundation model capable of accelerating discovery across the life sciences.

We are seeking a highly skilled AI Research Engineer to join our team and advance our AI-driven scientific initiatives. You will build methods for supervised adaptation of pre-trained microscopy vision models and cross-modality representation learning/ alignment. You will build robust pipelines that adapt foundation models to specialized microscopy tasks and develop algorithms that align image level embeddings across modalities (e.g., fluorescence ↔ electron microscopy ↔ brightfield ↔ …).

In collaboration with other engineers and scientists, you will use these models for scalable vision tasks, instance segmentation, tracking, classification, and more. You will utilize probabilistic models to produce uncertainty-aware predictions across scales. This role requires deep knowledge of the underlying models and practical implementation skills to maximize biological impact. You will lead rigorous model evaluations, implement novel architectures, and ensure all work meets the highest standards of reproducible open science. Success in this role requires close collaboration with our microscopy experts, cellular biologists, neuroscientists, and computer scientists to ensure models can be deployed in large data real-world scenarios.

Strong programming skills in Python, PyTorch, and/ or JAX are required, along with the ability to reason about neural network behavior from first principles. The role also requires knowledge of microscopy data formats and tools such as Zarr and Neuroglancer. We seek candidates who can think critically about model design, understand how architectural choices and regularization affect model behavior, and design rigorous experiments to evaluate models. Domain expertise in microscopy image analysis is not necessary, but will be highly valued. Because this is a team project, we value a clean shared code base and git-based collaborative workflows. Familiarity with state-of-the-art vision frameworks such as DinoV3, SAM, CellPose, or Vision Transformers is required. We are looking for candidates with experience in ML model deployment, workflow orchestration, and high-throughput data processing, as well as experience working with large biological datasets in scalable GPU-based computing environments.

What we provide:

  • A competitive compensation package, with comprehensive health and welfare benefits.

  • A supportive team environment that promotes collaboration and knowledge sharing.

  • Access to a world-class computational infrastructure and large, high-quality datasets.

  • The opportunity to engage with world-class researchers, software engineers and AI/ML experts, contribute to impactful science, and be part of a dynamic community committed to advancing humanity’s understanding of fundamental scientific questions.

  • Amenities that enhance work-life balance such as on-site childcare, free gyms, available on-campus housing, social and dining spaces, and convenient shuttle bus service to Janelia from the Washington D.C. metro area.

  • Opportunity to partner with frontier AI labs on scientific applications of AI (see https://www.anthropic.com/news/anthropic-partners-with-allen-institute-and-howard-hughes-medical-institute).

What you’ll do:

  • You’ll be responsible for the post-training process of our multi-modal microscopy vision foundation model, which includes model fine tuning on annotated data, and aligning learned representations across multiple microscopy modalities.

  • Design and execute rigorous experiments to evaluate model performance on a wide distribution of microscopy images and model architectures.

  • Collaborate with Scientists at Janelia and the broader academic community to integrate our model into their workflows across a wide variety of vision tasks.

  • Collaborate with interdisciplinary teams, potentially mentor junior engineers, and direct or assist in directing the work of others to meet project goals while advising stakeholders on data strategies and best practices.

What you bring:

  • Master's or PhD degree in Computer Science, Applied Mathematics, Computational Neuroscience, or a related field—or an equivalent combination of education and relevant experience.

  • 3+ years of experience fine-tuning spatial transformer networks, contrastive learning, model distillation, RLHF and/or cross-modal alignment methods.

  • Familiarity with state of the art vision fine tuning methods, such as low-rank adaptation (LoRA), linear probing etc.

  • Strong programming skills in Python, PyTorch, and JAX. Skills in Javascript are a plus.

  • Familiarity with computational tools in microscopy and connectomics data (Cellpose, CAVE, Flood Filling Networks, Neuroglancer, Zarr).

  • Experience with ML model deployment, workflow orchestration, and high-throughput data processing and model training.

  • Keen interest to work in a truly interdisciplinary environment and learn about cellular/molecular biology (e.g. transcriptomics) & neuroscience.

Physical Requirements:

Remaining in a normal seated or standing position for extended periods of time; reaching and grasping by extending hand(s) or arm(s); dexterity to manipulate objects with fingers, for example using a keyboard; communication skills using the spoken word; ability to see and hear within normal parameters; ability to move about workspace. The position requires mobility, including the ability to move materials weighing up to several pounds (such as a laptop computer or tablet).

Persons with disabilities may be able to perform the essential duties of this position with reasonable accommodation. Requests for reasonable accommodation will be evaluated on an individual basis.

Please Note:

This job description sets forth the job’s principal duties, responsibilities, and requirements; it should not be construed as an exhaustive statement, however.  Unless they begin with the word “may,” the Essential Duties and Responsibilities described above are “essential functions” of the job, as defined by the Americans with Disabilities Act.

Compensation Range

AI Engineer I: $96,325.60 (minimum) - $120,407.00 (midpoint) - $156,529.10 (maximum)

AI Engineer II: $123,125.60 (minimum) - $153,907.00 (midpoint) - $200,079.10 (maximum)

AI Engineer III: $149,515.20 (minimum) - $186,894.00 (midpoint) - $242,962.20 (maximum)

AI Engineer IV: $184,453.60 (minimum) - $230,567.00 (midpoint) - $299,737.10 (maximum)

Pay Type: Salary

HHMI’s salary structure is developed based on relevant job market data. HHMI considers a candidate's education, previous experiences, knowledge, skills and abilities, as well as internal consistency when making job offers. Typically, a new hire for this position in this location is compensated between the minimum and the midpoint of the salary range.

#LI-BG1

Compensation and Benefits

Our employees are compensated from a total rewards perspective in many ways for their contributions to our mission, including competitive pay, exceptional health benefits, retirement plans, time off, and a range of recognition and wellness programs. Visit our Benefits at HHMI site to learn more. 

HHMI is an Equal Opportunity Employer

We use E-Verify to confirm the identity and employment eligibility of all new hires.

Top Skills

Cave
Cellpose
Dinov3
Flood Filling Networks
Git
Gpu
JavaScript
Jax
Lora
Neuroglancer
Python
PyTorch
Sam
Vision Transformers
Zarr
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: Chevy Chase, MD
1,800 Employees
Year Founded: 1953

What We Do

For 60 years, HHMI has been moving science forward. We’re an independent, ever-evolving philanthropy that supports basic biomedical scientists and science educators with the potential for transformative impact. We invest in people, not projects. We encourage collaborative and results-driven working styles and offer an adaptable environment where employees can function at their highest level. As HHMI scientists continue to push boundaries in laboratories and classrooms, you can be sure that your contributions while working at HHMI are making a difference. To move science forward, we need experts in areas such as communications, finance, human resources, information technology, investments, and law as well as scientists. Visit our website at http://www.hhmi.org

Similar Jobs

IMC Trading Logo IMC Trading

US PhD Talent Community

Fintech • Machine Learning • Software • Financial Services
Hybrid
2 Locations
1954 Employees

Lessen LLC Logo Lessen LLC

Vendor Sourcing Specialist

Cloud • Real Estate • Software • PropTech
Hybrid
Chicago, IL, USA
713 Employees

Lessen LLC Logo Lessen LLC

Escalation Specialist

Cloud • Real Estate • Software • PropTech
Hybrid
Chicago, IL, USA
713 Employees
45K-50K Annually

Caterpillar Logo Caterpillar

Manager Digital Product - Used Equipment

Artificial Intelligence • Cloud • Internet of Things • Software • Cybersecurity • Industrial
Hybrid
Peoria, IL, USA
100000 Employees
148K-240K Annually

Similar Companies Hiring

Formation Bio Thumbnail
Pharmaceutical • Healthtech • Biotech • Big Data • Artificial Intelligence
New York, NY
140 Employees
SOPHiA GENETICS Thumbnail
Software • Healthtech • Biotech • Big Data • Artificial Intelligence
Boston, MA
450 Employees
Pfizer Thumbnail
Pharmaceutical • Natural Language Processing • Machine Learning • Healthtech • Biotech • Artificial Intelligence
New York, NY
121990 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account