Sort by
Refine Your Search
-
Listed
-
Category
-
Field
-
for Science @ Scale: Pretraining, instruction tuning, continued pretraining, Mixture-of-Experts; distributed training/inference (FSDP, DeepSpeed, Megatron-LM, tensor/sequence parallelism); scalable evaluation
-
, Mixture-of-Experts; distributed training/inference (e.g. FSDP, DeepSpeed, Megatron-LM, tensor/sequence parallelism); scalable evaluation pipelines for reasoning and agents. Federated & Collaborative
-
Infiniband networks and diagnostics. Extensive experience with High Performance Parallel File Systems (Lustre, WEKA, GPFS, etc). Experience with performance and diagnostic tools for benchmarking, analysis and
Searches related to parallel
Enter an email to receive alerts for parallel positions