Sort by
Refine Your Search
-
optimization and LLM alignment: design preference-based training and fine-tuning methods (RLHF, PPO, DPO, reward modeling) for medical and multilingual LLMs. Agentic and tool-augmented AI systems: develop
Enter an email to receive alerts for agent-based-modelling "https:" positions