πWorldwide
AI Research Engineer

Tether.to
πRemote - Worldwide
Please let Tether.to know you found this job on JobsCollider. Thanks! π
Summary
Join Tether's AI model team and drive innovation in architecture development for cutting-edge models. You will enhance intelligence, improve efficiency, and introduce new capabilities to advance the field. Deep expertise in LLM architectures and pre-training optimization is crucial, with a hands-on, research-driven approach. Explore and implement novel techniques and algorithms to push the limits of AI performance. Tether offers a global, remote work environment with opportunities to collaborate with bright minds and make a mark in the fintech space. The role involves working with large-scale LLM training on distributed servers with thousands of NVIDIA GPUs. This is a chance to contribute to a leading company in the industry.
Requirements
- A degree in Computer Science or related field
- Ideally PhD in NLP, Machine Learning, or a related field, complemented by a solid track record in AI R&D (with good publications in A* conferences)
- Hands-on experience contributing to large-scale LLM training runs on large, distributed servers equipped with thousands of NVIDIA GPUs, ensuring scalability and impactful advancements in model performance
- Familiarity and practical experience with large-scale, distributed training frameworks, libraries and tools
- Deep knowledge of state-of-the-art transformer and non-transformer modifications aimed at enhancing intelligence, efficiency and scalability
- Strong expertise in PyTorch and Hugging Face libraries with practical experience in model development, continual pretraining, and deployment
Responsibilities
- Conduct pre-training AI models on large, distributed servers equipped with thousands of NVIDIA GPUs
- Design, prototype, and scale innovative architectures to enhance model intelligence
- Independently and collaboratively execute experiments, analyze results, and refine methodologies for optimal performance
- Investigate, debug, and improve both model efficiency and computational performance
- Contribute to the advancement of training systems to ensure seamless scalability and efficiency on target platforms
Share this job:
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.
Similar Remote Jobs
πWorldwide
πWorldwide
πWorldwide
πWorldwide
πWorldwide
πWorldwide
πWorldwide
πWorldwide