The Core Generative AI team at our company Research is a unified group of researchers and engineers dedicated to developing our generative foundational models that serve LTX Studio, our AI-based video creation platform. Our focus is on creating a controllable, cutting-edge video generative model by merging cutting-edge algorithms with exceptional engineering. This involves enhancing machine learning components within our sophisticated internal training framework, crucial for developing advanced models. We specialize in both research and engineering that enable efficient and scalable training and inference, allowing us to deliver state-of-the-art AI-generated video models.
As a Large Scale Training Engineer, you will play a key role in enhancing the training throughput of our internal framework and enabling researchers to pioneer new model concepts. This role demands excellent engineering skills for designing, implementing, and optimizing cutting-edge AI models, alongside writing robust machine learning code and deeply understanding supercomputer performance. Your expertise in performance optimization, understanding distributed systems, and bug elimination will be crucial, as our framework supports extensive computations across numerous virtual machines.
This role is designed for individuals who are not only technically proficient but also deeply passionate about pushing the boundaries of AI and machine learning through innovative engineering and collaborative research.
Key Responsibilities
Profile and optimize the training process to ensure efficiency and effectiveness, including optimizing multimodal data pipelines and data storage methods.
Develop high-performance TPU/CPU kernels and integrate advanced techniques into our training framework to maximize hardware efficiency.
Utilize knowledge of hardware features to make aggressive optimizations and advise on hardware/software co-designs.
Collaboratively develop model architectures with researchers that facilitate efficient training and inference.
As a Large Scale Training Engineer, you will play a key role in enhancing the training throughput of our internal framework and enabling researchers to pioneer new model concepts. This role demands excellent engineering skills for designing, implementing, and optimizing cutting-edge AI models, alongside writing robust machine learning code and deeply understanding supercomputer performance. Your expertise in performance optimization, understanding distributed systems, and bug elimination will be crucial, as our framework supports extensive computations across numerous virtual machines.
This role is designed for individuals who are not only technically proficient but also deeply passionate about pushing the boundaries of AI and machine learning through innovative engineering and collaborative research.
Key Responsibilities
Profile and optimize the training process to ensure efficiency and effectiveness, including optimizing multimodal data pipelines and data storage methods.
Develop high-performance TPU/CPU kernels and integrate advanced techniques into our training framework to maximize hardware efficiency.
Utilize knowledge of hardware features to make aggressive optimizations and advise on hardware/software co-designs.
Collaboratively develop model architectures with researchers that facilitate efficient training and inference.
Requirements:
Industry experience with small to large-scale ML experiments and multi-modal ML pipelines.
Strong software engineering skills, proficiency in Python, and experienced with modern C++.
Deep understanding of GPU, CPU, TPU, or other AI accelerator architectures.
Enjoy diving deep into system implementations to enhance performance and maintainability.
Passion for driving ML accuracy with low-precision formats and optimizing compute kernels.
Background in JAX/Pallas, Triton, CUDA, OpenCL, or similar technologies is a plus.
Industry experience with small to large-scale ML experiments and multi-modal ML pipelines.
Strong software engineering skills, proficiency in Python, and experienced with modern C++.
Deep understanding of GPU, CPU, TPU, or other AI accelerator architectures.
Enjoy diving deep into system implementations to enhance performance and maintainability.
Passion for driving ML accuracy with low-precision formats and optimizing compute kernels.
Background in JAX/Pallas, Triton, CUDA, OpenCL, or similar technologies is a plus.
This position is open to all candidates.