Job Description
<h3>π Description</h3> β’ As an ML Engineer specializing in Inference at Symbl.ai, you will play a key role in the development and optimization of our Nebula large language model, as well as other Conversation AI projects.
β’ You will be responsible for implementing and deploying deep learning models to enable efficient and accurate inference in real-time communication scenarios.
β’ Joining Symbl.ai as an ML Engineer focused on Inference offers the opportunity to work on cutting-edge technologies in the Conversation AI space.
β’ You will be at the forefront of shaping the future of communication with our innovative product, Nebula, and other exciting projects.
β’ Design and implement efficient algorithms and models for real-time inference in Conversation AI applications, with a focus on Nebula.
β’ Collaborate with cross-functional teams to integrate machine learning models into production systems, ensuring scalability, reliability, and performance.
β’ Optimize and fine-tune machine learning models for resource-constrained environments, such as edge devices or cloud-based platforms.
β’ Develop monitoring and evaluation mechanisms to assess the performance and effectiveness of inference models in real-world scenarios.
β’ Stay updated on the latest advancements in machine learning inference techniques and methodologies, incorporating new approaches into our projects as needed.
β’ Contribute to the documentation and dissemination of best practices for implementing and deploying machine learning inference solutions. <h3>π― Requirements</h3> β’ Possess a strong background in machine learning, with hands-on experience in developing and deploying inference models in production environments.
β’ Demonstrate proficiency in Python and relevant machine learning frameworks such as TensorFlow or PyTorch.
β’ Have experience with optimization techniques for machine learning models, including quantization, pruning, and model compression.
β’ Exhibit strong problem-solving skills and the ability to troubleshoot and debug complex technical issues related to inference.
β’ Possess excellent communication and collaboration skills, with the ability to work effectively in a remote team environment.
β’ Show a passion for learning and staying updated on advancements in machine learning inference technologies, with a keen interest in applying these technologies to Conversation AI.
β’ Be experienced in fundamental libraries for accelerating ML workflows, like CUDNN/TensorRT, ROCm, OpenVino, or OpenPPL. Understanding of one or more ML communication frameworks like NCCL is an advantage. <h3>ποΈ Benefits</h3> β’ 100% covered health coverage for you, and 90% for your dependents.
β’ 100% covered Life & AD&D, and short-term disability coverage for you.
β’ 401(k) with 3% matching.
β’ Continued education and professional development.
β’ Fixed leave policy of 18 Planned vacation days, 8 sick days, generous maternity, paternity and 16 annual holidays.