Responsibilities:

  • Enable model scaling and productionization: Provide critical support to scale frontier research into industry-leading next-generation models, including large-scale training data acquisition, reinforcement learning environment construction, and training efficiency optimization.
  • Build evaluation systems: Design and implement comprehensive, automated evaluation systems for next-generation models to better understand model capability boundaries and inform future research priorities.
  • Drive real-world applications: Apply theoretical breakthroughs to real product challenges, accelerating the practical deployment and real-world impact of AI technologies.

Qualifications:

  • Bachelor’s degree or higher (or equivalent practical experience) in Computer Science, Software Engineering, or a related field.
  • Strong programming skills: Proficient in Python and C/C++ in Linux environments; experienced with PyTorch and mainstream large-model training and fine-tuning frameworks; capable of independently implementing complex deep learning models and system components, with strong debugging and performance optimization abilities.
  • Data processing expertise: Experience with large-scale data preprocessing, data generation, and data augmentation; solid understanding of data-driven model iteration workflows.
  • Familiarity with large-scale model training pipelines, including distributed training, model parallelism, and training efficiency optimization.
  • Strong analytical and problem-solving skills, with a collaborative mindset and effective communication abilities.
If you are interested in these job openings, please submit your resume and cover letter to shandahr@shanda.com. We also welcome assistance from recruitment agencies.