Logo
ByteDance

Research Scientist Graduate (Foundation Model, Video Generation) - 2025 Start (P

ByteDance, Seattle

Save Job

Research Scientist Graduate (Foundation Model, Video Generation) - 2025 Start (PhD)

Seattle Regular R&D PhD Graduates - 2025 Start Job ID: A242820

Responsibilities

Welcome to the Doubao Vision team, where we lead multi-modality foundation models focusing on visual understanding and generation. Our mission is to solve visual intelligence challenges in AI by conducting cutting-edge research in areas like vision and language, large vision models, and generative foundation models. Our team comprises experienced research scientists and engineers working to push the boundaries of foundation models and apply these technologies to diverse applications, creating a feedback loop for continuous improvement. Join us in shaping the future of AI and revolutionizing our products for global users.

We are seeking talented individuals for 2025. As a graduate, you will have unparalleled opportunities to start your career, pursue innovative ideas, and explore growth avenues. Contribute to a future driven by your insights with ByteDance. Candidates must be able to commit to onboarding by the end of 2025. Please clearly state your availability and graduation date in your resume.

Applications are reviewed on a rolling basis. Early application is encouraged. Candidates may apply for up to two positions, considered in the order of application. This limit applies globally across ByteDance and its affiliates.

Responsibilities

  • Conduct cutting-edge research and development in foundation models and multimodal machine learning, especially in generative AI (e.g., image, video, 3D generation).
  • Innovate in video generation technology.
  • Develop foundation models to enhance ByteDance products.
  • Explore new downstream AI-based products.

Qualifications

Minimum Qualifications:

  • Final year Ph.D. or recent Ph.D. in Computer Science or related engineering fields.
  • Strong skills in algorithms and programming; proficiency in Python and PyTorch.
  • Good teamwork and collaboration skills.

Preferred Qualifications:

  • Publications in top-tier venues like CVPR, ECCV, ICCV, NeurIPS, ICLR, ICML, SIGGRAPH, or Multimedia.
  • Hands-on experience with deep learning frameworks and large-scale training.
  • Experience with large-scale image/video processing and foundation models.
#J-18808-Ljbffr