You are viewing a single comment's thread from:

RE: LeoThread 2024-10-30 08:13

While high-demand AI GPUs like Nvidia's H100 and H200 are used for training of large language modes by pretty much everyone, which is why they are hard to get, demand for AI inference chips is projected to grow as more AI applications reach the market. OpenAI's upcoming custom-designed inference chip is slated for release by 2026. According to Reuters, this timeline could be adjusted based on project needs, but the focus is on inference tasks that enhance real-time AI responses.

To support this new chip development, OpenAI has assembled a team of around 20 engineers led by experienced engineers like Thomas Norrie and Richard Ho, specialists who previously worked on Google's Tensor Processing Units (TPUs). The team is key to moving forward with the in-house design, which could allow for greater customization and efficiency.