首页 News 正文

Nvidia Announces "Super AI Chip" H200 Expected to Start Supply in the Second Quarter of Next Year

六月清晨搅
219 0 0

Huang Renxun has upgraded his "equipment" again. On November 14th, a reporter from Daily Economic News learned from NVIDIA that on November 13th local time, NVIDIA announced the launch of NVIDIA HGX H200 (AI chip model, hereinafter referred to as "H200"). It is reported that H200 is the first to use HBM3e GPU (memory, faster and larger than before), further accelerating generative AI and large language models, while promoting scientific computing for HPC (high-performance computing) workloads. It can provide 141GB of display memory with a transmission speed of 4.8 TB/s, nearly doubling the capacity and bandwidth compared to the previous generation architecture of NVIDIA A100.
In the view of Ian Buck, Vice President of NVIDIA's Ultra Large Scale and High Performance Computing, in order to create intelligence through generative AI and HPC applications, it is necessary to use large, fast GPU graphics memory to process massive amounts of data quickly and efficiently. When H200 is used in conjunction with NVIDIA Grace CPUs using ultra fast NVLink C2C interconnect technology, it forms the GH200 Grace Hopper superchip with HBM3e - a computing module designed specifically for large-scale HPC and AI applications.
From the perspective of specifications, H200 will provide options for four and eight way H200 server motherboards, which are compatible with the hardware and software of the HGX H100 system; It can also be used for the NVIDIA GH200 Grace Hopper superchip using HBM3e released in August this year. These configurations enable H200 to be deployed in various data centers, including local, cloud, hybrid cloud, and edge; It can provide the highest performance for various application workloads, including LLM training and inference for super large models with parameters above 175 billion.
In terms of architecture, compared to the previous generation, the NVIDIA Hopper architecture has achieved a performance leap, such as nearly doubling the inference speed on a 70 billion parameter LLM - Llama 2 compared to the H100 (NVIDIA AI chip).
According to NVIDIA, H200 will be available through global system manufacturers and cloud service providers starting in the second quarter of 2024; Server manufacturers and cloud service providers are also expected to start offering systems equipped with H200 at the same time.
CandyLake.com 系信息发布平台,仅提供信息存储空间服务。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

  •   21世纪经济报道记者韩利明上海报道当地时间7月8日,礼来(NYSE:LLY)宣布将以每股57美元的现金,收购Morphic的所有流通股(总计约32亿美元),以获得治疗炎症性肠病(IBD)和其他慢性疾病的实验性疗法,拓展免疫 ...
    浪無月
    昨天 18:13
    支持
    反对
    回复
    收藏
  •   记者今日获悉,金融壹账通近日成功签约广州农村商业银行资产配置系统项目。据了解,金融壹账通将助力广州农村商业银行推进其财富管理业务的数字化转型,为全行财富客户提供专业化、场景化、个性化的精准服务,推 ...
    dongtianya
    前天 19:08
    支持
    反对
    回复
    收藏
  •   本报讯 (记者李冰)7月8日,在支付宝开放日上,支付宝宣布升级条码支付体验,推出“支付宝碰一下”,用户无需展示付款码,解锁手机碰一下商家收款设备,最快一步完成支付。据介绍,“碰一下”和“扫一下”都属 ...
    WJ1127H
    前天 20:17
    支持
    反对
    回复
    收藏
  •   7月8日,在支付宝开放日上,支付宝宣布升级条码支付体验,推出“支付宝碰一下”,用户无需展示付款码,解锁手机碰一下商家收款设备,最快一步完成支付。   碰一碰支付有很多优势   支付宝“碰一下”最快3 ...
    yearn1985
    前天 17:14
    支持
    反对
    回复
    收藏
六月清晨搅 注册会员
  • 粉丝

    0

  • 关注

    0

  • 主题

    30