본문 바로가기
조회 수 39 추천 수 0 댓글 0

단축키

Prev이전 문서

Next다음 문서

단축키

Prev이전 문서

Next다음 문서

NVIDIA가 호퍼 아키텍처 GPU와 HBM3e 메모리를 탑재한 H200, 그리고 HGX H200을 발표했습니다. 

 

메모리 대역폭은 4.8TB/s, 용량은 141GB로 H100보다 대역폭이 1.4배, 용량이 2배로 늘었습니다. 그래서 Llama2 70B는 1.9배, GPT-3 175B는 1.6배의 성능 향상이 있습니다. 

 

SC23—NVIDIA today announced it has supercharged the world’s leading AI computing platform with the introduction of the NVIDIA HGX™ H200. Based on NVIDIA Hopper™ architecture, the platform features the NVIDIA H200 Tensor Core GPU with advanced memory to handle massive amounts of data for generative AI and high performance computing workloads.

The NVIDIA H200 is the first GPU to offer HBM3e — faster, larger memory to fuel the acceleration of generative AI and large language models, while advancing scientific computing for HPC workloads. With HBM3e, the NVIDIA H200 delivers 141GB of memory at 4.8 terabytes per second, nearly double the capacity and 2.4x more bandwidth compared with its predecessor, the NVIDIA A100.

H200-powered systems from the world’s leading server manufacturers and cloud service providers are expected to begin shipping in the second quarter of 2024.

“To create intelligence with generative AI and HPC applications, vast amounts of data must be efficiently processed at high speed using large, fast GPU memory,” said Ian Buck, vice president of hyperscale and HPC at NVIDIA. “With NVIDIA H200, the industry’s leading end-to-end AI supercomputing platform just got faster to solve some of the world’s most important challenges.”

Perpetual Innovation, Perpetual Performance Leaps
The NVIDIA Hopper architecture delivers an unprecedented performance leap over its predecessor and continues to raise the bar through ongoing software enhancements with H100, including the recent release of powerful open-source libraries like NVIDIA TensorRT™-LLM.

The introduction of H200 will lead to further performance leaps, including nearly doubling inference speed on Llama 2, a 70 billion-parameter LLM, compared to the H100. Additional performance leadership and improvements with H200 are expected with future software updates.

NVIDIA H200 Form Factors
NVIDIA H200 will be available in NVIDIA HGX H200 server boards with four- and eight-way configurations, which are compatible with both the hardware and software of HGX H100 systems. It is also available in the NVIDIA GH200 Grace Hopper™ Superchip with HBM3e, announced in August.

With these options, H200 can be deployed in every type of data center, including on premises, cloud, hybrid-cloud and edge. NVIDIA’s global ecosystem of partner server makers — including ASRock RackASUS, Dell Technologies, Eviden, GIGABYTE, Hewlett Packard Enterprise, Ingrasys, Lenovo, QCT, Supermicro, Wistron and Wiwynn — can update their existing systems with an H200.

Amazon Web Services, Google Cloud, Microsoft Azure and Oracle Cloud Infrastructure will be among the first cloud service providers to deploy H200-based instances starting next year, in addition to CoreWeaveLambda and Vultr.

Powered by NVIDIA NVLink™ and NVSwitch™ high-speed interconnects, HGX H200 provides the highest performance on various application workloads, including LLM training and inference for the largest models beyond 175 billion parameters.

An eight-way HGX H200 provides over 32 petaflops of FP8 deep learning compute and 1.1TB of aggregate high-bandwidth memory for the highest performance in generative AI and HPC applications.

When paired with NVIDIA Grace™ CPUs with an ultra-fast NVLink-C2C interconnect, the H200 creates the GH200 Grace Hopper Superchip with HBM3e — an integrated module designed to serve giant-scale HPC and AI applications.

Accelerate AI With NVIDIA Full-Stack Software
NVIDIA’s accelerated computing platform is supported by powerful software tools that enable developers and enterprises to build and accelerate production-ready applications from AI to HPC. This includes the NVIDIA AI Enterprise suite of software for workloads such as speech, recommender systems and hyperscale inference.

Availability
The NVIDIA H200 will be available from global system manufacturers and cloud service providers starting in the second quarter of 2024.

Watch Buck’s SC23 special address on Nov. 13 at 6 a.m. PT to learn more about the NVIDIA H200 Tensor Core GPU.

 

01_o.jpg

 

https://nvidianews.nvidia.com/news/nvidia-supercharges-hopper-the-worlds-leading-ai-computing-platform




List of Articles
번호 분류 제목 글쓴이 조회 수 날짜
공지 덕질 공통 이용규칙 및 안내 (업데이트중+ 2024-04-13) 😀컴덕824 5132 2024.04.14
공지 1000P를 모으면 다이소 상품권 1000원을 신청할 수 있습니다. file Private 5117 2024.02.14
344 정보/소식 화웨이, 미국의 제재를 우회해 비밀 팹 네트워크를 구축 컴덕098 336 2023.08.24
343 정보/소식 화웨이 토크밴드 B7, 스마트밴드+이어셋 file 😀익명063 1970 2023.03.30
342 정보/소식 화웨이 스마트폰에 SK 하이닉스 메모리 탑재, 미국의 수출 제한 위반? file 😀무명의컴덕416 438 2023.09.11
341 정보/소식 하드디스크 판매량은 2028년까지 강세를 유지할 것 file 😀컴덕378 461 2023.08.21
340 정보/소식 필립스, 44.5인치 듀얼 QHD 커브드 디스플레이 file 😀익명161 1669 2023.03.30
339 정보/소식 피닉스 젠4/RDNA3를 위한 라이젠 7000G 데스크탑 출시 file 😀무명의컴덕392 376 2023.10.08
338 정보/소식 피닉스 2 다이의 라이젠 8000G, PCIe 대역폭 제한 file 😀컴덕660 624 2024.01.15
337 정보/소식 프레임워크 노트북 13은 라이젠 7 7840U 탑재, 96GB 메모리 지원 file 😀컴덕232 486 2023.08.10
336 정보/소식 펜타곤 근처에서 폭발이 일어났다는 가짜 이미지가 SNS에서 퍼지는 중 😀익명511 273 2023.05.29
335 정보/소식 페이스북, 미국으로 유럽 사용자 데이터를 전송해 13억 달러의 벌금 부과 😀익명871 293 2023.05.29
334 정보/소식 펌) 점보 프레임뽕이 와서 과연 유용한가 테스트를 해보았습니다 😀컴덕814 421 2024.04.06
333 정보/소식 파이슨, 컴퓨텍스 2023에서 E31T, E27T SSD 컨트롤러 발표 file 😀익명327 388 2023.06.09
332 정보/소식 파워컬러, Radeon RX 7000 Devil 시리즈 티저 공개 file 😀익명047 1652 2023.04.02
331 정보/소식 팀그룹의 M.2 SSD용 액티브 쿨러 file 😀익명158 272 2023.05.29
330 정보/소식 트위터, 개발자용 API의 세로운 요금 체계 발표 file 😀익명868 1637 2023.03.31
329 정보/소식 텔 14세대 메테오 레이크 : 16 코어 42MB 캐시의 4nm 공정 file 😀컴덕079 476 2023.08.10
328 정보/소식 테슬라, 트위터 광고 시작 😀익명753 268 2023.05.29
327 정보/소식 쿨러마스터 마스터에어 MA824 듀얼타워 CPU 쿨러 file 😀익명574 318 2023.06.09
326 정보/소식 코어 울트라 9 185H의 성능 관련 소문 file 😀익명498 309 2023.11.09
325 정보/소식 코어 i9가 탑재된 4x4 폼펙터 NUC file 😀컴덕780 432 2023.08.21
목록
Board Pagination Prev 1 2 3 4 5 6 7 8 9 10 ... 18 Next
/ 18