The H200 features 141GB of HBM3e and a 4.8 TB/s memory bandwidth, a substantial step up from Nvidia’s flagship H100 data center GPU. ‘The integration of faster and more extensive memory will ...
The AI chip giant says the open-source software library, TensorRT-LLM, will double the H100’s performance for running inference on leading large language models when it comes out next month.
Results that may be inaccessible to you are currently showing.
Hide inaccessible results