posted on 2023-05-07 21:26 read(431) comment(0) like(10) collect(0)
-1)GeForce RTX 4090
GPU Engine Specifications: | NVIDIA CUDA® Core Count | 16384 |
Acceleration Frequency(GHz) | 2.52 | |
Base Frequency(GHz) | 2.23 | |
Memory specifications: | Standard memory configuration | 24 GB GDDR6X |
memory bandwidth | 384 bits | |
Technical Support: | Ray Tracing Core | 3rd generation |
Tensor Cores | 4th generation | |
NVIDIA architecture | There's Lovelace | |
Supports NVIDIA DLSS | 3 | |
Support for NVIDIA Reflex | yes | |
Support NVIDIA Broadcast | yes | |
Supports PCI Express Gen 4 | yes | |
Support Resizable BAR | yes | |
Supports NVIDIA® GeForce Experience™ | yes | |
Support for NVIDIA Ansel | yes | |
Supports NVIDIA FreeStyle | yes | |
Support for NVIDIA ShadowPlay | yes | |
Support for NVIDIA Highlights | yes | |
Supports NVIDIA G-SYNC® | yes | |
Game Ready Driver | yes | |
Support NVIDIA Studio driver | yes | |
NVIDIA Omniverse | yes | |
Supports Microsoft DirectX® 12 Ultimate | yes | |
Supports NVIDIA GPU Boost™ | yes | |
Supports NVIDIA NVLink™ (SLI-Ready) | no | |
Supports Vulkan RT API, OpenGL 4.6 | yes | |
NVIDIA Encoder (NVENC) | 2x 8th Generation | |
NVIDIA Codec (NVDEC) | 5th Generation | |
AV1 encoding | yes | |
AV1 decoding | yes | |
CUDA-capable | 8.9 | |
Support VR Ready | yes | |
Display support: | Highest digital resolution and refresh rate (1) | 4K 240Hz or with the support of DSC technology, display 8K 60Hz HDR effect |
Standard Display Interface | HDMI(2), 3x DisplayPort(3) | |
Number of multi-monitors supported | 4(5) | |
HDCP | 2.3 | |
Graphics card size: | length | 304 mm |
width | 137 mm | |
slot | 3 slots (61mm) | |
Temperature and Power Specifications: | Maximum GPU temperature (°C) | 90 |
Graphics card power (W) | 450W | |
Required System Power (W) (4) | 850W | |
Auxiliary Power Interface | 3 x PCIe 8-pin adapter cable (adapter included) or 1 x PCIe Gen 5 interface power cable supporting 450W and higher rated power |
0) GeForce RTX 3090 graphics card, it seems that there is no single or double precision data?
GeForce RTX 3090 Ti | GeForce RTX 3090 | ||
---|---|---|---|
GPU Engine Specifications: | NVIDIA CUDA® Core Count | 10752 | 10496 |
Acceleration Frequency(GHz) | 1.86 | 1.70 | |
Base Frequency(GHz) | 1.67 | 1.40 | |
Memory specifications: | Standard memory configuration | 24 GB GDDR6X | 24 GB GDDR6X |
memory bandwidth | 384 bits | 384 bits | |
Technical Support: | RT Core | 2nd generation | 2nd generation |
Tensor Cores | 3rd generation | 3rd generation | |
NVIDIA architecture | Ampere | Ampere | |
Microsoft DirectX® 12 Ultimate | yes | yes | |
NVIDIA DLSS | yes | yes | |
NVIDIA Reflex | yes | yes | |
NVIDIA Broadcast | yes | yes | |
PCI Express Generation 4 | yes | yes | |
Resizable BAR | yes | yes | |
NVIDIA® GeForce Experience™ | yes | yes | |
NVIDIA Ansel | yes | yes | |
NVIDIA FreeStyle | yes | yes | |
NVIDIA ShadowPlay | yes | yes | |
NVIDIA Highlights | yes | yes | |
NVIDIA G-SYNC® | yes | yes | |
Game Ready Driver Support | yes | yes | |
NVIDIA Studio Driver | yes | yes | |
NVIDIA Omniverse | yes | yes | |
NVIDIA GPU Boost™ | yes | yes | |
NVIDIA NVLink™ (SLI-Ready) | yes | yes | |
Vulkan RT API、OpenGL 4.6 | yes | yes | |
HDMI 2.1 | yes | yes | |
DisplayPort 1.4a | yes | yes | |
NVIDIA Encoder | 7th generation | 7th generation | |
NVIDIA codec | 5th generation | 5th generation | |
CUDA-capable | 8.6 | 8.6 | |
VR Ready | yes | yes | |
Show support: | Highest Digital Resolution (1) | 7680x4320 | 7680x4320 |
Standard Display Interface | HDMI(2), 3x DisplayPort(3) | HDMI(2), 3x DisplayPort(3) | |
Number of multi-monitors supported | 4 | 4 | |
HDCP | 2.3 | 2.3 | |
Founders Edition Graphics Card Dimensions: | length | 12.3" (313 mm) | 12.3" (313 mm) |
width | 5.4" (138 mm) | 5.4" (138 mm) | |
high | 3 slots | 3 slots | |
Founders Edition thermal power specifications: | Maximum GPU temperature (°C) | 92 | 93 |
Graphics card power (W) | 450 | 350 | |
Recommended System Power (W) (2) | 850 | 750 | |
Auxiliary Power Interface | 3x PCIe 8-Pin auxiliary power supply interface (box adapter) or fifth-generation PCIe interface with load up to 450W or higher | 2 PCIe 8-Pin ports |
1)
A40:
The NVIDIA A40 accelerates the most demanding visual computing workloads from the data center, combining the latest NVIDIA Ampere architecture RT Cores, Tensor Cores, and CUDA® Cores with 48 GB of graphics memory. From powerful virtual workstations accessible from anywhere to dedicated render nodes, NVIDIA A40 brings nextgeneration NVIDIA RTX™ technology to the data center for the most advanced professional visualization workloads.
There is no single or double precision.
Neither does bandwidth.
2)
A30:
Built for AI inference at scale, the same compute resource can rapidly re-train AI models with TF32, as well as accelerate high-performance computing (HPC) applications using FP64 Tensor Cores. Multi-Instance GPU (MIG) and FP64 Tensor Cores combine with fast 933 gigabytes per second (GB/s) of memory bandwidth in a low 165W power envelope, all running on a PCIe card optimal for mainstream servers.
3)A100
NVIDIA A100 GPU adopts the super-large core GA100 of the new Ampere architecture, 7nm process, 54.2 billion transistors, an area of 826 square millimeters, 6912 cores, equipped with 5120-bit 40/80GB HBM2 memory, a bandwidth of nearly 1.6TB/s, and a power consumption of 400W.
NVIDIA A100 Tensor Core GPUs deliver outstanding acceleration for AI, data analytics, and high-performance computing (HPC) applications at every scale, powering high-performance, elastic data centers around the world. As the engine of NVIDIA's data center platform, the A100 delivers up to 20x higher performance than the previous-generation NVIDIA Volta™. Efficiently scalable, the A100 can also be partitioned into seven independent GPU instances with Multi-Instance GPU (MIG) technology, providing a unified platform that enables elastic data centers to dynamically adjust to changing workload demands. NVIDIA A100 Tensor Core technology supports a wide range of mathematical precision, enabling a single accelerator for each workload. The latest generation of A100 80GB doubles the GPU memory, providing 2TB/s of global ultra-fast memory bandwidth, which can accelerate the processing of very large models and massive data sets. The A100 is part of a complete NVIDIA data center solution consisting of a stack of hardware, networking, software, libraries, and optimized AI models and applications from the NGC™ catalog. As a powerful end-to-end AI and HPC platform for the data center, the A100 empowers researchers to achieve real-world results and deploy solutions into production environments at scale.
It's a pity that it is not allowed to be sold in China, the evil ghost.
4)V100
The NVIDIA® V100 Tensor Core GPU is a powerful accelerator for deep learning, machine learning, high-performance computing (HPC), and graphics computing. Powered by the NVIDIA Volta™ architecture, the V100 Tensor Core GPU can deliver the performance of nearly 32 CPUs in a single GPU, enabling researchers to tackle previously inaccessible challenges. The V100 has already topped MLPerf, the industry's first AI benchmark, proving it to be the world's most powerful computing platform with massive scalability and versatility.
Not worse than A100, but better than A30.
American brains are also funny
5)A800
NVIDIA will launch a new A800 GPU for Chinese users to replace the A100, which is in line with the US government's export control policy. According to the policy of the US government, the data transfer rate of chips for export shall not exceed 600GB/s, and NVIDIA A800 is set at 400GB/s. But other specifications are unclear, such as the number of cores, operating frequency, video memory, power consumption, etc.
, Nvidia and AMD said that products including Nvidia's data center chips A100 and H100 were included in the export control list by the US Department of Commerce.
And according to Nvidia, the new A800 can replace the A100, both of which are GPU (graphics processing unit) processors. According to the NVIDIA A800 GPU information introduced on the official website of chip dealer OMNISKY Rongtian, the data transfer rate of the new chip is 400GB per second, which is lower than the 600GB per second of A100, which represents a significant decline in the performance of the data center. Moreover, the A800 supports a memory bandwidth of up to 2TB/s, with little change in other parameters.
Author:gfg
link:http://www.pythonblackhole.com/blog/article/357/b46de47d9337b94ca55b/
source:python black hole net
Please indicate the source for any form of reprinting. If any infringement is discovered, it will be held legally responsible.
name:
Comment content: (supports up to 255 characters)
Copyright © 2018-2021 python black hole network All Rights Reserved All rights reserved, and all rights reserved.京ICP备18063182号-7
For complaints and reports, and advertising cooperation, please contact vgs_info@163.com or QQ3083709327
Disclaimer: All articles on the website are uploaded by users and are only for readers' learning and communication use, and commercial use is prohibited. If the article involves pornography, reactionary, infringement and other illegal information, please report it to us and we will delete it immediately after verification!