More In-Depth Details of Floating Point Precision - NVIDIA CUDA - PyTorch Dev Discussions
TensorFlow Performance with 1-4 GPUs -- RTX Titan, 2080Ti, 2080, 2070, GTX 1660Ti, 1070, 1080Ti, and Titan V | Puget Systems
NVIDIA RTX 3090 FE OpenSeq2Seq FP16 Mixed Precision - ServeTheHome
Hopperアーキテクチャ|NTTPCのGPU+|NVIDIA Eliteパートナー
Hardware Recommendations for Machine Learning / AI | Puget Systems
Deep Dive Into Nvidia's “Hopper” GPU Architecture
TITAN RTX Benchmarks for Deep Learning in TensorFlow 2019: XLA, FP16, FP32, & NVLink | Exxact Blog
Titan V Deep Learning Benchmarks with TensorFlow
Mixed-Precision Programming with CUDA 8 | NVIDIA Technical Blog
後藤弘茂のWeekly海外ニュース】NVIDIAが次世代グラフィックスのために作ったGPU「GeForce RTX」ファミリー - PC Watch
Hardware for Deep Learning. Part 3: GPU | by Grigory Sapunov | Intento
NVIDIAが語るVoltaとTuring、最新GPUはこう使おう:GTC Japan 2018(3/4 ページ) - EE Times Japan
Using Tensor Cores for Mixed-Precision Scientific Computing | NVIDIA Technical Blog
Nvidia V100 | The Linux Cluster
Accelerating TensorFlow on NVIDIA A100 GPUs - Edge AI and Vision Alliance
NVIDIA's 80-billion transistor H100 GPU and new Hopper Architecture will drive the world's AI Infrastructure - HardwareZone.com.sg
Caffe2 adds 16 bit floating point training support on the NVIDIA Volta platform | Caffe2
Benchmarking GPUs for Mixed Precision Training with Deep Learning
Apache MXNet リリースに追加された新しい NVIDIA Volta GPU と Sparse Tensor のサポート | Amazon Web Services ブログ
AMD FidelityFX Super Resolution FP32 fallback tested, native FP16 is 7% faster - VideoCardz.com
FP16 Throughput on GP104: Good for Compatibility (and Not Much Else) - The NVIDIA GeForce GTX 1080 & GTX 1070 Founders Editions Review: Kicking Off the FinFET Generation