Sza 发表于 2019-8-29 18:15

2060级别的TensorCore的算力在预期的将来能满足什么应用?

之前看别人讨论还只是张大饼的AI补帧时,楼下有人提醒以图灵的算力不一定能够满足实时计算的需求。

于是,我这个外行想问一下2060这个级别的卡,它的AI算力对个人用户而言将来可以做些什么吗?

其实是我想更块2000~3000左右的显卡,因为对图形性能要求不高,对DLSS也没有需求。就想知道20系的TensorCore能不能战未来,不然对我来说都是和光追模块一样都只是昂贵的电热丝

qwased 发表于 2019-8-29 18:20

However, the TensorCore performance of Geforce game graphics is severely limited.The peak FP16 Tensor TFLOPS with FP32 Accumulate is only 43.6% of NVIDIA Quadro RTX6000.This is very abnormal, obviously an artificial limit.However, at least this generation of Geforce RTX gaming graphics hardware supports FP16 computing.There are many requirements for the use of TensorCore. Please refer to the NVIDIA Developer website for details.
According to my test, the RTX2080TI uses FP16 with 10-40% improvement over FP32.(use tensorflow/benchmark).In addition, the use of CUDA 10.0 can also bring a small performance boost.

游戏卡的TensorCore是残疾的,没什么卵用

albertfu 发表于 2019-8-29 18:56

光追不算电热丝吧,dlss效果还行的游戏开光追还可以

abcbuzhiming 发表于 2019-8-29 22:19

Sza 发表于 2019-8-29 23:07

abcbuzhiming 发表于 2019-8-29 22:19
貌似现有AI的算力衡量单位里貌似不存在个人用户的PC这个概念,要玩AI随便都是四路泰坦起 ...

他们训练模型要用这些昂贵的生产力工具,个人用户没这种需求吧。手机上的NPU算力没那么高都能够后期修个月亮。

startraveller 发表于 2019-8-30 09:33

Sza 发表于 2019-8-29 23:07
他们训练模型要用这些昂贵的生产力工具,个人用户没这种需求吧。手机上的NPU算力没那么高都能够后期修个 ...

那你就当是电热丝好了

ada_ovo 发表于 2019-8-30 11:23

只能怪amd不给力啊

DeepFishing 发表于 2019-8-30 11:25

qwased 发表于 2019-8-29 18:20
However, the TensorCore performance of Geforce game graphics is severely limited.The peak FP16 Tenso ...

跑正向推断有tensor core强很多,然后老黄这玩意游戏里面深度相关的应该是更低精度去跑,int8甚至int4,增益就明显了

—— 来自 Sony H8296, Android 9上的 S1Next-鹅版 v2.1.2

KanaiYuu 发表于 2019-8-30 14:07

tensor core也就是在混合精度计算上比较强,为混合精度的矩阵乘法做了优化

mimighost 发表于 2019-8-30 15:36

除非你做混合精度的深度学习训练,否则这个对你来说没啥用

诚司 发表于 2019-8-30 15:42

PC上的只算AI的算力现在基本一点关系都没有,手机NPU还能拿来超分辨率,人脸识别……
PC本来也没这种软件啊,github代码倒是随便跑,拿来换头之类
除非………………你vtuber出道

你的全家 发表于 2019-8-30 16:12

DeepFishing 发表于 2019-8-30 11:25
跑正向推断有tensor core强很多,然后老黄这玩意游戏里面深度相关的应该是更低精度去跑,int8甚至int4, ...

其实反向也用的

AMP做法是fp 用16算 loss算的时候乘2^15 bp 然后cast 32 除回来修改

—— 来自 vivo NEX S, Android 9上的 S1Next-鹅版 v2.1.0-play

中村隆太郎 发表于 2019-8-30 16:14

TL;DR advice

Best GPU overall: RTX 2070
GPUs to avoid: Any Tesla card; any Quadro card; any Founders Edition card; Titan RTX, Titan V, Titan XP
Cost-efficient but expensive: RTX 2070
Cost-efficient and cheap:RTX 2060, GTX 1060 (6GB).
I have little money: GTX 1060 (6GB)
I have almost no money: GTX 1050 Ti (4GB).Alternatively: CPU (prototyping) + AWS/TPU (training); or Colab.
I do Kaggle: RTX 2070. If you do not have enough money go for a GTX 1060 (6GB) or GTX Titan (Pascal) from eBay for prototyping and AWS for final training. Use fastai library.
I am a competitive computer vision or machine translation researcher: GTX 2080 Ti with the blower fan design. If you train very large networks get RTX Titans.
I am an NLP researcher: RTX 2080 Ti use 16-bit.
I want to build a GPU cluster: This is really complicated, you can get some ideas from my multi-GPU blog post.
I started deep learning and I am serious about it: Start with an RTX 2070. Buy more RTX 2070 after 6-9 months and you still want to invest more time into deep learning. Depending on what area you choose next (startup, Kaggle, research, applied deep learning) sell your GPU and buy something more appropriate after about two years.
I want to try deep learning, but I am not serious about it: GTX 1050 Ti (4 or 2GB). This often fits into your standard desktop and does not require a new PSU. If it fits, do not buy a new computer!
Which GPU(s) to Get for Deep Learning: My Experience and Advice for Using GPUs in Deep Learning
https://timdettmers.com/2019/04/03/which-gpu-for-deep-learning/

这是没有出super系列之前写的,那么现在最佳推荐应该是RTX 2070=2060s
页: [1]
查看完整版本: 2060级别的TensorCore的算力在预期的将来能满足什么应用?