Platform Challenges & Exploratio

2022-11-07  本文已影响0人  MatrixOnEarth

@(Deep Learning)[Engineering]

姚伟峰

2017年旧文

Deep Learning Helps Medical Image Diagnosis

More and more work are showing deep learning can help medical images diagnosis, which not only saves health-care costs but also accelerate health-care democratization especially in developing and undeveloped regions.


天池医疗AI大赛

DL Medical Image Diagnosis brings new challenges to platform

Medical image analysis is different from well-studied image analysis problems like ImageNet/Microsoft COCO.

Based on above, Medical AI is not only a computation intensive but also a memory intensive workload.

The Limitation of GPU-centric solution

For GPU-centric solution’s memory limitation, currently researchers must make compromises, and these compromises finally hurts effects.

CPU-centric Platforms unleash Medical AI Explorations

Intel Xeon & Xeon Phi supply best capability to handle computation- & memory-intensive workloads and make best flexibility on Medical AI explorations.


Case Study

  • Resolution Matters in Medical AI



  • Batch Size Matters in Medical AI

    *Nitish Shirish Keskar, Dheevatsa Mudigere, etc. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima, 2017.2, https://arxiv.org/abs/1609.04836

A bigger picture

Segmented AI Workloads Call for Diversified Platforms

As AI workloads are getting more and more segmented and thus more and more performance characteristics are disclosed, more and more diversified platforms need to be considered in order to get best effect/performance/cost.

AI Cloud Services Can and Should Leverage a Broader Platform Portfolio

CSPs have convergence power to handle the broader platform platform. It's an opportunity. In PC era, it's hard to do that because every user need handle the diversity of the HWs and distributions, now CSPs can handle it and deliver an unified service to users transparently.


后记 (2022/11/08)
5年之后再回头看这个PPT,整个领域似乎变了,又似乎没变。
变了的是,NV GPU在内存带宽和内存容量上持续改善,到H100已经演变成聚合带宽为3.35 TB/s的80GB大容量HBM3了。这些改变,其实是因应推荐系统、3D分析、大规模语言模型(LLM)这些应用的需求,逐渐拓宽GPU的对这些领域的适用性区间的努力。变了的还有,在短短5年的时间里,CUDA生态已经发展成深度学习甚至是高性能计算领域的by-default,现在很多data scientists也会读一些、写一些甚至改一些CUDA代码了,围绕着这个生态也催生了Triton这种更利于小白data scientist的CUDA代码生成工具。传统的围绕CPU编程的高性能计算生态围墙被跨过了。可见生态从来是power的附庸,只要你有压倒性的power,整个生态就愿意去适应你、完善你,最后变成你的生态。Intel因其庞大的组织和决策结构,拖累了其在新领域快速创新并建立护城河的能力,被颠覆性创新者推进了创新者的窘境,成为创新者的窘境的另一个生动实例。

没有变的是,application还在那儿,application的需求还在那儿,NV似乎正在从屠龙少年变成龙,我们正见证旧瓶新酒的商业宣传,关于“GPU can do all”,这个故事Intel也讲过,当年的主语是CPU。这符合一鸡多吃的商业利润最大化的动机,是个无可厚非的商业故事。



但回到技术的语境里,当我们抛弃其他的上下文,有时候我们会觉得目前NV的努力可能只是缓解措施,并没有从根本上解决问题。在Xeon Phi被Intel cancel之后,业界依旧认可"scale + vector + tensor + spatial"的聚合微架构是有前途的最终解决路径之一。之前,我们寄希望于Xeon Phi用"Atom + AVX512QFMA"及其后续演进在原编程模型的框架下完成这次新计算IP的纳入,最终完成统一与收敛,维护昔日帝国的辉煌。而如今,新的先驱如Tenstorrent以及Esperanto们,以"RISC-V + domain specific extensions"为新的武器,继续前行在这条道路上。

如果我们是牧村浪潮(Makimoto’s Wave)的信徒,那我们会相信converge,是谁摘得桂冠,我们需要耐心等待。



是为记。

References

  1. How Makimoto’s Wave Explains the Tsunami of New AI Processors
上一篇 下一篇

猜你喜欢

热点阅读