"Path towards better deep learning models and implications for hardware"

#Deep #Learning
Share


ABSTRACT: Advances in Deep Learning over the past several years have demonstrated two paths to better models: scale and algorithmic innovation. Brute-force scaling of model parameter count increases model capacity, and when presented with enough training data, has shown better performance in many domains. But it requires more compute than can be delivered by a single traditional processor – clusters of 10s to even 1000s of general purpose processors are commonly used today for neural network training. This approach to scaling is not sustainable. We need algorithmic innovations to find more efficient neural network architectures and training methods. This requires more flexible hardware to develop and test novel approaches. In this presentation, we will look at the trends of recent deep learning models, discuss implications for hardware, and share how the Cerebras CS-1 addresses these requirements for both scale and flexibility of compute.



  Date and Time

  Location

  Hosts

  Registration



  • Date: 14 Sep 2020
  • Time: 07:30 AM to 08:30 AM
  • All times are (GMT-08:00) US/Pacific
  • Add_To_Calendar_icon Add Event to Calendar
If you are not a robot, please complete the ReCAPTCHA to display virtual attendance info.
  • Contact Event Host
  • Starts 11 September 2020 08:08 PM
  • Ends 14 September 2020 07:00 AM
  • All times are (GMT-08:00) US/Pacific
  • No Admission Charge