Interpreting Machine Learning Models

#Machine #Learning #interpretability
Share

Have you ever wondered why your Machine Learning (ML) model predicted some specific values? With the growing potential and adoption of ML, making ML models and their decisions interpretable is becoming more and more important, which in turn drives the need to develop various frameworks to help explain, understand, diagnose, and refine ML models. In this session, you will learn about some techniques and frameworks for interpreting ML models.



  Date and Time

  Location

  Hosts

  Registration



  • Add_To_Calendar_icon Add Event to Calendar
If you are not a robot, please complete the ReCAPTCHA to display virtual attendance info.
  • London, Ontario
  • Canada

  • Contact Event Host
  • Co-sponsored by Optimized Computing and Communications (OC2) Laboratory
  • Starts 21 October 2020 07:00 PM UTC
  • Ends 30 October 2020 07:00 PM UTC
  • No Admission Charge


  Speakers

Mohamad Kalil Mohamad Kalil of IBM

Biography:

Mohamad Kalil is a Data Scientist at IBM Analytics. He obtained his Ph.D. in Electrical and Computer Engineering from Western University. He is currently focused on infusing AI into Business Intelligence applications to help users uncover patterns hidden in their data and explore and extract data insights with less effort.