Skip to content
CIFAR header logo
fr
menu_mobile_logo_alt
  • News
  • Events
    • Public Events
    • Invitation-only Meetings
  • Programs
    • Research Programs
    • Pan-Canadian AI Strategy
    • Next Generation Initiatives
    • Global Call for Ideas
  • People
    • Fellows & Advisors
    • CIFAR Azrieli Global Scholars
    • Canada CIFAR AI Chairs
    • AI Strategy Leadership
    • Solution Network Members
    • Leadership
  • Support Us
  • About
    • Our Story
    • CIFAR 40
    • Awards
    • Partnerships
    • Publications & Reports
    • Careers
    • Staff Directory
    • Equity, Diversity & Inclusion
  • fr
  • Home
  • Bio

Follow Us

Samira Ebrahimi Kahou-BW_F

Samira Ebrahimi Kahou

Appointment

Canada CIFAR AI Chair

Pan-Canadian AI Strategy

Connect

Personal Page

Google Scholar

About

Samira Ebrahimi Kahou is a Canada CIFAR AI Chair at Mila, an associate professor at École de technologie supérieure (ÉTS), and an adjunct professor at the School of Computer Science at McGill University.

Ebrahimi Kahou’s current research interests are multimodal learning, reasoning across modalities, metalearning and in general efficient representation learning. She also works on deep learning methods for dialogue systems and humanitarian AI.

Awards

  • Member of team that won the second place in the PROBA-V Super Resolution Challenge, European Space Agency, 2019
  • Best Thesis Award in the Department of Computer Engineering, Polytechnique Montréal, 2017
  • Leader of team that won the third place in the Emotion Recognition in the Wild Challenge, ICMI, 2015
  • Best Paper Award, ECCV workshop on computer vision with local binary patterns, 2014
  • Leader of team that won the first place in the Emotion Recognition in the Wild Challenge, 2013

Relevant Publications

  • Kim, S., Kim, H., Lee, J., Yoon, S., Kahou, S. E., Kashinath, K., & Prabhat, M. (2019). Deep-hurricane-tracker: Tracking and forecasting extreme climate events. In 2019 IEEE Winter Conference on Applications of Computer Vision (WACV) (pp. 1761-1769). IEEE.

  • Goyal, R., Ebrahimi Kahou, S., Michalski, V., Materzynska, J., Westphal, S., Kim, H., … & Memisevic, R. (2017). The” something something” video database for learning and evaluating visual common sense. In Proceedings of the IEEE international conference on computer vision (pp. 5842-5850).

  • The Theano Development Team., Al-Rfou, R., Alain, G., Almahairi, A., Angermueller, C., Bahdanau, D., … & van Tulder, G. (2016). Theano: A Python framework for fast computation of mathematical expressions.Kahou, S. E.,

  • Bouthillier, X., Lamblin, P., Gulcehre, C., Michalski, V., Konda, K., … & Bengio, Y. (2016). Emonets: Multimodal deep learning approaches for emotion recognition in video. Journal on Multimodal User Interfaces, 10(2), 99-111.

  • Kahou, S. E., Pal, C., Bouthillier, X., Froumenty, P., Gülçehre, Ç., Memisevic, R., … & Wu, Z. (2013). Combining modality specific deep neural networks for emotion recognition in video. In Proceedings of the 15th ACM on International conference on multimodal interaction (pp. 543-550).

Institution

École de technologie supérieure (ÉTS)

McGill University

Mila

Department

Software and Information Technology Engineering, School of Computer Science

Education

  • PhD (Computer Engineering), Polytechnique Montréal

Country

Canada

Support Us

CIFAR is a registered charitable organization supported by the governments of Canada, Alberta and Quebec, as well as foundations, individuals, corporations and Canadian and international partner organizations.

Donate Now
CIFAR header logo

MaRS Centre, West Tower
661 University Ave., Suite 505
Toronto, ON M5G 1M1 Canada

Contact Us
Media
Careers
Accessibility Policies
Supporters
Financial Reports
Subscribe

  • © Copyright 2023 CIFAR. All Rights Reserved.
  • Charitable Registration Number: 11921 9251 RR0001
  • Terms of Use
  • Privacy
  • Sitemap

Subscribe

Stay up to date on news & ideas from CIFAR.

This website stores cookies on your computer. These cookies are used to collect information about how you interact with our website and allow us to remember you. We use this information in order to improve and customize your browsing experience and for analytics and metrics about our visitors both on this website and other media. To find out more about the cookies we use, see our Privacy Policy.
Accept Learn more