google-research

Форк
0

README.md

IME: Interpretable Mixture of Experts

IME consists of a group of interpretable experts and an assignment module which puts weights on different experts. The assignment module can either be an interpretable model like a linear model or a black box model like an LSTM. During inference time each sample is assigned to a single expert.

Data

The datasets can be obtained and put into data/ folder in the following way:

  • [Electricity dataset] in "data/ECL" run process_ecl.py
  • [Rossmann dataset] download train.csv and store.csv from https://www.kaggle.com/c/rossmann-store-sales/data insert dataset in folder "data/Rossmann" and than run process_rossmann.py in "data/Rossmann"

Requirements

Dependencies can be installed using the following command:

pip install -r requirements.txt

Usage

  • Commands for training and testing the model IME with white box assignment module and black box assignment module respectively on Rossman:
python main.py --model IME_WW --data Rossmann --num_experts 20 --learning_rate 0.0001 --learning_rate_gate 0.001 --utilization_hp 1 --smoothness_hp 0.01 --diversity_hp 0
python main.py --model IME_BW --data Rossmann --num_experts 20 --learning_rate 0.0001 --learning_rate_gate 0.001 --utilization_hp 1 --smoothness_hp 0.01 --diversity_hp 0
  • Scripts for Rossmann and electricity experiments can be found in scripts/
  • To run Rossmann experiments use
 bash scripts/Rossman.sh

This is not an officially supported Google product.

Использование cookies

Мы используем файлы cookie в соответствии с Политикой конфиденциальности и Политикой использования cookies.

Нажимая кнопку «Принимаю», Вы даете АО «СберТех» согласие на обработку Ваших персональных данных в целях совершенствования нашего веб-сайта и Сервиса GitVerse, а также повышения удобства их использования.

Запретить использование cookies Вы можете самостоятельно в настройках Вашего браузера.