齋藤優太
齋藤優太
ホーム
論文・出版
ブログ
連絡先
履歴書
英語
Light
Dark
Automatic
日本語
English
1
Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Off-policy evaluation (OPE) aims to accurately evaluate the performance of counterfactual policies using only offline logged data. …
Takuma Udagawa
,
Haruka Kiyohara
,
Yusuke Narita
,
Yuta Saito
,
Kei Tateno
引用
コード
スライド
arXiv
Fair Ranking as Fair Division: Impact-Based Individual Fairness in Ranking
Rankings have become the primary interface of many two-sided markets. Many have noted that the rankings not only affect the …
Yuta Saito
,
Thorsten Joachims
引用
コード
動画
スライド
arXiv
会議録
Off-Policy Evaluation for Large Action Spaces via Embeddings
Off-policy evaluation (OPE) in contextual bandits has seen rapid adoption in real-world systems, since it enables offline evaluation of …
Yuta Saito
,
Thorsten Joachims
引用
コード
動画
スライド
arXiv
Proceedings
Towards Resolving Propensity Contradiction in Offline Recommender Learning
We study offline recommender learning from explicit rating feedback in the presence of selection bias. A current promising solution for …
Yuta Saito
,
Masahiro Nomura
引用
Proceedings
Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
In real-world recommender systems and search engines, optimizing ranking decisions to present a ranked list of relevant items is …
Haruka Kiyohara
,
Yuta Saito
,
Tatsuya Matsuhiro
,
Yusuke Narita
,
Nobuyuki Shimizu
,
Yasuo Yamamoto
引用
コード
arXiv
会議録
Open Bandit Dataset and Pipeline: Towards Realistic and Reproducible Off-Policy Evaluation
Off-policy evaluation (OPE) aims to estimate the performance of hypothetical policies using data generated by a different policy. …
Yuta Saito
,
Shunsuke Aihara
,
Megumi Matsutani
,
Yusuke Narita
引用
コード
データセット
会議録
arXiv
A Real-World Implementation of Unbiased Lift-based Bidding System
Daisuke Moriwaki
,
Yuta Hayakawa
,
Isshu Munemasa
,
Yuta Saito
,
Akira Matsui
,
Masashi Shibata
引用
Efficient Hyperparameter Optimization under Multi-Source Covariate Shift
A typical assumption in supervised machine learning is that the train (source) and test (target) datasets follow completely the same …
Masahiro Nomura
,
Yuta Saito
引用
コード
会議録
arXiv
Evaluating the Robustness of Off-Policy Evaluation
Off-policy Evaluation (OPE), or offline evaluation in general, evaluates the performance of hypothetical policies leveraging only …
Yuta Saito
,
Takuma Udagawa
,
Haruka Kiyohara
,
Kazuki Mogi
,
Yusuke Narita
,
Kei Tateno
引用
コード
スライド
会議録
arXiv
Optimal Off-Policy Evaluation from Multiple Logging Policies
We study off-policy evaluation (OPE) from multiple logging policies, each generating a dataset of fixed size, i.e., stratified …
Nathan Kallus
,
Yuta Saito
,
Masatoshi Uehara
引用
コード
会議録
arXiv
»
引用
×