Ø Contextual BAI with a fixed budget. n 𝑲 treatment arms 𝐾 = 1,2, … , 𝐾 : Treatment arms: alternatives of medicine, policy, and advertisements. • Each arm 𝑎 ∈ [𝐾] has a potential outcome 𝑌! ∈ ℝ. n Context 𝑋 ∈ 𝒳 ⊆ ℝ". n Policy 𝜋: 𝐾 ×𝒳 → 0,1 such that ∑!∈[%] 𝜋(𝑎 ∣ 𝑥) = 1 ∀𝑥 ∈ 𝒳. • A policy is a function that returns a estimated best arm given a context 𝑥. • Policy class Π: a set of policies 𝜋. Ex. Neural networks. n Distribution 𝑃 of (𝑌*, … , 𝑌%, 𝑋). • 𝒫: a set of 𝑃. • 𝔼! and ℙ! : expectation and probability laws under 𝑃. • 𝜇"(𝑃) = 𝔼! 𝑌# " and 𝜇"(𝑃)(𝑥) n Policy value 𝑄 𝑃 𝜋 ≔ 𝔼4 ∑!∈ % 𝜋 𝑎 𝑋 𝜇!(𝑃)(𝑋) . • Expected outcome under a policy 𝜋 and a distribution 𝑃. n Optimal policy 𝜋∗(𝑃) ∈ arg max6∈7 𝑄 𝑃 𝜋 . n Experimenter’s interest: training 𝜋∗ using experimental data. n Adaptive experiment with 𝑻 rounds: • In each round 𝑡 ∈ 𝑇 ≔ 1,2, … , 𝑇 , an experimenter 1. observes 𝑑-dimensional context 𝑋8 ∈ 𝒳 ⊆ ℝ". 2. assigns treatment 𝐴8 ∈ [𝐾], and 3. observes outcome 𝑌8 ≔ ∑!∈ % 1 𝐴8 = 𝑎 𝑌8 ! . • After the final round 𝑇, an experimenter trains a policy 𝜋. • Denote the trained policy by L 𝜋 ∈ Π. n AS: adaptive sampling during an experiment. n PL: policy learning at the end of the experiment. Ø Procedure of the PLAS strategy. • Define the target treatment-assignment ratio as follows: - 𝐾 = 2: 𝑤∗(𝑎 ∣ 𝑥) = 9!(:) 9"(:);9#(:) # ∀𝑎 ∈ [2] and ∀𝑥 ∈ 𝒳. - 𝐾 ≥ 3: 𝑤∗ 𝑎 𝑥 = 9! : # ∑ $∈ & 9$ : # ∀𝑎 ∈ 𝐾 and ∀𝑥 ∈ 𝒳. 𝜎"(𝑥) $: Conditional variance of 𝑌# " given 𝑥. → Our designed policy adaptively estimates the ratio and assigns arms following the probability. • (AS) In each round 𝒕 ∈ 𝑻 , an experimenter 1. observes 𝑋8 , 2. estimates 𝑤∗(𝑎 ∣ 𝑋8 ) by replacing 𝜎!(𝑋8 ) > with its estimator L 𝜎!(𝑋8 ) >, and 3. assigns an arm 𝑎 following the probability R 𝑤8 (𝑎 ∣ 𝑋8 ), an estimator of 𝑤∗(𝑎 ∣ 𝑋8 ). • (PL) At the end, the experimenter trains 𝜋 as ! 𝜋!"#$ ≔ arg max %∈' ) (∈[*] 1 𝑇 ) ,∈[-] 𝜋(𝑎 ∣ 𝑋, ) 1[𝐴, = 𝑎](𝑌, − ! 𝜇, (𝑎 ∣ 𝑋, )) 8 𝑤, (𝑎 ∣ 𝑋, ) + ! 𝜇, (𝑎 ∣ 𝑋, ) . • ? 𝑤#(𝑎 ∣ 𝑋#) and D 𝜇#(𝑎 ∣ 𝑋#) are estimators of 𝑤∗ and 𝜇(𝑃)(𝑥) using samples until 𝑡. v n Evaluate the strategy by using the worst-case analysis. n Lower and upper bounds depend on the policy class complexity. • Use the Vapnik-Chervonenkis dimension when 𝐾 = 2 and the Natarajan dimension when 𝐾 ≥ 3 for the complexity. • Denote the complexity of Π by 𝑀. n Make several restrictions on the policy class and distribution. Ø Worst-case regret lower bounds. • If 𝐾 = 2, any strategy with a trained policy L 𝜋 satisfies sup !∈𝒫 𝑇𝔼! 𝑅$ 𝑃 ( 𝜋 ≥ 1 8 𝔼% 𝑀 𝜎&(𝑋) + 𝜎'(𝑋) ' + 𝑜 1 𝑎𝑠 𝑇 → ∞. • If 𝐾 ≥ 3, any strategy with a trained policy L 𝜋 satisfies sup !∈𝒫 𝑇 𝔼! 𝑅$ 𝑃 , 𝜋 ≥ 1 8 𝔼% 𝑀 2 &∈ ' 𝜎&(𝑋) ( + 𝑜 1 𝑎𝑠 𝑇 → ∞. • Note that 𝔼% 𝑀 ∑ (∈ ) 𝜎((𝑋) ' ≥ 𝔼% 𝑀 𝜎&(𝑋) + 𝜎'(𝑋) ' . Ø Worst-case regret upper bounds of the PLAS strategy. • If 𝐾 = 2, the PLAS strategy satisfies 𝑇 𝔼4 𝑅F 𝑃 L 𝜋GHIJ ≤ 272 𝑀 + 870.4 𝔼K 𝜎*(𝑋) + 𝜎>(𝑋) > + 𝑜 1 𝑎𝑠 𝑇 → ∞. • If 𝐾 ≥ 3, the PLAS strategy satisfies 𝑇 𝔼! 𝑅$ 𝑃 , 𝜋)*+, ≤ 𝐶 log 𝑑 𝑀 + 870.4 𝔼% 𝑀 2 &∈ ' 𝜎&(𝑋) ( + 𝑜 1 𝑎𝑠 𝑇 → ∞, where 𝐶 > 0 is universal constant. Ø The leading factor depending on 𝑀, 𝑇, and 𝜎! 𝑥 > of the upper bounds aligns with that of the lower bounds. • 𝔼% 𝑀 𝜎&(𝑋) + 𝜎'(𝑋) ' and 𝔼% 𝑀 ∑ (∈ ) 𝜎((𝑋) ' . → Minimax (rate-)optimal for the expected simple regret. Masahiro Kato, Mizuho-DL Financial Technology Co., Ltd. Kyohei Okumura, Northwestern University Takuya Ishihara, Tohoku University Toru Kitagawa, Brown University ü New framework: Contextual best arm identification (BAI). ü Algorithm: Adaptive experimental design for policy learning. ü Results: minimax optimality of the proposed algorithm. 2. Problem Setting n Performance measure of L 𝜋 : Expected simple regret. • Simple regret: 𝑟F 𝑃 (L 𝜋)(𝑥) = 𝜇!∗ 4 𝑃 (𝑥) − 𝜇 D !( under 𝑃 • Expected simple regret: 𝑅F 𝑃 L 𝜋 ≔ 𝔼4 𝑟8 𝑃 L 𝜋 𝑋 = 𝑄 𝑃 𝜋∗ 𝑃 − 𝑄 𝑃 L 𝜋 . n Our goal: • Design of an algorithm (strategy) of an experimenter. • Under the strategy, the experimenter trains 𝜋 while minimizing the expected simple regret 𝔼4 𝑟8 𝑃 . → We design the Adaptive-Sampling (AS) and Policy-Learning (PL) strategy and show its minimax optimality for 𝑅F 𝑃 L 𝜋 . 3. Performance Measure 1. Contribution 4. PLAS Strategy v v v v 5. Minimax Optimality v v v