8
Stack Exchange
votes

Can these asymptotic conditional expectations be bounded from above?

Posted on 5/4/2016
UQ ID:466
Science
Cross Validated
Original

Problem Setup

Let be a dimensional zero-mean, i.i.d. random variables. Let be

Let be a zero-mean normal random variable with covariance

The paper by Bentkus, for a given class , has the following result:

  • The constant when is all the convex sets in and
  • when is all the Euclidian balls in .

Question Can we upperbound any of the following when grows with ?

Here is the component dimensional vector and are the first the components of .

Possible Method

I have a method which might possibly give some upper-bound but I am unsure of this idea and it requires some nuanced arguments (which I am unable to make).

Issues:

  1. I am not sure the conditions under which the last equation is valid.
  2. In the above method I am using the Lebesgue measure, can we get better bounds using other measures? (Note : when is all the Euclidean balls in .)
  3. How can one apply similar ideas for question 2?
Keywords & Tags
normal-distribution
conditional-probability
conditional-expectation
asymptotics
Model Answers (9)
Submitted answers from various AI models
Unsolved
No reviews yet

K2-Think

Pending UQ Validation
Needs Human Verification
MBZUAI-IFM
Qwen2.5-32B
Sep 18, 2025, 01:50 PM

To address the questions regarding bounding the asymptotic conditional expectations, we analyze each part as follows:

1. Bound for ( \big|\mathbb{E}\big[S^d_n \big] - \mathbb{E}\big[Y^d\big]\big| )

Since ( {X_i^d} ) are zero-mean i.i.d. random variables, the sum ( \sum_{i=1}^n X_i^d ) has mean zero. Normalizing by ( \sqrt{n} ), the mean of ( S_n^d ) remains zero: ( \mathbb{E}[S_...

0 model reviews
0 human reviews
No reviews yet

GPT-5

UQ Validated
Needs Human Verification
OpenAI
Sep 2, 2025, 06:56 AM

Short answer

  1. It is exactly 0. Because E[Xi^d]=0, we have E[Sn^d]=(1/√n)∑i E[Xi^d]=0. You also assumed Y^d has mean 0. Hence |E[Sn^d]−E[Y^d]|=0 for all n,d.

  2. As written, there is no general useful upper bound. The two conditional expectations are random variables defined on different probability spaces; if you interpret them as conditional expectation functions mS(x)=E[Sn(i)|Sn(1:i−1)=x] and...

0 model reviews
0 human reviews
No reviews yet

Claude 3.7 Sonnet

Failed UQ Validation
Needs Human Verification
Anthropic
Aug 1, 2025, 12:00 AM

Bounding Asymptotic Conditional Expectations

This question deals with bounding differences between expectations of a normalized sum of i.i.d. random variables and its normal approximation, particularly when the dimension grows with the sample size.

Question 1:

For the first question, we actually don't need a complicated bound because:

$\mathbb{E}[S^...

0 model reviews
0 human reviews
No reviews yet

Claude Opus 4

Failed UQ Validation
Needs Human Verification
Anthropic
Aug 1, 2025, 12:00 AM

I'll analyze this problem about bounding asymptotic conditional expectations when dimension grows with .

Analysis of Question 1

Your proposed method has merit but needs refinement. The key issue is the last inequality where you try to bound the distance between densities using Bentkus's result.

Issues with the Current Approach

  1. Density vs. Probability Bound: Bentkus's...
0 model reviews
0 human reviews
No reviews yet

o3 Pro

Failed UQ Validation
Needs Human Verification
OpenAI
Aug 1, 2025, 12:00 AM

Let us write the –dimensional random variables in block form

where etc. Throughout we assume

 • (hence ),
 • is non-singular,
 • .

QUESTION 1
|E[S^d_n]-E[Y^d]| ? ...

0 model reviews
0 human reviews
No reviews yet

o4 mini

Failed UQ Validation
Needs Human Verification
OpenAI
Aug 1, 2025, 12:00 AM

Here is a sketch of how one can control both

(1)  |\E [S⁽ᵈ⁾ₙ] − \E [Y⁽ᵈ⁾]|,
(2)  sup_{z∈ℝⁱ⁻¹} | \E [S⁽ᵈ⁾ₙ(i) | S⁽ᵈ⁾ₙ(1:i−1)=z ] − \E [Y⁽ᵈ⁾(i) | Y⁽ᵈ⁾(1:i−1)=z ] |,

when d grows with n. We will rely on Bentkus’ inequality in total‐variation (TV) distance, together with a few simple Cauchy–Schwarz or Lipschitz‐test‐function arguments.

––––––––––––––––––––––––––––––––

  1. Th...
0 model reviews
0 human reviews
No reviews yet

DeepSeek R1

Failed UQ Validation
Needs Human Verification
DeepSeek
deepseek-ai/DeepSeek-R1
Aug 1, 2025, 12:00 AM
<think> We are given a sequence of i.i.d. d-dimensional random vectors {X_i^d} with mean zero and covariance matrix Σ^d. We define S_n^d as the normalized sum: S_n^d = (sum_{i=1}^n X_i^d) / sqrt(n). Let Y^d be a d-dimensional normal random vector with mean zero and covariance Σ^d.

The paper by Bentkus provides a bound on the supremum over a class A of sets of the absolute difference between the p...

0 model reviews
0 human reviews
No reviews yet

Gemini 2.5 Pro

Failed UQ Validation
Needs Human Verification
Google
Aug 1, 2025, 12:00 AM

This is an excellent and insightful question that touches on some deep results in probability theory. Let's break down each part.

Analysis of Question 1:

This part has a surprisingly simple answer. The bound is exactly zero.

  1. Expectation of : The random variables are i.i.d. with mean zero, i.e., $\mat...
0 model reviews
0 human reviews
No reviews yet

o3

Failed UQ Validation
Needs Human Verification
OpenAI
Aug 1, 2025, 12:00 AM

Below we denote by
Sₙ = Sₙᵈ : = n^{-1/2}∑_{k=1}^{n}X_k  and Y = Yᵈ∼N(0,Σ), 
β : = E|X¹|³ (the same β is used in Bentkus’ inequality) and we keep the assumption E X=0, Σ = EXXᵗ non–singular and β<∞.


1 The difference of the unconditional means

Because eve...

0 model reviews
0 human reviews
Submit Your Model's Answer
Your submission will be validated by our LLM validator and then by human experts

You need to be logged in to submit an answer