Jump to content

Wikipedia:Reference desk/Archives/Mathematics/2018 April 12

From Wikipedia, the free encyclopedia
Mathematics desk
< April 11 << Mar | April | May >> April 13 >
Welcome to the Wikipedia Mathematics Reference Desk Archives
The page you are currently viewing is a transcluded archive page. While you can leave answers for any questions shown below, please ask new questions on one of the current reference desk pages.


April 12

[edit]

size of largest gap in partition

[edit]

Suppose I take the integers 1,2,...8000 and pick a random subset of size 1000. Define and so you've got a random partition with its endpoints. Call M(A) the largest gap in A, i.e. M(A) could be as large as 7001, but that is unlikely. I'd like to find an approximate smallest S such that I can be pretty sure () that for random A, while not worrying about "extreme" () outliers.

I don't actually care too much about the exact answer to the above, since numerical simulation is good enough for my immediate purpose (it relates to compressed sensing). What I'm wondering is if this (finding the distribution of gap sizes in partitions) is a well known problem and if there's a standard way to solve it. Thanks! 173.228.123.166 (talk) 00:48, 12 April 2018 (UTC)[reply]

  • To recover a classic problem, remove the endpoint breaks. This makes a "rod" of length 8001 broken into 999 pieces and we want the size of the largest chunk. Now, the fact that we break at integer positions might affect the result slightly, but probably not much because the rod is very large. MSE came to my help for remembering the solution (for real-positioned break points), and with your numbers it yields an average length of maximum of . However this does not tell us the standard deviation (or the full probability distribution function) that we would need.
I will be back with some simulation results shortly. TigraanClick here to contact me 16:44, 12 April 2018 (UTC)[reply]
Done. The code is certainly suboptimal but it does the job.
Python code
import random
from statistics import mean, stdev

def onesimu(N, K, withreplacement=True):
    if withreplacement:
        lchosen = []
        for _ in range(K):
            lchosen.append(random.randint(1,N))
    else:
        lchosen = random.sample(range(1,N+1), K)

    lchosen = [0] + sorted(lchosen) + [N+1]

    return max([lchosen[i+1]-lchosen[i] for i in range(len(lchosen)-1)])

def domultiplesimus(N, K, Ntries, wr=True):
    results = []
    for _ in range(Ntries):
        results.append(onesimu(N, K, withreplacement=wr))
    results.sort()

    print('Result for {} tries (N={}, K={}):'.format(Ntries, N, K))
    print('Using replacement: {}'.format(wr))
    print('Mean and standard deviation: {} ± {}'.format(mean(results),stdev(results)))
    print('0.1% largest results:', results[Ntries-1-(Ntries//1000):])

if __name__ == '__main__':
    domultiplesimus(8000,1000,10000,True)
    domultiplesimus(8000,1000,10000,False)
The above results in a (rounded) 60±10 maximum interval length in the first case (selection with replacement, i.e. the a_i may not be unique), with the 0.1% "largest largest" (your S value, in effect) starting at 109, and a 57±10 maximum interval length in the second case (without replacement, i.e. the a_i are unique), with the 0.1% "largest largest" starting at 107. Of course your numbers may vary but that's the general idea. A little Python knowledge would of course allow you to modify the script so as to retrieve the full histograms. TigraanClick here to contact me 17:20, 12 April 2018 (UTC)[reply]


Gap number i goes from element number to element number , for . The length of this gap is . The total length of the gaps is . The number of gaps is 1001. The mean length of the gaps is . Every gap length g satisfies . The distribution of gap lengths is approximately a binomial distribution having and , because this distribution satisfies the above conditions. The standard deviation is . Let . Bo Jacoby (talk) 20:43, 12 April 2018 (UTC).[reply]

(What is IIRC?) . Approximations are kind of errors even when they provide useful results. Surely there is room for improvement. Bo Jacoby (talk) 19:43, 13 April 2018 (UTC).[reply]
IIRC = If I Remember Correctly. (FYI, it's in Wiktionary.)
No, approximations are not "kinds of errors" and for them to be useful they need a theoretical background, not just "here's some calculations I made up, maybe they work." That you think otherwise is horrifying (although emblematic of the terribleness of the RD). Things are approximately binomial because there is a convergence theorem, not just because that would be convenient! This kind of garbage answer is a disservice to the question-asker. --JBL (talk) 12:29, 14 April 2018 (UTC)[reply]
The problem I have with your reasoning is not that for finite N the binomial distribution is slightly different from a Gaussian. The problem I have is that the binomial distribution is weakly convergent to the Gaussian, and it converges "near the center", and it is not valid even for very large N to apply it to the events on the tails.
I recognize this is not a rigorous way of phrasing it but my math is lacking to make the point more precisely. For an example: if you flip a coin N times (binomial distribution with p=0.5) the probability of "all tails" is 0.5^N or 2^(-N), whereas in the gaussian approximation with mean = 0.5N, variance = 0.25N, then the PDF falls quicker: . I.e. the approximation gets worse as N grows larger, roughly by an exponential factor that grows as (e/2)^N. Now again I am not sure this is the correct benchmark against which to compare your solution, but I am sure it deserves a bit of care. TigraanClick here to contact me 20:24, 15 April 2018 (UTC)[reply]

improvement

[edit]

The size, , of a randomly chosen gap is approximately poisson distributed with mean value 7. The probability, that assumes the value , is for . The probability, that is not greater than , is . The probability, that no sizes of 1001 randomly chosen gaps are greater than , is . The OP wants .

.
.
.
.
.

Now so the answer is . Bo Jacoby (talk) 06:01, 14 April 2018 (UTC).[reply]

There is no independence here, so your claim that the probability is a certain 1001th power is completely false, and none of the subsequent calculations have value. --JBL (talk) 12:33, 14 April 2018 (UTC)[reply]
The dependence when some gaps are very big is rare, and so the approximation is good. Please improve the calculation rather than merely critisizing. Bo Jacoby (talk) 13:26, 14 April 2018 (UTC).[reply]
There is no way to "improve" this calculation, because it is just pseudomathematics. The improvement is to discard it and do something that makes sense. When someone answers a question with nonsense, it is a service to the question-asker to point out that the answer is nonsense. What would be even better is if you did not post nonsense answers. --JBL (talk) 14:16, 14 April 2018 (UTC)[reply]

Then do something that makes sense. Solve the problem if you can. The sum of the sizes of the 1001 gaps is exactly , while the sum of the poisson distributed variables is . So there is indeed dependence. Taking this dependence into account complicate the calculations without improving the result. Bo Jacoby (talk) 14:58, 14 April 2018 (UTC).[reply]

There is already a completely adequate discussion by Tigraan above! --JBL (talk) 17:59, 14 April 2018 (UTC)[reply]

Tigraan objected against using the gaussian distribution, which is consequently not used in the improvement section. Can JBL or Tigraan solve the problem? Bo Jacoby (talk) 18:38, 14 April 2018 (UTC).[reply]

Tigraan gave a clear and correct answer before you ever commented on this thread! I know they say that the internet is write-only, but really. --JBL (talk) 21:12, 14 April 2018 (UTC)[reply]

From 8000 consecutive numbers remove 1000 random numbers to get 7000 numbers divided into 1001 chunks (or gaps). Tigraan's formulation: 'This makes a "rod" of length 8001 broken into 999 pieces and we want the size of the largest chunk' should read: 'This makes a "rod" of length 7000 broken into 1001 pieces and we want the size of the largest chunk', and 'the average length of maximum' should read: . I redid Tigraan's brute force simulation in J.

     A=.1+(10000$1000)?8000 NB. 10000 random subsets 
     A=.(/:{])"1 A          NB. sort
     A=.0,.A,.8001          NB. append endpoints
     A=.(}.-}:)"1 A         NB. compute gap sizes
     A=.>./"1 A             NB. max gap size
     E=.+/%#                NB. Expected value
     sd=.E&.:*:&(-E)        NB. standard deviation
     (E,sd)A                NB. compute E and sd
56.6289 9.56994

This calculation does not answer the question posed by the OP.

Once in 1000 times the maximum gap length was 117.

>./>./(}.-}:)"1]0,.8001,.~(/:{])"1]1+(1000$1000)?8000 
117

I was aggrieved by JBL's insolence, but I deserved it. Bo Jacoby (talk) 04:33, 15 April 2018 (UTC).[reply]

I'm coming into this late, and I haven't read all of the discussion, but I ran 1 million trials, and if I haven't made a mistake, I get 38. Bubba73 You talkin' to me? 06:40, 16 April 2018 (UTC)[reply]

Whoops, I got it backwards. 99.9% of the time M < 38. 99.9% of the time M < 69. Bubba73 You talkin' to me? 07:53, 16 April 2018 (UTC)[reply]
Actually, < 70. Bubba73 You talkin' to me? 14:08, 16 April 2018 (UTC)[reply]

I wrote above:

The size, , of a randomly chosen gap is approximately poisson distributed.

This is a serious mistake. The simulation indicates that

The size, , of a randomly chosen gap is approximately geometically distributed.

Then the argument goes as before.

The probability, that assumes the value , is , where and . The probability, that is not greater than , is . The probability, that no sizes of the 1001 (approximately independent) gaps are greater than , is . The OP wants .

.
.
.
.

Bo Jacoby (talk) 14:52, 16 April 2018 (UTC).[reply]

I checked my program and reran it on 1,000,000 simulations.

The smallest max gap was 34. The largest max gap was 157. 99.9% of the max gaps are <= 69, so S = 70. Bubba73 You talkin' to me? 17:53, 16 April 2018 (UTC)[reply]

The question is not that 99.9% of all gaps on a large number of attempts fall below S. The question is that for one given attempt, the probability of at least one gap being >S is lower than 0.1%. That is not the same thing because there are multiple gaps in each attempt. Basically that is a confusion between median and mean. TigraanClick here to contact me 11:58, 17 April 2018 (UTC)[reply]
That is the way I interpreted it - 99.9% of the time the largest gap is < 70. Bubba73 You talkin' to me? 17:20, 18 April 2018 (UTC)[reply]
@Bubba73: something is definitely wrong with your experiment or how you've implemented it. --JBL (talk) 11:14, 19 April 2018 (UTC)[reply]

My above simulation of 1000 cases gave the max gap size 117. This indicates that in one out of 1000 cases the max gap size is like that. Repeating the simulation gave the numbers 99 101 100 95 103 103 122 111. This is in variance with the analytic attempt (S=89) and Bubbas simulation (S=70). A mystery needs to be clarified. Bo Jacoby (talk) 12:41, 17 April 2018 (UTC).[reply]

If it's a vote, then I get 9990491/10000000 tests with maxgap < 104 (1213/10000000 with maxgap = 104, 1273/10000000 with maxgap;= 103). --catslash (talk) 23:23, 17 April 2018 (UTC)[reply]
Repeating that with a 'better' pseudo-random number generator gives 9990195/10000000 tests with maxgap < 104 (1255/10000000 with maxgap = 104, 1414/10000000 with maxgap;= 103) - and it makes the tails noticeably smoother. --catslash (talk) 00:24, 18 April 2018 (UTC)[reply]

Bo Jacoby (talk · contribs), Tigraan (talk · contribs), and everyone else who replied: thanks very much for these answers, which I found enlightening and useful. Sorry I wasn't able to reply earlier. 173.228.123.166 (talk) 03:02, 19 April 2018 (UTC)[reply]