What's the meaning of "random" in Mathematics?

In summary: But everyone is welcome to share their knowledge and learn from others, which makes this community a great place to discuss and learn about different topics.
  • #1
fbs7
345
37
Physics, Economists, Biologists, Astronomers and my brother all love the word "Random", as that allows allows them to get out of clockwork processes and allow for variations due to unknowns or whatever else.

But, how does a Mathematician reconcile itself with the idea of random? There's no axiom for "choice", no function for "random value", no explanation of what "chance" is.

Meanwhile I heard that someone spent 500 pages of logic to prove that 1+1 = 2 (or something like that), so how is it possible that mathematicians and logicians spend all that trouble to prove some really basic stuff, while at the same time just accept theories around probabilities and random numbers without (as least from my untrained point of view) an axiomatic foundation for choice?
 
  • Like
Likes FactChecker, TachyonLord, Demystifier and 1 other person
Mathematics news on Phys.org
  • #2
fbs7 said:
Physics, Economists, Biologists, Astronomers and my brother all love the word "Random", as that allows allows them to get out of clockwork processes and allow for variations due to unknowns or whatever else.

But, how does a Mathematician reconcile itself with the idea of random? There's no axiom for "choice", no function for "random value", no explanation of what "chance" is.

Meanwhile I heard that someone spent 500 pages of logic to prove that 1+1 = 2 (or something like that), so how is it possible that mathematicians and logicians spend all that trouble to prove some really basic stuff ...
As a programmer you should know that ##1+1=0##, so basic stuff is quite relative here.
... while at the same time just accept theories around probabilities and random numbers without (as least from my untrained point of view) an axiomatic foundation for choice?
Random are the possible values of a measurable function on a probability space ##(\Omega,\Sigma,P)## to a measure space ##(\Omega',\Sigma')##.

Axiomatic enough?
 
  • Like
Likes Matt Benesi and suremarc
  • #3
fresh_42 said:
As a programmer you should know that ##1+1=0##, so basic stuff is quite relative here.

Random are the possible values of a measurable function on a probability space ##(\Omega,\Sigma,P)## to a measure space ##(\Omega',\Sigma')##.

Axiomatic enough?
Well, f(x) = 1/(1+x) has possible values on [0,1] on a space defined by [0,∞]... how is that different than random?

Also, if that "measurable function" is a random value, then can someone write a formula for a truly random function (as opposed to pseudo-random)? How would it define a value, if the value is random?

I know one can write a formula to express probability, and that's fine. Say, probably = 1/(2+x) or something like that. But does that doesn't define a random value, it just gives an average, over a large enough number of samples, for the sampled values to fall within a range.

Like, probability of a duck to be male, if you count 1 million ducks, is I guess 50%. But then the "choice" that the duck is male or female is not random, it's defined by a well-known biological process.
 
Last edited:
  • #4
fbs7 said:
Well, f(x) = 1/(1+x) has possible values on [0,1] on a space defined by [0,∞]... how is that different than random?
You need a bit more than just the intervals: sigma algebras (the measures) on ##[0,1]## and ##[0,\infty)##, plus a probability measure to get the usual properties on ##[0,1]##: ##P:\Sigma \longrightarrow [0,1], P(\emptyset)=0, P(\Omega)=1, P(\dot{\cup} A_i)=\sum P(A_i)##.
Also, if that "measurable function" is a random value, then what is the function?
The function isn't random, it's a random variable ##X\, : \,(\Omega, \Sigma ,P) \longrightarrow (\Omega',\Sigma')\,.##
How would it define a value, if the value is random?
##P## by relating a certain real value to each element of ##\Sigma \subseteq \mathbb{P}(\Omega)##. The values themselves aren't random, they are determined. They represent a probability, and this is fixed by the function, aka random variable. Randomness is just the common word for probability. It is a confusing word, as the probability value itself is not random.

You can handle the subject as ordinary real analysis. Randomness is it's interpretation, not the method.
 
  • Like
Likes fbs7
  • #5
Oh... randomness is interpretation...

So a "random" variable is really just another variable, just like "time" is just a variable without anything different than say a "mass" variable. Our interpretation of "time" makes it special, but that meaning of something flowing over.. whatever... is not mathematical, it's just human interpretation.

Likewise, the "random" nature of a variable is just human interpretation, right?

Does that mean that, because "random" is a semantic thing that we use to interpret the variable, that "random" is not really a mathematical concept (any more than "time" is)? After all, a variable is just a variable, and mathematically nothing makes them any different from one another?
 
  • #6
Well, I would agree here. Of course the setup is made for probabilistic calculations and to grasp random processes. But the machinery is real analysis, although on special domains and with certain restrictions. The art of all is, and this cannot be solved by the machinery, to determine which specific setup describes a given random outcome. You can calculate your motions relativistic or classic, the equations cannot decide for you which one to use. I think the separation of a given experiment from the calculus necessary to describe it is actually an achievement which wasn't obvious from the start. It helps to avoid confusion and to encapsulate the discussion. And of course we won't need to talk about sigma algebras, if - as at school - some colored balls are selected from some pots. But in the end it led to a false understanding of randomness by equaling it with uncertainty. It's a bit as in physics, where old metaphors from a century ago are still in the heads of people.
 
  • Like
Likes fbs7
  • #7
Once again thank you so much for the kind and patient explanations to this old bloke here.

I see that many people here cherish teaching and explaining as the highest form of discourse, and for my part I try to understand as much as I can, to not waste even a word from such wonderful mentoring -- for which you have my dear thanks!

I guess that many people here are teachers in their areas, is that a reasonable assumption?
 
  • Like
Likes morix
  • #8
fbs7 said:
I guess that many people here are teachers in their areas, is that a reasonable assumption?
We have all kind of members, including teachers and current, future or former professors. Many others just have studied the fields they answer in. So yes, the general level of competence is significantly higher than on other internet forums. That does not mean that people wouldn't mistakes, at least they happen to me, usually due to bad reading, but those are normally quickly corrected. Me, too, finds it refreshing to read about things I have no expertise in, but can be sure it's on a scientific level.
 
  • #9
fresh_42 said:
As a programmer you should know that 1+1=0, so basic stuff is quite relative here.
In a one-bit adder, this is true, but only because we've lost the carry digit. In any other context, 1 + 1 = 2, in any base higher than 2, and 1 + 1 = 10, in base-2.
 
  • #10
Mark44 said:
In a one-bit adder, this is true, but only because we've lost the carry digit. In any other context, 1 + 1 = 2, in any base higher than 2, and 1 + 1 = 10, in base-2.
I thought of a Boolean variable. In many languages "IF X = true" can also be written "IF X = 1". In any case, I wanted to stress that even ##1+1=2## isn't automatically a given truth.
 
  • #11
fresh_42 said:
I thought of a Boolean variable. In many languages "IF X = true" can also be written "IF X = 1". In any case, I wanted to stress that even ##1+1=2## isn't automatically a given truth.
But strictly speaking, the operation on Boolean variables that corresponds to addition is OR, in which case true OR true = true. Using 1 for true, we have 1 + 1 = 1. If we take this further to include AND, we would have true AND true = true, as well. In neither case do we have 1 + 1 = 0.
 
  • #12
You're right. My only excuse is: far too many COBOL and RPG switches ...
 
  • Like
Likes jedishrfu and Mark44
  • #13
fbs7 said:
Oh... randomness is interpretation...

So a "random" variable is really just another variable, just like "time" is just a variable without anything different than say a "mass" variable. Our interpretation of "time" makes it special, but that meaning of something flowing over.. whatever... is not mathematical, it's just human interpretation.
I wouldn't say that. Whether a variable is truly random or is determined by unknown factors is a physics concern, not a mathematical concern. Mathematically, once a variable is assumed to be random, it is treated rigorously as such. It has a probability distribution, etc. The subject is well established and there are many good texts on probability theory.
 
  • Like
Likes StoneTemplePython
  • #14
FactChecker said:
I wouldn't say that. Whether a variable is truly random or is determined by unknown factors is a physics concern, not a mathematical concern. Mathematically, once a variable is assumed to be random, it is treated rigorously as such. It has a probability distribution, etc. The subject is well established and there are many good texts on probability theory.

Hmm... my mind screws were more in place with the idea that "random" is an interpretation thing... :-(

If I say x ∈ X, how do I know if this is a random variable or not? What makes this set X different than all other sets that can be defined in mathematics, that makes it a "set of random values" (or whatever the proper terminology for that), if not human semantic that attaches the word "random" to X?
 
  • #15
My point is that mathematics does not know or care why something has been identified as random, as long as it has been and a probability distribution of its values is defined. It may be "random" due to a human decision to rely on probabilities rather than attempt to figure out all the necessary physics to make it deterministic. We treat a coin toss as random because the idea of making it deterministic is inconcievable. There are innumerable examples like that. The location of raindrops on a square foot of ground would be random for all practical purposes.
 
  • #16
As is often the case, I look to Feller volume 1 for inspiration.

Feller said:
A function defined on a sample space is called a random variable... The term random variable is somewhat confusing; random function would be more appropriate (the independent variable being a point in the sample space, that is, the outcome of an experiment).

which directly contradicts this:
fbs7 said:
Oh... randomness is interpretation...
So a "random" variable is really just another variable, just like "time" is just a variable without anything different than say a "mass" variable...
Likewise, the "random" nature of a variable is just human interpretation, right?

The issue is: I don't think there is a satisfying B level answer to this thread.
 
  • Like
Likes FactChecker
  • #17
fresh_42 said:
You're right. My only excuse is: far too many COBOL and RPG switches ...
There is one Boolean operation that works like what you described: XOR. true XOR true = false. In symbols, 1 ⊕ 1 = 0.
 
  • #18
fbs7 said:
So a "random" variable is really just another variable, just like "time" is just a variable without anything different than say a "mass" variable.
FactChecker said:
I wouldn't say that.
I do agree. My point of view is
Orodruin said:
In general, I think the connection between probability theory and the measure theory is typically underemphasised in introductory courses on probability (at least for non-mathematicians). Also, just for OP's reference: https://en.wikipedia.org/wiki/Measure_(mathematics)
Once we setup the mathematical framework, we are in the middle of measure theory and the word random is obsolete. Distributions and other probability specific quantities are merely more properties and well defined functions.

To transport the term randomness from the experiment into the math does in my opinion more harm than good. It is neither necessary nor does it provide additional insights. Randomness is coded by the choice of specific measure spaces, a probability measure, resp. a distribution function. Once we arrived there, it will only be used to interpret the results in terms of the experiment again, but cannot affect calculations itself. Within mathematics, randomness, which we discuss here to be axiomatically defined, is in my opinion simply a synonym for a specific and deterministic calculus. Thus randomness is left behind as a property of the experiment only, and in this regard as a property of a variable same as time or mass would be.
 
  • #19
fresh_42 said:
I do agree. My point of view is

Once we setup the mathematical framework, we are in the middle of measure theory and the word random is obsolete. Distributions and other probability specific quantities are merely more properties and well defined functions.
I see your point. But general measure theory does not include the requirement that the total measure remain 1. That is the essential property that allows one to interpret it as a probability of a random variable. The scaling of Bayes' rule is to retain it as a probability.
 
  • #20
FactChecker said:
I see your point. But general measure theory does not include the requirement that the total measure remain 1.
This is why the random variable is from a domain ##(\Omega,\Sigma,P)## where ##P## is the probability measure, a bit more explicit on the German version of Wikipedia or on nLab.
 
Last edited:
  • #21
StoneTemplePython said:
As is often the case, I look to Feller volume 1 for inspiration.

which directly contradicts this:

The issue is: I don't think there is a satisfying B level answer to this thread.

Fair enough. So, in the world beyond B-level, if I have an independent variable x ∈ A and a function f(x) = ## \frac 1 { \sqrt { 2 \pi }} e ^ { - \frac { x^2 } 2 } ##, and I have another variable y ∈ B and a second function g(y) = ## \frac 1 { \sqrt { 2 \pi }} e ^ { - \frac { y^2 } 2 } ##, if I don't attach some human interpretation to the variables and formulas, how would I know if f(x) is a random function ( that for example expresses a normal probability distribution for the number of customers in a shop based on the amount of rain), and g(y) is a regular explicit formula ( that for example describes the exact, non-random, number of items some clockwork machine will build in 1 hour based on the hardness of the raw materials fed to it)?

That is, what are the mathematical qualities of the domains A and B (or the functions f(x) and g(x)) that make one related to "random" and "probability" and the other just another explicit formula?
 
Last edited:
  • #22
fbs7 said:
That is, what are the mathematical qualities of the domains A and B (or the functions f(x) and g(x)) that make one related to "random" and "probability" and the other just another explicit formula?
A random variable does not have a representation as a deterministic function. There is no "y=f(x)" giving a value of the variable y. The probability density function of a random variable does not give you the value of the variable; it gives the probability that the variable will have the value x.
 
  • Like
Likes Delta2 and Klystron
  • #23
As the discussion meanwhile reflects more of the time and school dependent interpretations, though still axiomatics, I will kindly ignore the "B" level, the more as the OP has received his answers. However, I think the debate itself is a fruitful one, as it appears that even the knowing differ on the interpretations. If so, then the dispute should take place. I'm almost certain the OP agrees with this htjack, especially as it is not a distractive subject but merely a distracted level.

FactChecker said:
A random variable does not have a representation as a deterministic function.
It has according to Wikipedia:
A random variable ##{\displaystyle f \colon \Omega \to \Omega'}## is a measurable function from a set of possible outcomes ##{\displaystyle \Omega }## to a measurable space ##{\displaystyle \Omega'}##. ##^*)##
Let ##{\displaystyle (\Omega ,\Sigma ,P)}## be a probability space and ##{\displaystyle (\Omega ',\Sigma ')}## a measurable space. A ##{\displaystyle (\Sigma ,\Sigma ')}-##measurable function ##{\displaystyle X\colon \Omega \to \Omega '}## is then a ##{\displaystyle \Omega '}##-random variable on ##{\displaystyle \Omega }##.
and nLab:
The formalization of this idea in modern probability theory (Kolmogorov 33, III) is to take a random variable to be a measurable function ##f## on a probability space ##(\Omega,P)## (e.g. Grigoryan 08, 3.2, Dembo 12, 1.2.1). ...##^*)##

So the random variable is a function on a configuration space and as such it is deterministic.

However
One thinks of ##\Omega## as the space of all possible configurations (all the “possible worlds” with respect to the idealized situation under consideration), thinks of the measure ##P(A)## of any subset of it as the probability that one of the configurations ##x\in A \subseteq \Omega ## is randomly realized, and thinks of ##f(x)## as the value of the given random variable in the situation of that configuration [##A##].
##^*)##

*) Variable names changed in accordance to previous posts. Emphasis mine.

Personally, I appreciate this modern view very much and wished I would have learned it this way. An analytical approach would have been far easier for me to understand as this mambo jumbo probability gibberish about ##X##, which I actually had encountered - friendly confused with combinatorics. In this sense I admit that there are different views around, especially historically and if distribution (probability measure ##P\,##), random variable (measurable function ##f\, : \,\Omega \longrightarrow \Omega'\,##) and randomness (so to say the sigma algebra ##\Sigma## over ##\Omega\,##) are not properly defined, or distinguished. But I definitely like the deterministic approach within a once set up calculus. ##f(A)## is different from ##P(A)##. So whether a random variable ##X## is considered to be ##X=f## or ##X=P## makes a difference here. I stay with Kolmogoroff and consider ##X=f## and ##P## the evaluation of ##A \in \Sigma##.
 
Last edited:
  • #24
I stand corrected. If one talks about functions on a very specialized space, then a random variable can be defined as a function on the set of possible outcomes. But I think this is a very specific setup designed to make it a deterministic function and is not at all what the OP would consider a general deterministic function. So IMHO, to imply that mathematics does not consider it a special case is misleading.

EDIT: Actually @fresh_42 's answer may, indeed, be what the OP was looking for. I may have underestimated the sophistication of his question since I have never thought of it this way.
 
Last edited:
  • #25
fbs7 said:
Fair enough. So, in the world beyond B-level, if I have an independent variable x ∈ A and a function f(x) = ## \frac 1 { \sqrt { 2 \pi }} e ^ { - \frac { x^2 } 2 } ##, and I have another variable y ∈ B and a second function g(y) = ## \frac 1 { \sqrt { 2 \pi }} e ^ { - \frac { y^2 } 2 } ##

my view is it's inappropriate for you to jump straight into continuous random variables. Start with coin tossing / Bernouli's. You can achieve remarkably sophisticated results with 0s and 1s. Moreover if you don't know what a Dedekind cut is (adjacent thread) you can't possibly understand what's going on with general random variables.

Speaking of coin tossing, there's probably a joke in here given the earlier discussion of bits, XORs, etc. and some of the comments made by @fresh_42 @fresh_##\mathbb F_2##
- - - -
As for the rest of the posts here, I think introducing measures right away is a mistake. Start with a discrete sample space and tease out information. Don't introduce random variables even in this setting until much later. Focus on the sample space and events, over and over. Really this is the core OP's question -- to understand the mathematical treatment of "randomness" you need to get your head around what's going on with these idealized experiments that are defined by sample space(s) -- that's where the "randomness" is modeled.

- - - -
A common theme in my posts is to use basic lightweight machinery, and only use heavier machinery if absolutely needed. It's part of the reason I use ##\text{GM}\leq \text{AM}## over and over. There's a similar idea with Feller vol 1.

fresh_42 said:
So the random variable is a function on a configuration space and as such it is deterministic.

fair but I already said this... I'll restate it with different underlining for others benefit:

Feller said:
A function defined on a sample space is called a random variable... The term random variable is somewhat confusing; random function would be more appropriate (the independent variable being a point in the sample space, that is, the outcome of an experiment).

again the 'randomness' lurks in the sample space.

There are a lot of people on PF who seem to say and think that probability is merely a special case of measure theory. (I'm not sure whether Fresh is one per se, but a forum search will see many others). I find this humorous as it seems to miss the point. Here's a nice zinger from a favorite blogger:

Tao said:
At a purely formal level, one could call probability theory the study of measure spaces with total measure one, but that would be like calling number theory the study of strings of digits which terminate. At a practical level, the opposite is true: just as number theorists study concepts (e.g. primality) that have the same meaning in every numeral system that models the natural numbers, we shall see that probability theorists study concepts (e.g. independence) that have the same meaning in every measure space that models a family of events or random variables. And indeed, just as the natural numbers can be defined abstractly without reference to any numeral system (e.g. by the Peano axioms), core concepts of probability theory, such as random variables, can also be defined abstractly, without explicit mention of a measure space; we will return to this point when we discuss free probability later in this course.

https://terrytao.wordpress.com/2010/01/01/254a-notes-0-a-review-of-probability-theory/

so for starting out: why not focus on probabilistic concepts as opposed to representation in terms of measures? If we have a discrete sample space we do have this choice and this is exactly where Feller vol 1 fits in.

(outside the scope thought: even in a discrete setting, dominated convergence can help streamline an awful lot arguments with stochastic processes... I just don't want to put the cart in front of the horse here)
 
  • Like
Likes Auto-Didact
  • #26
StoneTemplePython said:
There are a lot of people on PF who seem to say and think that probability is merely a special case of measure theory. (I don't think Fresh is one per se, but a forum search will see many others).
Meanwhile, I am. I find it far more transparent than counting colored balls! I had to learn stochastic in my second year by ...
StoneTemplePython said:
As for the rest of the posts here, I think introducing measures right away is a mistake. Start with a discrete sample space and tease out information.
... with the result, that it was incredibly tough to form a calculus. The terms likelihood, probability or random always remained foggy, badly defined terms and those in the know appeared to me like Merlin.
fresh_42 said:
An analytical approach would have been far easier for me to understand than this mumbo jumbo probability gibberish about ##X##, which I actually had encountered - friendly confused with combinatorics.
So in my opinion, this ...
StoneTemplePython said:
I think introducing measures right away is a mistake. Start with a discrete sample space and tease out information.
... is the mistake. However, it hits a notorious weakness of mine. I'm no friend of the common pedagogic concepts which proceed along the lines:
  1. It <insert a content of your choice, e.g. calculating 3-5 or introducing partial differentials, or complex numbers etc.> is impossible.
  2. It is too difficult for you.
  3. We will deal with it later.
  4. Btw., it is now possible, not difficult at all, and now is the time. :-p
I really hate this approach. It is based on the assumption of stupidity, and it fools students. In my opinion we should start to teach actual mathematics instead of procrastinate content over and over again. No wonder that people think ##17 - 25 \cdot 0## is mathematics!

I do not see any difficulties in the introduction of sigma algebras. Group theory and topology, too, are second year stuff and of comparable difficulty in my mind. Langauge may have to be adapted, e.g. by easy examples, but not content. As far as I see it, discrete and continuous random variables do not require a different treatment. The separation of randomness as part of the experiment, and not part of the calculus is rather appealing to me. I know that this might not be common sense, but as far as I'm concerned, it should be. The old fashioned methods didn't work well enough.
 
  • #27
Wow, so much brilliant and deep discussion here! Thank you! Either I'm going to learn something, or my brain will fry away! :biggrin:

I want to learn, so I went to read about Kolmogorov, de Finetti and Cox; Kolmogorov is way out of my league with "measures" and "σ-algebra" are way too abstract for me.

But I found Cox's postulates charming! They sound rather intuitive! Gotta love them 5 Cox postulates (as described in https://itschancy.wordpress.com/201...bayesian-statistics-part-two-coxs-postulates/)

(a) Cox-plausibilities are real numbers
(b) If two claims are equal in Boolean algebra, they have the same Cox-plausibility
(c) If two claims A and B and prior information X, then there exists a conjunction function f such that

A^B | X = f( A|X, B|A^X )

(d) etc...

That's brilliant! So this Cox guy didn't worry about a "random" nature of anything; he instead assumed "claims" that have a "plausibility". I can dig that! For example if I assume the claim "a 6-sided die can roll a 1" is "equal" (whatever that means) to the claim "a 6-sided die can roll a 2", etc..., then from (b) it follows that the plausibility that a 6-sided die can roll a 1 will be 1/6. No random dimgby-domgby or jingty-jumpty! I'll see if I can prove from (c) that the claim that two die rolls will add to 8 has a plausibility of 5/36!

Cox seemed to me to be a more accessible theory of probably than Kolmogorov, but from what I read most people seems to love Kolmogorov more than Cox, is that right?
 
  • #28
StoneTemplePython said:
- - - -
A common theme in my posts is to use basic lightweight machinery, and only use heavier machinery if absolutely needed. It's part of the reason I use ##\text{GM}\leq \text{AM}## over and over. There's a similar idea with Feller vol 1.
Regardless of the technicalities of this thread, reading Feller is a good recommendation in all cases.
 
  • Like
Likes StoneTemplePython
  • #29
StoneTemplePython said:
.. Moreover if you don't know what a Dedekind cut is (adjacent thread) you can't possibly understand what's going on with general random variables. ...

Yes, them Dedekind cuts got me good! And are still getting me. I kinda got (more or less) the example in Wikipedia on why √2 ∈ ℝ using Dedekind cuts -- that is to prove that whatever a/b such that (a/b)2 < 2, then there exists another rational p/q such that (p/q)2 < 2, therefore there is no upper rational value to the set of ℚ such that (a/b)2 < 2, therefore √2 ∈ ℝ.

I can read the formulas the Dedekind thingie, but understanding them is a different level. Like, when I read them, I immediately thought that whatever r ∈ ℝ and whatever ε, there will be a p/q ∈ ℚ such that (p/q - r) <= ε (because that's the construction of the Dedekind cut thingie)... so if I take the limit... lim ε->0 (p/q - r ) <= lim ε->0 ε = 0.. therefore in the lim ε->0 p/q = r... which is obviously wrong... brain blows up :frown:
 
  • #30
fbs7 said:
Yes, them Dedekind cuts got me good!
You might want to look for the alternative. It requires Cauchy sequences and equivalence classes. At least those are useful anyway, whereas Dedekind cuts are just this. Google "real numbers as Cauchy limits" or so.
 
  • Like
Likes fbs7
  • #31
fresh_42 said:
... is the mistake. However, it hits a notorious weakness of mine. I'm no friend of the common pedagogic concepts which proceed along the lines:
  1. It <insert a content of your choice, e.g. calculating 3-5 or introducing partial differentials, or complex numbers etc.> is impossible.
  2. It is too difficult for you.
  3. We will deal with it later.
  4. Btw., it is now possible, not difficult at all, and now is the time. :-p
I really hate this approach. It is based on the assumption of stupidity, and it fools students. In my opinion we should start to teach actual mathematics instead of procrastinate content over and over again. No wonder that people think ##17 - 25 \cdot 0## is mathematics!

I do not see any difficulties in the introduction of sigma algebras...

Yes, I picked up on this in a spat over prime numbers in the pre-calc forum. I actually am (semi) sympathetic to this in general.

I don't think it applies here though, in particular the underlined part.

Feller vol 1 does not assume stupidity on the part of the reader, is rigorous, is the book that got probability accepted by mathematicians outside the USSR, explicitly constrains itself to denumerable sample spaces to focus on probabilistic, not-analytic challenges (it even tells us that the sequel, vol 2 introduces measures to generalize the settings), contains an awful lot of analysis, includes original difficult results (e.g. Feller-Erdos-Pollard), and spawned real research (e.g. one I like: subsequent to publishing the first edition of vol 1, KL Chung pointed out that using countable state Markov Chain results from Kolmogorov, and a very well chosen chain, implies Feller-Erdos-Pollard) that was incorporated in version 2 and 3 of vol 1, either directly or as footnotes.

My approach of start simple and build really is close to how Polya would proceed (I think).
 
Last edited:
  • #32
I was just rolling in my bed, unable to sleep, and now I know the reason for that -- that's because the concept of "random" is finally getting in order in my head. What a trip this was! What makes sense to me is that there are really 3 concepts of "random", and people refer with the same word for completely different things:

(a) One is the is a well-structured, axiomatic, abstract mathematical structure that defines and studies "probability". This is not based on any actual dice rolling or some ghost taking cards out of a deck, but it is instead a logical system built around abstract concepts like a probability space and a measurement space. As such, it's rather beautiful. Here "random" doesn't really mean anything, as it's an axiom, and we could just as well call "bananility" instead of "probability" and the mathematical structure would be exactly the same.

(b) Another is the mysterious realm of quantum mechanics, where for some crazy odd reason real objects do seem to exactly follow laws derived from the abstractions above. Here "random" really means random, there's no other way to describe. Why quantum objects behave so is a mind-blowing question, and I suspect it's one of the greatest mysteries of physics, but thankfully everyday dudes like me don't have to worry about it and have no use for that, we can just have faith in physicists to get stuff to work by using those rules, and hopefully not blow the planet to pieces while doing that.

(c) Another is the macroscopic realm that we all handle everyday. Here the "random" really means unknown. There's no real random in macroscopic. We think the cards from the deck as random just because they are turned face down, and if we could calculate exactly all the forces acting in the dice we could deterministically predict which number would be rolled. One could despair with the unknown, but by making assumptions (like the deck is not missing cards and the cards are equally probable) and by applying that mathematical framework, we can make guesses and estimates of outcomes, which, if we assumed right will, in large numbers, be close to the mathematical predictions.

What's awesome is that (c) is routinely used by billions of people. It's actually very amazing if one considers of it - regular joes use it everyday, for example saying "wow, what a hail-mary pass -- he'll never be able to repeat that!" to express how unlikely p(x)2 is, without really knowing why that's correct. We joes don't care, and do not use, anything about tensors or Hilbert matrices or the 350-millionth digit on pi, but we use probability as commonly as we use algebra to verify the bills.

For that reason I now have renewed respect for probability theory, due to widespread use, and now I think that field is one of the Titans of mathematics, with the same practical utility as algebra and geometry! Once again thanks all for this very inspiring discussion!
 
  • Like
Likes Auto-Didact and FactChecker
  • #33
That is a good overview of the situation. One comment I would make is that the different ways of looking at it are really all compatible and complement each other. The axiomatic view really does describe the view of (c), even though the translation between the two may appear difficult. In a famous disagreement between Einstein and Neils Bohr, Einstein contended that (c) was the only situation and there was always a hidden, unknown cause for every "random" event. (I hope that I am not butchering this). A famous Einstein quote is "God does not play dice." At the quantum level, Bohr is considered the winner of that disagreement because of experiments by John Stewart Bell. I highly recommend that you look at the volumes "An Introduction to Probability Theory and its Applications" by Feller. They may be expensive and hard to find, but they are classics.

In the axiomatic view, it is irrelevant whether there is fundamental randomness or just unknown deterministic causes. The mathematics doesn't care. But a great strength of mathematics is that the logic and validity holds in many different applications.
 
Last edited:
  • Like
Likes StoneTemplePython and fbs7
  • #34
fbs7 said:
I was just rolling in my bed, unable to sleep, and now I know the reason for that -- that's because the concept of "random" is finally getting in order in my head. What a trip this was! What makes sense to me is that there are really 3 concepts of "random", and people refer with the same word for completely different things:

(a) One is the is a well-structured, axiomatic, abstract mathematical structure that defines and studies "probability". This is not based on any actual dice rolling or some ghost taking cards out of a deck, but it is instead a logical system built around abstract concepts like a probability space and a measurement space. As such, it's rather beautiful. Here "random" doesn't really mean anything, as it's an axiom, and we could just as well call "bananility" instead of "probability" and the mathematical structure would be exactly the same.

(b) Another is the mysterious realm of quantum mechanics, where for some crazy odd reason real objects do seem to exactly follow laws derived from the abstractions above. Here "random" really means random, there's no other way to describe. Why quantum objects behave so is a mind-blowing question, and I suspect it's one of the greatest mysteries of physics, but thankfully everyday dudes like me don't have to worry about it and have no use for that, we can just have faith in physicists to get stuff to work by using those rules, and hopefully not blow the planet to pieces while doing that.

(c) Another is the macroscopic realm that we all handle everyday. Here the "random" really means unknown. There's no real random in macroscopic. We think the cards from the deck as random just because they are turned face down, and if we could calculate exactly all the forces acting in the dice we could deterministically predict which number would be rolled. One could despair with the unknown, but by making assumptions (like the deck is not missing cards and the cards are equally probable) and by applying that mathematical framework, we can make guesses and estimates of outcomes, which, if we assumed right will, in large numbers, be close to the mathematical predictions.

What's awesome is that (c) is routinely used by billions of people. It's actually very amazing if one considers of it - regular joes use it everyday, for example saying "wow, what a hail-mary pass -- he'll never be able to repeat that!" to express how unlikely p(x)2 is, without really knowing why that's correct. We joes don't care, and do not use, anything about tensors or Hilbert matrices or the 350-millionth digit on pi, but we use probability as commonly as we use algebra to verify the bills.

For that reason I now have renewed respect for probability theory, due to widespread use, and now I think that field is one of the Titans of mathematics, with the same practical utility as algebra and geometry! Once again thanks all for this very inspiring discussion!

The late physicist E.T Jaynes wrote a provocative book "Probability Theory: the Logic of Science", Cambridge University Press, 2003, in which he essentially rejects the very idea of "randomness". That's right, a large probability book by somebody who does not believe in randomness! For Jaynes (and several others---maybe mostly physicists), probability is associated with a "degree of plausibility". He shows that using some reasonable axioms about how plausibilities combine, you can end up with multiplication laws like P(A & B) = P(A) P(B|A), etc. His book essentially tries to stay away from the whole "Kolmogorov" measure-theoretic way of doing probability, and so can only treat problems that do not involve things like ##P(\lim_{n \to \infty} A_n)## (but can certainly deal with things like ##\lim_{n \to \infty} P(A_n)##).

In his third chapter entitled "Elementary Sampling Theory", he says on pp. 73-74 (after developing the basic probability distributions):
"In the case of sampling with replacement, we apply this strategy as follows.
(1) Suppose that, after tossing the ball in, we shake up the urn. However complicated the problem was initially, it now becomes many orders of magnitude more complicated, because the solution now depends on every detail of the precise way we shake it, in addition to all the factors mentioned above.
(2) We now assert that the shaking has somehow made all these details irrelevant, so that the problem reverts back to the simple one where the Bernoulli urn rule applies.
(3) We invent the dignified-sounding word randomization to describe what we have done. This term is, evidently, a euphemism whose real meaning is: deliberately throwing away relevant information when it becomes too complicated for us to handle."

"We have described this procedure in laconic terms, because an antidote is needed for the impression created by some writers on probability theory, who attach a kind of mystical significance to it. For some, declaring a problem to be "randomized" is an incantation with the same purpose and effect as those uttered by an exorcist to drive out evil spirits; i.e., it cleanses the subsequent calculations and renders them immune to criticism. We Agnostics often envy the True Believer, who thus acquires so easily that sense of security which is forever denied to us."

Jaynes goes on some more about this issue, often revisiting it in subsequent chapters. Lest you think that his book is just "hand-waving", be assured that it is satisfying technical, presenting most of the usual equations that you will find in other books at the senior undergraduate and perhaps beginning graduate level (at least in "applied" courses). The man is highly opinionated and I do not subscribe to all he posits, but I find the approach interesting and refreshing, even though it is one I, personally, would not embrace. He does end the book with a long appendix outlining other approaches to probability, including the usual measure-theoretic edifice.
 
Last edited:
  • Like
Likes Auto-Didact, fbs7, jim mcnamara and 1 other person
  • #35
fbs7 said:
If I say x ∈ X, how do I know if this is a random variable or not?
Hi fbs:

This seems like a very strange question to me. If you say x ∈ X, you know something about x and X. Presumably you would know if x is a random variable if someone you believe to be knowledgeable tells you it is a random variable. What is needed by someone with the appropriate knowledge is that the process for obtaining values for x is a random process. So the randomness of a variable is determined by whether the process for obtaining values for the variable is a random process.

I am guessing you have some uncertainty about what it means for a process to be random. A random process is a process for which it is impossible by any means to know in advance what a particular value will be. This is the distinction between a random process and a pseudo-random process. If the process is pseudo-random, and you know the nature of this process and its initial conditions, in principle you can calculate the next value it will generate.

I hope this is helpful.

Regards,
Buzz
 

Similar threads

Replies
5
Views
614
Replies
31
Views
5K
Replies
7
Views
3K
Replies
7
Views
3K
Replies
1
Views
2K
Replies
7
Views
3K
Replies
9
Views
3K
Replies
4
Views
3K
Back
Top