Science topic

Joint Probability Distribution - Science topic

Explore the latest questions and answers in Joint Probability Distribution, and find Joint Probability Distribution experts.
Questions related to Joint Probability Distribution
  • asked a question related to Joint Probability Distribution
Question
1 answer
I want to understand how to configure and apply such a model to identify significant or popular locations in a geographic dataset.
Relevant answer
Answer
Bayesian networks offer a powerful way to extract Points of Interest (POIs) from geographic datasets. They work by modeling the probabilistic relationships between different factors that hint at the presence of a POI. For instance, if a location is frequently visited during lunchtime, with users staying there for a significant duration, it's likely a restaurant. A Bayesian network can capture this "pattern" and use it to identify potential restaurants across a dataset. To configure a Bayesian network for POI extraction, we first need to conceptualize our model. Think of the network as a map of variables and their connections. Variables might represent the location itself (which we can divide into a grid), the time of day, the dwell time (how long someone stays), and the hidden variable, the actual POI we want to uncover. The connections between these variables are defined by probabilities. For example, there's a higher probability of a long dwell time at a true POI compared to just passing through a location.
Before training the model, we feed it some initial probabilities. These probabilities might be quite basic (perhaps a general sense of where different types of POIs tend to cluster). After this, the real magic happens. Using algorithms like Expectation-Maximization, the Bayesian network learns from the GPS data, adjusting its probabilities to best explain the observed patterns. Once trained, we can ask the Bayesian network to do the inference. Given a location, time, and dwell time, it can tell us the most probable type of POI present. By looking at how often a location is flagged as different POIs across a multitude of users, we can even get insights into the popularity of those locations.
  • asked a question related to Joint Probability Distribution
Question
4 answers
I understand how Gibbs sampling works to allow estimations of joint probability distributions in a simple sense. For example, using a simple two dimensional example (A and B are two different tennis players whose results are not independent, and 0=losing, 1 = winning), you might start with the following conditional probabilities:
P(A=0| B=0) = 0.5,
P(A=0| B=1) = 0.125
P(B=0|A=0) = 0.5
P(B=0| A =1) = 0.125
To get things moving, you would then suppose a starting value for A (let it be 0). We can then use that value of A (A=0) for the next iteration for B, where we therefore look at P(B=0|A=0). As shown above, there is a 0.5 probability that B=0 in that case. Let's run a random number (0-1) generator, which happens to yield 0.67. As this is greater than the probability of 0.5, we take B=1 (using the rule that if the random number is lower than the conditional probability, 0 is yielded, and if the random number is greater than the conditional probability, 1 is yielded). This gives us the first pair of joint values: A=0 and B=1 [which can be written as 0,1]. We now run the next iteration for A using that last value of B (B=1). P(A=0|B=1) is 0.125. The random number generator yields 0.28, so we take A =1. We then look at P(B=0|A=1) [as the last value of A yielded was 1], which is 0.125. The random number generator yields 0.34, which means we take B to be 1. So our second pair of values is: A=1, B=1 [or, 1,1]. We can repeat this process for a very large number of iterations, and if we then count the numbers of paired values that are 0,0; 0,1; 1,0; and 1,1, we should be able to estimate the joint probability distribution. I have attached a simple excel program that carries out such a Gibbs sampling. It can be seen that the estimation of the joint probability distribution is very close to the actual joint probability distribution from which the conditional probabilities were calculated (in practice, of course you wouldn't have access to the true joint probabilities as then you'd have no reason to do the Gibbs sampling). I have largely based my example on an excellent YouTube video by Ben Lambert.
However, this is where I need advice and help. I do not understand how the above example relates to network meta-analyses (NMAs). For example, imagine a network meta-analysis of three treatments A, B and C. How do the data from these studies relate to conditional probabilities? For example, if the odds ratio of outcome X is 0.5 for the comparison of A vs B, the odds ratio of outcome X is 0.2 for the comparison of B vs C, and the odds ratio of outcome X is 0.1 for the comparison of A versus C (clearly no incoherence here!), how do we proceed? I have a vague idea that we could use the odds ratios to calculate conditional probabilities, but can't quite grasp exactly what should happen. I have looked at most of the relevant documents (like the DSU document TSD2) but these don't explain exactly what occurs in the Gibbs sampling itself. Can anyone describe, in simple terms, how the sampling would proceed in an NMA, in relation to the model of Gibbs sampling I have given earlier?
Relevant answer
Answer
Many thanks Mohammad!
  • asked a question related to Joint Probability Distribution
Question
6 answers
Could any expert try to examine our novel approach for multi-objective optimization?
The brand new approch was entitled "Probability - based multi - objective optimization for material selection", and published by Springer available at https://link.springer.com/book/9789811933509,
DOI: 10.1007/978-981-19-3351-6.
Relevant answer
Answer
  • asked a question related to Joint Probability Distribution
Question
5 answers
I would like to fit a trivariate joint probability distribution using the nested three-dimensional Copula in MATLAB. To this end, their respective marginal distribution and two dimensional joint distribution have been modelled. I wonder if there is any Copula code or toolbox available to model a multivariate (three or more) joint probability distribution in MATLAB. Thank you in advance!
Relevant answer
Hi Hao
I suggest you MVCAT toolbox which s provided by @Mojtaba Sadegh
the link below you can download it and the corresponding manual
  • asked a question related to Joint Probability Distribution
Question
9 answers
Suppose we have statistics N(m1, m2), where m1 is the value of the first factor, m2 is the value of the second factor, N(m1, m2) is the number of observations corresponding to the values ​​of factors m1 and m2. In this case, the probability P(m1, m2) = N(m1, m2) /K, where K is the total number of observations. In real situations, detailed statistics N(m1, m2) is often unavailable, and only the normalized marginal values ​​S1(m1) and S2(m2) are known, where S1(m1) is the normalized total number of observations corresponding to the value m1 of the first factor and S2(m2) is the normalized total number of observations corresponding to the value m2 of the second factor. In this case P1(m1) = S1(m1)/K and P2(m2) = S2(m2)/K. It is clear that based on P1(m1) and P2(m2) it is impossible to calculate the exact value of P(m1, m2). But how to do this approximately with the best confidence? Thanks in advance for any advice.
Relevant answer
Answer
For your normalising constant or marginal distribution of vector case, you may use saddle point approximation, see e.g., http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.123.1487&rep=rep1&type=pdf
  • asked a question related to Joint Probability Distribution
Question
6 answers
How to find the distance distribution of a random point in a cluster from the origin? I have uniformly distributed cluster heads following the Poisson point process and users are deployed around the cluster head, uniformly, following the Poisson process. I want to compute the distance distribution between a random point in the cluster and the origin. I have attached the image as well, where 'd1', 'd2', and 'theta' are Random Variables. I want to find the distribution of 'r'.
Relevant answer
Answer
It gets a bit messy algebraically, but I recommend that you transform your three random variables (d_1, d_2, theta) into three new variables (r, r_c, phi_1). the variable "r" is computed using the Law of Cosines, r_c is just d_2 and phi_1 is the angle formed between r and d_2. Using the Law of Sines, we have
d_1 / sin(phi_1) = d_2 / sin(phi_2).
phi_2 is the angle between r and d_1.
Realizing that phi_2 = pi - theta - phi_1, we can eliminate phi_2 from the Law of Sines and after some messy manipulation, find phi_1 as a function of d_1, d_2 and theta.
You then construct the Jacobian matrix d(r, rc, phi_1)/d(d_1, d_2, theta). The transformed PDF is given by
p(r, rc, phi_1) = p(d_1, d_2, theta) * det(J)^{-1}
where det(J) is the determinant of the Jacobian J.
The PDF for r requires you to integrate out the rc, phi_1 variables:
p(r) = int(p(r, rc, phi_1)*det(J)^{-1} d.rc d.phi_1)
Hope this helps,
Cheers!
  • asked a question related to Joint Probability Distribution
Question
7 answers
I have several random variables X = [x1, x2,...... xn] represented by columns in a Data matrix and rows are representing random samples. I also have the marginal Probability Density Functions as f(x1), f(x2), ... f(xn) for individual random variables. I would like to calculate their joint PDF as f(x1,x2,....xn).
Relevant answer
Answer
Hello Look for the probability density function in help, it depends of course on the distribution Good luck
  • asked a question related to Joint Probability Distribution
Question
5 answers
Im working with Nataf model trying to fit a joint probabilistic model for circular and Linear variables, but I have some difficulties in calculating the correlation matrix because, I could find an equation for calculating the equivalent correlation between two circular variables or between a circular variable and a linear variable.
Relevant answer
Answer
  • asked a question related to Joint Probability Distribution
Question
4 answers
What is the best way to calculate joint probability distributions from multiple discrete probability distributions?
I need to calculate the combined or joint probability distribution of a number of discrete probability distributions.
Xn={x1;x2;x3...}; P(Xn)={P1;P2;P3...} for n>1.
With regards to the discrete probability distributions: Values must be summed; Probabilities must be multiplied.
Is there a faster way than merely multiplying each discrete probability distribution with each other?
Relevant answer
Answer
Hi Brian Barnard,
Generally speaking, your question is difficult. However, if the correlations between variables are small, or the structure of the discretized random variables are not too complicated, there are some references for you, e.g.
This paper provides you a thorough consideration about how to construct joint probability from discrete marginal ones and given correlations. You need to verify that such a construction from your case gives a real joint probability distribution.
Thanks,
Biao
  • asked a question related to Joint Probability Distribution
Question
7 answers
Assume that X, Y, and Z are identical independent Gaussian random variables. I'd like to compute the mean and variance of S=min{P, Q}, where :
Q=(X-Y)2,
P=(X-Z)2.
Any help is appreciated.
Relevant answer
Answer
Dear Professor Joachim Domsta ,
Thank you for your suggestion, I am agree with your solution except in the first line. The attachment is what I've done so far. Please take a look at it and correct me if I'm wrong
  • asked a question related to Joint Probability Distribution
Question
1 answer
I have three sets of variables. Two of them are continuous random variable and the other one is discrete in nature. Is there any matlab function which returns the joint probability distribution of these three random variables? Kindly explain with an example.
Relevant answer
Answer
Hi Debanjan,
Did you find a good answer for your question?
  • asked a question related to Joint Probability Distribution
Question
4 answers
Hi there,
I'm trying to fit rainfall data to incomplete gamma distribution. So, I don't know how to proceed.
Should I estimate the shape and scale parameters before the Lelleifors test? (i.g. using my data set to fit gamma to it).
Or should I first find out Lilliefors results? I'm confuse.
Kind regards,
Jefferson
Relevant answer
Answer
Dear Jefferson,
this was a misunderstanding. I meant that you could estimate the parameters of the gamma distribution by means of the Kolmogorov-Smirnov minimum distance estimator using R package distrMod and then use these parameters in combination with function ks.test to perform the Kolmogorov-Smirnov test.
Best,
Matthias
  • asked a question related to Joint Probability Distribution
Question
14 answers
is it just possibility versus probability?
Relevant answer
Answer
It is not the probability that has a distribution. Probability is a normalized measure over a sampling space (with an associated sigma algebra). The sampling space is the "set of all possible outcomes", thus giving all "possibilities" that are considered as "measurable". The underlying algebra defines that outcomes can itself be sets of outcomes. The outcomes that contain only a single element are called "elementary outcomes (or events)".
The probability distribution is a feature of a random variable. A random variable is a function returns a numeric value. In contrast to a "usual" function, the value returned from a random variable is not a simple single number. It returns a whole set of numbers, one for each elementary outcome. The key is that it all these values together with an associated probability.
Let's look at a sampling space with the elementary outcomes "male" and "female". They may be called the "possibilities", at it is possible that the attribute we observe maybe  male or female. We might further consider other possibilities, like "hermaphrodite", "intersex", "sexless". Deciding on this is based on subject knowledge. Here we keep it simple and focus on only two possibilities.
We can now define a random variable X that returns 0 for male and 1 for female (we could take other numeric values; it's only a convenient choice!). We must further define the probability distribution of this random variable, which associates the returned values with probabilities, like P(X=1)=0.3 and P(X=0)=0.7. For this random variable we can calculate moments like the expected value and the variance. Such a random variable is called a Bernoulli variable, and the probability distribution is called the Bernoulli distribution. This distribution is fully defined by a single value, giving the probability of X=1, because the other probability follows from the axioms. 
If there are many or even an infinite number of possibilities, like the counts of something (the outcomes can be any natural number), the random variable will return as many different values (in the case of counts it makes pretty much sense to make the random variable just return these numbers!). The individual assignment of probabilities in such cases is cumbersome or impossible, so that one assigns the probabilities as a function of the values the random variable can take. These functions are sometimes derived from simple assumptions. For counts one can imagine the interval as an infinite number of small sub-intervals in which at most one event may happen, so that these outcomes can be modelled by a Bernoulli distribution (with an infinitively small probability for X=1). A bit of caclulus will lead us to the solution for the counts, which is known as the Poisson distribution that is defined by the parameter lambda. Given the value of lambda we can calculate the probability assigned to any number the ransdom variable can take (0, 1, 2, 3,  ...).
When we allow an uncountably infinite number of outcomes, like real numbers we get from physical measurements, we facte the problem it is impossible to assign a finite positive probability value to all the values the random variable can take without violating the axioms. The trick here is to see that the integral of the values must be 1, and we can define a density that integrates to 1 over the domain of the random variable. Instead of assigning a (finite positive) probability value to a point within the domain we assign a probability density f(X=x), and probabilities are then given for integrable parts of the domain. The integral of f from -Inf to x, F(X<x), is called the probability function. Usually the probability distribution is given in form of the density, somtimes in form of the probability function. Such random variable are called continuous, they have a continuous domain and a continuous probability distribution, respectively.
If you measure the height of a person you get numeric values, like 1.6732 m etc. The values will be limited in precision, so you won't really have an continuum of possibilities. However, it is convenient to assume it were a continuum.It again makes pretty much sense to define a random variable that returns the height values. We may not restrict the domain, so that the random variable can return for instance negative values or values that correspond to impossibly large heights. The usefulness may still be assured by assigning a probability distribution that will have extremely low densities for such cases. A probability distribution that has these features is the normal distribution. One may also restrict the domain, either to just positive values, or to value above a lower limit, or between a given lower and upper limit. There are different density functions available for these cases, like the gamma- and the beta-distributions.
  • asked a question related to Joint Probability Distribution
Question
2 answers
Two systems give uncorrerlated or less correlated outputs while their inputs show some correlated behavior. So how to transform and find some complex relationship between inputs-outputs which can possibly give good correlation and copula dependence for outputs.
Relevant answer
Answer
Not sure I have understood the question as it seems to self contradict (which may be more my lack of insight) so i will rephrase how I think you mean. 
You start with data D1 and D2 as inputs into systems S1 and S2 respectively and you obtain outputs O1 and O2 respectively.
D1 and D2 are highly correlated. O1 and O2 are not well correlated. So S1 and S2 are diverging in their behaviour.
You expect S1 and S2 to behave similarly, therefore you were expecting O1 and O2 to be highly correlated.
You want to know how to mathematically transform S1 and S2 so that the produce better agreement.
Is this correct?
If not apologies for misunderstanding.
If yes, is your D1 and D2 signal based or at least consist of ordered variables where neighbouring variables have a strong covariance?
  • asked a question related to Joint Probability Distribution
Question
5 answers
Hello, I have a question regarding negative binomial (NB) regression. I am not sure if I can include predictor variables that are correlated with the exposure variable (say time)? I'm concerned that the predictor variable (VIF 3.8) is correlated to the exposure variable (VIF 9.8). I have carried out NB regression despite the collinearity and the results are significant. The overall likelihood ratio test is 6.377, df = 1 and sig = 0.012. Can I include the predictor variable in this case? 
Relevant answer
Answer
You can if you want. If there is a rationale behind having this predictor in the model: keep it. If not: why did you consider this variable as a predictor at all?
What is the model for? If, for instance, the model is for prediction, then test it (with and without the suspicios predictor) on a test set and see iwhich one performes better w.r.t. your requirements.
  • asked a question related to Joint Probability Distribution
Question
5 answers
What is the difference between joint distribution function and likelihood function?
Relevant answer
Answer
Let X be a random variable having probability density function f(.,theta) and X_1,X_2,...,X_n be a random sample from f(.), then joint distribution function of this sample be f(X_1,X_2,...,X_n,theta). If you loked this function as a function of theta i.e.f(theta; (X_1,X_2,...,X_n) , then this is called as likelihood.
The likelihood function is defined as the joint density function of the observed data treated as a functions of the parameter theta.
According to Lehmann, the likelihood function is a function of the parameter only, with the data held as a fixed constant. Here noted that the likelihood is not the probability of the parameter .
  • asked a question related to Joint Probability Distribution
Question
1 answer
My research is on coincident flooding using joint probability method but I have limited background knowledge of joint probability. I an expecting a sample of calculations.
Relevant answer
Answer
see this paper of a colleague of mine. It may help, 
  • asked a question related to Joint Probability Distribution
Question
12 answers
Using model f to obtain joint PDF of two parameters.
Relevant answer
Answer
@Adeyemi Aladejare
I think you should look the book “Bayesian Data Analysis” (2nd or 3rd edition) of Guelman et al.. It is really easy to follow and nice examples are discussed. You have a whole part in regression models that include Hierarchical and Generalize linear models. Please check it up the book’s homepage:
When you say “I have a regression model like E=aC+Ɛ.”, is C a random variable also? Or you mean E|c = a*c+ Ɛ, because this last exemple is just a case of a simple linear regression.
  • asked a question related to Joint Probability Distribution
Question
4 answers
X and Y are independent, identically distributed log-normal random variables. How can I get the PDF of Z where Z=abs(X-Y)?
Relevant answer
Answer
I am not an expert in this. To the best of my knowledge the standard quadrature formulas like trapezoidal rule (see http://en.wikipedia.org/wiki/Trapezoidal_rule ) are applicable to integrals over finite intervals, so one must first find an appropriate approximation of the integral over the semi-axis by the integral over large interval [0, R] and then to apply a quadrature formula. Just a minute ago I have made a Google search with words
``quadrature formulas for improper integral'' and it gives a lot of references. For example,
 S.Haber, O Shisha:  Improper integrals, simple integrals, and numerical quadrature, Journal of Approximation Theory Volume 11, Issue 1, May 1974, Pages 1–15
  • asked a question related to Joint Probability Distribution
Question
25 answers
Hi
I want to calculate joint probability P(x,y,z) based on P(x,y), P(x,z), P(z,y), P(x), P(y), P(z). How can I do?
x, y, z are three random variables and I compute their probability by estimation.
Relevant answer
Answer
Dear Mohammad
The best way to estimate joint probability density functions is to:
1) first estimate the marginal distributions one-by-one
2) Select a copula family and find the best parameters of the latter.
Gaussian and Student copulas are easy to parameter in any dimensions. Archimedean copula can be defined easily in 2D and extended to larger dimensions in some cases.  Selecting the best copula (from data) that "glues" the marginal distributions into a joint PDF is not an easy task since there are no simple goodness-of-fit tests. However the Kendall plot is a useful tool that allows to compare different fitting.
Good luck!
References: Nelsen, R. B. An introduction to copulas Springer, 2006.
Fermanian, J. D. Goodness-of-fit tests for copulas J. of Multivariate Analysis, 2005, 95, 119-152
  • asked a question related to Joint Probability Distribution
Question
3 answers
If anyone has this please give me some idea.
Relevant answer
Answer
Don't just dive straight into JPDA. Start with the nearest-neighbor Kalman filter, then look at Probabilistic Data association (PDA). Make sure you understand how all the events are enumerated and their weights evaluated in PDA (for the single-target case) before you move on to JPDA (multi-target case) - otherwise you will just get confused. In PDA it helps if you appreciate that a single Gaussian is being fitted to a linear combination of Gaussians (a 'mixture') at the end of each update. The fit is done so that the 1st and 2nd moments of the single Gaussian and the many-component mixture are the same.  
  • asked a question related to Joint Probability Distribution
Question
13 answers
Hello,
Maybe this is a very easy question, maybe not. I have a time-descrete stochastic process X = (Xt). Each Xt has a different pdf, so it is not i.i.d. All Xt have the same sample space and the pdfs are constructed of the same sample size. So now I don't want to have the joint probability function, I want to have the pdf of all realisations of all Xt collected together, as if there was no difference in time. How can I get this "summed up" pdf out of the separate pdfs?
Thanks in advance
Relevant answer
Answer
If I understand your question correctly, you do not want the pdf of the sum of the random variables. The previous responses are all related to how to find the pdf of the sum of several random variables. Your question (it seems to me) is how to find the mixture of two or more random variables - see attached figure. Let me know if my interpretation of your question is correct.
  • asked a question related to Joint Probability Distribution
Question
12 answers
I have a random vector whose joint probability distribution is known. However, I would like to sample this vector so that it lies within a convex polytope which can be represented by a set of linear inequalities, such as Ax= b , x>=0. I can naturally use rejection sampling, but the rate of acceptance is very small (<< 1% for practical size applications). I have come across a simple way of uniformly sampling from the unit simplex, and I think this may be adapted to a general convex polytope, but I can't figure out how.
Relevant answer
Answer
Any point in a convex polytope can be expressed as a convex linear combination of the n vertices. The coefficients of these convex linear combinations are positive and add up to 1. They are called barycentric coordinates. Therefore, they are in the n-part simplex S^n, which is (n-1)-dimensional. In compositional data analysis some families of logistic transformations from S^n into R^{n-1} are frequently used. They are called alr (additive-log-ratio, Aitchison 1986), ilr (isometric log-ratio, Egozcue et al. 2003) or even clr (centered log-ratio, Aitchison 1986). All of them are invertible, so that a simulated sample in R^{n-1} can be transformed, first to the simplex S^n via alr^{-1}, ilr^{-1} or clr^{-1}, and then into the polytope via barycentric coordinates.
For your specific case I would recommend to use alr transformation (also known as logistic transformation; its use is also frequent in logistic regression). If X=(x_1, x_2, ..., x_{n-1}) is a vector in R^{n-1}, Y=alr^{-1}(X)= C exp[x_1, x_2, ..., x_{n-1},0] , where C is a normalising constant such that the X's components add to 1 and Y is in S^n. The alr-transformation is defined as X = log[y_1/y_n, y_2/y_n, ..., y_{n-1}/y_n].
If you know the distribution you want to simulate in the coordinates Y the problem is solved. However, I guess that your distribution is defined and known on R^n and then restricted to the polytope (e.g. a multivariate normal distribution). This kind of choice is frequently based on the convenience. In this case there is no convenience at all. I would recommend to define a (n-1)-multivariate normal distribution on the alr-coordinates, simulate it, and then translate it into the polytope: R^{n-1}--alr^{n-1}--> S^n -- barycentric coord --> polytope.
  • asked a question related to Joint Probability Distribution
Question
1 answer
Suppose I have a random vector X = (X_1, X_2, ..., X_n) for which X_i ~ Po(\lambda_i), but the components are not independent, so that the joint distribution is not the product of the marginal distributions, and the covariance matrix of X is not diagonal. It seems the joint distribution does not have an explicit form. I would like to obtain some properties from this distribution, as mean, mode, and variance. How can I sample from it? Is there any good approximation to it other than the multivariate normal?
Relevant answer
Answer
since they are not independent joint distribution may be obtained through conditional distribution.
if X with pmf g and pmf of Y given X is h() then f(x,y)=g(x).h(y/x)