Using scipy, I'd like to get a measure of how likely it is that a random variable was generated by my log-normal distribution.
To do this I've considered looking at how far it is from the maximum of the PDF.
My approach so far is this: If the variable is r = 1.5, and the distribution σ=0.5, find the value from the PDF, lognorm.pdf(r, 0.5, loc=0). Given the result, (0.38286..), I would then like to look up what area of the PDF is below 0.38286...
How can this last step be implemented? Is this even the right way to approach this problem?
To give a more general example of the problem. Say someone tells me they have 126 followers on twitter. I know that Twitter followers are a log-normal distribution, and I have the PDF of that distribution. Given that distribution do I determine how believable this number of followers is?