How to understand special relativity

Yesterday I saw a video 400k reward winning video explaining SR to people have no physics background.
Here is my two sentences attempt for those who want to derived relativity on their own.

1. You, have to give a simpliest self consistent transform between two inertial system with light speed always the same for all observers in different inertial system.

2. You, don’t know about light speed invarience. But you need to try to make electromagnetic field law remain the same regardless of frame of reference.

This also explained the different between the discovery of SR and Lorentz transform, bcuz the reasoning process was slightly different.


quantum eraser and time related entanglement

the three facts make quantum eraser thought experiments easy enough for figuratively thinking .and thpse made it not surprising anymore.
btw the Pogaray?? qbit sphere is a very good thinking tool.

1 the which way information it self is an two level sys and
take right and left as 1 and 0. normal interference patter is the result of being perpandicular to this 1 and 0.

2 intanglement only has something to dp with the correlation of two measurements no mayter to space or time
3 the quantum state do have noth to do with it history. once *measured* it points to one direction of the pongaray sphere

Making Sense of the world … and other parts of the brain

It’s a paper about how the brain work in the light of Bayesian Brain Hypothesis, i.e. brain modules have hierarchical internal model of the input so that they infer the cause and back propagate this expectation.

It’s also a paper about how to think biological problems in a more physical way, i.e. transform and divide problem into smaller problems, and try to use simpler assumptions to go from phenomenological description to model intepretation.

And not least important, how the highly adaptive nature of neural network contribute to intelligence by the neocortex.

Daily Learning note – Why should we be Bayesian for Machine Learning



Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric?

author: Zoubin Ghahramani, Department of Engineering, University of Cambridge
published: Oct. 9, 2008,   recorded: September 2008,   views: 2789
  • Why be Bayesian?
    • we want to represent the strength of those beliefs numerical in the brain of any artifical agent
      • The Cox Axioms (Desiderata)
    • The Dutch Book Theorem (?? don’t really understand the example)
    • * Asymptotic Certainty
    • Asymptotic consensus
    • Bayesian Occam’s Razor and Model Comparison
    • Potential advantages
      • tries to be coherent and honest about uncertainty
      • easy to do model comparison, selection
      • rational process for model building and adding domain knowledge
      • easy to handle missing and hidden data
  • Where does the prior come from?
    • Objective Priors:
      • Non informative priors that attempt to capture ingnorance and have good frequentist properties.( not very good and helpful)
    • Subjective Priors:
      • Priors should capture our beliefs as well as possible . They are subjective but not arbitrary.
    • Hierarchical Priors:
      • multiple levels of priors:
      • (parameters and hypoparameters)
    • Empirical Priors:
      • Learn some of the parameters of the prior from the data
    • Two views of machine learning
      • The Black Box View: (general and user don’t need to think too much)
      • The Case study view: ( really try to understand the problem)
    • Bayesian Black Boxes?–(where is the prior??)
    • Parametric vs Non-parametric Models
      • Parametric models: model-based, have a finite fixed number of parameter {\theta}
      • Non-para: allow the number of “parameters” to grow with the data set size,  (also: memory-base learning) (e.g. kernel density estimation)
    • example: Infinite mixture models.
    • Is non-parametric the only way to go?
      • When do we really believe our parametric model?
      • But, when do we really believe or non-parametric model?
      • Is a non-parametric model (e.g. a DPM) really better than a large parametric model( e.g. a mixture of 100 components)?
    • The approximate inference Conundrum
      • All interesting models are intractable.
      • So we use approximate inference (MCMC, VB, EP , etc).
      • Since we often can’t control the effect of using approximate inference, are coherence arguments meaningless
      • Is Subjective Bayesianism pointless?
    • Reconciling Bayesian and Frequentist Views
      • Frequentist theory tends to focus on sampling properties of estimators, i.e. what would have happened had we observed other data sets from sets from our model. Also look at minimax performance of methods – i.e. what is the worst case performance if the environmnet is adversarial. Frequentist methods oftern optimize some penalized cost function.
  • How do we do these integrals?


* ELI5: frequentist and Bayesian:

Here is how I would explain the basic difference to my grandma:

I have misplaced my phone somewhere in the home. I can use the phone locator on the base of the instrument to locate the phone and when I press the phone locator the phone starts beeping.

Problem: Which area of my home should I search?

Frequentist Reasoning:

I can hear the phone beeping. I also have a mental model which helps me identify the area from which the sound is coming from. Therefore, upon hearing the beep, I infer the area of my home I must search to locate the phone.

Bayesian Reasoning:

I can hear the phone beeping. Now, apart from a mental model which helps me identify the area from which the sound is coming from, I also know the locations where I have misplaced the phone in the past. So, I combine my inferences using the beeps and my prior information about the locations I have misplaced the phone in the past to identify an area I must search to locate the phone.