Answer:
False. See te explanation an counter example below.
Step-by-step explanation:
For this case we need to find:
for all and for in the Markov Chain assumed. If we proof this then we have a Markov Chain
For example if we assume that then we have this:
Because we can only have if we have this:
, from definition given
With we have that
So based on these conditions would be 1 with probability 1/2 from the definition.
If we find a counter example when the probability is not satisfied we can proof that we don't have a Markov Chain.
Let's assume that for this case in order to satisfy the definition then
But on this case that means and on this case the probability , so we have a counter example and we have that:
for all so then we can conclude that we don't have a Markov chain for this case.