I think this is because we can write P(T1=t1, ... Tn=tn | W1=w1, ... Wn = wn) in the following way:
P(T1=t1, ... Tn=tn, W1=w1, ... Wn = wn) / P(W1=w1, ... Wn = wn)
For a fixed word sequence, P(W1=w1, ... Wn = wn) will be fixed. Thus, maximizing P(T1=t1, ... Tn=tn | W1=w1, ... Wn = wn) will be the same as maximizing P(T1=t1, ... Tn=tn,W1=w1, ... Wn = wn).
P(T1=t1, ... Tn=tn, W1=w1, ... Wn = wn) / P(W1=w1, ... Wn = wn)
For a fixed word sequence, P(W1=w1, ... Wn = wn) will be fixed. Thus, maximizing P(T1=t1, ... Tn=tn | W1=w1, ... Wn = wn) will be the same as maximizing P(T1=t1, ... Tn=tn,W1=w1, ... Wn = wn).