One of the main problems of topological localization in a real indoor environment is variations in the environment caused by dynamic objects and changes in illumination. Another problem arises from the sense of topological localization itself. Thus, a robot must be able to recognize observations at slightly different positions and angles within a certain topological location as identical in terms of topological localization. In this paper, a possible solution to these problems is addressed in the domain of global topological localization for mobile robots, in which environments are represented by their visual appearance. Our approach is formulated on the basis of a probabilistic model called the Bayes filter. Here, marginalization of dynamics in the environment, marginalization of viewpoint changes in a topological location, and fusion of multiple visual features are employed to measure observations reliably, and action-based view transition model and action-associated topological map are used to predict the next state. We performed experiments to demonstrate the validity of our proposed approach among several standard approaches in the field of topological localization. The results clearly demonstrated the value of our approach.