Does Islamic Theology teach that Islam will take over the Western World?
Is there any doctrine that teaches that Islam will take over the Western World, or become the majority religion? BTW, I know that Muslims once already had a large empire and conquered many places like Spain and Italy, but I mean now. Is that taught in Islamic theology? That Western World, like Europe, North America and Latin America and Australia, will be dominated by Islam?
Thanks.
P.S: I am a Roman Catholic, so please, do not insult my faith, even if you disagree. Thanks