What does Africa America mean?

Africa America meaning in Urban Dictionary

Any an element of the united states of america this is certainly entirely inhabited by black colored individuals in which HELPS, benefit, medications, physical violence and rape tend to be daily events. Much like Africa.