Meaning of Germany

Definition of Germany

(noun) a republic in central Europe; split into East Germany and West Germany after World War II and reunited in 1990

Other information on Germany

WIKIPEDIA results for Germany
Amazon results for Germany