Meaning Of The Word West Germany
What's the definition of West Germany? Find West Germany meanings, definitions and more at the Real Dictionary online.
West Germany Meaning
| West Germany Definition |
|---|
What's The Definition Of West Germany?
West Germany in British English
noun: a former republic in N central Europe, on the North Sea: established in 1949 from the zones of Germany occupied by the British, Americans, and French after the defeat of Nazi Germany; a member of the European Community; reunited with East Germany in 1990 |
More Definitions
A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z
Definitions Of The Day
- Congratulatory ‐ adjective: A congratulatory message expresses…
- Do your nut ‐ to become very angry about…
- Red meat ‐ meat that is red before cooking; esp., beef or…
- Kinder ‐ noun: (in Australia and New Zealand) a class or…
- Lateral bud ‐ noun
- Bosque ‐ a clump or…
- Buttercup squash ‐ noun: a small, usually dark-green squash that…
- Adenoid ‐ glandlike or glandular; adjective: of or pertaining…
- Contrabbasso ‐ noun: a…
- Comforter ‐ countable noun: A comforter is a person or thing…