Meaning of west coast

Definition of west coast

(noun) the western seaboard of the United States from Washington to southern California

Other information on west coast

WIKIPEDIA results for west coast
Amazon results for west coast