The cinema of the United States, primarily associated with major film studios collectively referred to as Hollywood, has significantly influenced the global film industry since the early 20th century.