In this paper, we present a novel approach to synthesize realistic images based on their semantic layouts. The core idea is that for objects with similar appearance, they also share similar representation. To this end, we first regress semantic layouts to natural images, with intermediate representation that contains both semantic and appearance information. Then we propose a dynamic weighted network that takes these intermediate representations as a condition to generate high-quality results. We demonstrate that our method gives the compelling generation performance qualitatively and quantitatively with extensive experiments on benchmarks.