In mimicry systems, receivers discriminate between the stimuli of models and mimics. Weber's Law of proportional processing states that receiver discrimination is based on proportional, not absolute, differences between stimuli. Weber's Law operates in a variety of taxa and modalities, yet it has largely been ignored in the context of mimicry, despite its potential relevance to whether receivers can discriminate models from mimics. Specifically, Weber's Law implies that for a given difference in stimulus magnitude between a model and mimic, as stimulus magnitudes increase, the mimic will be less discriminable from their model. This implies that mimics should benefit when stimulus magnitudes are high, and that high stimulus magnitudes will reduce selection for mimetic fidelity. Whether models benefit from high stimulus magnitudes depends on whether mimicry is honest or deceptive. We present four testable predictions about evolutionary trajectories of models and mimics based on this logic. We then provide a framework for testing whether receiver discrimination adheres to Weber's Law and illustrate it using coevolutionary examples and case studies from avian brood parasitism. We conclude that, when studying mimicry systems, researchers should consider whether receiver perception conforms to Weber's Law, because it could drive stimulus evolution in counterintuitive directions.