In everyday life, people often stand up and sit down. Unlike young, able-bodied individuals, older adults and those with disabilities usually stand up or sit down slowly, often pausing during the transition. It is crucial to design interfaces that accommodate these movements. Additionally, in public settings, protecting personal information is essential. Addressing these considerations, this paper presents a distance-based representation scheme for the motions of standing up and sitting down. This proposed scheme identifies both standing and sitting positions, as well as the transition process between these two states. Our scheme is based solely on the variations in distance between a sensor and the surfaces of the human body during these movements. Specifically, the proposed solution relies on distance as input, allowing for the use of a proximity sensor without the need for cameras or additional wearable sensor attachments. A single microcontroller is adequate for this purpose. Our contribution highlights that using a proximity sensor broadens the applicability of the approach while ensuring that personal information remains secure. Additionally, the scheme alleviates users’ mental burden, particularly regarding privacy concerns. Extensive experiments were performed on 58 subjects, including 19 people over the age of 70, to verify the effectiveness of the proposed solution, and the results are described in detail.