... Representational harms are abstract concepts that cannot be measured directly [38]-yet measuring such harms is important, as they can cause tangible negative outcomes, e.g., through the entrenchment of harmful social hierarchies, which may affect people's belief systems and psychological states [17,62,16]. To facilitate the measurement of representational harms, the NLP research community has produced and made publicly available numerous measurement instruments, 1 including tools [e.g., 40,14], datasets [e.g., 64,28,32,34,56], metrics [e.g., 11,15,10,58,43], benchmarks (consisting of both datasets and metrics) [e.g., 23,48,45,46,55,63,22,24,26,27,31], annotation instructions [e.g., 42], and other techniques [e.g., 36,52,61]. However, the research community lacks clarity about whether and to what extent these instruments meet the needs of practitioners tasked with developing and deploying LLM-based systems in the real world, and how the instruments could be improved. ...