The Way We Notice, That’s What Really Matters: Instantiating UI Components with Distinguishing Variations
AuthorsPriyan Vaithilingam, Alan Leung, Jeffrey Nichols, Titus Barik
The Way We Notice, That’s What Really Matters: Instantiating UI Components with Distinguishing Variations
AuthorsPriyan Vaithilingam, Alan Leung, Jeffrey Nichols, Titus Barik
Front-end developers author UI components to be broadly reusable by parameterizing visual and behavioral properties. While flexible, this makes instantiation harder, as developers must reason about numerous property values and interactions. In practice, they must explore the component’s large design space and provide realistic and natural values to properties. To address this, we introduce distinguishing variations: variations that are both mimetic and distinct. We frame distinguishing variation generation as design-space sampling, combining symbolic inference to identify visually important properties with an LLM-driven mimetic sampler to produce realistic instantiations from its world knowledge. We instantiate distinguishing variations in Celestial, a tool that helps developers explore and visualize distinguishing variations. In a study with front-end developers (n=12), participants found these variations useful for comparing and mapping component design spaces, reported that mimetic instantiations were domain-relevant, and validated that Celestial transformed component instantiation from a manual process into a structured, exploratory activity.
SQUIRE: Interactive UI Authoring via Slot QUery Intermediate REpresentations
April 6, 2026research area Human-Computer Interactionconference UIST
Frontend developers create UI prototypes to evaluate alternatives, which is a time-consuming process of repeated iteration and refinement. Generative AI code assistants enable rapid prototyping simply by prompting through a chat interface rather than writing code. However, while this interaction gives developers flexibility since they can write any prompt they wish, it makes it challenging to control what is generated. First, natural language on…
Soup-of-Experts: Pretraining Specialist Models via Parameters Averaging
July 4, 2025research area Methods and Algorithmsconference ICML
Large-scale models are routinely trained on a mixture of different data sources. Different data mixtures yield very different downstream performances. We propose a novel architecture that can instantiate one model for each data mixture without having to re-train the model. Our architecture consists of a bank of expert weights, which are linearly combined to instantiate one model. We learn the linear combination coefficients as a function of…