User Research Methods for IA in Technology Services
User research methods form the empirical backbone of information architecture practice, providing the structured evidence that distinguishes defensible structural decisions from intuition-driven design. In technology services — spanning enterprise platforms, SaaS products, mobile applications, and digital libraries — the choice of research method directly determines the quality of navigation systems, labeling, taxonomy, and findability. This page describes the major method categories used in IA-specific research, the mechanisms through which they produce actionable data, and the professional standards that govern their application.
Definition and scope
User research for IA refers to the systematic collection and analysis of data about how people seek, label, organize, and navigate information within a digital environment. The scope is narrower than general UX research: it focuses specifically on the mental models, vocabulary, and structural expectations users bring to an information system, rather than on task completion rates, visual aesthetics, or interaction patterns alone.
The Information Architecture Institute identifies three core research concerns that anchor IA-specific inquiry: how users categorize content, what terminology they apply to those categories, and how they traverse paths between content nodes. Methods that address these concerns fall into two broad families — generative methods (which produce structural hypotheses) and evaluative methods (which test existing or proposed structures against user behavior). Both families are active in technology services contexts, where the complexity of content environments frequently requires iterative cycling between generation and evaluation.
As described in Peter Morville and Louis Rosenfeld's Information Architecture for the Web and Beyond (O'Reilly Media, 4th edition), IA research is distinguished from general usability testing by its emphasis on organization systems and labeling systems rather than on interface mechanics.
How it works
IA user research operates through a structured sequence of phases:
-
Scope definition — The research question is anchored to a specific IA component: taxonomy structure, navigation labeling, search behavior, or content hierarchy. Scope definition prevents method misapplication, such as using a wayfinding evaluation tool when the actual problem is vocabulary mismatch.
-
Method selection — Based on whether the IA problem is generative or evaluative, practitioners select from the established method inventory (detailed below). The Usability.gov guidelines, maintained under the U.S. Department of Health and Human Services, provide a publicly accessible method classification framework.
-
Participant recruitment — Participants are screened to match the target user population for the information environment. Enterprise intranet IA research, for instance, typically requires domain-specific expertise in participants, while public-facing SaaS products may recruit from general consumer pools.
-
Data collection — Sessions are conducted under controlled or moderated conditions depending on method type. Remote unmoderated studies have become operationally standard for larger sample requirements.
-
Analysis and synthesis — Raw outputs (card groupings, click paths, verbal protocols) are coded and aggregated. Affinity mapping and dendrogram analysis are the primary synthesis tools for generative methods; task success rates and error analysis anchor evaluative outputs.
-
Structural recommendation — Findings are translated into specific IA artifacts: revised taxonomies, updated controlled vocabularies, reorganized site maps and hierarchies, or modified labeling systems.
Common scenarios
Card sorting is the most widely applied generative method in IA practice. Participants group content items (printed cards or digital equivalents) into categories and assign labels to those categories. Open card sorting produces user-generated category structures; closed card sorting tests whether users assign items to predefined categories consistently. The Nielsen Norman Group's published research across 50+ card sorting studies identifies optimal participant counts of 15 to 30 for open sorts when the goal is taxonomy development.
Tree testing is the primary evaluative counterpart to card sorting. Participants are presented with a text-only version of a proposed navigation hierarchy and asked to locate specific items without the support of visual design cues. Success rates and directness scores expose structural problems that visual prototypes would otherwise mask. Tree testing is particularly diagnostic for navigation design failures in enterprise and SaaS environments.
Contextual inquiry combines observation and interview techniques to capture how users seek information in their actual work environment. This method is especially relevant for IA for enterprise systems, where information-seeking behavior is embedded in complex workflows that laboratory methods fail to replicate.
Search log analysis examines query data from an existing system's search interface to identify vocabulary gaps, zero-result query patterns, and navigation abandonment signals. This method requires an operational system and produces quantitative evidence aligned with search systems in IA evaluation standards.
First-click testing measures where users click first when attempting a specific information-finding task. Research published by Bob Bailey and Cari Wolfson (Human Factors International) established that a correct first click predicts task completion at a rate exceeding 87 percent, making it a high-value, low-cost diagnostic tool.
Decision boundaries
Method selection is not arbitrary. Generative methods — card sorting and contextual inquiry — are appropriate when the IA problem is undefined or when an existing structure has failed without a clear diagnosis. Evaluative methods — tree testing and first-click testing — apply when a candidate structure exists and requires validation before implementation.
Sample size governs statistical reliability. Tree testing requires a minimum of 50 participants to produce directness and success scores with sufficient confidence intervals for organizational decision-making, per the Optimal Workshop methodology documentation. Card sorting reaches structural saturation at 15 to 20 participants for homogeneous user populations; heterogeneous populations across IA for omnichannel design contexts may require segmented recruitment of 30 or more per user group.
Measuring IA effectiveness after structural changes closes the research cycle, connecting pre-intervention research to post-deployment performance evidence and enabling systematic comparison across mental models in IA iterations.