Pillar guideSpeech-Language Pathology15 min read

How to Conduct a Language Sample: A Step-by-Step Protocol for SLPs

The hardest part of Language Sample Analysis is not the scoring — it is the 45 to 90 minutes of elicitation and transcription that happen before a single metric gets computed. This protocol walks through every step of a modern clinical language sample: room setup, materials, age-banded prompts, recording and consent, segmentation, and the three realistic transcription approaches for a high-caseload SLP. The goal is a sample that a clinician can actually finish in a school week, not a research-grade transcript that never gets done.

1. Why a written protocol matters

Every SLP who has ever sat down with a laptop, a preschooler, and a ten-minute slot between IEPs knows that "go conduct a language sample" is not a single skill — it is five distinct skills stacked on top of each other, and losing any one of them contaminates the whole sample. Elicitation that produces only two-word utterances gives you nothing to score. Recording equipment that drops out halfway through ruins an hour of clinic time. Segmentation done on the fly, in your head, while the child is still talking, reliably produces a different MLU than segmentation done from a clean transcript three days later. A written protocol is the only realistic way to make the whole workflow reproducible across children, clinicians, and caseload sizes.

The SUGAR (Sampling Utterances and Grammatical Analysis Revised) project at the University of Central Florida (Pavelko & Owens, 2017, 2019) was designed specifically around this reproducibility problem. Their published sampling protocol — 50-utterance conversational sample, elicited with a fixed set of prompts, transcribed into a standard format, segmented with clear rules — is the nearest thing the field has in 2026 to a clinical consensus on "what a language sample should look like". This pillar adapts that protocol for a school or outpatient clinician and integrates it with the free browser-based LSA tools on this site so that every step has a concrete tool attached to it.

The guiding principle throughout is clinical realism. A 50-utterance sample is not the research-grade ideal (Heilmann et al. 2010 show that longer samples are more stable), but it is the length that most school clinicians can actually collect and transcribe in a single IEP window. The protocol below is calibrated so that a full workflow — from consent to scored transcript — fits in roughly 60 to 90 minutes of total clinician time, which is the point at which language sampling starts being practical at scale rather than theoretical.

The 50-utterance rule

Every published LSA normative database from SUGAR to Heilmann to the SALT reference set assumes the sample has at least 50 complete, scorable utterances. Shorter samples produce unstable MLU and NDW values; longer samples plateau after roughly 75 utterances. Fifty is the clinical sweet spot, and it is the target this protocol is built around.

2. Before the session: consent, goals, and room setup

The ten minutes before the child walks into the room is where every avoidable language sample failure gets prevented. Write down the specific clinical question the sample is meant to answer before you plan the session — "does this 4-year-old’s MLU match age expectations?", "is this 6-year-old’s percent grammatical utterances in the clinical range?", or "is this 8-year-old’s narrative grammar structured enough for a school-age IEP?". The question determines the elicitation approach. A child whose question is MLU-focused needs a conversational sample; a child whose question is narrative structure needs a story-retell or story-generation task; a child whose question is grammaticality needs a sample long enough to give stable PGU.

Consent is the second step and it is the step that most commonly blocks the workflow three weeks later when you realise you never got written permission to audio-record. Every district has its own consent form; if yours does not, the minimum language to include is: (1) the purpose of the recording (clinical assessment), (2) who will hear the recording (the clinician and supervising SLP), (3) how long it will be retained, and (4) the parent’s right to withdraw consent and have the recording deleted. Get the signature before the session, not after.

Room setup is quick but non-negotiable. The room should be as quiet as you can realistically get — hallway door shut, HVAC vent not blasting directly over the microphone, no overlapping conversations at the other end of the SLP office. Sit at right angles to the child rather than directly across (right-angle seating keeps the mic away from direct blasts of air from the child’s mouth and reduces the "plosive pop" that makes disordered consonants hard to transcribe). A small table with age-appropriate materials in arm’s reach — picture books, a small set of figurines, a barrier game, a wordless picture book — is enough to carry a 20-minute session without the clinician having to reach for new props.

  • Write the specific clinical question down: MLU? PGU? Narrative grammar?
  • Get written parent consent to audio-record before the session, not after.
  • Shut the door. Turn off the HVAC if you can. No overlapping adult voices.
  • Sit at a right angle to the child, not directly across.
  • Stage 3-5 elicitation props in arm’s reach before the child enters.
  • Print the Language Sample Worksheet so prompts and tallies are on paper, not on your phone.

3. Elicitation prompts by age

The single biggest determinant of sample quality is prompt choice. A 3-year-old asked "tell me about your weekend" produces one-word answers; the same child shown a wordless picture book produces ten-word stretches of connected speech. Prompt-to-age matching is the difference between a 50-utterance sample collected in 20 minutes and a 50-utterance sample that never materialises. The prompts below come from the SUGAR protocol, the Heilmann lab recommendations (Heilmann, 2010), and decades of clinical-room practice — they are calibrated to each developmental band to maximise both quantity and grammatical variety.

For children aged 3 to 4, concrete-object play and simple picture-book prompts outperform open questions. Favourites include the "What’s in the bag?" barrier game with 5-8 small objects, a wordless picture book like Mercer Mayer’s Frog Where Are You?, and a simple pretend-play setup with a dollhouse or farm animals. Avoid "tell me about..." prompts at this age because they consistently fail to elicit multi-word utterances in typically developing children, let alone children with language delay.

For children aged 5 to 6, narrative tasks become productive and are worth prioritising because they stress the syntactic complexity metrics (DSS, IPSyn) that are the dominant clinical targets at this age. Story-retell from a wordless picture book works well: read the book through once silently pointing at the pictures, then hand it to the child and ask them to "tell the story back". Story-generation from a single stimulus picture (for example, the SBS test stimuli or a National Geographic Kids magazine cover) also works but produces shorter samples. Conversational prompts ("what did you do at recess?", "tell me about your pet") start being productive at this age but are still not as reliable as structured tasks.

For children aged 7 to 11, conversational sampling is the default. School-age children produce long stretches of connected speech on topics they care about, and the clinician’s job shifts from "prompt harder" to "get out of the way". High-value prompts include recent movies or games the child likes, how to play a favourite video game (this is a goldmine for procedural discourse and complex syntax), and "tell me the plot of..." prompts on books or shows the child has recently watched. Narrative generation tasks still work at this age and are the right choice if your clinical question is specifically about story grammar.

For children aged 12 and up, expository tasks become the best source of complex syntax. Ask a teenager to explain how to play a sport, how their favourite video game works, or how to do something they are good at (bake a cake, skateboard, solve a Rubik’s cube). Expository sampling produces the longest T-units and the highest subordination indices of any elicitation task in this age range, which is precisely what you want for older-student LSA. Story-retell from a single short film or a short story still works but produces shorter samples than expository tasks for most teens.

  • Ages 3-4: wordless picture books, barrier games, simple pretend play with figurines.
  • Ages 5-6: story-retell from wordless books, story-generation from a single stimulus picture.
  • Ages 7-11: conversational sampling on video games, recess, favourite books or shows.
  • Ages 12+: expository tasks — "explain how to play..." or "explain how... works".
  • Avoid "tell me about your weekend" before age 7 — it under-produces at every younger age.
  • The printable Language Sample Worksheet on this site has all of these prompts pre-listed.

4. Recording equipment and audio quality

A clean recording is the cheapest insurance a clinician can buy against wasted clinic time. The recording does not need to be studio-grade — a modern smartphone or a $30 USB lavalier microphone clipped to the child’s shirt produces audio that transcribes cleanly in 2026, as long as the room is quiet and the mic is within 12 inches of the child’s mouth. The specific device matters less than three simple rules: proximity, redundancy, and pre-session test.

Proximity means the microphone should be close enough that the child’s voice is louder than the room noise on the recording. If you play the first ten seconds back and the HVAC is competing with the child, move the mic closer or turn the HVAC off. Redundancy means you should record on two devices in parallel whenever possible — phone plus laptop microphone, phone plus a cheap USB recorder, whatever you have. Audio files get corrupted, phones run out of storage, and a dropped recording after a 30-minute session is a cost you cannot afford on a school-clinician schedule. Pre-session test means you press record, say "test one two three" in the child’s seat, play it back, and confirm you can hear clearly before the child walks in. Every experienced clinician has a story about the recording that looked fine on the meter and turned out to be 30 minutes of silence.

For the actual file format, a lossless or near-lossless recording (WAV, M4A, MP3 at 192kbps or above) is what you want. Voice memo apps on iOS and Android default to M4A at sufficient quality for LSA transcription. Do not use speech-to-text-to-file pipelines during the session itself — you want the raw audio, and any transcription step runs after the session when you have the time to edit it. Save the file with a child identifier that matches your district’s naming convention (never the child’s full name in the filename) and move it to your encrypted clinical drive within 24 hours.

  • Modern smartphone or $30 USB lavalier is sufficient in 2026.
  • Mic within 12 inches of the child’s mouth; room as quiet as possible.
  • Record on two devices in parallel whenever you can — redundancy saves clinic time.
  • Pre-session test: record 10 seconds, play back, confirm before child enters.
  • Save as WAV or M4A; 192kbps+ for MP3; no live speech-to-text during the session.
  • Filename uses child identifier, never the child’s full name. Move to encrypted drive within 24 hours.

The two-device rule

Record every session on at least two devices in parallel. The cost is near zero (your phone plus your laptop mic counts), and the upside is that when one device fails — and eventually one will — you have not wasted 30 minutes of clinic time and the child’s trust for nothing.

5. Running the 20-minute session

The session itself is where clinician experience compounds. The target is 50 scorable utterances from the child, which typically takes 15 to 25 minutes of clock time for a conversational sample and slightly longer for narrative or expository tasks (because the clinician has to set up the stimulus first). The clinician’s job during the session is to keep the child talking without contaminating the sample with adult-scaffolding utterances that the child then parrots back.

A few concrete rules help. First, minimise yes/no questions — they elicit yes/no answers, and yes/no answers are not scorable for MLU, DSS, or most grammatical-accuracy metrics. Open-ended prompts ("tell me more", "what happened next", "and then?") consistently outperform closed questions for sample length. Second, resist the urge to expand or recast the child’s utterances during the sample. Recasting is a valuable therapy technique but it contaminates the assessment sample because the child may repeat your scaffold verbatim. Third, avoid giving the child multiple-choice prompts ("was it red or blue?") because they elicit one-word answers. Fourth, allow silences. Children fill conversational silences with their own speech — an adult who rushes to fill every pause will produce a sample that is mostly adult turns, which is exactly the opposite of what you want.

As the session runs, the clinician should keep a loose tally of how many child utterances have been produced so that you do not run long. Fifty utterances is the floor; if you are at 20 utterances after ten minutes and the child is giving you useful material, keep going for another ten. If you are at 10 utterances after 15 minutes because the child is closing down, switch prompts or change the stimulus object before forcing more — a sample collected from an increasingly reluctant child will have MLU and PGU values that reflect the child’s reluctance rather than their underlying competence.

Mark the end of the session clearly for the recording: say something like "okay, thank you, we’re all done" at a normal volume. This gives you a clear audio landmark when you transcribe later and it signals to the child that the assessment portion is over. If you are continuing with other clinical activities afterwards (play, games, standardised testing), stop the recording first so you do not contaminate the file with non-sample audio.

  • Target: 50 scorable child utterances in 15-25 minutes of clock time.
  • Avoid yes/no questions and multiple-choice prompts — they kill utterance length.
  • Do not recast or expand the child’s utterances during the sample.
  • Allow silences; children fill them with their own speech.
  • Keep a loose tally during the session so you know when you can stop.
  • Mark a clear verbal end-of-session landmark for the transcript.

6. Segmentation: where does one utterance end and the next begin?

Segmentation is the least-glamorous and most error-prone step in LSA. Every MLU, DSS, IPSyn, and PGU calculation depends on where the clinician draws the boundaries between utterances, and small segmentation changes can shift MLU by half a morpheme in either direction on a 50-utterance sample. The good news is that the rules are deterministic once you pick a framework; the bad news is that every framework picks slightly different rules, so you have to commit to one and use it consistently.

The C-unit (Communication Unit) framework is the default for SUGAR-style school-age sampling. A C-unit is an independent clause plus any modifiers and subordinate clauses attached to it. The rules: a C-unit ends at a main-clause boundary, a coordinating conjunction that starts a new main clause, a clear intonational fall, a long pause, or a speaker change. Run-on sentences connected with "and then... and then... and then..." are segmented into separate C-units at each "and then". The T-unit (Terminable Unit) framework, developed by Hunt for school-age writing analysis, is similar but slightly more permissive about embedded coordination. Brown’s original framework for preschool data (Brown, 1973) is still the right choice for ages 3-5 and uses utterance boundaries defined primarily by pauses and intonation rather than grammatical completeness.

A small number of conventions apply to all three frameworks and are worth committing to memory. First, repetitions and stutters inside an utterance are excluded from the word count but the utterance itself is still scored — a child who says "I-I-I went to the store" has an utterance of length four ("I went to the store"), not six. Second, mazes (the child’s false starts, fillers, and self-corrections) are excluded from MLU counts but tracked separately as a fluency measure. Third, unintelligible utterances are excluded from MLU and NDW but kept for the intelligibility calculation — you cannot score what you cannot transcribe. Fourth, imitations of the clinician’s immediately preceding utterance are excluded from the sample entirely because they reflect the clinician’s grammar, not the child’s.

The single most important habit is consistency. Pick one framework (SUGAR’s C-unit rules are the easiest to explain to new clinicians and match the published normative data on this site), write the framework name on the transcript header, and use it the same way every time. Clinicians who switch frameworks between children introduce scoring noise that swamps any real change in the child’s grammar.

  • C-unit framework: the default for SUGAR-style school-age sampling.
  • Exclude stutters and self-repetitions from word counts; score the repaired utterance.
  • Exclude mazes from MLU; track them separately as a fluency measure.
  • Exclude unintelligible utterances from MLU and NDW; keep them for the PIU calculation.
  • Exclude direct imitations of the clinician’s previous utterance entirely.
  • Pick one framework (C-unit, T-unit, or Brown) and commit to it across your caseload.

7. Three realistic transcription approaches

Once the session is recorded, you have to turn the audio into a typed transcript that a calculator can score. This is the single slowest step in LSA and it is the reason most caseloads run on standardised tests instead of language samples. Three transcription approaches are realistic in 2026: manual, AI-assisted, and fully-automated with clinician review. Each has different time costs, different error profiles, and a different place in a high-caseload clinician’s workflow.

Manual transcription is the gold standard for accuracy and the hardest to justify on a school caseload. A trained clinician transcribing a 50-utterance preschool sample at normal listening speed takes 45 to 60 minutes; a sample from a child with disordered speech can take 90 minutes or longer. Manual transcription captures the clinician’s ear for speech errors, mazes, and pragmatic context in a way that no automated tool currently matches. It is the right choice when you have a complex differential diagnosis question, when the recording is poor quality, or when the child’s speech is significantly disordered and an automated transcript would miss the clinically relevant errors.

AI-assisted transcription is the most common approach in 2026 for typical-speech samples. The workflow is: run the audio through a modern speech-to-text engine (OpenAI Whisper, Google Chirp, or a clinical-SLP tool like ConductSpeech that wraps a speech recogniser with LSA-specific post-processing), get a rough transcript in under two minutes, and then clean it manually — fix misrecognised words, mark the speaker turns, apply your segmentation framework, and exclude mazes. The total clinician time drops to 15-25 minutes per sample, which is the threshold at which language sampling starts being viable at scale. Accuracy on typical preschool speech is roughly 90-95% and the clinician’s edit pass closes the remaining gap.

Fully-automated transcription with clinician review is the newest approach and the one that ConductSpeech is built for. The workflow is: upload audio, the tool produces a transcript, segments it into utterances using the Brown/SUGAR rules, tags morphemes in obligatory contexts, computes MLU-m, MLU-w, NDW, and PGU, and drafts a present-levels paragraph. The clinician’s job shifts from typing to reviewing: skim the transcript for misrecognised words, confirm the segmentation, spot-check the morpheme tags, and edit the draft paragraph. Total clinician time on a typical-speech sample drops to 5-10 minutes. The approach does not replace clinical judgement — the transcript, the segmentation, and the scores are all editable — but it changes the calculus of "should I do a language sample this week?" from "no, I don’t have 45 minutes" to "yes, I have 10".

  • Manual transcription: 45-90 minutes per sample; gold standard for disordered speech.
  • AI-assisted (Whisper, Chirp): 15-25 minutes; the 2026 default for typical-speech samples.
  • Fully-automated with review (ConductSpeech): 5-10 minutes; audio-in, draft-out.
  • Every approach still requires clinician review — no transcript is unedited-ready.
  • Pick the approach by caseload size and transcription-time budget, not by ideology.

8. Scoring the transcript with free browser tools

Once the transcript is segmented and cleaned, scoring is the fastest step of the whole workflow. The free browser-based LSA calculators on this site are designed specifically for the paste-transcript-click-score pattern. The MLU Calculator takes a plain-text transcript (one utterance per line), applies Brown’s morpheme rules in code, and returns MLU-m, MLU-w, total morphemes, total utterances, and the matching Brown’s stage in one pass. The Lexical Diversity Calculator returns NDW, TTR, MATTR, and vocd-D from the same paste. The DSS Calculator implements the Lee (1974) eight-category syntactic scoring. The IPSyn Calculator implements the 60-item Index of Productive Syntax. The PGU Calculator returns percent grammatical utterances using the SUGAR rule.

Which calculator to run depends on the clinical question. For ages 3-5 with an MLU-focused question, run the MLU Calculator and cross-reference the Brown’s Stages Lookup to get the stage and expected morpheme mastery. For school-age children with a syntactic-complexity question, run the DSS Calculator or the IPSyn Calculator depending on which metric your district report template expects. For older students with a grammaticality-focused question, run the PGU Calculator to get percent grammatical utterances and cross-reference the SUGAR Norms Lookup. For any child, the Lexical Diversity Calculator gives you NDW in under a second.

The scoring step should take no more than 5 minutes of clinician time once the transcript is clean, because every calculator runs in the browser and requires only a paste-and-click. The clinical interpretation takes longer — matching the scores against published norms, writing the present-levels paragraph, and drafting the IEP goals — but the arithmetic itself is effectively free. This is the single biggest reason the free-tools approach is viable for a high-caseload school SLP: it is not that the scoring is faster than SALT (it is comparable), it is that there is no licence cost, no installation step, and no transcription-format learning curve. You paste, you click, you get the numbers.

  • MLU Calculator: MLU-m, MLU-w, total morphemes, utterances, Brown’s stage in one pass.
  • Lexical Diversity Calculator: NDW, TTR, MATTR, vocd-D.
  • DSS Calculator: Lee (1974) developmental sentence scoring.
  • IPSyn Calculator: Scarborough (1990) 60-item productive syntax.
  • PGU Calculator: SUGAR-rule percent grammatical utterances.
  • Brown’s Stages Lookup and SUGAR Norms Lookup for age-banded reference data.
  • Total scoring time once the transcript is clean: under 5 minutes per child.

9. Common pitfalls and how to avoid them

The five most common failure modes in clinical language sampling are not clinical mistakes — they are workflow mistakes, and they are all preventable with a written protocol. They come up at every district training and they are worth spelling out so that new clinicians do not have to discover them by burning clinic time.

  • "The sample is too short." Fewer than 50 utterances means unstable MLU and NDW. Solution: keep a tally during the session and extend elicitation until you hit 50, or split the assessment across two short sessions if the child fatigues.
  • "The recording is unusable." Solution: the two-device rule. Record on phone and laptop in parallel every time.
  • "My MLU numbers do not match the other clinician in my district." Solution: agree on a single segmentation framework (C-unit is the easiest default) and write it on every transcript header.
  • "I do not have time to transcribe." Solution: switch from manual to AI-assisted transcription for typical-speech samples. The accuracy gap is small and the time savings are 30+ minutes per child.
  • "I do not know which metric to run." Solution: pick the metric that answers the specific clinical question you wrote down before the session. If the question is "within age expectations?", run MLU and cross-reference the age-banded norm. If the question is "syntactically mature?", run DSS or IPSyn. If the question is "grammatically accurate?", run PGU.
  • "The child gave me only yes/no answers." Solution: switch to open-ended "tell me more" prompts and props the child has to manipulate. Children talk when their hands are busy.
  • "The parent report says the child talks constantly at home but I got ten utterances." Solution: the child is in an assessment context with an unfamiliar adult. Schedule a second session, use a more familiar play partner (a sibling or the parent), or switch to a recorded home sample with parent consent.

10. A recommended 2026 workflow

The workflow recommendation that comes out of this protocol is calibrated for a school or outpatient clinician with a caseload of 20 to 50 students and a district that expects ecologically valid LSA at IEP reviews. It combines the free tools on this site, a modern recording approach, and either AI-assisted or fully-automated transcription depending on caseload size. The recommendation is not the research-grade ideal — longer samples, manual transcription, and multiple coders would all be better in isolation — but it is the stack that actually fits the clinical week.

The recommended workflow, start to finish, is: (1) write the clinical question down, (2) get parent consent to audio-record, (3) print the Language Sample Worksheet on this site for the age band, (4) set up a quiet room with props in arm’s reach, (5) record the session on two devices in parallel, (6) target 50 scorable utterances in 15-25 minutes, (7) transcribe with AI assistance (Whisper, Chirp, or ConductSpeech) and clean manually, (8) segment using the C-unit framework, (9) paste into the MLU Calculator and the metric calculator matching your clinical question, (10) cross-reference the Brown’s Stages Lookup or SUGAR Norms Lookup for the age-banded norm, and (11) use the IEP Goal Generator to draft the present-levels paragraph and annual goals. The whole pipeline fits in 60 to 90 minutes of clinician time per child, which is the threshold where LSA stops being a luxury and starts being the default assessment.

For clinicians whose caseload or time budget makes even that pipeline tight, ConductSpeech is built specifically to automate steps 7 through 11 — audio in, scored transcript and draft IEP goals out — so the clinician’s role is to record, review, and edit. The honest framing is not that ConductSpeech replaces clinical judgement but that it reallocates clinician time from the slow transcription step to the fast interpretation and goal-writing step. For every caseload size, though, the protocol above is the frame: the specific transcription approach changes with caseload, but the elicitation, segmentation, and scoring steps stay the same.

Free tools and reference pages

Every link below stays on conductscience.com. Open any tool in a new tab and come back here for the protocol context.

Free tools

Language Sample Worksheet

Printable elicitation prompts and tally sheet calibrated to the age bands in this protocol — take it into the clinic room.

Open

MLU Calculator

Paste a cleaned transcript; get MLU-m, MLU-w, total morphemes, utterances, and the matching Brown's stage in one click.

Open

Lexical Diversity Calculator

NDW, TTR, MATTR, and vocd-D from the same paste — the vocabulary panel of the protocol.

Open

Developmental Sentence Score Calculator

Lee (1974) eight-category syntactic scoring for school-age samples where syntactic complexity is the clinical question.

Open

IPSyn Calculator

The 60-item Index of Productive Syntax from Scarborough (1990) — the second syntactic-complexity option for school-age samples.

Open

PGU Calculator

Percent Grammatical Utterances using the SUGAR rule — the grammaticality metric for older-child samples.

Open

Brown's Stages Lookup

Map an MLU value or child age onto Brown's five stages with example utterances for the age bands in this protocol.

Open

SUGAR Norms Lookup

Pavelko & Owens (2017, 2019) age-banded MLU, NDW, and TPU norms — the cross-reference step after scoring.

Open

IEP Goal Generator

Drafts measurable annual IEP goals from the morpheme, syntax, and lexical-diversity baselines produced by this protocol.

Open

Narrative Scoring Scheme Calculator

Heilmann (2010) seven-component narrative scoring — the scoring tool for story-retell samples collected with this protocol.

Open

Story Grammar Scorer

Classic story grammar scoring for narrative samples elicited with the school-age and older-child prompts in this protocol.

Open

Conversation Turn Analyzer

Quantifies speaker balance, turn length, and topic maintenance — the pragmatic complement to MLU and PGU.

Open

Frequently asked questions

How long should a clinical language sample be?
Fifty scorable child utterances is the clinical floor and the number every published LSA normative database assumes. Shorter samples produce unstable MLU and NDW values; longer samples plateau after roughly 75 utterances. A 50-utterance conversational sample typically takes 15-25 minutes of recording time, and this protocol is calibrated to hit that target in a single IEP window.
What age-appropriate elicitation prompts work best?
For ages 3-4, wordless picture books, barrier games, and simple pretend play outperform open questions. For ages 5-6, story-retell tasks are most productive. For ages 7-11, conversational sampling on recess or video games works well. For ages 12+, expository tasks ("explain how to play...") produce the longest utterances and the most complex syntax. Avoid "tell me about your weekend" before age 7 — it under-produces at every younger age.
What recording equipment do I need?
A modern smartphone or a $30 USB lavalier microphone is sufficient in 2026. The specific device matters less than three rules: proximity (mic within 12 inches of the child’s mouth), redundancy (record on two devices in parallel), and pre-session test (record 10 seconds, play back, confirm audio quality before the child enters the room). Save as WAV or M4A; MP3 at 192kbps or higher is also fine.
Do I need parent consent to audio-record a language sample?
Yes, every time. The consent form should cover the purpose of the recording (clinical assessment), who will have access, how long the recording will be retained, and the parent’s right to withdraw consent. Every district has its own form; get the signature before the session, not after. The recording is part of the clinical record and should be stored on your encrypted clinical drive with a non-identifying filename.
Which segmentation framework should I use?
The C-unit (Communication Unit) framework is the easiest default for school-age samples and matches the SUGAR published normative data on this site. A C-unit is an independent clause plus any modifiers and subordinate clauses attached to it, segmented at main-clause boundaries, intonational falls, and speaker changes. For preschool samples, the original Brown (1973) framework using pauses and intonation is still standard. The important rule is to pick one framework and use it consistently across your caseload.
How long does manual transcription take?
A trained clinician transcribing a 50-utterance preschool sample at normal listening speed takes 45-60 minutes of uninterrupted time. Disordered-speech samples take 90 minutes or longer. This is the reason most school caseloads run on standardised tests instead of language samples — the transcription bottleneck, not the scoring bottleneck, is what kills clinical LSA at scale.
Can AI transcription replace manual transcription?
For typical preschool and school-age speech in 2026, AI-assisted transcription (Whisper, Chirp, ConductSpeech) produces transcripts that are roughly 90-95% accurate out of the box and closes the remaining gap with a 5-15 minute clinician edit pass. Total clinician time drops from 45-90 minutes per sample to 15-25 minutes. For significantly disordered speech, manual transcription remains the gold standard because automated tools miss the clinically relevant error patterns. Pick the approach by caseload size and by how disordered the child’s speech is.
What do I do if the child only gives me one-word answers?
Switch prompts. Yes/no questions and multiple-choice prompts reliably produce one-word answers at every age. Switch to open-ended "tell me more" and "what happened next" prompts, hand the child something to manipulate (a book, a figurine, a barrier-game object), and allow silences — children fill silences with their own speech. If the child continues to under-produce after 10 minutes, consider whether the assessment context itself is the problem: schedule a second session with a more familiar play partner or collect a parent-facilitated home sample.
How do I handle mazes and repetitions in the transcript?
Mazes — false starts, fillers, and self-corrections — are excluded from MLU word counts but tracked separately as a fluency measure. A child who says "I-I-I went to the store" has an utterance of length four ("I went to the store"), not six. Stutters and repetitions inside a single utterance are also excluded from the word count but the utterance is still scored. Direct imitations of the clinician’s immediately preceding utterance are excluded from the sample entirely because they reflect the clinician’s grammar, not the child’s.
How often should I conduct a language sample?
At minimum, at initial evaluation and at every annual IEP review. Mid-year progress-monitoring samples are optional but useful for children whose goal areas are syntactic or morphological. The cadence is constrained by transcription time more than by clinical judgement — which is exactly the bottleneck that AI-assisted and fully-automated transcription approaches are designed to loosen. At a high enough automation level, mid-year samples become routine rather than exceptional.

References

  1. Brown, R. (1973). A First Language: The Early Stages. Cambridge, MA: Harvard University Press.
  2. Pavelko, S. L., & Owens, R. E. (2017). Sampling Utterances and Grammatical Analysis Revised (SUGAR): New normative values for language sample analysis measures. Language, Speech, and Hearing Services in Schools, 48(3), 197-215.
  3. Pavelko, S. L., & Owens, R. E. (2019). Diagnostic accuracy of the Sampling Utterances and Grammatical Analysis Revised (SUGAR) measures for identifying children with language impairment. Language, Speech, and Hearing Services in Schools, 50(2), 211-223.
  4. Heilmann, J., Nockerts, A., & Miller, J. F. (2010). Language sampling: Does the length of the transcript matter? Language, Speech, and Hearing Services in Schools, 41(4), 393-404.
  5. Heilmann, J., Miller, J. F., Nockerts, A., & Dunaway, C. (2010). Properties of the Narrative Scoring Scheme using narrative retells in young school-age children. American Journal of Speech-Language Pathology, 19(2), 154-166.
  6. Miller, J. F., & Iglesias, A. (2008). Systematic Analysis of Language Transcripts (SALT) [Computer software]. SALT Software, LLC.
  7. Lee, L. L. (1974). Developmental Sentence Analysis: A Grammatical Assessment Procedure for Speech and Language Clinicians. Evanston, IL: Northwestern University Press.
  8. Scarborough, H. S. (1990). Index of Productive Syntax. Applied Psycholinguistics, 11(1), 1-22.
  9. Rice, M. L., Smolik, F., Perpich, D., Thompson, T., Rytting, N., & Blossom, M. (2010). Mean length of utterance levels in 6-month intervals for children 3 to 9 years with and without language impairments. Journal of Speech, Language, and Hearing Research, 53(2), 333-349.
  10. Eisenberg, S. L., & Guo, L. (2013). Differentiating children with and without language impairment based on grammaticality. Language, Speech, and Hearing Services in Schools, 44(1), 20-31.
  11. Hunt, K. W. (1970). Syntactic maturity in schoolchildren and adults. Monographs of the Society for Research in Child Development, 35(1), 1-67.
  12. Owens, R. E. (2014). Language Disorders: A Functional Approach to Assessment and Intervention (6th ed.). Boston, MA: Pearson.

This article is a clinical protocol reference, not a substitute for individual clinical judgement. The elicitation prompts and segmentation rules are adapted from published normative databases and should be used alongside your district’s assessment policies and supervising SLP’s guidance.

Get the full analysis

Run the full protocol in 10 minutes instead of 90

ConductSpeech handles the transcription and scoring steps automatically so the clinician owns elicitation and interpretation. Free trial; no installation required.

Get the full analysis with ConductSpeech