Word | Count | Copy |
---|---|---|
{{ item.word }} | {{ item.count }} |
N-gram | Count | Copy |
---|---|---|
{{ g.token }} | {{ g.count }} |
Text statistics describe how writing is structured and how it reads in practice. A word counter and readability checker helps you see length, variety, and pace so you can edit with intent.
Counts and clarity scores arrive together, then frequency lists and ngrams reveal which words carry most weight. Paste text or drop a file, choose options that fit your draft, and read results that are easy to compare.
A short paragraph may show a quick reading time while a dense passage can lift the grade level and lower reading ease. Use consistent settings across drafts so changes reflect your edits rather than different rules.
Results are estimates shaped by language and punctuation, so unusual formatting or very small samples can mislead. When comparing versions, look for direction rather than single point differences.
The analysis observes words, sentences, characters, whitespace, and syllables in a snapshot of text. From these quantities it derives average word length, average sentence length, lexical density, estimated reading minutes, and estimated speaking minutes. Lexical density is the share of unique words within all words, sometimes called a type–token ratio.
Two readability indices are computed from sentence length and syllables per word. Reading ease increases as sentences shorten and syllables per word fall, while grade level rises with longer sentences or more syllables per word. These indices summarize effort, not writing quality.
Tokens are built from letters with diacritics and internal apostrophes or hyphens, with an option to include numbers. Sentence boundaries follow terminal punctuation. Syllables are estimated with a vowel‑group heuristic and common ending adjustments.
Comparisons are most meaningful within the same language and with the same options for case handling, stop‑words, accent normalization, minimum word length, and n‑gram rules. Extremely short texts can produce noisy indices.
Symbol | Meaning | Unit/Datatype | Source |
---|---|---|---|
W | Total words | count | Derived |
S | Total sentences | count | Derived |
SYL | Total syllables | count | Estimated |
ASL | Average sentence length | words/sentence | Derived |
ASW | Average syllables per word | syllables/word | Derived |
FRE | Reading ease score | 0–100 | Computed |
FKGL | Grade level estimate | grade | Computed |
In this case the reading ease is moderate while the grade level is about tenth grade. Values are rounded to two decimals.
Parameter | Meaning | Unit/Datatype | Typical Range | Sensitivity | Notes |
---|---|---|---|---|---|
Case sensitive | Treat uppercase and lowercase as distinct | boolean | false or true | Affects unique words and frequencies | Lowercasing improves comparability |
Include numbers | Count numeric tokens as words | boolean | false or true | Shifts counts in data‑heavy text | Recognizes decimals with “.” or “,” |
Remove stop‑words | Exclude common function words | boolean | false or true | Raises lexical density | Fixed English list |
Normalize accents | Fold diacritics to base letters | boolean | false or true | Unifies spellings | Applied before tokenizing |
Minimum word length | Discard short tokens | integer | 1–∞ | Small shift to counts | Default 1 |
N‑gram size | Length of contiguous token sequences | integer | 1–3 | Reveals phrases | 1 words, 2 bigrams, 3 trigrams |
Across sentences | Allow n‑grams to cross sentence ends | boolean | false or true | Changes phrase counts | When off, reset at punctuation |
Field | Type | Min | Max | Step/Pattern | Notes |
---|---|---|---|---|---|
Top N words | number | 1 | — | integer | Controls frequency table and chart |
Top N n‑grams | number | 1 | — | integer | Controls n‑gram table |
Minimum word length | number | 1 | — | step 1 | Shorter tokens are discarded |
N‑gram size | select | 1 | 3 | — | 1 words, 2 bigrams, 3 trigrams |
File input | file | — | — | text read | Reads content of selected file as text |
The reading ease and grade formulas are widely attributed to Rudolf Flesch and to work by J. Peter Kincaid and colleagues on grade level. Type–token ratio is a common lexical diversity measure in linguistics.
No data is transmitted or stored server‑side. Clipboard and downloads operate locally for your session.
Text analysis measures counts, readability, and frequent terms to guide revision.
Example: After removing stop‑words and setting minimum length to 3, frequent terms shift to content words that better represent the topic.
Use the same settings to compare drafts confidently.
No. Text is processed in your browser, and exports are created locally for your session.
No server‑side storage.They use a heuristic that is strong for common English words yet may miss rare names or unusual endings.
Treat as estimates.Word and sentence counts are integers. Averages and scores use two decimals. Reading and speaking minutes are whole numbers.
Period as decimal separator.Yes, once loaded it operates entirely in the page without remote requests for analysis.
Charts render from local data.There is no charge indicated by the package metadata. If licensing changes, follow the site’s terms.
Check project terms if published.Enable case folding, keep numbers off if not needed, set a minimum length, and consult the Unique Words figure or frequency table.
Stop‑words can be removed.Scores near your target suggest minor edits may flip direction. Short sentences or simpler words usually increase ease and lower grade.
Use trends across drafts.