TYPECHA Details for certificate ID taU0t6nh

Title Ironies of LLMs
Status approved
Submission Date July 7, 2024
Confidence 97.8
TYPECHA version 1.2.0

Writing Velocity Canvas

Writing Velocity Canvas

The Writing Velocity Canvas is a graphical representation of the document's revision history. Blue shades mark insertions and red marks deletions. The brighter the color, the larger the insertion or deletion.


Deletions Colorbar Insertions

Ironies of LLMs There’s an irony baked into Large Language Models. In fact, most technologies that automate or enhance processes that involve humans contain “ironies of automation”. For example, an assembly line is created to automate a manufacturing process. However, when something goes wrong, the entire thing must come to a standstill. All production may need to immediately stop to resolve the issue. This is typically not the case when humans build things. Another example could be automating monitoring software for a control room. The software may reduce the cognitive burden of the humans responsible for the control room. However, when something does go wrong, it may take significantly longer for a human to then gain the situational awareness needed to resolve the issue as they have been out of the loop for some time. A third example are alarms meant to draw attention, such as those used by patient monitoring equipment in a hospital. At a certain threshold, such alarms may cause alarm fatigue, which leads to a decrease in a care provider’s response rate and an increase in their response time. In information technology, we have long depended on computers for the retrieval of accurate information. When a database is queried, whatever was stored can be reliably extracted. Any errors are bound to have happened at insertion or retrieval. This is barring a bit flip or other soft errors, but again, there are safeguards against these. Humans exercise their judgement but depend on the reliability of information retrieved from computer systems, often in automated ways. The irony is that through LLMs, we’re creating computer systems that generate information in inconsistent and reliable ways, but presented with a veneer of confidence by way of grammatically correct and tone appropriate language. We’ve created information technology that requires fact-checking.

Metric Value
words 302
sentences 15
syllables 496
wordtypes 193
characters 1546
long_words 98
paragraphs 1
unique words 183
complex_words 63
syll_per_word 1.6
rollover ratio 14.8
velocity - max 118.5
velocity - std 38.3
omission errors 0.1
bigram variation 222.0
complex_words_dc 118
confidence score 90.8
insertion errors 0.6
type_token_ratio 0.6
readability - ARI 12.7
readability - LIX 52.6
readability - RIX 6.5
rolling wpm - max 90.7
rolling wpm - std 52.4
velocity - median 66.2
words_per_sentence 20.1
characters_per_word 5.1
substitution errors 1.2
rolling wpm - median 111.9
word usage - auxverb 8
word usage - pronoun 16
readability - Kincaid 11.6
word usage - tobeverb 12
keypress duation - std 103.1
type-token ratio (TTR) 0.6
Yule's Characteristic K 77.5
interkey interval - std 46.8
neutral sentiment score 0.8
readability - SMOGIndex 14.2
sentences_per_paragraph 15.0
compound sentiment score 0.9
negative sentiment score 0.1
positive sentiment score 0.1
word usage - conjunction 9
word usage - preposition 38
keypress duation - median 149.4
Summer's lexical diversity 0.9
interkey interval - median 190.8
readability - Coleman-Liau 12.8
word usage - nominalization 13
readability - DaleChallIndex 10.8
root type-token ratio (RTTR) 10.6
readability - GunningFogIndex 16.4
sentence beginnings - article 2
sentence beginnings - pronoun 1
readability - FleschReadingEase 47.5
sentence beginnings - conjunction 0
sentence beginnings - preposition 2
sentence beginnings - interrogative 0
sentence beginnings - subordination 0
hypergeometric distribution diversity (HD-D) 0.9