Search This Blog

Thursday, March 27, 2025

Freeze-Dried Text Experiment

It is like instant coffee, or a shrunken pear: too dry to eat, but OK if you add water.  Meet "freeze-dried text" – concentrated idea nuggets waiting to be expanded by AI. Copy everything below this paragraph into any AI and watch as each transforms into real text. Caution: AI will hallucinate some references. Remember to type "NEXT" after each expansion to continue. Avoid activating any deep search features – it will slow everything down. This could be how we communicate soon – just the essence of our thoughts, letting machines do the explaining. Perhaps the textbooks of the future will be written that way. Note, the reader can choose how much explanation they really need - some need none, others plenty. So it is a way of customizing what you read. 

Mother Prompt

Expand each numbered nugget into a detailed academic paper section (approximately 500 words) on form-substance discrimination (FSD) in writing education. Each nugget contains a concentrated meaning that needs to be turned into a coherent text.

Maintain a scholarly tone while including:

Theoretical foundations and research support for the claims. When citing specific works, produce non-hallucinated real reference list after each nugget expansion.  

Practical implications with concrete examples only where appropriate.

Nuanced considerations of the concept's complexity, including possible objections and need for empirical research. 

Clear connections to both cognitive science and educational practice.

Smooth transitions that maintain coherence with preceding and following sections

Expand nuggets one by one, treating each as a standalone section while ensuring logical flow between sections. Balance theoretical depth with practical relevance for educators, students, and institutions navigating writing instruction in an AI-augmented landscape. Wait for the user to encourage each next nugget expansion. Start each Nugget expansion with an appropriate Subtitle 

Nuggets

1. Form-substance discrimination represents a capacity to separate rhetorical presentation (sentence structure, vocabulary, organization) from intellectual content (quality of ideas, logical consistency, evidential foundation), a skill whose importance has magnified exponentially as AI generates increasingly fluent text that may mask shallow or nonsensical content.
2. The traditional correlation between writing quality and cognitive effort has been fundamentally severed by AI, creating "fluent emptiness" where writing sounds authoritative while masking shallow content, transforming what was once a specialized academic skill into an essential literacy requirement for all readers.
3. Cognitive science reveals humans possess an inherent "processing fluency bias" that equates textual smoothness with validity and value, as evidenced by studies showing identical essays in legible handwriting receive more favorable evaluations than messy counterparts, creating a vulnerability that AI text generation specifically exploits.
4. Effective FSD requires inhibitory control—the cognitive ability to suppress automatic positive responses to fluent text—paralleling the Stroop task where identifying ink color requires inhibiting automatic reading, creating essential evaluative space between perception and judgment of written content.
5. The developmental trajectory of FSD progresses from "surface credibility bias" (equating quality with mechanical correctness) through structured analytical strategies (conceptual mapping, propositional paraphrasing) toward "cognitive automaticity" where readers intuitively sense intellectual substance without conscious methodological application.
6. Critical thinking and FSD intersect in analytical practices that prioritize logos (logical reasoning) over ethos (perceived authority) and pathos (emotional appeal), particularly crucial for evaluating machine-generated content that mimics authoritative tone without possessing genuine expertise.
7. The "bullshit detection" framework, based on Frankfurt's philosophical distinction between lying (deliberately stating falsehoods) and "bullshitting" (speaking without concern for truth), provides empirical connections to FSD, revealing analytical reasoning and skeptical disposition predict resistance to pseudo-profound content.
8. Institutional implementation of FSD requires comprehensive curricular transformation as traditional assignments face potential "extinction" in a landscape where students can generate conventional forms with minimal intellectual engagement, necessitating authentic assessment mirroring real-world intellectual work.
9. Effective FSD pedagogy requires "perceptual retraining" through comparative analysis of "disguised pairs"—conceptually identical texts with divergent form-substance relationships—developing students' sensitivity to distinction between rhetorical sophistication and intellectual depth.
10. The pedagogical strategy of "sloppy jotting" liberates students from formal constraints during ideation, embracing messy thinking and error-filled brainstorming that frees cognitive resources for substantive exploration while creating psychological distance facilitating objective evaluation.
11. Students can be trained to recognize "algorithmic fingerprints" in AI-generated text, including lexical preferences (delve, tapestry, symphony, intricate, nuanced), excessive hedging expressions, unnaturally balanced perspectives, and absence of idiosyncratic viewpoints, developing "algorithmic skepticism" as distinct critical literacy.
12. The "rich prompt technique" for AI integration positions technology as writing assistant while ensuring intellectual substance comes from students, who learn to gauge necessary knowledge density by witnessing how vague AI instructions produce sophisticated-sounding but substantively empty content.
13. Assessment frameworks require fundamental recalibration to explicitly privilege intellectual substance over formal perfection, with rubrics de-emphasizing formerly foundational skills rendered less relevant by AI while ensuring linguistic diversity is respected rather than penalized.
14. FSD serves as "epistemic self-defense"—equipping individuals to maintain intellectual sovereignty amid synthetic persuasion, detecting content optimized for impression rather than insight, safeguarding the fundamental value of authentic thought in knowledge construction and communication.
15. The contemporary significance of FSD extends beyond academic contexts to civic participation, as citizens navigate information ecosystems where influence increasingly derives from control over content generation rather than commitment to truth, making this literacy essential for democratic functioning.





Monday, March 24, 2025

Two Reactions to AI

A batch of student essays. About a third are clearly AI-generated. Two professors—same discipline, same university, same evidence—react in diametrically opposite ways. Rodrigo sighs with relief. Jane spirals into panic.

For Rodrigo, it is almost liberating. If his students can now write coherent college-level essays with the help of machines, then he is free to teach them something more ambitious. Argument structure, epistemology, stylistic nuance—areas where automation falters. He is not naïve; he knows AI is here to stay. But rather than fight it, he welcomes the detour. Less time marking the same intro-to-critical-writing dreck, more time pushing the intellectual envelope. Lucky him.

Jane, however, reads the situation as academic apocalypse. Her course was the product of years of iteration, finely tuned to teach writing through careful scoping, scaffolding, and feedback. Skip the process, she believes, and you skip the learning. The AI is not a tool in her eyes; it is a cheat code, one that threatens to render her teaching obsolete. She starts researching detection tools, imagining a future of surveillance, suspicion, and pedagogical collapse.

These are not just personality quirks or different thresholds for academic dishonesty. What really separates them is how they understand curriculum. For Rodrigo, curriculum is plastic—something owned, revised, improved. He feels empowered to tinker. If a foundational skill can be outsourced, then the baseline has shifted, and he can raise the stakes. A change in student capability is an opportunity, not a crisis.

Jane sees curriculum differently. For her, it is an infrastructure. Complex, interdependent, and not easily re-routed. Learning outcomes, general education requirements, accreditation standards—these are not suggestions, they are fixtures. If a key skill like essay-writing becomes an unreliable indicator of mastery, the whole sequence threatens to unravel. You cannot simply skip a floor in the building and hope the roof holds.

There is a quiet tragedy here. Not because Jane is wrong—her concerns are deeply valid—but because she feels disempowered by a system she herself has worked so hard to uphold. The larger structures of academia—its bureaucracies, its resistance to rapid change—amplify the sense of threat. It is not just that students are using ChatGPT; it is that there is no institutional plan, no curricular pivot, no workflow update to guide faculty through this transition. So each professor is left to improvise, bringing their own philosophies and tolerances to bear.

And that is where the real tension lies. Technology does not just disrupt skills—it exposes fault lines in our educational ideologies. Are we guardians of a process or facilitators of progress? Should we protect the sequence, or adjust the map when shortcuts appear?

Rodrigo shrugs and walks forward. Jane looks for the brakes. But maybe it is not about who is right. Maybe the more urgent task is to build a system where professors do not have to choose between clinging to the past and embracing a future they did not ask for. Because either way, the syllabus is no longer entirely ours to write.

UPD: Thanks to my colleague Hogan Hays for his thoughtful critique of this blog


Wednesday, March 19, 2025

RAG and the Tyranny of Text

Writing and reading are, at their core, terribly inefficient. To communicate knowledge, we take  complex non-linear understanding and flatten it into a linear string of symbols—words, sentences, paragraphs—then expect someone else to decode those symbols one by one to reconstruct the original meaning. For every piece of information useful to us in a particular moment, we probably read thousands of unnecessary words. Laws, academic research, instruction manuals—entire professions exist solely to interpret and summarize the large texts, and find the bits useful for a particular case.

We are so accustomed to this system that we barely question it. We assume that knowledge must be buried in thick books, endless PDFs, or jargon-laden policies, and that extracting value from them is simply the price we pay. The reality is that text, as a technology, is painfully exclusionary. It filters out those who do not have the time, education, or patience to wade through its inefficiencies. The result? A world where information is not truly accessible—it is just available, locked behind barriers of expertise and labor. The problem only growth with the increase of information. We can search now, but search you need to know what exactly the thing you're searching is called. 

Enter Retrieval-Augmented Generation (RAG). This technology upends the whole premise of reading as a necessary struggle. Instead of requiring humans to sift through dense documents, a RAG-powered AI can scan, understand, and extract the exact information you need. It will understand you even you're not sure what to look for. No more endless searching, skimming, or cross-referencing. You ask, it finds and explains at whatever level of difficulty you are comfortable with, in any language.

The applications are obvious. College course materials, legal codes, corporate policies—things we must understand but rarely want to read—can now be accessed through AI assistants that do the heavy lifting. Medical test results, car repair manuals, tax codes—fields where knowledge has traditionally been mediated by experts—become directly intelligible to the people who need them. RAG doesn’t just speed up information retrieval; it removes the gatekeepers.

Despite the significance of this shift, most major AI companies have not fully embraced it. OpenAI is the only major player that has prioritized user-friendly RAG-based tools, allowing everyday users to create and share custom bots. The others—Anthropic, Google Gemini, Meta, Grok, Deep Seek— all offer API-based solutions that cater to developers, not the general public. Gemini allows its paid users to create custom bots, but somehow, inexplicably, does not allow to share them. It is a strange oversight. The AI race is usually about copying and outpacing competitors, yet here, OpenAI is sprinting ahead while others somehow hesitate.

The gap has created an opportunity. Startups are rushing in to offer the ease of use that the AI giants have neglected, sensing that the true power of AI is not just in intelligence but in revolutionary leap to accessibility. AI is, by nature, a democratic technology—relatively cheap, scalable, and available to almost anyone. And yet, its most transformative application—RAG—is still frustratingly out of reach for many. 

What we are witnessing is the beginning of a fundamental shift. For centuries, knowledge has been tied to advanced literacy (the ability to create and understand long texts), to institutions, to intermediaries who dictate who gets to understand what. RAG challenges that structure. It does not just improve search; it changes who gets to find answers in the first place. If AI is truly to fulfill its promise, it won’t be by making people read faster—it will be by making linear reading largely obsolete. We will always always read novels and poems word by word, because humans created art out of the terrible technology of writing. But those are only small portion of written information we consume. 



Wednesday, March 12, 2025

The Modern Confessional: AI Disclosure as Ritual

 

Organizations across academia and publishing now routinely demand confession of AI use. Publishers require authors to disclose whether AI tools assisted in manuscript preparation. Funding agencies insert checkboxes for AI utilization. Academic journals add disclosure statements to submission forms. None adequately explain their rationale or how this information shapes evaluation.

This peculiar practice reveals our creation of a new moral domain around AI use in knowledge production. The requirement to disclose functions precisely as Michel Foucault described confessional practices in "The History of Sexuality." Foucault argued that confession itself produces the very notion of sin it purports to address. The act of requiring disclosure creates the impression of transgression where none inherently exists.

Medieval confession did not merely document pre-existing sins - it manufactured them through the very apparatus of confession. Similarly, disclosure requirements around AI use manufacture a transgressive quality around technologies that have no inherent moral valence.

The mechanics operate almost identically. Both create categories of behavior requiring special scrutiny. Both position authority figures as arbiters of acceptability. The confessing subject experiences this manufactured transgression viscerally - the academic disclosing AI use feels compelled toward contrition without clear understanding of what offense they have committed.

Authors find themselves in impossible positions, uncertain how much assistance constitutes meaningful use. Did grammar checking through Microsoft Editor count? What about Grammarly's suggestions? The lack of clear standards transforms disclosure into guesswork.

Rather than focusing on tools, we might evaluate outputs based on established academic standards regardless of production methods. This approach acknowledges that quality, originality and intellectual integrity manifest in final products, not production processes. Technical assistance has always existed across academic work - from statistical software to citation managers to editorial help from colleagues.

Current disclosure requirements function primarily as modern confession, manufacturing transgression through the very apparatus designed to reveal it. By recognizing this dynamic, we might reimagine our approach to technological assistance in ways that foster genuine integrity rather than performative disclosure.


Why Education Clings to Irrelevance

There’s a particular irony in the way schools prepare us for life by arming us with tools designed for a world that no longer exists. Latin,...