top of page

AI Readiness in Higher Education: Why Document Workflow Is the First Problem to Solve

  • Writer: David Holstein
    David Holstein
  • 15 minutes ago
  • 10 min read

TLDR: AI readiness in higher education is misunderstood as a model question, a vendor question, or a budget question. It is actually a document question. Every R1 institution we have worked with has the same first problem: decades of institutional knowledge trapped in PDFs, scanned forms, policy archives, and accommodation letters that no AI agent can reason against. We call it the PDF Graveyard. Three Bettera-built agents (alumni, student, research) demonstrate what becomes possible when institutions escape the graveyard. This piece walks through what the graveyard contains, what it does to AI agents, and the four-step sequence to escape it.


AI readiness higher education: the institutional PDF Graveyard inventory showing the document categories that block AI agents at most R1 institutions (alumni transcripts, accommodation letters, grant documents, accreditation evidence, policy archives, procurement records)

AI readiness in higher education is misunderstood as a model question, a vendor question, or a budget question. It is actually a document question.


Every R1 CIO has seen the same demo from a different vendor. Element451. Mainstay. Ocelot. Cayuse. EAB Navigate. Microsoft Copilot. Each one shows an AI agent that answers student questions, processes faculty requests, or surfaces alumni data in ways that look genuinely impressive. Each one also stalls in the same place when the institution tries to deploy it.


The agent works on the demo data. The institutional data is different. The institutional data lives in PDFs.


We call it the PDF Graveyard. Decades of institutional knowledge trapped in scanned forms, archived policies, accommodation letters, grant documents, accreditation evidence, procurement records, and advising notes. The graveyard is large at every R1 institution we have worked with. It is also the first problem to solve, before any AI agent can produce value at institutional scale.


Bettera is the only ServiceNow consulting partner exclusively focused on higher education, and we have built agents that operate against institutional documents rather than against vendor demo data. Three of those agents (alumni, student, research) are embedded in the body of this piece as proof. Each one shows the same pattern: an AI agent that reads institutional PDFs, cross-references against institutional records, and produces governed output the institution can defend.


This piece walks through what the PDF Graveyard contains, what it does to AI agents that try to navigate it, and the four-step sequence to escape it.


Why most higher ed AI deployments fail before they start

Three observations from Bettera's engagements with R1 institutions.


First: AI vendors demo on clean data. Every vendor demo shows the agent operating on a curated dataset where the documents are already structured, the records are already connected, and the institutional context is already encoded. The demo is genuinely impressive. The demo data is also not what the institution actually has.


Second: institutional data lives in documents the AI cannot read. Transcripts from 1995-2018 are scanned PDFs sitting in storage. Accommodation letters arrive as PDFs from outside physicians. Grant documents are 80-page PDFs combining technical narrative, budget tables, and signature blocks. Accreditation evidence is a folder structure of mixed document types accumulated over a decade. The institution has the knowledge. The AI cannot reach it.


Third: the integration work is harder than the AI work. Most AI deployment failures in higher ed are not failures of the AI model. They are failures to give the AI agent a structured representation of the institutional knowledge it needs to do its job. The agent works. The data path to the agent is broken.

We call this institutional condition the PDF Graveyard. The name is intentionally vivid. The reality is genuinely funereal: institutional knowledge buried in formats the AI cannot exhume.


Solving the PDF Graveyard is not optional. It is the first problem to solve before any AI deployment can produce value at institutional scale. Institutions that skip this step end up with expensive AI projects that demo well in IT presentations and stall when they hit real institutional workloads.


The next sections walk through what the graveyard contains, what it does to agents, and how to escape it.


What lives in the institutional PDF Graveyard

The inventory. Six document categories every R1 institution maintains, each blocking a different class of AI agent.


Alumni records. Transcripts, donation history, event attendance, mentorship participation, employment updates. Most institutions have decades of alumni records in mixed formats: scanned PDFs from the pre-digital era, structured database records from the digital era, and a long transition period in between with inconsistent quality.


Student records beyond the SIS. Accommodation letters from outside physicians, advising notes, behavioral intervention records, academic appeal documents, study abroad approvals. The structured student data lives in the SIS. The unstructured context lives in PDFs that travel through email, get filed in shared drives, and are nearly impossible for an AI agent to navigate.


Faculty research records. Grant proposals, IRB approvals, conflict-of-interest disclosures, sponsored research agreements, principal investigator histories, lab safety documentation. Research administration generates enormous PDF volumes that flow through email and SharePoint with minimal structure.


Accreditation evidence. Multi-year folder structures accumulated by accreditation cycle. SACSCOC, HLC, WSCUC, MSCHE, NEASC institutions all maintain large libraries of evidence documents the accreditation team can navigate manually but no AI agent can reason against.


Policy and procedure archives. Faculty handbooks, student conduct codes, administrative procedures, HR policies, financial procedures. The institution updates these periodically. Old versions sit in folders. The current version may exist in multiple conflicting copies. The AI agent does not know which is authoritative.


Procurement and contracts. Vendor agreements, RFP responses, sole-source justifications, software licenses, professional services agreements. The institution's contractual landscape is almost entirely PDF.


Each of these categories is its own burial ground. Each blocks a different class of AI agent.


What the PDF Graveyard does to AI agents

The operational consequence. What happens when AI agents encounter the graveyard.


Confabulation. The agent generates plausible-sounding answers that are not grounded in actual institutional knowledge. The student asks "what is the deadline to add a class this semester?" The agent answers based on training data that may or may not reflect the institution's current academic calendar. The answer sounds confident. It is wrong in a percentage of cases.


Refusal. The agent declines to answer because it has no access to the relevant document. The user gets an unhelpful response and loses confidence in the system. Each refusal is a credibility loss.


Escalation cascade. The agent escalates to a human staff member who has to look up the same PDF the agent could not parse. Every escalation is a failure to deflect, and the institution does not see the staff-time savings the AI deployment was supposed to produce.


Slow rollouts. Pilot deployments succeed on cherry-picked use cases where the documents happen to be cleaner. Expansion stalls when the use cases hit the unstructured majority of institutional data. The pilot becomes a permanent pilot.


Erosion of institutional confidence in AI. Each failure compounds. Faculty, staff, and students learn that the institutional AI cannot answer the questions that actually matter. The institution's AI brand becomes "the thing that does not work" rather than "the thing that helps."


This is the failure mode that solving the PDF Graveyard prevents.


Three Bettera agents that escape the graveyard

Three Bettera-built agents demonstrate the same pattern in three different institutional domains. Each one reads institutional documents that previously lived only in the PDF Graveyard. Each one cross-references against institutional records. Each one produces governed output the institution can defend.


Bettera Agents in Action — The Triptych

Agent 1: Alumni Portal — Transcript and Engagement

Sarah is an alumna requesting her digital transcript. The Alumni Agent generates the transcript, surfaces a $15 outstanding balance, and routes a Stripe payment. All in one conversation.



The agent navigates an alumni record graveyard that previously required three different staff members (registrar, bursar, alumni relations) to coordinate. The institution recovers the outstanding balance. The alumna gets a transcript faster than the manual workflow would have produced.


Agent 2: Student Portal — Medical Accommodation

Alex is a sick student requesting an academic accommodation. The Student Agent reads a medical certificate PDF from an outside physician, extracts the diagnosis and recommended accommodation duration, cross-references against the academic record, and asks Alex to confirm the interpretation.



The agent processes a document type that traditionally requires manual review by an accommodation coordinator. The student gets a faster response. The accommodation coordinator focuses on harder cases. The interpretation is preserved as a record the institution can defend.


Agent 3: Research Workflow — Grant-Triggered Staffing

Dr. Miller has just secured research funding (Grant #882). The Research Agent triggers a staffing approval flow with the principal investigator, processed via mobile one-tap approval.



The agent connects grant administration to faculty hiring workflow in a domain where research administrators traditionally spend weeks chasing approvals. The grant team gets to staff faster. The principal investigator gets the workflow out of email and into a structured approval surface.


The common pattern across all three demos: institutional documents that previously lived only in PDFs are now structured input for an AI agent. Outputs are governed by AI Control Tower. Records are preserved for FERPA and audit purposes. The institution captures value from knowledge that was previously inaccessible.


The four-step sequence to escape the PDF Graveyard


AI readiness higher education: the four-step sequence to escape the PDF Graveyard (inventory, prioritize, structure, deploy) with deliverables and exit criteria for each step

Bettera's operational framework for escaping the PDF Graveyard. Four steps. Each one depends on the prior step being complete.


Step 1: Inventory the graveyard. Catalog the document categories the institution maintains and the AI use cases the institution wants to enable. Map the intersections. Most institutions discover the graveyard is larger than they realized. The inventory itself is informative: it surfaces document categories the institution did not know it was maintaining.


Step 2: Prioritize by frequency of agent encounter. Not every document category needs immediate attention. The categories that AI agents will encounter most often (transcripts for alumni and registrar use cases, accommodation letters for student services, advising notes for academic affairs) get prioritized. Long-tail categories (twenty-year-old procurement contracts) can wait. The principle is operational, not exhaustive.


Step 3: Convert highest-frequency documents to structured representation. This is the operational work. OCR, document parsing, entity extraction, structured representation generation. The output is a structured data layer that AI agents can reason against, while the original documents remain preserved for legal and audit purposes. We discuss the deployment side of this work in our piece on knowledge management in higher education on ServiceNow.


Step 4: Deploy agents that read the structured representation. Once the structured data layer exists, agents can operate against it the way the vendor demos did. The three Bettera agents shown in the triptych above are the result of completing this sequence for alumni, student, and research domains.


The sequence is not optional. Each step depends on the prior step being complete.


Skipping any step produces the failure modes described in the previous section:

confabulation, refusal, escalation cascade, slow rollouts, erosion of institutional confidence.


Where AI readiness in higher education connects to the broader governance posture

This piece works as a standalone but also as the operational layer underneath the strategic and legal pieces in the Bettera AI series.


The institution's AI investment has four layers:


Strategic posture. Our piece on Now Assist in higher education AI governance walks through the three governance pillars (visibility, compliance, accountability) and the four FERPA edge cases. Without governance, AI deployment is risk without yield.


Platform deployment. Our piece on the ServiceNow AI Control Tower for higher education walks through the four-layer institutional configuration overlay and the five-phase deployment sequence. Without the platform, governance is policy-hoped-for rather than platform-enforced.


Legal posture. Our piece on FERPA and AI compliance in higher education walks through the four FERPA edge cases that counsel needs to interpret. Without legal posture, AI deployment is institutional exposure.


AI readiness. The piece you are reading. The document workflow foundation that makes the rest of the investment pay off. Without AI readiness, the governance, the platform, and the legal posture are scaffolding around an empty building.


All four layers matter. Skipping any of them produces predictable failure modes. The institutions that do all four are the institutions that achieve defensible AI deployment at scale.


The knowledge management piece linked in Step 3 above (knowledge management in higher education on ServiceNow) is where the AI readiness work gets the deployment walkthrough. AI readiness is fundamentally a knowledge-management problem with an AI consumer at the end of the pipeline. Solving the PDF Graveyard is what makes the AI investment work.


Frequently asked questions

What is AI readiness in higher education?

AI readiness in higher education is the institution's capacity to deploy AI agents that produce value at institutional scale. It depends on four layers: strategic governance posture, platform deployment (typically ServiceNow's AI Control Tower for institutions running ServiceNow), legal posture under FERPA, and document workflow readiness. The fourth layer (document workflow) is what most institutions underestimate. We call the unsolved version of this problem the PDF Graveyard, and solving it is the precondition for the other three layers paying off.


What is the PDF Graveyard?

The PDF Graveyard is Bettera's term for the institutional knowledge that lives only in unstructured documents (PDFs, scanned forms, policy archives, accommodation letters, grant documents, accreditation evidence, procurement records). AI agents cannot reason against documents in this form. Most higher ed AI deployments fail because the institutional data lives in the PDF Graveyard and the AI agents cannot reach it.


Why do AI agents fail at higher education institutions?

The most common failure pattern is not AI model failure. It is the data path to the agent being broken because institutional knowledge lives in the PDF Graveyard. Five failure modes follow: confabulation (the agent invents plausible-sounding answers), refusal (the agent declines to answer), escalation cascade (the agent escalates to humans who look up the same PDF the agent could not parse), slow rollouts (pilots succeed on cherry-picked data and stall on real workloads), and erosion of institutional confidence in AI generally.


Do we need new AI vendors, or can ServiceNow agents work on our existing institutional data?

ServiceNow agents can work on institutional data once the data has been structured. The Bettera-built agents shown in the triptych (alumni, student, research) operate on the ServiceNow platform against institutional documents that have been moved out of the PDF Graveyard. New AI vendors are not the answer to the document workflow problem.


They face the same data-path challenge that any agent faces and typically come with additional integration and governance burden.


What is the relationship between knowledge management and AI readiness?

AI readiness is fundamentally a knowledge-management problem with an AI consumer at the end of the pipeline. The knowledge management work (structuring unstructured documents, building unified knowledge bases, generating AI-readable representations of institutional knowledge) is what makes AI agents work. We walk through the KM deployment work in our piece on knowledge management in higher education on ServiceNow.


How long does it take to make institutional documents AI-ready?

Timeline depends on the scope of the PDF Graveyard, the prioritization choices the institution makes, and the AI use cases targeted first. Institutions that focus on the highest-frequency document categories first (transcripts, accommodation letters, advising notes) typically see early-use-case readiness in months rather than years. The full graveyard takes longer to clear, but full clearance is not necessary to deploy useful agents.


What is the first step to escaping the PDF Graveyard?

The inventory. Catalog the document categories the institution maintains and the AI use cases the institution wants to enable. Map the intersections. The inventory itself is informative: most institutions discover the graveyard is larger than they realized, and the inventory surfaces document categories the institution did not know it was maintaining. The next step is prioritization by frequency of agent encounter.


Where this leaves the institution

The PDF Graveyard is real at every R1 institution we have worked with. Solving it is not optional. The institutions that solve it produce AI deployments that work at institutional scale. The institutions that skip it produce expensive AI projects that demo well in IT presentations and stall when they hit real institutional workloads.


If your institution is exploring AI deployment and the documents are the question you have not yet been able to answer, that is the working session we facilitate at Bettera.


Contact us and we will walk through your institution's PDF Graveyard together using the four-step escape sequence.


Bettera is the only ServiceNow consulting partner exclusively focused on higher education, and AI readiness is the first conversation we have with every R1 institution scoping AI deployment.


Public Sources Cited


Beyond the Help Desk White Paper

bottom of page