The "Invisible" AI Layer

Unlocking Your Firm's Knowledge Without Changing Workflows

KNOWLEDGE MANAGEMENTAIRAGLLM

Dr. Alina Pukhovskaya

1/21/20263 min read

Today, your firm already works mostly in SharePoint (or Google Drive, etc.).

You don’t want to move everything somewhere else — you just want the AI to “understand” what’s there, stay up-to-date, and respect permissions.

The solution is to install a secure Organizational LLM paired with a Retrieval-Augmented Generation (RAG) pipeline.

In this architecture, the Large Language Model acts as the reasoning engine, while the RAG system acts as the retrieval mechanism. Instead of training the model on your data (which is static and expensive), the RAG system indexes your content in real-time. This allows the AI to reference your specific files to generate accurate, citation-backed answers without ever moving the source data.

Here’s how it works in practice:

1. Where files live — nothing changes for employees

Employees keep saving and updating files exactly where they always do:

  • Project folders in SharePoint (“Client X > Deliverables > Final Report.docx”)

  • Team spaces in Confluence for meeting notes or procedures

  • Personal folders on OneDrive or Drive if already part of your governance policy


They don’t need to upload or move files anywhere new. The system never replaces SharePoint or Confluence; it simply reads from them in the background.

2. How the system stays up-to-date

A lightweight connector or sync job runs on your internal server (or Azure tenant). It uses each system’s secure API to watch for file changes:

  • When a document is created or updated, the connector re-processes that single file through the local pipeline, extracts its text again, and updates the vector index.

  • When a document is deleted, the system removes its corresponding record and embeddings.


This sync typically runs on a schedule (e.g., every 12 hours or nightly) or near-real-time for high-value folders.

So, if someone uploads a new client proposal at 5 p.m., it’s searchable in the AI assistant the next morning.

3. What happens behind the scenes

Inside your private environment, you now have two internal “layers”:

Primary storage: The original files remain in SharePoint, Confluence, etc. Example: “Client X > Deliverables > Final Report.docx”

Knowledge base (index): A local database containing only cleaned text and metadata. Example: “Chunk #218 – Client X Final Report.docx, page 3, topic: ROI analysis”

When an employee asks the AI: “What were the ROI assumptions in Client X’s project?”

The LLM doesn’t open the file itself — it retrieves the right chunks from the knowledge base and shows the source link back to the live SharePoint file.

So users always see (and can open) the original, most recent file.

4. Managing updates and versioning

Because SharePoint and Confluence already handle version control, we don’t duplicate that logic. The connector simply checks each file’s last-modified timestamp or version ID:

  • If newer → re-parse and replace the old entry in the index.

  • If same → skip (no wasted processing).

  • If deleted → remove from index.


This keeps your AI knowledge base light and clean even if thousands of files change over time.

5. Handling new files

Whenever someone adds a new document in an existing folder that’s being monitored, the connector automatically picks it up.

If a new team or department launches a new SharePoint library, your admin just adds that path to the sync configuration once — after that, it’s automatic.

6. What users experience

From an employee’s point of view:

  • They continue working in SharePoint/Confluence as usual.

  • They can open the internal “Knowledge Assistant” web page and ask questions in plain language.

  • The system responds with answers and links to the relevant source documents.

  • When they click, they go straight to the document in SharePoint with the correct permissions.


No new habits, no new folders — the AI layer quietly sits on top of existing systems.

7. Governance and control
  • Access rights are mirrored from the source systems (if someone can’t open a SharePoint file, they won’t see its content in search results).

  • IT or Knowledge Management defines which repositories are indexed.

  • Periodic audits ensure obsolete or sensitive folders are excluded.

8. Future scalability

Later, if you add new tools (Teams chats, ServiceNow tickets, Jira issues), you simply add another connector — still within your private infrastructure.

The system grows horizontally without changing employee workflows.

Ready to Build Predictable Growth?