ibl.ai AI Education Blog

Explore the latest insights on AI in higher education from ibl.ai. Our blog covers practical implementation guides, research summaries, and strategies for AI tutoring platforms, student success systems, and campus-wide AI adoption. Whether you are an administrator evaluating AI solutions, a faculty member exploring AI-enhanced pedagogy, or an EdTech professional tracking industry trends, you will find actionable insights here.

Topics We Cover

Featured Research and Reports

We analyze key research from leading institutions including Harvard, MIT, Stanford, Google DeepMind, Anthropic, OpenAI, McKinsey, and the World Economic Forum. Our premium content includes audio summaries and detailed analysis of reports on AI impact in education, workforce development, and institutional strategy.

For University Leaders

University presidents, provosts, CIOs, and department heads turn to our blog for guidance on AI governance, FERPA compliance, vendor evaluation, and building AI-ready institutional culture. We provide frameworks for responsible AI adoption that balance innovation with student privacy and academic integrity.

Interested in an on-premise deployment or AI transformation? Call or text 📞 (571) 293-0242
Back to Blog

Software Bill of Materials (SBOM) for the ibl.ai Platform

Miguel AmigotJune 2, 2025
Premium

SBOM, software bill of materials, generative AI platform, LLM-agnostic, LangChain, Langfuse, Flowise, OpenAI GPT-4, Google Gemini, Azure OpenAI, Anthropic Claude, AWS Bedrock, open-source LMS, OpenAPI, Python SDK, JavaScript SDK, OAuth2, OIDC, SAML, LTI 1.3, ReactJS, Next.js, React Native, MentorAI, university CIO, edtech, AI tutor, permissive licenses, vendor lock-in avoidance, cost control, enterprise security, higher education technology

The ibl.ai platform is a generative-AI-powered learning system built on an open-source LMS foundation, extended with a flexible LLM layer and fully exposed through OpenAPI-compliant services. All core components use permissive licenses (MIT / Apache 2.0), ensuring zero hidden licensing costs and no vendor lock-in for the institution. The architecture is modular, standards-based, and designed for secure campus deployment on-prem or in any cloud.


1 · Generative AI Engine & Frameworks

ComponentRole in PlatformLicense
LangChainFramework for building and chaining LLM-powered tools; powers tutoring agents, content generation, and multi-model orchestration.MIT
LangfuseObservability & tracing layer for LLM calls; enables prompt/response logging, performance dashboards, and debugging.MIT
FlowiseNo-code visual builder for LLM workflows and agents; accelerates rapid prototyping and custom AI flows.Apache 2.0
OpenAI SDK (Python & Node)Official libraries for GPT's latest models; supports streaming, fine-tuning, and advanced usage analytics.MIT
Google Gemini SDKUnified client for Gemini models on Vertex AI; offers multimodal (text + image) generation and enterprise controls.MIT

2 · Supported Large Language Models

ProviderExample ModelsHighlights
OpenAILatest Available ModelsLeading accuracy, broad ecosystem, coding & conversation excellence.
Google Cloud AIGeminiNative multimodal reasoning, Vertex AI integration, fine-tune workflows.
Microsoft AzureAzure OpenAIEnterprise compliance, regional data residency, Azure AD integration.
AnthropicLatest Claude ModelsSafety-focused “Constitutional AI,” 100k-token context for long documents.
AWS BedrockAmazon Titan (+ third-party models)Flexible mix-and-match models under AWS security and cost controls.

Model-agnostic: Administrators may choose, combine, or swap models without code changes.


3 · Platform Core (LMS & API)

ComponentDescriptionLicense
Open-Source LMS CoreFull course delivery, enrollment, grading, and analytics engine. Mature, scalable, and extensible to meet university requirements.Permissive OSS
REST API (OpenAPI)100 % feature coverage via OpenAPI-defined endpoints; supports content, tutoring, analytics, and admin operations.
Python SDKAuto-generated client; simplifies server-side integrations and data pipelines.MIT
JavaScript / TypeScript SDKAuto-generated client for web/mobile apps and serverless functions.MIT
Auth LayerOAuth2 / OIDC & SAML2 for SSO, plus LTI 1.3 for cross-LMS embedding; supports server-to-server or client-initiated flows.

4 · Front-End & Application Ecosystem

Framework / AppPurposeLicense
ReactJSCore library for dynamic web UIs (dashboards, portals).MIT
Next.jsFull-stack React framework with server-side rendering and API routes.MIT
React NativeCross-platform mobile framework (iOS & Android).MIT
MentorAI (reference app)Pre-built AI tutor (web + mobile) showcasing best practices; code shareable with the university.Source available
Custom Partner AppsAny partner-built web/mobile apps leveraging the OpenAPI & SDKs; authenticate via OAuth2/SSO.OSS frameworks

5 · Licensing & Cost Perspective

  • All components use permissive licenses (MIT / Apache 2.0 / AGPLv3).
  • No proprietary runtime fees; the university retains full code ownership.
  • Costs arise only from optional usage-based AI model calls with chosen providers.
  • Modular design allows on-prem, cloud, or hybrid deployment while meeting security and compliance requirements.

6 · At-a-Glance Benefits for CIOs

PillarValue Delivered
Open & ExtensibleOpenAPI endpoints, open-source code, flexible SDKs.
Vendor-Neutral AISwap or mix LLM providers without lock-in.
Enterprise SecurityOAuth2/OIDC, SAML, LTI 1.3, and role-based access.
Future-ProofRapid adoption of new models via LangChain & Flowise.
Cost ControlNo platform license fees; pay only for chosen AI usage and infra.

Related Articles

Microsoft Copilot + ibl.ai: Building an AI stack universities actually own

Microsoft Copilot excels as a GPT-4 assistant baked into Microsoft 365, yet it lacks the course-grounding, data residency, and model flexibility campuses require. ibl.ai’s open, LLM-agnostic mentorAI backend supplies that secure layer—RAG over syllabus content, multi-tenant SOC 2/FERPA controls, analytics, and big cost savings—so universities keep Copilot’s front-line productivity while owning the AI core.

Jaione AmigotMay 7, 2025

How mentorAI Integrates with Anthropic

mentorAI lets universities route each task to Anthropic’s Claude 3 family through their own Anthropic API key or AWS Bedrock endpoint, sending high-volume chats to Haiku (≈ 21 k tokens per second), deeper tutoring to Sonnet, and 200 k-context research queries to Opus—no code changes required. The platform logs every token, enforces safety filters, and keeps transcripts inside the institution’s cloud, while Anthropic’s commercial-API policy of not using customer data for training protects FERPA/GDPR compliance.

Jeremy WeaverMay 7, 2025

How ibl.ai Supercharges Khan Academy’s Mission—Without Competing

Khanmigo offers GPT-4-powered, student-friendly tutoring on top of Khan Academy’s content, but campuses still need secure ownership, LMS/SIS integration, and model flexibility. ibl.ai’s mentorAI supplies that backend—open code, LLM-agnostic orchestration, compliance tooling, analytics, and cost control—letting universities embed Khanmigo today, swap models tomorrow, and run everything inside their own cloud without vendor lock-in.

Jaione AmigotMay 7, 2025

Claude + ibl.ai: A Blueprint for AI-Native Universities

Anthropic’s new Claude for Education supplies the guarded, Socratic chat front end, while ibl.ai’s share-the-code MentorAI delivers the back-office muscle—LLM-agnostic orchestration, SSO/LTI, audit logs, and faculty overrides—inside a university-owned cloud. Together they ground Claude in syllabus files, blend models, monitor costs, and swap engines at will, eliminating lock-in.

Jaione AmigotMay 7, 2025

See the ibl.ai AI Operating System in Action

Discover how leading universities and organizations are transforming education with the ibl.ai AI Operating System. Explore real-world implementations from Harvard, MIT, Stanford, and users from 400+ institutions worldwide.

View Case Studies

Get Started with ibl.ai

Choose the plan that fits your needs and start transforming your educational experience today.