top of page

AIcademy

Build a RAG system with internal data sources

Description


In this training you will learn how to set up a Retrieval-Augmented Generation (RAG) system that allows you to have large language models (LLMs) answer based on your own documents, knowledge sources or internal systems. By combining document search with text generation you create AI applications that are more up-to-date, reliable and domain-specific than generic language models alone.

You will discover how to convert documents to embeddings, store them in a vector database and link them to an LLM via retrieval. We work with tools such as LangChain , LlamaIndex , Hugging Face Transformers and vector databases such as FAISS , Weaviate , Pinecone or Chroma . You will also learn how to use PDFs, Word files, web pages or internal reports as input and how to build a smart interface or API on top of your RAG pipeline.


Learning objectives


  • Understanding what a RAG architecture is and when to use it (vs. fine-tuning or prompt engineering)

  • Process and chunk documents in a way that fits your use case

  • Generating Embeddings with Models via Hugging Face or OpenAI

  • Using vector databases such as FAISS, Weaviate or Pinecone for semantic searches

  • Apply metadata filtering, ranking and hybrid search for more accurate retrieval

  • Building RAG pipelines with LangChain or LlamaIndex

  • Integrate LLMs via Hugging Face Transformers, OpenAI API or Anthropic API

  • Building a working application with a front-end (e.g. Streamlit) or REST API


Approach and working methods


The training is completely hands-on. You build a working RAG solution step by step: from document processing to integration with a language model and a working interface. We cover all layers of the architecture and you learn how to make choices based on data volume, user needs and scalability.

Examples of applications you can build:

  • A legal or policy assistant based on your own documents

  • An HR bot that provides answers based on personnel information and regulations

  • An IT search assistant that answers helpdesk questions based on manuals and internal documentation

  • A project assistant who interprets and summarizes reports, minutes and decisions


You work with:

  • Python and frameworks such as LangChain or LlamaIndex

  • Vector databases : FAISS, Chroma, Weaviate, Pinecone

  • Embeddings of OpenAI or Hugging Face (e.g. BGE, Instructor, MiniLM)

  • LLMs via OpenAI, Hugging Face Hub or own hosted models

  • Frontend/API via Streamlit, FastAPI or Flask


For whom


This course is intended for AI engineers, ML specialists, developers and data scientists who want to develop smart AI applications based on their own documents or knowledge sources. Experience with Python and basic knowledge of LLMs and vector search is recommended.


Interested in this training?


Feel free to contact us. We are happy to tailor the training to your specific use case, type of documents and technical preferences.


Build a RAG system with internal data sources

Description


In this training you will learn how to set up a Retrieval-Augmented Generation (RAG) system that allows you to have large language models (LLMs) answer based on your own documents, knowledge sources or internal systems. By combining document search with text generation you create AI applications that are more up-to-date, reliable and domain-specific than generic language models alone.

You will discover how to convert documents to embeddings, store them in a vector database and link them to an LLM via retrieval. We work with tools such as LangChain , LlamaIndex , Hugging Face Transformers and vector databases such as FAISS , Weaviate , Pinecone or Chroma . You will also learn how to use PDFs, Word files, web pages or internal reports as input and how to build a smart interface or API on top of your RAG pipeline.


Learning objectives


  • Understanding what a RAG architecture is and when to use it (vs. fine-tuning or prompt engineering)

  • Process and chunk documents in a way that fits your use case

  • Generating Embeddings with Models via Hugging Face or OpenAI

  • Using vector databases such as FAISS, Weaviate or Pinecone for semantic searches

  • Apply metadata filtering, ranking and hybrid search for more accurate retrieval

  • Building RAG pipelines with LangChain or LlamaIndex

  • Integrate LLMs via Hugging Face Transformers, OpenAI API or Anthropic API

  • Building a working application with a front-end (e.g. Streamlit) or REST API


Approach and working methods


The training is completely hands-on. You build a working RAG solution step by step: from document processing to integration with a language model and a working interface. We cover all layers of the architecture and you learn how to make choices based on data volume, user needs and scalability.

Examples of applications you can build:

  • A legal or policy assistant based on your own documents

  • An HR bot that provides answers based on personnel information and regulations

  • An IT search assistant that answers helpdesk questions based on manuals and internal documentation

  • A project assistant who interprets and summarizes reports, minutes and decisions


You work with:

  • Python and frameworks such as LangChain or LlamaIndex

  • Vector databases : FAISS, Chroma, Weaviate, Pinecone

  • Embeddings of OpenAI or Hugging Face (e.g. BGE, Instructor, MiniLM)

  • LLMs via OpenAI, Hugging Face Hub or own hosted models

  • Frontend/API via Streamlit, FastAPI or Flask


For whom


This course is intended for AI engineers, ML specialists, developers and data scientists who want to develop smart AI applications based on their own documents or knowledge sources. Experience with Python and basic knowledge of LLMs and vector search is recommended.


Interested in this training?


Feel free to contact us. We are happy to tailor the training to your specific use case, type of documents and technical preferences.


1.jpg

Description:
Learn how to set up a Retrieval-Augmented Generation (RAG) system that lets AI answer questions based on your own documents and knowledge sources.


Learning objectives:

  • Working with vector databases such as FAISS and Weaviate.

  • Chunking strategies, metadata filtering, and hybrid search methods.

  • Setting up RAG pipelines with LangChain or LlamaIndex.

  • Connecting document data to a front-end or API.


For whom: AI engineers and developers who want to connect AI to internal knowledge.

Build a RAG system with internal data sources

Description


In this training you will learn how to set up a Retrieval-Augmented Generation (RAG) system that allows you to have large language models (LLMs) answer based on your own documents, knowledge sources or internal systems. By combining document search with text generation you create AI applications that are more up-to-date, reliable and domain-specific than generic language models alone.

You will discover how to convert documents to embeddings, store them in a vector database and link them to an LLM via retrieval. We work with tools such as LangChain , LlamaIndex , Hugging Face Transformers and vector databases such as FAISS , Weaviate , Pinecone or Chroma . You will also learn how to use PDFs, Word files, web pages or internal reports as input and how to build a smart interface or API on top of your RAG pipeline.


Learning objectives


  • Understanding what a RAG architecture is and when to use it (vs. fine-tuning or prompt engineering)

  • Process and chunk documents in a way that fits your use case

  • Generating Embeddings with Models via Hugging Face or OpenAI

  • Using vector databases such as FAISS, Weaviate or Pinecone for semantic searches

  • Apply metadata filtering, ranking and hybrid search for more accurate retrieval

  • Building RAG pipelines with LangChain or LlamaIndex

  • Integrate LLMs via Hugging Face Transformers, OpenAI API or Anthropic API

  • Building a working application with a front-end (e.g. Streamlit) or REST API


Approach and working methods


The training is completely hands-on. You build a working RAG solution step by step: from document processing to integration with a language model and a working interface. We cover all layers of the architecture and you learn how to make choices based on data volume, user needs and scalability.

Examples of applications you can build:

  • A legal or policy assistant based on your own documents

  • An HR bot that provides answers based on personnel information and regulations

  • An IT search assistant that answers helpdesk questions based on manuals and internal documentation

  • A project assistant who interprets and summarizes reports, minutes and decisions


You work with:

  • Python and frameworks such as LangChain or LlamaIndex

  • Vector databases : FAISS, Chroma, Weaviate, Pinecone

  • Embeddings of OpenAI or Hugging Face (e.g. BGE, Instructor, MiniLM)

  • LLMs via OpenAI, Hugging Face Hub or own hosted models

  • Frontend/API via Streamlit, FastAPI or Flask


For whom


This course is intended for AI engineers, ML specialists, developers and data scientists who want to develop smart AI applications based on their own documents or knowledge sources. Experience with Python and basic knowledge of LLMs and vector search is recommended.


Interested in this training?


Feel free to contact us. We are happy to tailor the training to your specific use case, type of documents and technical preferences.


How It All Started

This is a space to share more about the business: who's behind it, what it does and what this site has to offer. It's an opportunity to tell the story behind the business or describe a special service or product it offers. You can use this section to share the company history or highlight a particular feature that sets it apart from competitors.

Let the writing speak for itself. Keep a consistent tone and voice throughout the website to stay true to the brand image and give visitors a taste of the company's values and personality.

bottom of page