Quickstart
Build your first RAG pipeline in minutes
This quickstart gets you up and running with GraphorLM as quickly as possible. It introduces you to the platform’s interface and core RAG pipeline components without diving into technical details or complex configurations.
In this tutorial, you will:
- Load your data into GraphorLM
- Configure a basic RAG pipeline with intelligent chunking
- Deploy your pipeline
Step one: Set up GraphorLM Project
GraphorLM Cloud provides a fully-managed environment where you can build, test, and deploy RAG pipelines without worrying about infrastructure or complex setup. The free trial gives you full access to all platform features.
- Sign up for a GraphorLM Cloud account
- Login to your account Login
- Once logged in, you’ll see the GraphorLM dashboard
- Select New project and name it as you want
Once your project is created, you’ll be taken to the sources dashboard where you can begin building your RAG pipeline.
Step two: Load your data
GraphorLM allows you to easily upload your own documents to use in your RAG pipeline.
- In your project, navigate to Sources in the left sidebar
- Click Add Sources to open the data upload interface
- Choose your preferred import method:
- Local files: Upload files from your computer (PDF, MD, PNG, JPG, DOCX, TXT, CSV and others are supported)
- URL: Import content directly from a web address
- Select or drag and drop your files into the upload area
- Click Finish to begin the ingestion process
During processing, GraphorLM will automatically:
- Extract text using advanced OCR for images and scanned documents
- Identify document structure and metadata
- Prepare your content for chunking in the pipeline
You can monitor the ingestion progress on the Sources dashboard.
After upload, you can click on any file to:
- Customize the parsing method and configure element classification to improve document structure recognition
- Preview the extracted content before pipeline processing
Step three: Create a RAG pipeline
Now let’s create a RAG flow to process your documents and build an intelligent question-answering system.
- In the left sidebar, click Flows to navigate to the Flows dashboard
- Click New flow to start creating your pipeline
- Enter a descriptive name for your flow (e.g., “DocumentQA”)
- Click Create to open the visual flow builder
You’ll see the visual flow builder interface with components that can be connected to form your RAG pipeline:
- Dataset: Select the documents you uploaded in the previous step
- Chunking: Configure how your documents are segmented for optimal retrieval
- Retrieval: Define how relevant information is searched and ranked
- Testset: Create evaluation scenarios with ground-truth answers
- Questions: Define sample queries to test your pipeline
- LLM: Connect to language models for generating responses
- Evaluation: Measure the quality and accuracy of your pipeline
- Response: View the final output of your RAG system
The flow builder uses a drag-and-drop interface where you can connect these components to create your custom RAG pipeline.
Step four: Configure pipeline components
Now let’s build your RAG pipeline by adding and connecting the essential components:
-
Add the Dataset component
- Drag the Dataset component from the sidebar onto the canvas
- Double-click on the component to open its configuration panel
- Select the sources you imported earlier from the files menu
- Close the configuration panel
- This component provides your documents to the pipeline
-
Add the Chunking component
- Drag the Chunking component onto the canvas
- Connect the output of the Dataset component to the input of the Chunking component
- Double-click to configure with these recommended settings:
- Embedding / Indexer: text-embedding-3-small (provides semantic understanding)
- Elements to remove: None (or select elements to exclude)
- Splitter: Smart chunking (maintains semantic coherence)
- Chunk Size: 5000 (adjust based on your document complexity)
- Close the configuration panel
- This component divides your documents into smaller pieces for efficient retrieval
-
Add the Retrieval component
- Drag the Retrieval component onto the canvas
- Connect the output of the Chunking component to the input of the Retrieval
- Double-click to configure:
- Search Type: Similarity (recommended for semantic relevance)
- Top K: 5 (number of chunks to retrieve)
- Score Threshold: 0.5 (minimum relevance score)
- Close the configuration panel
- This component searches your chunked documents for relevant information
-
Add the Response component
- Drag the Response component onto the canvas
- Connect the output of the Retrieval component to the input of the Response
- This component formats and displays the final output of your pipeline
After configuring each component, you can test individual nodes by:
- Double-clicking the component to open its configuration panel
- Clicking Update results to see the component’s output based on current settings
- Examining the processed data to verify it meets your expectations
Important Note: To fully simulate a flow with retrieval nodes, you must add either:
- A Question node: For testing individual queries
- A Testset node: For batch testing multiple questions
These input nodes provide the questions that drive the retrieval process in your RAG pipeline.
Step five: Deploy your RAG pipeline
After building and testing your flow, it’s time to deploy it for production use. Deployment makes your RAG pipeline accessible via API and MCP (Model Context Protocol), allowing integration with applications, websites, or other systems.
To deploy your pipeline:
- Click on the Deploy new revision button in the top-right corner of the flow builder
- In the deployment modal, add a detailed Tool description for MCP Server:
- This description helps LLMs understand when and how to use your flow
- Write a clear, concise explanation of what questions your flow can answer
- Example: “This tool answers questions about technical documentation and product specifications”
- Under Traffic Configuration:
- Select the checkbox to display this revision immediately
- This sets 100% of traffic to use this revision, replacing any existing traffic splits
- Click Create to deploy the revision
Congratulations! Your pipeline is now online.
Final step: Connect to your flow
Click the Connect to flow button to view integration options:
- REST API: Direct HTTP endpoint for custom applications
- MCP Server: Model Context Protocol integration for LLM tools
Your RAG pipeline is now ready for production use with real users and data!
You can test your online pipeline by clicking on the chat button at the bottom right corner of the GraphorLM interface. This built-in chat interface allows you to:
- Send questions directly to your deployed flow
- View retrieved context from your documents
- Experience the pipeline from an end-user perspective
- Verify that your deployment is working correctly
Next steps
Congratulations! You’ve built and tested your first RAG pipeline with GraphorLM. Here’s what you can explore next:
Integrate Workflow
Connect your RAG systems to applications via REST API and MCP Server integration
Data Ingestion
Transform unstructured documents into processable data using advanced OCR and classification
Chunking
Optimize document segmentation with intelligent chunking for maximum retrieval relevance
Evaluation
Measure and improve your RAG pipeline performance with comprehensive metrics