LogoLogo
  • Introduction
    • What Is an Agent Application?
    • Core Concepts
    • Getting Started
  • GLIK Cloud
    • Getting Started
      • GLIK Cloud vs. GLIK Studio
    • Prompt Orchestration Interface
    • What Is a Workflow?
    • Workflow Editor
    • App Types (Overview)
    • Workspaces & Permissions
    • Enterprise Readiness & Compatibility
    • Security & Data Handling
  • Enterprise SaaS
    • Expense Policy Decision Engine
    • Compliance & Audit Automation
    • Compliance Advisor Copilot
  • App Types
    • Overview
    • Chatbot
    • Agent
    • Advanced Chat
    • Workflow
  • Templates
    • Overview
    • Policy Automation
      • Overview
        • Expense Policy Validator — Summary Sheet
        • Enterprise Policy Intelligence
      • Expense Policy Decision Engine
        • Expense Policy Decision Engine (Starter)
          • Practice Exercises
        • Expense Policy Decision Engine (Threshold Memory)
          • Practice Exercises
          • Threshold Agent Response Tuning Exercises
        • Expense Policy Decision Engine (Policy API Integration)
          • Practice Exercises
        • Expense Policy Decision Engine (Escalation & Conflict Resolution)
        • Expense Policy Decision Engine (Audit & Logging)
    • Compliance and Audit Automation
      • KYC/AML Review Copilot
        • Learning Track
      • Compliance Copilot – MiCA Reporting
        • Workflow Phases
        • Block-by-Block Guide
    • Compliance Advisors
      • Global Control Copilot – Cross-Jurisdiction Policy Interpreter
        • Input Combinations & Workflow Outcomes
        • Reference Input Payloads
        • Policy Retrieval via Input Routing
        • Input Logic & Routing Behavior
        • Predefined Policy Thresholds
    • Knowledge Systems
      • Overview
      • Compliance SOP Agent
    • Process Automation
    • Task Resolution Agents
      • Why Agentify Task Resolution
    • Work Coordination Agents
      • Escrow Agent Orchestration
    • Embedded Operational Copilots
    • Expense & ERP Agents
      • ERP Vendor AI Copilots and Agents
      • Custody Approval Workflow for Token Issuance
    • Inventory & Logistics Agents
    • Sales & Forecasting Agents
    • Plugin-Based Agent Platforms
  • Marketplace
    • Overview
    • Publishing Templates
  • System Architecture
    • Overview
    • Blocks & Nodes
      • Utilities
        • Start Block
        • End Node
        • HTTP Request
        • List Operator
      • Classifier Nodes
        • Question Classifier
      • Logic Blocks
        • IF/ELSE Branch
        • Iteration
        • Loop
      • Transform Blocks
        • Variable Assigner
        • Variable Aggregator
        • Parameter Extractor
        • Data Enrichment
        • Prompt Template
        • Code
      • Input & Extraction
        • Doc Extractor
        • Knowledge Retrieval
        • LLM Block
          • LLM Reasoning
          • Fallback to LLM Reasoning
        • Tool Node
        • Agent
        • Answer
    • GLIK Knowledge
      • Creating & Managing Knowledge
      • GLIK Knowledge Retrieval
      • Writing to Knowledge
      • Scoped Memory & Access Control
      • Injection & Variable Binding
      • Performance & Limits
    • Execution Model
      • Workflow Architecture
      • Flow Engine
      • Node Lifecycle
      • Protocol Compatibility & Schema Interoperability
      • Input Binding & Value Resolution
    • Memory & Variable Scope
      • Conversation Variables
      • Memory Layers (User, App, Org)
      • Memory Slot Injection
      • Memory Retention Policy
    • Decision Routing
      • Conditional Logic Engine
      • LLM Fallthrough Patterns
      • Policy Enforcement & Escalation Paths
    • Enterprise Orchestration
      • Policy-Driven Automation
      • Enterprise Modularity
      • Auditability & Governance
      • Explainability & Decision Transparency
    • Agentifying Legacy Systems
      • Why Legacy Systems Resist Change
      • Best Practices for Agentifying ERP Workflows
      • GLIK’s Wrap-Around Model
      • Agent Surfaces (PDF, OCR, API, UI)
      • No-API Memory-Based Control
      • Compliance & Risk Considerations
    • System Observability
      • Execution Logs
      • Save Points & Snapshots
      • Variable Debugging
      • Session Trace Viewer
  • Developers
    • Overview
    • GLIK Open Core
      • Deployment & Installation
      • CLI Reference
      • Security & Compliance
      • Customization Guide
      • Versioning & Updates
  • GLIK Roadmap
  • Deprecation
    • Orchestration Interface
      • Node Orchestration
        • Node
          • Start
          • End
          • Direct Reply
          • LLM
          • Question Classifier
          • Knowledge Retrieval
          • Code Execution
          • Doc Extractor
          • HTTP Request
          • Conditional Branch IF/ELSE
          • Iteration
          • List Operator
          • Parameter Extraction
          • Template
          • Tools
          • Variable Aggregator
          • Variable Assigner
      • Variables
      • Application Toolkits
      • File Upload
    • Chatbot Features
    • Dataset
      • Dataset Creation
      • Text Preprocessing and Cleaning
        • Advanced Configuration
      • Retrieval Test/Citation
    • Studio
  • Brand Kit & Identity
    • Logos & Visual Assets
    • Typography & Colors
    • Messaging Pillars
    • Product Screenshots
    • Diagrams & Icons
    • Company Boilerplate
    • Downloads (.zip)
  • Legal
    • Terms of Service
    • Privacy Policy
    • Cookie Policy
    • Trademark Notice
    • Acceptable Use Policy
    • Open Core License
Powered by GitBook

Platform

  • Open GLIK Cloud
  • Getting Started
  • Templates

Documentation

  • Core Concepts
  • GLIK Open Core
  • Security & Data Handling
  • Workspaces & Permissions

Company

  • RIvalz AI
  • Contact Support
  • Status Page

© 2023–2025 Rivalz Technologies Ltd.

On this page
  • Definition
  • Scenarios
  • How to Configure
  • Writing Prompts
  • Explanation of Special Variables
  • Context Variables
  • Conversation History
  • Model Parameters
  • Advanced Features
  • Use Cases
  • Reading Knowledge Base Content
  • Reading Document Files

Was this helpful?

  1. Deprecation
  2. Orchestration Interface
  3. Node Orchestration
  4. Node

LLM

PreviousDirect ReplyNextQuestion Classifier

Last updated 3 months ago

Was this helpful?

Definition

Invokes the capabilities of large language models to process information input by users in the "Start" node (natural language, uploaded files, or images) and provide effective response information.

Scenarios

LLM is the core node of Advanced Agent, utilizing the conversational/generative/classification/processing capabilities of large language models to handle a wide range of tasks based on given prompts and can be used in different stages of workflows.

  • Intent Recognition: In customer service scenarios, identifying and classifying user inquiries to guide downstream processes.

  • Text Generation: In content creation scenarios, generating relevant text based on themes and keywords.

  • Content Classification: In email batch processing scenarios, automatically categorizing emails, such as inquiries/complaints/spam.

  • Text Conversion: In translation scenarios, translating user-provided text into a specified language.

  • Code Generation: In programming assistance scenarios, generating specific business code or writing test cases based on user requirements.

  • RAG: In knowledge base Q&A scenarios, reorganizing retrieved relevant knowledge to respond to user questions.

  • Image Understanding: Using multimodal models with vision capabilities to understand and answer questions about the information within images.

By selecting the appropriate model and writing prompts, you can build powerful and reliable solutions within Chatflow/Workflow.

How to Configure

  1. Select a Model: Glik supports major global models, including OpenAI's GPT series, Anthropic's Claude series, and Google's Gemini series. Choosing a model depends on its inference capability, cost, response speed, context window, etc. You need to select a suitable model based on the scenario requirements and task type.

  2. Configure Model Parameters: Model parameters control the generation results, such as temperature, TopP, maximum tokens, response format, etc. To facilitate selection, the system provides three preset parameter sets: Creative, Balanced, and Precise.

  3. Write Prompts: The LLM node offers an easy-to-use prompt composition page. Selecting a chat model or completion model will display different prompt composition structures.

  4. Advanced Settings: You can enable memory, set memory windows, and use the Jinja-2 template language for more complex prompts.

If you are using Glik for the first time, you need to complete the model configuration in System Settings-Model Providers before selecting a model in the LLM node.

Writing Prompts

In the LLM node, you can customize the model input prompts. If you select a chat model, you can customize the System/User/Assistant sections.

Prompt Generator

If you're struggling to come up with effective system prompts (System), you can use the Prompt Generator to quickly create prompts suitable for your specific business scenarios, leveraging AI capabilities.

In the prompt editor, you can call out the variable insertion menu by typing / or { to insert special variable blocks or upstream node variables into the prompt as context content.

Explanation of Special Variables

Context Variables

Context variables are a special type of variable defined within the LLM node, used to insert externally retrieved text content into the prompt.

In common knowledge base Q&A applications, the downstream node of knowledge retrieval is typically the LLM node. The output variable result of knowledge retrieval needs to be configured in the context variable within the LLM node for association and assignment. After association, inserting the context variable at the appropriate position in the prompt can incorporate the externally retrieved knowledge into the prompt.

This variable can be used not only as external knowledge introduced into the prompt context for LLM responses but also supports the application's citation and attribution feature due to its data structure containing segment reference information.

If the context variable is associated with a common variable from an upstream node, such as a string type variable from the start node, the context variable can still be used as external knowledge, but the citation and attribution feature will be disabled.

Conversation History

To achieve conversational memory in text completion models (e.g., gpt-3.5-turbo-Instruct), Glik designed the conversation history variable in the original Prompt Expert Mode (discontinued). This variable is carried over to the LLM node in Chatflow, used to insert chat history between the AI and the user into the prompt, helping the LLM understand the context of the conversation.

The conversation history variable is not widely used and can only be inserted when selecting text completion models in Chatflow.

Model Parameters

The parameters of the model affect the output of the model. Different models have different parameters. The following figure shows the parameter list for GPT-4.

The main parameter terms are explained as follows:

  • Temperature: Usually a value between 0-1, it controls randomness. The closer the temperature is to 0, the more certain and repetitive the results; the closer it is to 1, the more random the results.

  • Top P: Controls the diversity of the results. The model selects from candidate words based on probability, ensuring that the cumulative probability does not exceed the preset threshold P.

  • Presence Penalty: Used to reduce the repetitive generation of the same entity or information by imposing penalties on content that has already been generated, making the model inclined to generate new or different content. As the parameter value increases, greater penalties are applied in subsequent generations to content that has already been generated, lowering the likelihood of repeating content.

  • Frequency Penalty: Imposes penalties on words or phrases that appear too frequently by reducing their probability of generation. With an increase in parameter value, greater penalties are imposed on frequently occurring words or phrases. Higher parameter values reduce the frequency of these words, thereby increasing the lexical diversity of the text.

If you do not understand what these parameters are, you can choose to load presets and select from the three presets: Creative, Balanced, and Precise.

Advanced Features

  • Memory: When enabled, each input to the intent classifier will include chat history from the conversation to help the LLM understand the context and improve question comprehension in interactive dialogues.

  • Memory Window: When the memory window is closed, the system dynamically filters the amount of chat history passed based on the model's context window; when open, users can precisely control the amount of chat history passed (in terms of numbers).

  • Conversation Role Name Settings: Due to differences in model training stages, different models adhere to role name instructions differently, such as Human/Assistant, Human/AI, Human/Assistant, etc. To adapt to the prompt response effects of multiple models, the system provides conversation role name settings. Modifying the role name will change the role prefix in the conversation history.

Use Cases

Reading Knowledge Base Content

To enable workflow applications to read "Knowledge Base" content, such as building an intelligent customer service application, please follow these steps:

  1. Add a knowledge base retrieval node upstream of the LLM node;

  2. Fill in the output variable result of the knowledge retrieval node into the context variable of the LLM node;

  3. Insert the context variable into the application prompt to give the LLM the ability to read text within the knowledge base.

The result variable output by the Knowledge Retrieval Node also includes segmented reference information. You can view the source of information through the Citation and Attribution feature.

Regular variables from upstream nodes can also be filled into context variables, such as string-type variables from the start node, but the Citation and Attribution feature will be ineffective.

Reading Document Files

To enable workflow applications to read document contents, such as building a ChatPDF application, you can follow these steps:

  • Add a file variable in the "Start" node;

  • Add a document extractor node upstream of the LLM node, using the file variable as an input variable;

  • Fill in the output variable text of the document extractor node into the prompt of the LLM node.

For more information, please refer to File Upload.

Jinja-2 Templates: The LLM prompt editor supports Jinja-2 template language, allowing you to leverage this powerful Python template language for lightweight data transformation and logical processing. Refer to the .

official documentation