LLM Chat
You have generated your LLM Dataset and you have trained your LLM Model. Now we can use our LLM Chat and interact with the trained expert.
- Start by creating a new Canvas and drag the LLM Chat element onto the Canvas.
- Open the LLM Chat Element setting and and make the following adjustments:
- Select a Trained Artifact from the drop down. This is where the models you’ve trained with the LLM Trainer are stored. If you don’t have a trained artifact, you will chat with the pre-trained base model.
- Base Model Architecture: This is the core model you’ll be fine tuning. Select any model from the dropdown.
- Model System Prompt: This default prompts your model to act as an assistant. You can leave this setting alone, or play with it to change how the model responds.
For example: ask the model to talk like a Pirate or Rock Star, speak in a casual tone, or speak in like a poet. If you have brand voice guidelines, this is the place to add them.
- Max token: 256 is recommended for testing
- Model Storage Path: Using the “Select Directory” button, choose the folder where you want to save the base model for the chat.
- Model Adapter Folder Path: A backup to the built in model artifact registry. If you have a model that is not in your registry, you can plug it in here.
- Max Tokens: A setting limiting the number of tokens from the LLMs output. By default this is loaded from the Trained Artifact.
- Temperature: A balance between predictability and creativity. Lower settings prioritize learned patterns, giving more deterministic outputs. Higher temperatures encourage creativity and diversity.