TABLE OF CONTENTS
- Where to find the Preview panel
- Send a test message
- What to look for in a response
- Test workflows in the Preview panel
- Test fallback messages and human handover
- Start a new conversation
- Iterate based on preview results
- Best practices
Before you deploy an agent to employees, test how it behaves in real conversations. The Preview panel gives you an interactive chat environment where you can send messages and see how the agent responds—in the same interface your employees will use. Use it to verify that your knowledge, workflows, and instructions are all working together the way you intended.
Where to find the Preview panel
The Preview panel is available from anywhere inside the Build section of AI Agent Studio. You do not need to navigate to a separate page to open it.
Open AI Agent Studio and select the agent you want to test.
In the left navigation, expand Build.
Click any item under Build—Knowledge, Workflows, Instructions, or Configurations.
At the bottom of the page, click Preview AI agent. The Preview panel opens on the right side of the screen.
Tip The Preview panel stays open as you move between Build sections. You can adjust a setting in Configurations, then switch back to the panel to test the change without reopening it.
Send a test message
Use the message input field to send questions or requests to your agent, just as an employee would. The agent responds in the conversation area.
In the message input field, type a question or request.
Press Enter or click the Send icon.
Review the agent’s response in the conversation area.
Continue the conversation by sending follow-up messages. The agent retains context within the same conversation.
Tip: Test with realistic employee phrasing, including informal language and incomplete sentences. Employees rarely phrase questions the same way twice, and testing variations helps you identify gaps in knowledge coverage.
What to look for in a response
When reviewing an agent response in the Preview panel, evaluate the following:
Accuracy: Does the response reflect the correct information from your knowledge sources?
Completeness: Does the response fully address the question, or does it leave out relevant details?
Tone: Does the language match the style you defined in Instructions?
Workflow execution: If the message should trigger a workflow, did the agent initiate it and collect the required inputs?
Fallback handling: If the agent could not answer, did it display the correct fallback message and offer to create a ticket?
Human handover: If the employee requested a human or the agent could not resolve the issue, did the handover initiate correctly?
Test workflows in the Preview panel
Workflows define actions the agent can take on behalf of employees—such as resetting a password, provisioning access, or creating a ticket. Use the Preview panel to confirm that each workflow triggers correctly, collects required inputs, and returns the expected outcome.
Send a message that is designed to trigger the workflow you want to test.
Respond to any follow-up questions the agent asks to collect required inputs.
Verify that the agent confirms the action was completed and provides the correct output.
Test error scenarios by providing invalid inputs, such as a ticket number that does not exist, and confirm the agent responds with a helpful error message.
Note: Feedback for queries that involve a workflow is collected after all steps in that workflow are completed, not when the question is first asked. When testing feedback collection, complete the full workflow before evaluating the feedback prompt.
Test fallback messages and human handover
The Preview panel lets you verify that your fallback message appears when the agent cannot find an answer, and that human handover works as configured.
Test the fallback message
Ask a question that falls outside your agent’s knowledge coverage. Confirm that the agent responds with the fallback message you configured in Configurations > Conversation behavior, and that it offers the option to create a ticket if you enabled Suggest creating a ticket.
Test human handover
Initiate a handover by sending a message that requests a human agent, such as:
“I want to speak to someone.”
“Transfer me to an agent.”
“This isn’t helping.”
Confirm that the agent displays the handover message and initiates the ticket creation process. Verify that the ticket is created with the correct details in Freshservice.
Important Ensure your Freshservice ticket routing and assignment rules are configured before testing handover in Preview. The agent creates the ticket, but your existing Freshservice workflows determine which team or agent receives it.
Start a new conversation
Each conversation in the Preview panel has its own context. The agent does not carry information from one conversation into another. Starting a new conversation resets this context so you can test a different scenario from a clean state.
In the Preview panel, close the current conversation or click the new conversation control.
A fresh conversation opens. The agent greets you as it would greet a new employee.
Send your first message for the new test scenario.
Tip: Start a new conversation whenever you change a setting or update knowledge sources. Testing in the same conversation after making changes can produce misleading results because the agent may draw on context from earlier messages.
Iterate based on preview results
Preview is most effective when you use it as part of a continuous cycle: test, identify an issue, make a change, and test again. Use the following approach to work through issues systematically.
Identify the issue. Note the exact message you sent, the response the agent gave, and what you expected instead.
Go to the relevant Build section. Knowledge issues require changes to your knowledge sources. Workflow issues require changes in Workflows. Tone or behavior issues require changes in Instructions. Fallback or handover issues require changes in Configurations.
Make the change and save it.
Return to the Preview panel, start a new conversation, and retest the scenario that previously failed.
Test related scenarios to confirm no regressions.
Best practices
Test after every configuration change. Open Preview immediately after saving any change to verify the updated behavior before moving on.
Use realistic language. Phrase test messages the way employees actually write—informally, with abbreviations, and sometimes with errors. Avoid using exact phrasing from your knowledge sources.
Cover edge cases. In addition to common questions, test out-of-scope requests, ambiguous inputs, and scenarios where workflows fail. These are the situations most likely to surface in production.
Test each change in isolation. Make one change at a time and retest before making the next. This makes it easier to identify what fixed or introduced a problem.
Have someone else test. Ask a colleague who did not build the agent to run test conversations. They are more likely to discover gaps because they approach it the same way employees will.
What’s next
After verifying your agent’s behavior in Preview, you are ready to:
Deploy the agent to the support portal, Microsoft Teams, or Slack so employees can start using it.
Review conversation logs and feedback ratings after the first few weeks to identify areas for improvement.
Return to Configurations to refine fallback messages, feedback collection, and handover settings based on real usage patterns.

