Deploying Your Rasa Assistant
This page explains when and how to deploy an assistant built with Rasa. It will allow you to make your assistant available to users and set you up with a production-ready environment.
When to Deploy Your Assistant
The best time to deploy your assistant and make it available to test users is once it can handle the most important happy paths or is what we call a minimum viable assistant.
The recommended deployment methods described below make it easy to share your assistant with test users via the share your assistant feature in Rasa X. Then, when you're ready to make your assistant available via one or more Messaging and Voice Channels, you can easily add them to your existing deployment set up.
Recommended Deployment Methods
The recommended way to deploy an assistant is using either the Server Quick-Install or Helm Chart options we support. Both deploy Rasa X and your assistant. They are the easiest ways to deploy your assistant, allow you to use Rasa X to view conversations and turn them into training data, and are production-ready. For more details on deployment methods see the Rasa X Installation Guide.
Server Quick-Install
The Server Quick-Install script is the easiest way to deploy Rasa X and your assistant. It installs a Kubernetes cluster on your machine with sensible defaults, getting you up and running in one command.
Default: Make sure you meet the OS Requirements, then run:
curl -s get-rasa-x.rasa.com | sudo bashCustom: See Customizing the Script and the Server Quick-Install docs docs.
Helm Chart
For assistants that will receive a lot of user traffic, setting up a Kubernetes or Openshift deployment via our Helm charts is the best option. This provides a scalable architecture that is also straightforward to deploy. However, you can also customize the Helm charts if you have specific requirements.
Default: Read the Helm Chart Installation docs.
Custom: Read the above, as well as the Advanced Configuration documentation, and customize the open source Helm charts to your needs.
Deploying a Rasa Open Source Assistant
While the above deployment methods involve deploying an assistant with Rasa X, the following instructions describe how to deploy a Rasa Open Source server only by using the Rasa Helm Chart in a scalable cluster environment using OpenShift or Kubernetes (K8S).
Cluster Requirements
To install the Rasa Helm chart, you need an existing Kubernetes cluster or OpenShift cluster. Setting up a Kubernetes / OpenShift cluster can be tedious, hence we recommend to get a managed cluster from a cloud provider like Google Cloud, DigitalOcean, Microsoft Azure, or Amazon EKS.
note
The Rasa Helm chart is open source and available in the helm-charts repository. Please create an issue in this repository if you discover bugs or have suggestions for improvements.
Installation Requirements
Check that you have installed the Kubernetes or OpenShift command line interface (CLI). You can check this using the following command:
- Kubernetes
- OpenShift
kubectl version --short --client# The output should be similar to this# Client Version: v1.19.11If this command resulted in an error, please install the Kubernetes CLI or the OpenShift CLI depending on the cluster you’re using.
Make sure that the Kubernetes / OpenShift CLI is correctly connected to your cluster. You can do so by using the following commands:
- Kubernetes
- OpenShift
kubectl version --short# The output should be similar to this# Client Version: v1.19.11# Server Version: v1.19.10If you get an error when executing the command, you are not connected to your cluster. To get the command to connect to the cluster please consult your cluster’s admin or the documentation of your cloud provider.
Make sure you have the Helm CLI installed. To check this, run:
helm version --short# The output should be similar to this# v3.6.0+g7f2df64If this command leads to an error, please install the Helm CLI.
In case you are using a version
<3.5
of Helm, please update to Helm version>=3.5
.
Installation
1. Create Namespace
We recommend installing Rasa Open Source in a separate namespace to avoid interfering with existing cluster deployments. To create a new namespace run the following command:
- Kubernetes
- OpenShift
2. Create Values File
Prepare an empty file called rasa-values.yml
which will include all your custom
configuration for the installation with Helm.
All available values you can find in the Rasa helm chart repository.
note
The default configuration of the Rasa chart deploys a Rasa Open Source Server, downloads a model, and serves the downloaded model. Visit the Rasa helm chart repository to check out more examples of configuration.
3. Loading an initial model
The first time you install Rasa, you may not have a model server available yet, or you may want an lightweight model for testing the deployment. For this purpose, you can choose between training or downloading an initial model. By default, the Rasa chart downloads an example model from GitHub. To use this option, you don't have to change anything.
If you want to define an existing model to download from a URL you define instead, update your rasa-values.yaml
with the URL according to the following configuration:
note
The URL for the initial model download has to point to a tar.gz file and must not require authentication.
If you want to train an initial model you can do this by setting the applicationSettings.trainInitialModel
to true
.
It creates a init container that trains a model based on data located in the /app
directory. If the /app
directory is empty it creates a new project.
You can find an example that shows how to download data files from a git repository and train an initial model in the Rasa helm charts examples.
4. Deploy Rasa Open Source Assistant
Run the following commands:
note
OpenShift only: If the deployment fails and oc get events
returns
1001 is not an allowed group spec.containers[0].securityContext.securityContext.runAsUser
,
re-run the installation command with the following values:
Then wait until the deployment is ready. If you want to check on its status, the following command will block until the Rasa deployment is ready:
- Kubernetes
- OpenShift
5. Access Rasa Open Source Assistant
By default the Rasa deployment is exposed via the rasa
(<release name>
) service and accessible only within a Kubernetes cluster. You can get
the IP address using this command:
- Kubernetes
- OpenShift
You can then access the deployment on http://127.0.0.1:${SERVICE_PORT}
Visit the Rasa helm chart README to learn other ways to expose your deployment.
Next Steps
Visit the Rasa helm chart repository where you can find examples of configuration and learn how to e.g. integrate your Rasa Open Source deployment with Rasa X.
Alternative Deployment Methods
Docker Compose
You can also run Rasa X in a Docker Compose setup, without the cluster environment. We have an install script for doing so, as well as manual instructions for any custom setups.
Default: Read the Docker Compose Install Script docs or watch the Masterclass Video on deploying Rasa X.
Custom: Read the Docker Compose Manual Install documentation for full customization options.
Rasa Open Source Only Deployment
It is also possible to deploy a Rasa assistant without Rasa X using Docker Compose. To do so, you can build your Rasa Assistant locally or in Docker. Then you can deploy your model in Docker Compose.
Deploying Your Action Server
Building an Action Server Image
If you build an image that includes your action code and store it in a container registry, you can run it
as part of your deployment, without having to move code between servers.
In addition, you can add any additional dependencies of systems or Python libraries
that are part of your action code but not included in the base rasa/rasa-sdk
image.
Automating your Action Server Image Builds
In addition to a manually creating a new Action Server image, you can use the Rasa Action Server GitHub Action to automate image builds. If GitHub Actions are new to you, it might be helpful to get familiar with GitHub Actions Documentation.
The following steps assume that you already created a GitHub repository and you have a DockerHub account.
To create a workflow for building and pushing a Docker image into a DockerHub registry:
Add GitHub Secrets with your DockerHub login name and password. You can find details on how to create encrypted secrets for a repository in the Github docs
The example uses the following secrets:
DOCKER_HUB_LOGIN
- a login name for DockerHubDOCKER_HUB_PASSWORD
- a password for DockerHub
In your GitHub repository create a file
.github/workflows/action_server.yml
.
The GitHub Action workflow below builds a new docker image every time files inside the actions/
directory have changed and the changes are pushed into the main
branch.
Push your changes to the
main
branch. After changes are pushed, the workflow will build and push a new image into the DockerHub registry.Now, you can use your new brand docker image.
You can also extend your workflow, so that you do not have to manually update your Rasa X deployment. The example below shows how to extend your workflow with an additional step that updates a Rasa X Helm Chart deployment.
As you can see it's possible to use output variables from the action_server
step. The steps.action_server.outputs.docker_image_name
variable returns
a docker image name and the steps.action_server.outputs.docker_image_tag
variable returns a docker image tag.
More examples on how to use and customize Rasa GitHub Actions you can find in the Rasa GitHub Actions repository.
Manually Building an Action Server
To create your image:
Make sure your actions are defined in
actions/actions.py
. Therasa/rasa-sdk
image will automatically look for the actions in this file.If your actions have any extra dependencies, create a list of them in a file,
actions/requirements-actions.txt
.Create a file named
Dockerfile
in your project directory, in which you'll extend the official SDK image, copy over your code, and add any custom dependencies (if necessary). For example:# Extend the official Rasa SDK imageFROM rasa/rasa-sdk:3.1.1# Use subdirectory as working directoryWORKDIR /app# Copy any additional custom requirements, if necessary (uncomment next line)# COPY actions/requirements-actions.txt ./# Change back to root user to install dependenciesUSER root# Install extra requirements for actions code, if necessary (uncomment next line)# RUN pip install -r requirements-actions.txt# Copy actions folder to working directoryCOPY ./actions /app/actions# By best practices, don't run the code with root userUSER 1001
You can then build the image via the following command:
The <custom_image_tag>
should reference how this image will be different from others. For
example, you could version or date your tags, as well as create different tags that have different code for production
and development servers. You should create a new tag any time you update your code and want to re-deploy it.
Using your Custom Action Server Image
If you're building this image to make it available from another server, for example a Rasa X or Rasa Enterprise deployment, you should push the image to a cloud repository.
This documentation assumes you are pushing your images to DockerHub. DockerHub will let you host multiple public repositories and one private repository for free. Be sure to first create an account and create a repository to store your images. You could also push images to a different Docker registry, such as Google Container Registry, Amazon Elastic Container Registry, or Azure Container Registry.
You can push the image to DockerHub via:
To authenticate and push images to a different container registry, please refer to the documentation of your chosen container registry.
How you reference the custom action image will depend on your deployment. Pick the relevant documentation for your deployment: