Replace hours spent searching policy documents with higher-value relational work.
The Assistant to the HR Representative is a lightweight quickstart designed to give HR Representatives in Financial Services a trusted sounding board for discussions and decisions. Chat with this assistant for quick insights and actionable advice.
This quickstart was designed for environments where GPUs are not available or necessary, making it ideal for lightweight inference use cases, prototyping, or constrained environments. By making the most of vLLM on CPU-based infrastructure, this Assistant to the HR Representative can be deployed to almost any OpenShift AI environment.
This quickstart includes a Helm chart for deploying:
- An OpenShift AI Project.
- vLLM with CPU support running an instance of TinyLlama.
- AnythingLLM, a versatile chat interface, running as a workbench and connected to the vLLM.
Use this project to quickly spin up a minimal vLLM instance and start serving models like TinyLlama on CPU—no GPU required. 🚀
- No GPU needed! 🤖
- 2 cores
- 4 Gi
- Storage: 5Gi
- No GPU needed! 🤖
- 8 cores
- 8 Gi
- Storage: 5Gi
Note: This version is compiled for Intel CPU's (preferably with AWX512 enabled to be able to run compressed models as well, but optional).
Here's an example machine from AWS that works well: https://instances.vantage.sh/aws/ec2/m6i.4xlarge
- Red Hat OpenShift 4.16.24 or later
- Red Hat OpenShift AI 2.16.2 or later
- Dependencies for Single-model server:
- Red Hat OpenShift Service Mesh
- Red Hat OpenShift Serverless
- Standard user. No elevated cluster permissions required.
Follow the below steps to deploy and test the HR assistant.
git clone https://github.com/rh-ai-quickstart/llm-cpu-serving.git && \
cd llm-cpu-serving/
PROJECT="hr-assistant"
oc new-project ${PROJECT}helm install ${PROJECT} helm/ --namespace ${PROJECT}
oc -n ${PROJECT} get pods -w
(Output)
NAME READY STATUS RESTARTS AGE
anythingllm-0 3/3 Running 0 76s
anythingllm-seed-lchf6 0/1 Completed 0 76s
tinyllama-1b-cpu-predictor-544bdf75f9-x9fwh 2/2 Running 0 75s
You can get the OpenShift AI Dashboard URL by:
oc get routes rhods-dashboard -n redhat-ods-applicationsOnce inside the dashboard, navigate to Data Science Projects -> tinyllama-cpu-demo (or what you called your ${PROJECT} if you changed from default).
Inside the project you can see Workbenches, open up the one for AnythingLLM.
Finally, click on the Assistant to the HR Representative Workspace that's pre-created for you and you can start chatting with your Assistant to the HR Representative! :)
Try for example asking it:
Hi, one of our employees is going to get a raise, what do I need to keep in mind for this?
It will provide you a reply and some citations related to the question.
helm uninstall ${PROJECT} --namespace ${PROJECT}
- The runtime is built from vLLM CPU
- Runtime image is pushed to quay.io/repository/rh-aiservices-bu/vllm-cpu-openai-ubi9
- Code for Runtime image and deployment can be found on github.com/rh-aiservices-bu/llm-on-openshift
- Industry: Banking and securities
- Product: OpenShift AI
- Use case: Productivity




