This tutorial shows you how to run a simple pipeline (ASR + NLU) where Whisper transcribes your speech and feeds it into the NLU component to run inference
Since the NLU component is not yet available on PyPI, for now, we will need to clone the repository and install it locally.
Clone the SIC repo
git clone https://github.com/Social-AI-VU/social-interaction-cloud.git
Switch to the nlu_component
branch:
git checkout nlu_component
Create and activate a virtual environment:
If you are using pure Python environment
python -m venv venv_sic source venv_sic/bin/activate |
If you are using Anaconda environment
conda create -n venv_sic python=3.12 conda activate venv_sic |
Install SIC, nlu and whisper dependencies from local repo:
cd social-interaction-cloud pip install ."[whisper-speech-to-text,nlu]" |
Run the NLU and Whisper components in separate terminals (Don’t forget to run a redis server. See the details here https://socialrobotics.atlassian.net/wiki/spaces/CBSR/pages/2180415493/Getting+started#Step-1%3A-starting-Redis-on-your-laptop ):
# Start the Redis server redis-server conf/redis/redis.conf # Start the NLU and Whisper components separately run-whisper run-nlu |
Open a new terminal and activate the same virtual environment you created earlier
Clone the sic_applications
repo
git clone https://github.com/Social-AI-VU/sic_applications.git
Add (trained) nlu model and ontology to the configuration of sic_applications
Run the demo (Don’t forget to put your ontology file and an already trained model in sic_application_conf/nlu folder, the default name is ontology.)
cd sic_applications/demos/desktop
python demo_desktop_asr_nlu.py