Skip to content

Jinchao-Yu/rasa_nlu

 
 

Repository files navigation

Rasa NLU

Join the chat at https://gitter.im/RasaHQ/rasa_nlu Build Status Coverage Status PyPI version Documentation Status roadmap badge

Rasa NLU (Natural Language Understanding) is a tool for intent classification and entity extraction. You can think of Rasa NLU as a set of high level APIs for building your own language parser using existing NLP and ML libraries. Find out more on the homepage of the project, where you can also sign up for the mailing list.

Extended documentation:

If you are new to Rasa NLU and want to create a bot, you should start with the tutorial.

Install

Via Docker Image From docker hub:

docker run -p 5000:5000 rasa/rasa_nlu:0.9.1-full

(for more docker installation options see Advanced Docker Installation)

Via Python Library From pypi:

pip install rasa_nlu
python -m rasa_nlu.server &

(for more python installation options see Advanced Python Installation)

Basic test

The below command can be executed for either method used above.

curl 'http://localhost:5000/parse?q=hello'

Example use

Get the Server Status

curl 'http://locahost:5000/status'

Check the Server Version

curl 'http://locahost:5000/version'

Training New Models

Examples and Documentation of the training data format are provided. But as a quick start execute the below command to train a new model

curl 'https://raw.githubusercontent.com/RasaHQ/rasa_nlu/master/data/examples/rasa/demo-rasa.json' | \
curl --request POST --header 'content-type: application/json' -d@- --url localhost:5000/train?name=test_model

The above command does the following:

  1. It Fetches some of the example data in the repo
  2. It POSTS that data to the /train endpoint and names the model /name=test_model

Parsing New Requests

Make sure the above command has finished before executing the below. You can check with the /status command above.

curl 'http://localhost:5000/parse?q=hello&model=test_model'

FAQ

Who is it for?

The intended audience is mainly people developing bots, starting from scratch or looking to find a a drop-in replacement for wit, LUIS, or api.ai. The setup process is designed to be as simple as possible. Rasa NLU is written in Python, but you can use it from any language through a HTTP API. If your project is written in Python you can simply import the relevant classes. If you're currently using wit/LUIS/api.ai, you just:

  1. Download your app data from wit, LUIS, or api.ai and feed it into Rasa NLU
  2. Run Rasa NLU on your machine and switch the URL of your wit/LUIS api calls to localhost:5000/parse.

Why should I use Rasa NLU?

  • You don't have to hand over your data to FB/MSFT/GOOG
  • You don't have to make a https call to parse every message.
  • You can tune models to work well on your particular use case.

These points are laid out in more detail in a blog post. Rasa is a set of tools for building more advanced bots, developed by the company Rasa. Rasa NLU is the natural language understanding module, and the first component to be open sourced.

What languages does it support?

Short answer: English, German, and Spanish currently. Longer answer: If you want to add a new language, the key things you need are a tokenizer and a set of word vectors. More information can be found in the language documentation.

How to contribute

We are very happy to receive and merge your contributions. There is some more information about the style of the code and docs in the documentation.

In general the process is rather simple:

  1. create an issue describing the feature you want to work on (or have a look at issues with the label help wanted)
  2. write your code, tests and documentation
  3. create a pull request describing your changes

You pull request will be reviewed by a maintainer, who might get back to you about any necessary changes or questions.

Advanced installation

Advanced Python

From github:

git clone [email protected]:RasaHQ/rasa_nlu.git
cd rasa_nlu
pip install -r requirements.txt

To test the installation use (this will run a very stupid default model. you need to train your own model to do something useful!):

Advanced Docker

Before you start, ensure you have the latest version of docker engine on your machine. You can check if you have docker installed by typing docker -v in your terminal.

To see all available builds go to the Rasa docker hub, but to get up and going the quickest just run:

docker run -p 5000:5000 rasa/rasa_nlu:latest-full

There are also three volumes, which you may want to map: /app/models, /app/logs, and /app/data. It is also possible to override the config file used by the server by mapping a new config file to the volume /app/config.json. For complete docker usage instructions go to the official docker hub readme.

To test run the below command after the container has started. For more info on using the HTTP API see here

curl 'http://localhost:5000/parse?q=hello'

Docker Cloud

Warning! setting up Docker Cloud is quite involved - this method isn't recommended unless you've already configured Docker Cloud Nodes (or swarms)

Deploy to Docker Cloud

Development Internals

Steps to release a new version

Releasing a new version is quite simple, as the packages are build and distributed by travis. The following things need to be done to release a new version

  1. update rasa_nlu/version.py to reflect the correct version number
  2. edit the CHANGELOG.rst, create a new section for the release (eg by moving the items from the collected master section) and create a new master logging section
  3. edit the migration guide to provide assistance for users updating to the new version
  4. commit all the above changes and tag a new release, e.g. using
    git tag -f 0.7.0 -m "Some helpful line describing the release"
    git push origin master --tags
    
    travis will build this tag and push a package to pypi
  5. only if it is a major release, a new branch should be created pointing to the same commit as the tag to allow for future minor patches, e.g.
    git checkout -b 0.7.x
    git push origin 0.7.x
    

License

Licensed under the Apache License, Version 2.0. Copyright 2017 Rasa Technologies GmbH. Copy of the license.

As a reference, the following contains a listing of the licenses of the different dependencies as of this writing. Licenses of minimal dependencies:

required package License
gevent MIT
klein MIT
boto3 Apache License 2.0
typing PSF
future MIT
six MIT
jsonschema MIT
matplotlib PSF
requests Apache Licence 2.0

Licenses of optional dependencies (only required for certain components of Rasa NLU. Hence, they are optional):

optional package License
MITIE Boost Software License 1.0
spacy MIT
scikit-learn BSD 3-clause
scipy BSD 3-clause
numpy BSD 3-clause
duckling Apache License 2.0
sklearn-crfsuite MIT
cloudpickle BSD 3-clause
google-cloud-storage Apache License 2.0

About

turn natural language into structured data

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 99.4%
  • Shell 0.6%