This is the Ruby client for the NLP Cloud API. See the documentation for more details.
NLP Cloud serves high performance pre-trained or custom models for NER, sentiment-analysis, classification, summarization, dialogue summarization, paraphrasing, intent classification, product description and ad generation, chatbot, grammar and spelling correction, keywords and keyphrases extraction, text generation, image generation, code generation, question answering, automatic speech recognition, machine translation, language detection, semantic search, semantic similarity, tokenization, POS tagging, embeddings, and dependency parsing. It is ready for production, served through a REST API.
You can either use the NLP Cloud pre-trained models, fine-tune your own models, or deploy your own models.
If you face an issue, don't hesitate to raise it as a Github issue. Thanks!
Install via gem.
gem install nlpcloud
Here is a full example that summarizes a text using Facebook's Bart Large CNN model, with a fake token:
require 'nlpcloud'
client = NLPCloud::Client.new("bart-large-cnn","4eC39HqLyjWDarjtT1zdp7dc")
client.summarization("One month after the United States began what has become a
troubled rollout of a national COVID vaccination campaign, the effort is finally
gathering real steam. Close to a million doses -- over 951,000, to be more exact --
made their way into the arms of Americans in the past 24 hours, the U.S. Centers
for Disease Control and Prevention reported Wednesday. That s the largest number
of shots given in one day since the rollout began and a big jump from the
previous day, when just under 340,000 doses were given, CBS News reported.
That number is likely to jump quickly after the federal government on Tuesday
gave states the OK to vaccinate anyone over 65 and said it would release all
the doses of vaccine it has available for distribution. Meanwhile, a number
of states have now opened mass vaccination sites in an effort to get larger
numbers of people inoculated, CBS News reported.")
Here is a full example that does the same thing, but on a GPU:
require 'nlpcloud'
client = NLPCloud::Client.new("bart-large-cnn","4eC39HqLyjWDarjtT1zdp7dc", gpu: true)
client.summarization("One month after the United States began what has become a
troubled rollout of a national COVID vaccination campaign, the effort is finally
gathering real steam. Close to a million doses -- over 951,000, to be more exact --
made their way into the arms of Americans in the past 24 hours, the U.S. Centers
for Disease Control and Prevention reported Wednesday. That s the largest number
of shots given in one day since the rollout began and a big jump from the
previous day, when just under 340,000 doses were given, CBS News reported.
That number is likely to jump quickly after the federal government on Tuesday
gave states the OK to vaccinate anyone over 65 and said it would release all
the doses of vaccine it has available for distribution. Meanwhile, a number
of states have now opened mass vaccination sites in an effort to get larger
numbers of people inoculated, CBS News reported.")
Here is a full example that does the same thing, but on a French text:
require 'nlpcloud'
client = NLPCloud::Client.new("bart-large-cnn","4eC39HqLyjWDarjtT1zdp7dc", gpu: true, lang: "fra_Latn")
client.summarization("Sur des images aériennes, prises la veille par un vol de surveillance
de la Nouvelle-Zélande, la côte d’une île est bordée d’arbres passés du vert
au gris sous l’effet des retombées volcaniques. On y voit aussi des immeubles
endommagés côtoyer des bâtiments intacts. « D’après le peu d’informations
dont nous disposons, l’échelle de la dévastation pourrait être immense,
spécialement pour les îles les plus isolées », avait déclaré plus tôt
Katie Greenwood, de la Fédération internationale des sociétés de la Croix-Rouge.
Selon l’Organisation mondiale de la santé (OMS), une centaine de maisons ont
été endommagées, dont cinquante ont été détruites sur l’île principale de
Tonga, Tongatapu. La police locale, citée par les autorités néo-zélandaises,
a également fait état de deux morts, dont une Britannique âgée de 50 ans,
Angela Glover, emportée par le tsunami après avoir essayé de sauver les chiens
de son refuge, selon sa famille.")
A json object is returned:
{
"summary_text": "Over 951,000 doses were given in the past 24 hours. That's the largest number of shots given in one day since the rollout began. That number is likely to jump quickly after the federal government gave states the OK to vaccinate anyone over 65. A number of states have now opened mass vaccination sites."
}
Pass the model you want to use and the NLP Cloud token to the client during initialization.
The model can either be a pre-trained model like en_core_web_lg
, bart-large-mnli
... but also one of your custom models custom_model/<model id>
(e.g. custom_model/2568
).
Your token can be retrieved from your NLP Cloud dashboard.
require 'nlpcloud'
client = NLPCloud::Client.new("<model>", "<your token>")
If you want to use a GPU, pass gpu: true
.
require 'nlpcloud'
client = NLPCloud::Client.new("<model>", "<your token>", gpu: true)
If you want to use the multilingual add-on in order to process non-English texts, pass lang: "<your language code>"
. For example, if you want to process French text, you should set lang: "fra_Latn"
.
require 'nlpcloud'
client = NLPCloud::Client.new("<model>", "<your token>", lang: "your language code")
If you want to make asynchronous requests, pass asynchronous: true
.
require 'nlpcloud'
client = NLPCloud::Client.new("<model>", "<your token>", asynchronous: true)
If you are making asynchronous requests, you will always receive a quick response containing a URL. You should then poll this URL with async_result()
on a regular basis (every 10 seconds for example) in order to check if the result is available. Here is an example:
client.async_result("https://api.nlpcloud.io/v1/get-async-result/21718218-42e8-4be9-a67f-b7e18e03b436")
The above command returns a JSON object when the response is ready. It returns nil
otherwise.
Call the asr()
method and pass the following arguments:
- (Optional: either this or the encoded file should be set)
url
: a URL where your audio or video file is hosted - (Optional: either this or the url should be set)
encoded_file
: a base 64 encoded version of your file - (Optional)
input_language
: the language of your file as ISO code
client.asr(url:"Your url")
The above command returns a JSON object.
Call the chatbot()
method and pass your input. As an option, you can also pass a context and a conversation history that is a list of hashes. Each hash is made of an input
and a response
from the chatbot.
client.chatbot("<Your input>", context: "<Your context>", history: [{"input"=>"input 1","response"=>"response 1"}, {"input"=>"input 2","response"=>"response 2"}, ...])
The above command returns a JSON object.
Call the classification()
method and pass the following arguments:
- The text you want to classify, as a string
- The candidate labels for your text, as a list of strings
- (Optional)
multi_class
: Whether the classification should be multi-class or not, as a boolean
client.classification("<Your block of text>", labels: ["label 1", "label 2", "..."])
The above command returns a JSON object.
Call the code_generation()
method and pass the instruction for the program you want to generate:
client.code_generation("<Your instruction>")
The above command returns a JSON object.
Call the dependencies()
method and pass the text you want to perform part of speech tagging (POS) + arcs on.
client.dependencies("<Your block of text>")
The above command returns a JSON object.
Call the embeddings()
method and pass a list of blocks of text that you want to extract embeddings from.
client.embeddings(["<Text 1>", "<Text 2>", "<Text 3>", ...])
The above command returns a JSON object.
Call the entities()
method and pass the text you want to perform named entity recognition (NER) on.
client.entities("<Your block of text>")
The above command returns a JSON object.
Call the generation()
method and pass the following arguments:
- The block of text that starts the generated text. 256 tokens maximum for GPT-J on CPU, 1024 tokens maximum for GPT-J and GPT-NeoX 20B on GPU, and 2048 tokens maximum for Fast GPT-J and Finetuned GPT-NeoX 20B on GPU.
- (Optional)
max_length
: Optional. The maximum number of tokens that the generated text should contain. 256 tokens maximum for GPT-J on CPU, 1024 tokens maximum for GPT-J and GPT-NeoX 20B on GPU, and 2048 tokens maximum for Fast GPT-J and Finetuned GPT-NeoX 20B on GPU. Iflength_no_input
is false, the size of the generated text is the difference betweenmax_length
and the length of your input text. Iflength_no_input
is true, the size of the generated text simply ismax_length
. Defaults to 50. - (Optional)
length_no_input
: Whethermin_length
andmax_length
should not include the length of the input text, as a boolean. If false,min_length
andmax_length
include the length of the input text. If true, min_length andmax_length
don't include the length of the input text. Defaults to false. - (Optional)
end_sequence
: A specific token that should be the end of the generated sequence, as a string. For example if could be.
or\n
or###
or anything else below 10 characters. - (Optional)
remove_input
: Whether you want to remove the input text form the result, as a boolean. Defaults to false. - (Optional)
num_beams
: Number of beams for beam search. 1 means no beam search. This is an integer. Defaults to 1. - (Optional)
num_return_sequences
: The number of independently computed returned sequences for each element in the batch, as an integer. Defaults to 1. - (Optional)
top_k
: The number of highest probability vocabulary tokens to keep for top-k-filtering, as an integer. Maximum 1000 tokens. Defaults to 0. - (Optional)
top_p
: If set to float < 1, only the most probable tokens with probabilities that add up to top_p or higher are kept for generation. This is a float. Should be between 0 and 1. Defaults to 0.7. - (Optional)
temperature
: The value used to module the next token probabilities, as a float. Should be between 0 and 1. Defaults to 1. - (Optional)
repetition_penalty
: The parameter for repetition penalty, as a float. 1.0 means no penalty. Defaults to 1.0. - (Optional)
bad_words
: List of tokens that are not allowed to be generated, as a list of strings. Defaults to null. - (Optional)
remove_end_sequence
: Optional. Whether you want to remove theend_sequence
string from the result. Defaults to false.
client.generation("<Your input text>")
The above command returns a JSON object.
Call the gs_correction()
method and pass the text you want correct:
client.gs_correction("<Your block of text>")
The above command returns a JSON object.
Call the image_generation()
method and pass the text you want to use to generate your image:
client.image_generation("<Your block of text>")
The above command returns a JSON object.
Call the intent_classification()
method and pass the text you want to extract intents from:
client.intent_classification("<Your block of text>")
The above command returns a JSON object.
Call the kw_kp_extraction()
method and pass the text you want to extract keywords and keyphrases from:
client.kw_kp_extraction("<Your block of text>")
The above command returns a JSON object.
Call the langdetection()
method and pass the text you want to analyze in order to detect the languages.
client.langdetection("<The text you want to analyze>")
The above command returns a JSON object.
Call the paraphrasing()
method and pass the text you want to paraphrase.
client.paraphrasing("<Your text to paraphrase>")
The above command returns a JSON object.
Call the question()
method and pass the following:
- Your question
- A context that the model will use to try to answer your question
client.question(question: "<Your question>",context: "<Your context>")
The above command returns a JSON object.
Call the semantic_search()
method and pass your search query.
client.semantic_search("Your search query")
The above command returns a JSON object.
Call the semantic_similarity()
method and pass a list made up of 2 blocks of text that you want to compare.
client.semantic_similarity(["<Block of text 1>", "<Block of text 2>"])
The above command returns a JSON object.
Call the sentence_dependencies()
method and pass a block of text made up of several sentences you want to perform POS + arcs on.
client.sentence_dependencies("<Your block of text>")
The above command returns a JSON object.
Call the sentiment()
method and pass the text you want to analyze the sentiment of:
client.sentiment("<Your block of text>")
The above command returns a JSON object.
Call the speech_synthesis()
method and pass the text you want to convert to audio:
client.speech_synthesis("<Your block of text>")
The above command returns a JSON object.
Call the summarization()
method and pass the text you want to summarize.
client.summarization("<Your text to summarize>")
The above command returns a JSON object.
Call the tokens()
method and pass the text you want to tokenize.
client.tokens("<Your block of text>")
The above command returns a JSON object.
Call the translation()
method and pass the text you want to translate.
client.translation("<Your text to translate>")
The above command returns a JSON object.