Deprecated |
---|
This Dialogflow client library and Dialogflow API V1 have been deprecated and will be shut down on October 23th, 2019. Please migrate to Dialogflow API V2. |
The API.AI Objective-C(Cocoa) SDK makes it easy to integrate speech recognition with API.AI natural language processing API on Apple devices. API.AI allows using voice commands and integration with dialog scenarios defined for a particular agent in API.AI.
Run pod update
in the ApiAiDemo project folder.
Open ApiAIDemo.xworkspace in Xcode.
In ViewController -viewDidLoad insert API key.
configuration.clientAccessToken = @"YOUR_CLIENT_ACCESS_TOKEN";
Note: an agent in api.ai should exist. Keys could be obtained on the agent's settings page.
Define sample intents in the agent.
Run the app in Xcode. Inputs are possible with text and voice (experimental).
Run pod install
in your project folder.
Update Podfile to include:
pod 'ApiAI'
pod update
In the AppDelegate.h
, add ApiAI.h import and property:
#import <ApiAI/ApiAI.h>
@property(nonatomic, strong) ApiAI *apiAI;
In the AppDelegate.m, add
self.apiAI = [[ApiAI alloc] init];
// Define API.AI configuration here.
id <AIConfiguration> configuration = [[AIDefaultConfiguration alloc] init];
configuration.clientAccessToken = @"YOUR_CLIENT_ACCESS_TOKEN_HERE";
self.apiAI.configuration = configuration;
...
// Request using text (assumes that speech recognition / ASR is done using a third-party library, e.g. AT&T)
AITextRequest *request = [apiai textRequest];
request.query = @[@"hello"];
[request setCompletionBlockSuccess:^(AIRequest *request, id response) {
// Handle success ...
} failure:^(AIRequest *request, NSError *error) {
// Handle error ...
}];
[_apiAI enqueue:request];
Please read and follow the steps in the CONTRIBUTING.md.
See LICENSE.
Your use of this sample is subject to, and by using or downloading the sample files you agree to comply with, the Google APIs Terms of Service.
This is not an official Google product.