Home

Awesome

Platform

Use the cognitive capabilities of your iOS App with Core ML and Watson | With Watson

This iOS application uses the Watson Visual Recognition service to analyze images taken on iPhone or album photos using a user-trained model on the Watson Studio platform (also known as the Data Platform). In this way, you can understand how to integrate a Watson API into a Swift project with the use of the Swift SDK.

Access Dropbox to download the available Datasets, or create your own set of images.

If you want to read this content in Brazilian Portuguese, click here.

Components and technologies

How to configure and run locally

To install the application it is necessary that you have installed the latest version of Xcode and Cocoapods (dependency manager for Swift and Objective-C projects).

1. Get the app

git clone https://github.com/victorshinya/with-watson.git
cd with-watson/

2. Download all app dependencies

pod install

3. Open the "With Watson.xcworkspace" file

open With\ Watson.xcworkspace

4. Open the Constants.swift file

Open the file and fill it in with the Watson Visual Recognition credential and the trained Visual Recognition template - done within Watson Studio.

5. Execute with the CMD + R command (or press the play button)

Now run the project on the Xcode simulator or on your own iPhone / iPod / iPad. Remember that it is not possible to open the camera inside the simulator.

License

MIT License

Copyright (c) 2018 Victor Kazuyuki Shinya