Big ML |
service |
-
https://bigml.com/developers/sources
- CSV
- ARFF
- <= 64GB, can be gzipped (4TB if the file is stored on Amazon S3)
- can create sources from remote locations: supported protocols include azure, https, odata, s3, and dropbox
- Creating a source is a service, as is updating a source, so this can be easily automated.
|
-
https://bigml.com/developers/models
- Service call: send dataset's id
- creation is async: you must query to determine if it's finished
- Can also use a service call to update the model, but that only updates the model's metadata, it does not actually retrain it
- you can easily create ensemble models, which is pretty cool
|
-
https://bigml.com/developers/predictions
- async batch predictions are also available
- send model id and input data as JSON, plus username and api key
- get back prediction plus a bunch of metadata, including a confidence score, as JSON
- can statically refer to this prediction later by a unique ID
|
Google Predictions |
service |
- Upload a CSV to Google Cloud Storage
- Can call a service endpoint to update the data set of a model (I don't think this auto-retrains the model, though)
|
|
-
https://developers.google.com/prediction/docs/reference/v1.6/trainedmodels/predict
- you can batch requests, but that does not seem to be an async operation
- send model and project ids, along with a list of input features (JSON structure mostly consisting of a JSON list)
- get back JSON with output value. If task was classification, also get back a list of all class labels with their associated probabilities.
- also gets back a url to re-request the same prediction again
|
Comments
Please sign in to leave a comment.