[
https://issues.apache.org/jira/browse/PIO-105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Donald Szeto updated PIO-105:
-----------------------------
Target Version/s: 0.12.0-incubating
> Batch Predictions
> -----------------
>
> Key: PIO-105
> URL: https://issues.apache.org/jira/browse/PIO-105
> Project: PredictionIO
> Issue Type: New Feature
> Components: Core
> Reporter: Mars Hall
> Assignee: Mars Hall
> Fix For: 0.12.0-incubating
>
>
> Implement a new {{pio batchpredict}} command to enable massive, fast, batch
> predictions from a trained model. Read a multi-object JSON file as the input
> format, with one query object per line. Similarly, write results to a
> multi-object JSON file, with one prediction result + its original query per
> line.
> Currently getting bulk predictions from PredictionIO is possible with either:
> * a {{pio eval}} script, which will always train a fresh, unvalidated model
> before getting predictions
> * a custom script that hits the {{queries.json}} HTTP API, which is a serious
> bottleneck when requesting hundreds-of-thousands or millions of predictions
> Neither of these existing bulk-prediction hacks are adequate for the reasons
> mentioned.
> It's time for this use-case to be a firstclass command :D
> Pull request https://github.com/apache/incubator-predictionio/pull/412
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)