[ 
https://issues.apache.org/jira/browse/PIO-105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16093567#comment-16093567
 ] 

ASF GitHub Bot commented on PIO-105:
------------------------------------

Github user dszeto commented on the issue:

    https://github.com/apache/incubator-predictionio/pull/412
  
    LGTM. Great work @mars !
    
    Agree with @takezoe 's comments. I think there are a few follow up tasks we 
should do.
    
    1. Refactor common code that exists in both `CreateServer` and 
`BatchPredict`.
    2. Brush up Scala coding style as pointed out by @takezoe .
    3. Document the new feature in the doc site.


> Batch Predictions
> -----------------
>
>                 Key: PIO-105
>                 URL: https://issues.apache.org/jira/browse/PIO-105
>             Project: PredictionIO
>          Issue Type: New Feature
>          Components: Core
>            Reporter: Mars Hall
>            Assignee: Mars Hall
>
> Implement a new {{pio batchpredict}} command to enable massive, fast, batch 
> predictions from a trained model. Read a multi-object JSON file as the input 
> format, with one query object per line. Similarly, write results to a 
> multi-object JSON file, with one prediction result + its original query per 
> line.
> Currently getting bulk predictions from PredictionIO is possible with either:
> * a {{pio eval}} script, which will always train a fresh, unvalidated model 
> before getting predictions
> * a custom script that hits the {{queries.json}} HTTP API, which is a serious 
> bottleneck when requesting hundreds-of-thousands or millions of predictions
> Neither of these existing bulk-prediction hacks are adequate for the reasons 
> mentioned.
> It's time for this use-case to be a firstclass command :D
> Pull request https://github.com/apache/incubator-predictionio/pull/412



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to