Re:Re:Re: [Discuss] Prepare to release 1.1

2022-06-29 Thread Mingyu Chen
Hi all,
The outer join bug has been fixed in dev-1.0.1.
And also fix some other bugs.


This thread is closed and I will start vote for the v1.1 release.




--

此致!Best Regards
陈明雨 Mingyu Chen

Email:
morning...@apache.org




在 2022-06-27 09:52:53,"陈明雨"  写道:
>Hi all,
>The dev-1.0.1 is now closed. I'm just waiting for one last PR to resolve the 
>outer join bug.
>
>
>
>
>--
>
>此致!Best Regards
>陈明雨 Mingyu Chen
>
>Email:
>morning...@apache.org
>
>
>
>
>
>在 2022-06-24 10:06:05,"zhg yang"  写道:
>>+1 we should maintain a version with bug fix but without new features
>>Thanks
>>Yang Zhengguo
>>
>>
>>陈明雨  于2022年6月2日周四 23:27写道:
>>
>>> Hi all,
>>> It has been more than two months since version 1.0 was released.
>>> During this period, the Doris community has updated a lot of functions,
>>> and made a lot of bug fixes and stability and performance improvements.
>>> And version 1.0 also has some serious bug that need to be fixed.
>>> So I recommend starting work on the 1.1 release. And a final release date
>>> would be June 15th.
>>>
>>>
>>> Version 1.1 will be produced from the dev-1.0.1[1] branch, mainly as an
>>> optimized version of version 1.0, which fixes and optimizes a large number
>>> of vectorization engine-related problems.
>>> Compared with version 1.0, both performance and stability are greatly
>>> improved.
>>>
>>>
>>> All changes related to version 1.1 can be seen here [2]. At the same time,
>>> we also have some PRs waiting to be merged in this version[3].
>>> And I drafted a Release Note[4].
>>>
>>>
>>> If you have a PR that needs to be merged into this release, please reply
>>> in this thread, or simply label the PR with dev-1.0.1.
>>>
>>>
>>> [1] https://github.com/apache/incubator-doris/tree/dev-1.0.1
>>> [2]
>>> https://github.com/apache/incubator-doris/issues?q=is%3Apr+is%3Aclosed+label%3Adev%2Fmerged-1.0.1
>>> [3]
>>> https://github.com/apache/incubator-doris/pulls?q=is%3Apr+is%3Aopen+label%3Adev%2F1.0.1
>>> [4] https://github.com/apache/incubator-doris/issues/9949
>>>
>>>
>>>
>>>
>>> --
>>>
>>> 此致!Best Regards
>>> 陈明雨 Mingyu Chen
>>>
>>> Email:
>>> chenmin...@apache.org


Re:Re:Re:Re: [Discuss] Prepare to release 1.1

2022-06-29 Thread Mingyu Chen
The first release candidate tag is: 
https://github.com/apache/doris/tree/1.1.0-rc01




--

此致!Best Regards
陈明雨 Mingyu Chen

Email:
morning...@apache.org





在 2022-06-29 15:05:44,"Mingyu Chen"  写道:
>Hi all,
>The outer join bug has been fixed in dev-1.0.1.
>And also fix some other bugs.
>
>
>This thread is closed and I will start vote for the v1.1 release.
>
>
>
>
>--
>
>此致!Best Regards
>陈明雨 Mingyu Chen
>
>Email:
>morning...@apache.org
>
>
>
>
>在 2022-06-27 09:52:53,"陈明雨"  写道:
>>Hi all,
>>The dev-1.0.1 is now closed. I'm just waiting for one last PR to resolve the 
>>outer join bug.
>>
>>
>>
>>
>>--
>>
>>此致!Best Regards
>>陈明雨 Mingyu Chen
>>
>>Email:
>>morning...@apache.org
>>
>>
>>
>>
>>
>>在 2022-06-24 10:06:05,"zhg yang"  写道:
>>>+1 we should maintain a version with bug fix but without new features
>>>Thanks
>>>Yang Zhengguo
>>>
>>>
>>>陈明雨  于2022年6月2日周四 23:27写道:
>>>
 Hi all,
 It has been more than two months since version 1.0 was released.
 During this period, the Doris community has updated a lot of functions,
 and made a lot of bug fixes and stability and performance improvements.
 And version 1.0 also has some serious bug that need to be fixed.
 So I recommend starting work on the 1.1 release. And a final release date
 would be June 15th.


 Version 1.1 will be produced from the dev-1.0.1[1] branch, mainly as an
 optimized version of version 1.0, which fixes and optimizes a large number
 of vectorization engine-related problems.
 Compared with version 1.0, both performance and stability are greatly
 improved.


 All changes related to version 1.1 can be seen here [2]. At the same time,
 we also have some PRs waiting to be merged in this version[3].
 And I drafted a Release Note[4].


 If you have a PR that needs to be merged into this release, please reply
 in this thread, or simply label the PR with dev-1.0.1.


 [1] https://github.com/apache/incubator-doris/tree/dev-1.0.1
 [2]
 https://github.com/apache/incubator-doris/issues?q=is%3Apr+is%3Aclosed+label%3Adev%2Fmerged-1.0.1
 [3]
 https://github.com/apache/incubator-doris/pulls?q=is%3Apr+is%3Aopen+label%3Adev%2F1.0.1
 [4] https://github.com/apache/incubator-doris/issues/9949




 --

 此致!Best Regards
 陈明雨 Mingyu Chen

 Email:
 chenmin...@apache.org


????streamload??????????????abort reason: coordinate BE is down

2022-06-29 Thread james
??


     FE??3??BE 


     ??streamload 
??
BE ?? ??


W0625 21:11:04.788328 19871 stream_load_executor.cpp:281] commit transaction 
failed, errmsg=errCode = 2, detailMessage = transaction [1638] is already 
aborted. abort reason: coordinate BE is down, 
id=d14f60f37fe29182-e2452ee346c991b8, job_id=-1, txn_id=1638, 
label=xxx_t_3e9877180e1f415686443124a9fb72b0, elapse(s)=34
W0625 21:11:04.788368 19871 stream_load.cpp:144] handle streaming load failed, 
id=d14f60f37fe29182-e2452ee346c991b8, errmsg=errCode = 2, detailMessage = 
transaction [1638] is already aborted. abort reason: coordinate BE is down
W0625 21:11:04.796619 19890 stream_load_executor.cpp:281] commit transaction 
failed, errmsg=errCode = 2, detailMessage = transaction [1640] is already 
aborted. abort reason: coordinate BE is down, 
id=bb4ac386050f365f-5697ff9a7614929d, job_id=-1, txn_id=1640, 
label=xxx_t_30bde0223a4c49bfae8855ea510e5157, elapse(s)=34
W0625 21:11:04.796648 19890 stream_load.cpp:144] handle streaming load failed, 
id=bb4ac386050f365f-5697ff9a7614929d, errmsg=errCode = 2, detailMessage = 
transaction [1640] is already aborted. abort reason: coordinate BE is down
W0625 21:11:04.796726 19872 stream_load_executor.cpp:281] commit transaction 
failed, errmsg=errCode = 2, detailMessage = transaction [1632] is already 
aborted. abort reason: coordinate BE is down, 
id=3e4bc3cd3e3e853f-cf8604ac349eb1b5, job_id=-1, txn_id=1632, 
label=xxx_t_a20283afde854b54b362b91141571445, elapse(s)=34
W0625 21:11:04.796748 19894 stream_load_executor.cpp:281] commit transaction 
failed, errmsg=errCode = 2, detailMessage = transaction [1635] is already 
aborted. abort reason: coordinate BE is down, 
id=0c4a7e043b34eb88-9d27887e66e1e392, job_id=-1, txn_id=1635, 
label=xxx_t_638ea9d5da994462b56674f4218aa95d, elapse(s)=34
W0625 21:11:04.796767 19872 stream_load.cpp:144] handle streaming load failed, 
id=3e4bc3cd3e3e853f-cf8604ac349eb1b5, errmsg=errCode = 2, detailMessage = 
transaction [1632] is already aborted. abort reason: coordinate BE is down
W0625 21:11:04.796799 19894 stream_load.cpp:144] handle streaming load failed, 
id=0c4a7e043b34eb88-9d27887e66e1e392, errmsg=errCode = 2, detailMessage = 
transaction [1635] is already aborted. abort reason: coordinate BE is down
W0626 21:07:20.208528 19894 stream_load_executor.cpp:281] commit transaction 
failed, errmsg=errCode = 2, detailMessage = transaction [1834] is already 
aborted. abort reason: coordinate BE is down, 
id=6c4fc4d91eead9d1-c7bd542607b427a5, job_id=-1, txn_id=1834, 
label=xxx_t_a9c563f5eb2f45ee8a299889704ee42e, elapse(s)=33
W0626 21:07:20.208611 19894 stream_load.cpp:144] handle streaming load failed, 
id=6c4fc4d91eead9d1-c7bd542607b427a5, errmsg=errCode = 2, detailMessage = 
transaction [1834] is already aborted. abort reason: coordinate BE is down

[Discuss][DSIP] Support buffering data and automatically committing in Be to improve Doris real-time push ability

2022-06-29 Thread 梅祎
Hi all,

In current data import ways,users generally use StreamLoad to push
real-time data to Doris, which need to buffer multiple records in the
client side, and then send them to Doris together. It is pretty complex for
users.

We propose to buffer the data in the server side(BE) and commit
automatically so that it is transparent to users and users can insert
records row by row.

I'll add detailed design later.


Spark????Doris 1.0???????????????? Unrecognized field "keysType"

2022-06-29 Thread james
??


      ??doris 1.0    1??FE,3??BE.


      sparkdoris






org.apache.doris.spark.exception.DorisException: Doris FE's response cannot map 
to schema. res: 
{"keysType":"DUP_KEYS","properties":[{"name":"id","aggregation_type":"","comment":"
 id","type":"BIGINT"},{"name":"name","aggregation_type":"NONE","comment":" 
","type":"VARCHAR"}],"status":200}
 at 
org.apache.doris.spark.rest.RestService.parseSchema(RestService.java:303)
 at org.apache.doris.spark.rest.RestService.getSchema(RestService.java:279)
 at 
org.apache.doris.spark.sql.SchemaUtils$.discoverSchemaFromFe(SchemaUtils.scala:51)
 at 
org.apache.doris.spark.sql.SchemaUtils$.discoverSchema(SchemaUtils.scala:41)
 at 
org.apache.doris.spark.sql.DorisRelation.lazySchema$lzycompute(DorisRelation.scala:48)
 at 
org.apache.doris.spark.sql.DorisRelation.lazySchema(DorisRelation.scala:48)
 at org.apache.doris.spark.sql.DorisRelation.schema(DorisRelation.scala:52)
 at 
org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:402)
 at 
org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:223)
 at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:211)
 at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:167)
 at 
com.vxdata.datacenter.dataquality.util.SparkSqlUtil$.loadTable(SparkSqlUtil.scala:178)
 at com.vxdata.buildmodel.computer.input.DBInput.compute(DBInput.scala:11)
 at 
com.vxdata.buildmodel.computer.service.JobFlowService.invokeCompute(JobFlowService.scala:239)
 at 
com.vxdata.buildmodel.computer.service.JobFlowService$$anonfun$exec$1.apply(JobFlowService.scala:81)
 at 
com.vxdata.buildmodel.computer.service.JobFlowService$$anonfun$exec$1.apply(JobFlowService.scala:78)
 at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
 at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
 at 
com.vxdata.buildmodel.computer.service.JobFlowService.exec(JobFlowService.scala:78)
 at 
com.vxdata.buildmodel.computer.service.JobFlowService.exec(JobFlowService.scala:34)
 at 
com.vxdata.etl.SparkETLMain$$anonfun$1.apply$mcV$sp(SparkETLMain.scala:64)
 at 
com.vxdata.datacenter.dataquality.common.TApplication$class.start(TApplication.scala:42)
 at com.vxdata.etl.SparkETLMain$.start(SparkETLMain.scala:12)
 at 
com.vxdata.etl.SparkETLMain$.delayedEndpoint$com$vxdata$etl$SparkETLMain$1(SparkETLMain.scala:20)
 at 
com.vxdata.etl.SparkETLMain$delayedInit$body.apply(SparkETLMain.scala:12)
 at scala.Function0$class.apply$mcV$sp(Function0.scala:34)
 at 
scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
 at scala.App$$anonfun$main$1.apply(App.scala:76)
 at scala.App$$anonfun$main$1.apply(App.scala:76)
 at scala.collection.immutable.List.foreach(List.scala:392)
 at 
scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35)
 at scala.App$class.main(App.scala:76)
 at com.vxdata.etl.SparkETLMain$.main(SparkETLMain.scala:12)
 at com.vxdata.etl.SparkETLMain.main(SparkETLMain.scala)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498)
 at 
org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:673)
 Caused by: org.codehaus.jackson.map.exc.UnrecognizedPropertyException: 
Unrecognized field "keysType" (Class 
org.apache.doris.spark.rest.models.Schema), not marked as ignorable
 at [Source: java.io.StringReader@7a1fa00e; line: 1, column: 14] (through 
reference chain: org.apache.doris.spark.rest.models.Schema["keysType"])
 at 
org.codehaus.jackson.map.exc.UnrecognizedPropertyException.from(UnrecognizedPropertyException.java:53)
 at 
org.codehaus.jackson.map.deser.StdDeserializationContext.unknownFieldException(StdDeserializationContext.java:267)
 at 
org.codehaus.jackson.map.deser.std.StdDeserializer.reportUnknownProperty(StdDeserializer.java:673)
 at 
org.codehaus.jackson.map.deser.std.StdDeserializer.handleUnknownProperty(StdDeserializer.java:659)
 at 
org.codehaus.jackson.map.deser.BeanDeserializer.handleUnknownProperty(BeanDeserializer.java:1365)
 at 
org.codehaus.jackson.map.deser.BeanDeserializer._handleUnknown(BeanDeserializer.java:725)
 at 
org.codehaus.jackson.map.deser.BeanDeserializer.deserializeFromObject(BeanDeserializer.java:703)
 at 
org.codehaus.jackson.map.deser.BeanDeserializer.deserialize(BeanDeserializer.java:580)
 at 
org.codehaus.jackson.map.ObjectMapper._readMapAndClose(ObjectMapper.java:2732)
 at org.codehaus.jackson.map.ObjectMapper.readValue(ObjectMapper.java:1863)
 at 
org.apache.doris.spark.rest.RestService.parseSchema(RestService.java:295) ... 
38 more ;

Re: 通过streamload导入数据报异常abort reason: coordinate BE is down

2022-06-29 Thread 张家峰
应该是你的be挂了
show backends  看看


james <23461...@qq.com.invalid>于2022年6月29日 周三15:44写道:

> 您好!
>
>
>      系统的部署情况说明:一台FE,3台BE 部署在一个局域网内。
>
>
>      在通过streamload导入大量数据时候,报上面的错误!
> 请问一般在什么情况下可能出现这样的问题,有什么解决方案没。
> BE 的 日志内容如下:
>
>
> W0625 21:11:04.788328 19871 stream_load_executor.cpp:281] commit
> transaction failed, errmsg=errCode = 2, detailMessage = transaction [1638]
> is already aborted. abort reason: coordinate BE is down,
> id=d14f60f37fe29182-e2452ee346c991b8, job_id=-1, txn_id=1638,
> label=xxx_t_3e9877180e1f415686443124a9fb72b0, elapse(s)=34
> W0625 21:11:04.788368 19871 stream_load.cpp:144] handle streaming load
> failed, id=d14f60f37fe29182-e2452ee346c991b8, errmsg=errCode = 2,
> detailMessage = transaction [1638] is already aborted. abort reason:
> coordinate BE is down
> W0625 21:11:04.796619 19890 stream_load_executor.cpp:281] commit
> transaction failed, errmsg=errCode = 2, detailMessage = transaction [1640]
> is already aborted. abort reason: coordinate BE is down,
> id=bb4ac386050f365f-5697ff9a7614929d, job_id=-1, txn_id=1640,
> label=xxx_t_30bde0223a4c49bfae8855ea510e5157, elapse(s)=34
> W0625 21:11:04.796648 19890 stream_load.cpp:144] handle streaming load
> failed, id=bb4ac386050f365f-5697ff9a7614929d, errmsg=errCode = 2,
> detailMessage = transaction [1640] is already aborted. abort reason:
> coordinate BE is down
> W0625 21:11:04.796726 19872 stream_load_executor.cpp:281] commit
> transaction failed, errmsg=errCode = 2, detailMessage = transaction [1632]
> is already aborted. abort reason: coordinate BE is down,
> id=3e4bc3cd3e3e853f-cf8604ac349eb1b5, job_id=-1, txn_id=1632,
> label=xxx_t_a20283afde854b54b362b91141571445, elapse(s)=34
> W0625 21:11:04.796748 19894 stream_load_executor.cpp:281] commit
> transaction failed, errmsg=errCode = 2, detailMessage = transaction [1635]
> is already aborted. abort reason: coordinate BE is down,
> id=0c4a7e043b34eb88-9d27887e66e1e392, job_id=-1, txn_id=1635,
> label=xxx_t_638ea9d5da994462b56674f4218aa95d, elapse(s)=34
> W0625 21:11:04.796767 19872 stream_load.cpp:144] handle streaming load
> failed, id=3e4bc3cd3e3e853f-cf8604ac349eb1b5, errmsg=errCode = 2,
> detailMessage = transaction [1632] is already aborted. abort reason:
> coordinate BE is down
> W0625 21:11:04.796799 19894 stream_load.cpp:144] handle streaming load
> failed, id=0c4a7e043b34eb88-9d27887e66e1e392, errmsg=errCode = 2,
> detailMessage = transaction [1635] is already aborted. abort reason:
> coordinate BE is down
> W0626 21:07:20.208528 19894 stream_load_executor.cpp:281] commit
> transaction failed, errmsg=errCode = 2, detailMessage = transaction [1834]
> is already aborted. abort reason: coordinate BE is down,
> id=6c4fc4d91eead9d1-c7bd542607b427a5, job_id=-1, txn_id=1834,
> label=xxx_t_a9c563f5eb2f45ee8a299889704ee42e, elapse(s)=33
> W0626 21:07:20.208611 19894 stream_load.cpp:144] handle streaming load
> failed, id=6c4fc4d91eead9d1-c7bd542607b427a5, errmsg=errCode = 2,
> detailMessage = transaction [1834] is already aborted. abort reason:
> coordinate BE is down

-- 
张家峰


?????? ????streamload??????????????abort reason: coordinate BE is down

2022-06-29 Thread james
??BEload??BE??




--  --
??: 
   "dev"


Re: Re: [Discuss][DSIP] Support Merge-On-Write implementation for UNIQUE KEY data model

2022-06-29 Thread Chen Zhang
Updated the scheduling
https://cwiki.apache.org/confluence/display/DORIS/DSIP-018%3A+Support+Merge-On-Write+implementation+for+UNIQUE+KEY+data+model

Best
Chen Zhang
在 2022年6月27日 +0800 11:59,Chen Zhang ,写道:
> Hi Devs, I've update the DISP last weekend, if you are interest on this 
> feature, welcome to review and comment, thanks
> https://cwiki.apache.org/confluence/display/DORIS/DSIP-018%3A+Support+Merge-On-Write+implementation+for+UNIQUE+KEY+data+model
>
> Best
> Chen Zhang
> 在 2022年6月24日 +0800 10:13,zhg yang ,写道:
> > @ Chen Zhang For the more important features, it is best to send a DISP
> > first to let everyone discuss the design
> > Thanks
> > Yang Zhengguo
> >
> >
> > Chen Zhang  于2022年6月23日周四 22:30写道:
> >
> > > @Minghong We'll use a multi-version delete bitmap, only save delta for
> > > each version.
> > > For example, we have a rowset with version [0-98], transaction 99 updated
> > > some row in that rowset, and so does transaction 100 and 101, there would
> > > be 3 delete bitmaps on that rowset, corresponding to rows updated by
> > > version 99, 100 and 101. A query with version x will only see the bitmap 
> > > up
> > > to version x. There's more details about space saving and cache
> > > acceleration, let's discuss it in DSIP.
> > >
> > > @Xiaoli, our team have finished most develop works for the basic function
> > > in our private repository, but there‘s still lots of works to do, welcome
> > > to get involve.
> > >
> > > @Mingyu, could you help to create a DISP doc? I don't seem to have
> > > permission.
> > >
> > > Best
> > > Chen Zhang
> > > On Jun 23, 2022, 21:41 +0800, Zhou Minghong ,
> > > wrote:
> > > > Hi Chen Zhang
> > > > one question about "and a delete bitmap (marks some rowid as deleted)”:
> > > > how to handle transaction information by a bitmap?
> > > > for example, transaction_100 delete a row, but this still visible to
> > > transaction_99, but not visible to trasanction_101. How to handle this 
> > > case?
> > > >
> > > >
> > > > Br/Minghong
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > At 2022-06-23 19:14:58, "Zhu,Xiaoli"  wrote:
> > > > > Hi Chen Zhang,
> > > > >
> > > > > I am very interested in this topic, and want to participate in the
> > > development.
> > > > >
> > > > > 在 2022/6/23 下午2:44,“Chen Zhang” 写入:
> > > > >
> > > > > Hi devs,
> > > > >
> > > > > Unique-Key data model is widely used in scenarios like Flink-CDC, user
> > > > > profile(用户画像), E-commerce orders, but the query performance for 
> > > > > current
> > > > > Merge-On-Read implementation is not good, due to the following 
> > > > > reasons:
> > > > >
> > > > > 1. Doris can't determine whether one row in a segment file is latest 
> > > > > or
> > > > > outdated, so it has to do some extra merge sort before getting the
> > > > > latest data, and key comparison is quite CPU-costive.
> > > > > 2. Aggregate function predicate push down is not supported by the
> > > > > Unique-Key data model due to reason(1).
> > > > >
> > > > > I'd like to propose to support a Merge-On-Write implementation for the
> > > > > Unique-Key data model, which leverages a new segment-file-level 
> > > > > primary
> > > > > key index (used for point lookup on write) and a delete bitmap (marks
> > > some
> > > > > rowid as deleted), which can optimize read performance significantly.
> > > > >
> > > > > At the beginning, we wanted to add another Primary-Key data model with
> > > > > Merge-On-Write implementation, but after a lot of discussion, we'd
> > > prefer
> > > > > to improve the Unique-Key data model rather than adding another one.
> > > > >
> > > > > I'll add detailed design and related research in the DSIP doc later.
> > > > >
> > > > >
> > > > > -
> > > > > To unsubscribe, e-mail: dev-unsubscr...@doris.apache.org
> > > > > For additional commands, e-mail: dev-h...@doris.apache.org
> > > > >
> > >


Re:Re:Re:Re:Re: [Discuss] Prepare to release 1.1

2022-06-29 Thread Mingyu Chen
There is a bug that may cause query stuck, which has been fix in master 
branch[1].
So I cherry-pick part of it to dev-1.0.1 branch[2] and make a new tag 
1.1.0-rc02[3]


[1] https://github.com/apache/doris/pull/10214
[2] 
https://github.com/apache/doris/commit/589b44eddf6eafcd55057468de828b2c222ceedd
[3] https://github.com/apache/doris/tree/1.1.0-rc02




--

此致!Best Regards
陈明雨 Mingyu Chen

Email:
morning...@apache.org





在 2022-06-29 15:09:25,"Mingyu Chen"  写道:
>The first release candidate tag is: 
>https://github.com/apache/doris/tree/1.1.0-rc01
>
>
>
>
>--
>
>此致!Best Regards
>陈明雨 Mingyu Chen
>
>Email:
>morning...@apache.org
>
>
>
>
>
>在 2022-06-29 15:05:44,"Mingyu Chen"  写道:
>>Hi all,
>>The outer join bug has been fixed in dev-1.0.1.
>>And also fix some other bugs.
>>
>>
>>This thread is closed and I will start vote for the v1.1 release.
>>
>>
>>
>>
>>--
>>
>>此致!Best Regards
>>陈明雨 Mingyu Chen
>>
>>Email:
>>morning...@apache.org
>>
>>
>>
>>
>>在 2022-06-27 09:52:53,"陈明雨"  写道:
>>>Hi all,
>>>The dev-1.0.1 is now closed. I'm just waiting for one last PR to resolve the 
>>>outer join bug.
>>>
>>>
>>>
>>>
>>>--
>>>
>>>此致!Best Regards
>>>陈明雨 Mingyu Chen
>>>
>>>Email:
>>>morning...@apache.org
>>>
>>>
>>>
>>>
>>>
>>>在 2022-06-24 10:06:05,"zhg yang"  写道:
+1 we should maintain a version with bug fix but without new features
Thanks
Yang Zhengguo


陈明雨  于2022年6月2日周四 23:27写道:

> Hi all,
> It has been more than two months since version 1.0 was released.
> During this period, the Doris community has updated a lot of functions,
> and made a lot of bug fixes and stability and performance improvements.
> And version 1.0 also has some serious bug that need to be fixed.
> So I recommend starting work on the 1.1 release. And a final release date
> would be June 15th.
>
>
> Version 1.1 will be produced from the dev-1.0.1[1] branch, mainly as an
> optimized version of version 1.0, which fixes and optimizes a large number
> of vectorization engine-related problems.
> Compared with version 1.0, both performance and stability are greatly
> improved.
>
>
> All changes related to version 1.1 can be seen here [2]. At the same time,
> we also have some PRs waiting to be merged in this version[3].
> And I drafted a Release Note[4].
>
>
> If you have a PR that needs to be merged into this release, please reply
> in this thread, or simply label the PR with dev-1.0.1.
>
>
> [1] https://github.com/apache/incubator-doris/tree/dev-1.0.1
> [2]
> https://github.com/apache/incubator-doris/issues?q=is%3Apr+is%3Aclosed+label%3Adev%2Fmerged-1.0.1
> [3]
> https://github.com/apache/incubator-doris/pulls?q=is%3Apr+is%3Aopen+label%3Adev%2F1.0.1
> [4] https://github.com/apache/incubator-doris/issues/9949
>
>
>
>
> --
>
> 此致!Best Regards
> 陈明雨 Mingyu Chen
>
> Email:
> chenmin...@apache.org


Re:[Discuss][DSIP] Support buffering data and automatically committing in Be to improve Doris real-time push ability

2022-06-29 Thread Mingyu Chen
Hi Meiyi,
Thanks for you proposal.
And for "buffer the data in the server side(BE)", Doris already support the 
"begin; insert; insert; insert; ...; commit" method,
which is also a kind of "buffer data in server side" feature for "insert into 
values" operations.


Maybe you can explain more about how it works with stream load and how to use 
it.
Please register an account here[1] and tell me your account name, I will create 
a DSIP for this proposal.


[1] https://cwiki.apache.org/confluence/display/DORIS/Home




--

此致!Best Regards
陈明雨 Mingyu Chen

Email:
morning...@apache.org





At 2022-06-29 15:54:34, "梅祎"  wrote:
>Hi all,
>
>In current data import ways,users generally use StreamLoad to push
>real-time data to Doris, which need to buffer multiple records in the
>client side, and then send them to Doris together. It is pretty complex for
>users.
>
>We propose to buffer the data in the server side(BE) and commit
>automatically so that it is transparent to users and users can insert
>records row by row.
>
>I'll add detailed design later.


Re:Re:[Discuss][DSIP] Support buffering data and automatically committing in Be to improve Doris real-time push ability

2022-06-29 Thread 蔡聪辉
Will there be too much performance drop if client insert records row by row? 
due to many rpcs?


At 2022-06-29 21:31:02, "Mingyu Chen"  wrote:
>Hi Meiyi,
>Thanks for you proposal.
>And for "buffer the data in the server side(BE)", Doris already support the 
>"begin; insert; insert; insert; ...; commit" method,
>which is also a kind of "buffer data in server side" feature for "insert into 
>values" operations.
>
>
>Maybe you can explain more about how it works with stream load and how to use 
>it.
>Please register an account here[1] and tell me your account name, I will 
>create a DSIP for this proposal.
>
>
>[1] https://cwiki.apache.org/confluence/display/DORIS/Home
>
>
>
>
>--
>
>此致!Best Regards
>陈明雨 Mingyu Chen
>
>Email:
>morning...@apache.org
>
>
>
>
>
>At 2022-06-29 15:54:34, "梅祎"  wrote:
>>Hi all,
>>
>>In current data import ways,users generally use StreamLoad to push
>>real-time data to Doris, which need to buffer multiple records in the
>>client side, and then send them to Doris together. It is pretty complex for
>>users.
>>
>>We propose to buffer the data in the server side(BE) and commit
>>automatically so that it is transparent to users and users can insert
>>records row by row.
>>
>>I'll add detailed design later.


Re: [Discuss][DSIP] Support buffering data and automatically committing in Be to improve Doris real-time push ability

2022-06-29 Thread 王博
Hi Meiyi,
Thanks for you proposal.
I would like to know why doris need this thing and what actually it is.
About what, it that you want to add something like kafka in Doris?
About why, can you explain more in what scenarios does the user need this?

梅祎  于2022年6月29日周三 15:54写道:

> Hi all,
>
> In current data import ways,users generally use StreamLoad to push
> real-time data to Doris, which need to buffer multiple records in the
> client side, and then send them to Doris together. It is pretty complex for
> users.
>
> We propose to buffer the data in the server side(BE) and commit
> automatically so that it is transparent to users and users can insert
> records row by row.
>
> I'll add detailed design later.
>


-- 
王博  Wang Bo


Re: [Discuss][DSIP] Support buffering data and automatically committing in Be to improve Doris real-time push ability

2022-06-29 Thread 梅祎
Hi all, thanks for your reply. Let me update details in DISP firstly.
@Mingyu, my account is meiyi, thanks.

王博  于2022年6月30日周四 09:18写道:

> Hi Meiyi,
> Thanks for you proposal.
> I would like to know why doris need this thing and what actually it is.
> About what, it that you want to add something like kafka in Doris?
> About why, can you explain more in what scenarios does the user need this?
>
> 梅祎  于2022年6月29日周三 15:54写道:
>
> > Hi all,
> >
> > In current data import ways,users generally use StreamLoad to push
> > real-time data to Doris, which need to buffer multiple records in the
> > client side, and then send them to Doris together. It is pretty complex
> for
> > users.
> >
> > We propose to buffer the data in the server side(BE) and commit
> > automatically so that it is transparent to users and users can insert
> > records row by row.
> >
> > I'll add detailed design later.
> >
>
>
> --
> 王博  Wang Bo
>


Re:Re: [Discuss][DSIP] Support buffering data and automatically committing in Be to improve Doris real-time push ability

2022-06-29 Thread Mingyu Chen
Done: 
https://cwiki.apache.org/confluence/display/DORIS/DSIP-020%3A+Support+buffering+data+and+automatically+committing+for+stream+load

--

此致!Best Regards
陈明雨 Mingyu Chen

Email:
morning...@apache.org





在 2022-06-30 11:31:15,"梅祎"  写道:
>Hi all, thanks for your reply. Let me update details in DISP firstly.
>@Mingyu, my account is meiyi, thanks.
>
>王博  于2022年6月30日周四 09:18写道:
>
>> Hi Meiyi,
>> Thanks for you proposal.
>> I would like to know why doris need this thing and what actually it is.
>> About what, it that you want to add something like kafka in Doris?
>> About why, can you explain more in what scenarios does the user need this?
>>
>> 梅祎  于2022年6月29日周三 15:54写道:
>>
>> > Hi all,
>> >
>> > In current data import ways,users generally use StreamLoad to push
>> > real-time data to Doris, which need to buffer multiple records in the
>> > client side, and then send them to Doris together. It is pretty complex
>> for
>> > users.
>> >
>> > We propose to buffer the data in the server side(BE) and commit
>> > automatically so that it is transparent to users and users can insert
>> > records row by row.
>> >
>> > I'll add detailed design later.
>> >
>>
>>
>> --
>> 王博  Wang Bo
>>