Hi, thats a known issue and unrelated:
https://issues.apache.org/jira/browse/SOLR-9120

M.
 
 
-----Original message-----
> From:Stephen Weiss <steve.we...@wgsn.com>
> Sent: Tuesday 17th May 2016 23:10
> To: solr-user@lucene.apache.org; Aleksey Mezhva <aleksey.mez...@wgsn.com>; 
> Hans Zhou <hans.z...@wgsn.com>
> Subject: Re: SolrCloud replicas consistently out of sync
> 
> I should add - looking back through the logs, we're seeing frequent errors 
> like this now:
> 
> 78819692 WARN  (qtp110456297-1145) [   ] o.a.s.h.a.LukeRequestHandler Error 
> getting file length for [segments_4o]
> java.nio.file.NoSuchFileException: 
> /var/solr/data/instock_shard5_replica1/data/index.20160516230059221/segments_4o
> 
> --
> Steve
> 
> 
> On Tue, May 17, 2016 at 5:07 PM, Stephen Weiss 
> <steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>> wrote:
> OK, so we did as you suggest, read through that article, and we reconfigured 
> the autocommit to:
> 
> <autoCommit>
> <maxTime>${solr.autoCommit.maxTime:30000}</maxTime>
> <openSearcher>false</openSearcher>
> </autoCommit>
> 
> <autoSoftCommit>
> <maxTime>${solr.autoSoftCommit.maxTime:600000}</maxTime>
> </autoSoftCommit>
> 
> However, we see no change, aside from the fact that it's clearly committing 
> more frequently.  I will say on our end, we clearly misunderstood the 
> difference between soft and hard commit, but even now having it configured 
> this way, we are still totally out of sync, long after all indexing has 
> completed (it's been about 30 minutes now).  We manually pushed through a 
> commit on the whole collection as suggested, however, all we get back for 
> that is o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping 
> IW.commit., which makes sense, because it was all committed already anyway.
> 
> We still currently have all shards mismatched:
> 
> instock_shard1   replica 1: 30788491 replica 2: 30778865
> instock_shard10   replica 1: 30973059 replica 2: 30971874
> instock_shard11   replica 2: 31036815 replica 1: 31034715
> instock_shard12   replica 2: 30177084 replica 1: 30170511
> instock_shard13   replica 2: 30608225 replica 1: 30603923
> instock_shard14   replica 2: 30755739 replica 1: 30753191
> instock_shard15   replica 2: 30891713 replica 1: 30891528
> instock_shard16   replica 1: 30818567 replica 2: 30817152
> instock_shard17   replica 1: 30423877 replica 2: 30422742
> instock_shard18   replica 2: 30874979 replica 1: 30872223
> instock_shard19   replica 2: 30917208 replica 1: 30909999
> instock_shard2   replica 1: 31062339 replica 2: 31060575
> instock_shard20   replica 1: 30192046 replica 2: 30190893
> instock_shard21   replica 2: 30793817 replica 1: 30791135
> instock_shard22   replica 2: 30821521 replica 1: 30818836
> instock_shard23   replica 2: 30553773 replica 1: 30547336
> instock_shard24   replica 1: 30975564 replica 2: 30971170
> instock_shard25   replica 1: 30734696 replica 2: 30731682
> instock_shard26   replica 1: 31465696 replica 2: 31464738
> instock_shard27   replica 1: 30844884 replica 2: 30842445
> instock_shard28   replica 2: 30549826 replica 1: 30547405
> instock_shard29   replica 2: 30637777 replica 1: 30634091
> instock_shard3   replica 1: 30930723 replica 2: 30926483
> instock_shard30   replica 2: 30904528 replica 1: 30902649
> instock_shard31   replica 2: 31175813 replica 1: 31174921
> instock_shard32   replica 2: 30932837 replica 1: 30926456
> instock_shard4   replica 2: 30758100 replica 1: 30754129
> instock_shard5   replica 2: 31008893 replica 1: 31002581
> instock_shard6   replica 2: 31008679 replica 1: 31005380
> instock_shard7   replica 2: 30738468 replica 1: 30737795
> instock_shard8   replica 2: 30620929 replica 1: 30616715
> instock_shard9   replica 1: 31071386 replica 2: 31066956
> 
> The fact that the min_rf numbers aren't coming back as 2 seems to indicate to 
> me that documents simply aren't making it to both replicas - why would that 
> have anything to do with committing anyway?
> 
> Something else is amiss here.  Too bad, committing sounded like an easy 
> answer!
> 
> --
> Steve
> 
> 
> On Tue, May 17, 2016 at 11:39 AM, Erick Erickson 
> <erickerick...@gmail.com<mailto:erickerick...@gmail.com>> wrote:
> OK, these autocommit settings need revisiting.
> 
> First off, I'd remove the maxDocs entirely although with the setting
> you're using it probably doesn't matter.
> 
> The maxTime of 1,200,000 is 20 minutes. Which means if you evern
> un-gracefully kill your shards you'll have up to 20 minutes worth of
> data to replay from the tlog.... or resynch from the leader. Make this
> much shorter (60000 or less) and be sure to gracefully kill your Solrs.
> no "kill -9" for intance....
> 
> To be sure, before you bounce servers try either waiting 20 minutes
> after the indexing stops or issue a manual commit before shutting
> down your servers with
> http://..../solr/collection/update?commit=true
> 
> I have a personal annoyance with the bin/solr script where it forcefully
> (ungracefully) kills Solr after 5 seconds. I think this is much too short
> so you might consider making it longer in prod, it's a shell script so
> it's easy.
> 
> <autoCommit>
> <maxTime>${solr.autoCommit.maxTime:1200000}</maxTime>
> <maxDocs>${solr.autoCommit.maxDocs:1000000000}</maxDocs>
> <openSearcher>false</openSearcher>
> </autoCommit>
> 
> 
> this is probably the  crux of "shards being out of sync". They're _not_
> out of sync, it's just that some of them have docs visible to searches
> and some do not since the wall-clock time these are triggered are
> _not_ the same. So you have a 10 minute window where two or more
> replicas for a single shard are out-of-sync.
> 
> 
> <autoSoftCommit>
> <maxTime>${solr.autoSoftCommit.maxTime:600000}</maxTime>
> </autoSoftCommit>
> 
> You can test all this one of two ways:
> 1> if you have a timestamp when the docs were indexed, do all
> the shards match if you do a query like
> q=*:*&timestamp:[* TO NOW/-15MINUTES]?
> or, if indexing is _not_ occurring, issue a manual commit like
> .../solr/collection/update?commit=true
> and see if all the replicas match for each shard.
> 
> Here's a long blog on commits:
> https://lucidworks.com/blog/2013/08/23/understanding-transaction-logs-softcommit-and-commit-in-sorlcloud/
> 
> Best,
> Erick
> 
> On Tue, May 17, 2016 at 8:18 AM, Stephen Weiss 
> <steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>> wrote:
> > Yes, after startup there was a recovery process, you are right.  It's just 
> > that this process doesn't seem to happen unless we do a full restart.
> >
> > These are our autocommit settings - to be honest, we did not really use 
> > autocommit until we switched up to SolrCloud so it's totally possible they 
> > are not very good settings.  We wanted to minimize the frequency of commits 
> > because the commits seem to create a performance drag during indexing.   
> > Perhaps it's gone overboard?
> >
> > <autoCommit>
> > <maxTime>${solr.autoCommit.maxTime:1200000}</maxTime>
> > <maxDocs>${solr.autoCommit.maxDocs:1000000000}</maxDocs>
> > <openSearcher>false</openSearcher>
> > </autoCommit>
> > <autoSoftCommit>
> > <maxTime>${solr.autoSoftCommit.maxTime:600000}</maxTime>
> > </autoSoftCommit>
> >
> > By nodes, I am indeed referring to machines.  There are 8 shards per 
> > machine (2 replicas of each), all in one JVM a piece.  We haven't specified 
> > any specific timestamps for the logs - they are just whatever happens by 
> > default.
> >
> > --
> > Steve
> >
> > On Mon, May 16, 2016 at 11:50 PM, Erick Erickson 
> > <erickerick...@gmail.com<mailto:erickerick...@gmail.com><mailto:erickerick...@gmail.com<mailto:erickerick...@gmail.com>>>
> >  wrote:
> > OK, this is very strange. There's no _good_ reason that
> > restarting the servers should make a difference. The fact
> > that it took 1/2 hour leads me to believe, though, that your
> > shards are somehow "incomplete", especially that you
> > are indexing to the system and don't have, say,
> > your autocommit settings done very well. The long startup
> > implies (guessing) that you have pretty big tlogs that
> > are replayed upon startup. While these were coming up,
> > did you see any of the shards in the "recovering" state? That's
> > the only way I can imagine that Solr "healed" itself.
> >
> > I've got to point back to the Solr logs. Are they showing
> > any anomalies? Are any nodes in recovery when you restart?
> >
> > Best,
> > Erick
> >
> >
> >
> > On Mon, May 16, 2016 at 4:14 PM, Stephen Weiss 
> > <steve.we...@wgsn.com<mailto:steve.we...@wgsn.com><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>>>
> >  wrote:
> >> Just one more note - while experimenting, I found that if I stopped all 
> >> nodes (full cluster shutdown), and then startup all nodes, they do in fact 
> >> seem to repair themselves then.  We have a script to monitor the 
> >> differences between replicas (just looking at numDocs) and before the full 
> >> shutdown / restart, we had:
> >>
> >> wks53104:Downloads sweiss$ php testReplication.php
> >> Found 32 mismatched shard counts.
> >> instock_shard1   replica 1: 30785553 replica 2: 30777568
> >> instock_shard10   replica 1: 30972662 replica 2: 30966215
> >> instock_shard11   replica 2: 31036718 replica 1: 31033547
> >> instock_shard12   replica 1: 30179823 replica 2: 30176067
> >> instock_shard13   replica 2: 30604638 replica 1: 30599219
> >> instock_shard14   replica 2: 30755117 replica 1: 30753469
> >> instock_shard15   replica 2: 30891325 replica 1: 30888771
> >> instock_shard16   replica 1: 30818260 replica 2: 30811728
> >> instock_shard17   replica 1: 30422080 replica 2: 30414666
> >> instock_shard18   replica 2: 30874530 replica 1: 30869977
> >> instock_shard19   replica 2: 30917008 replica 1: 30913715
> >> instock_shard2   replica 1: 31062073 replica 2: 31057583
> >> instock_shard20   replica 1: 30188774 replica 2: 30186565
> >> instock_shard21   replica 2: 30789012 replica 1: 30784160
> >> instock_shard22   replica 2: 30820473 replica 1: 30814822
> >> instock_shard23   replica 2: 30552105 replica 1: 30545802
> >> instock_shard24   replica 1: 30973906 replica 2: 30971314
> >> instock_shard25   replica 1: 30732287 replica 2: 30724988
> >> instock_shard26   replica 1: 31465543 replica 2: 31463414
> >> instock_shard27   replica 2: 30845514 replica 1: 30842665
> >> instock_shard28   replica 2: 30549151 replica 1: 30543070
> >> instock_shard29   replica 2: 30635711 replica 1: 30629240
> >> instock_shard3   replica 1: 30930400 replica 2: 30928438
> >> instock_shard30   replica 2: 30902221 replica 1: 30895176
> >> instock_shard31   replica 2: 31174246 replica 1: 31169998
> >> instock_shard32   replica 2: 30931550 replica 1: 30926256
> >> instock_shard4   replica 2: 30755525 replica 1: 30748922
> >> instock_shard5   replica 2: 31006601 replica 1: 30994316
> >> instock_shard6   replica 2: 31006531 replica 1: 31003444
> >> instock_shard7   replica 2: 30737098 replica 1: 30727509
> >> instock_shard8   replica 2: 30619869 replica 1: 30609084
> >> instock_shard9   replica 1: 31067833 replica 2: 31061238
> >>
> >>
> >> This stayed consistent for several hours.
> >>
> >> After restart:
> >>
> >> wks53104:Downloads sweiss$ php testReplication.php
> >> Found 3 mismatched shard counts.
> >> instock_shard19   replica 2: 30917008 replica 1: 30913715
> >> instock_shard22   replica 2: 30820473 replica 1: 30814822
> >> instock_shard26   replica 1: 31465543 replica 2: 31463414
> >> wks53104:Downloads sweiss$ php testReplication.php
> >> Found 2 mismatched shard counts.
> >> instock_shard19   replica 2: 30917008 replica 1: 30913715
> >> instock_shard26   replica 1: 31465543 replica 2: 31463414
> >> wks53104:Downloads sweiss$ php testReplication.php
> >> Everything looks peachy
> >>
> >> Took about a half hour to get there.
> >>
> >> Maybe the question should be - any way to get solrcloud to trigger this 
> >> *without* having to shut down / restart all nodes?  Even if we had to 
> >> trigger that manually after indexing, it would be fine.  It's a very 
> >> controlled indexing workflow that only happens once a day.
> >>
> >> --
> >> Steve
> >>
> >> On Mon, May 16, 2016 at 6:52 PM, Stephen Weiss 
> >> <steve.we...@wgsn.com<mailto:steve.we...@wgsn.com><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>>>>
> >>  wrote:
> >> Each node has one JVM with 16GB of RAM.  Are you suggesting we would put 
> >> each shard into a separate JVM (something like 32 nodes)?
> >>
> >> We aren't encountering any OOMs.  We are testing this in a separate cloud 
> >> which no one is even using, the only activity is this very small amount of 
> >> indexing and still we see this problem.  In the logs, there are no errors 
> >> at all.  It's almost like none of the recovery features that people say 
> >> are in Solr, are actually there at all.  I can't find any evidence that 
> >> Solr is even attempting to keep the shards together.
> >>
> >> There are no real errors in the solr log.  I do see some warnings at 
> >> system startup:
> >>
> >> http://pastie.org/private/thz0fbzcxgdreeeune8w
> >>
> >> These lines in particular look interesting:
> >>
> >> 16925 INFO  
> >> (recoveryExecutor-3-thread-4-processing-n:172.20.140.173:8983_solr 
> >> x:instock_shard15_replica1 s:shard15 c:instock r:core_node31) [c:instock 
> >> s:shard15 r:core_node31 x:instock_shard15_replica1] o.a.s.u.PeerSync 
> >> PeerSync: core=instock_shard15_replica1 
> >> url=http://172.20.140.173:8983/solr  Received 0 versions from 
> >> http://172.20.140.172:8983/solr/instock_shard15_replica2/ 
> >> fingerprint:{maxVersionSpecified=9223372036854775807, 
> >> maxVersionEncountered=1534492620385943552, maxInHash=1534492620385943552, 
> >> versionsHash=-6845461210912808581, numVersions=30888332, numDocs=30888332, 
> >> maxDoc=37699007}
> >> 16925 INFO  
> >> (recoveryExecutor-3-thread-4-processing-n:172.20.140.173:8983_solr 
> >> x:instock_shard15_replica1 s:shard15 c:instock r:core_node31) [c:instock 
> >> s:shard15 r:core_node31 x:instock_shard15_replica1] o.a.s.u.PeerSync 
> >> PeerSync: core=instock_shard15_replica1 
> >> url=http://172.20.140.173:8983/solr DONE. sync failed
> >> 16925 INFO  
> >> (recoveryExecutor-3-thread-4-processing-n:172.20.140.173:8983_solr 
> >> x:instock_shard15_replica1 s:shard15 c:instock r:core_node31) [c:instock 
> >> s:shard15 r:core_node31 x:instock_shard15_replica1] 
> >> o.a.s.c.RecoveryStrategy PeerSync Recovery was not successful - trying 
> >> replication.
> >>
> >> This is the first node to start up, so most of the other shards are not 
> >> there yet.
> >>
> >> On another node (the last node to start up), it looks similar but a little 
> >> different:
> >>
> >> http://pastie.org/private/xjw0ruljcurdt4xpzqk6da
> >>
> >> 74090 INFO  
> >> (recoveryExecutor-3-thread-1-processing-n:172.20.140.177:8983_solr 
> >> x:instock_shard25_replica2 s:shard25 c:instock r:core_node60) [c:instock 
> >> s:shard25 r:core_node60 x:instock_shard25_replica2] 
> >> o.a.s.c.RecoveryStrategy Attempting to PeerSync from 
> >> [http://172.20.140.170:8983/solr/instock_shard25_replica1/] - 
> >> recoveringAfterStartup=[true]
> >> 74091 INFO  
> >> (recoveryExecutor-3-thread-1-processing-n:172.20.140.177:8983_solr 
> >> x:instock_shard25_replica2 s:shard25 c:instock r:core_node60) [c:instock 
> >> s:shard25 r:core_node60 x:instock_shard25_replica2] o.a.s.u.PeerSync 
> >> PeerSync: core=instock_shard25_replica2 
> >> url=http://172.20.140.177:8983/solr START 
> >> replicas=[http://172.20.140.170:8983/solr/instock_shard25_replica1/] 
> >> nUpdates=100
> >> 74091 WARN  
> >> (recoveryExecutor-3-thread-1-processing-n:172.20.140.177:8983_solr 
> >> x:instock_shard25_replica2 s:shard25 c:instock r:core_node60) [c:instock 
> >> s:shard25 r:core_node60 x:instock_shard25_replica2] o.a.s.u.PeerSync no 
> >> frame of reference to tell if we've missed updates
> >> 74091 INFO  
> >> (recoveryExecutor-3-thread-1-processing-n:172.20.140.177:8983_solr 
> >> x:instock_shard25_replica2 s:shard25 c:instock r:core_node60) [c:instock 
> >> s:shard25 r:core_node60 x:instock_shard25_replica2] 
> >> o.a.s.c.RecoveryStrategy PeerSync Recovery was not successful - trying 
> >> replication.
> >>
> >> Every single replica shows errors like this (either one or the other).
> >>
> >> I should add, beyond the block joins / nested children & grandchildren, 
> >> there's really nothing unusual about this cloud at all.  It's a very basic 
> >> collection (simple enough it can be created in the GUI) and a dist 
> >> installation of Solr 6.  There are 3 independent zookeeper servers (again, 
> >> vanilla from dist), and there don't appear to be any zookeeper issues.
> >>
> >> --
> >> Steve
> >>
> >> On Mon, May 16, 2016 at 12:02 PM, Erick Erickson 
> >> <erickerick...@gmail.com<mailto:erickerick...@gmail.com><mailto:erickerick...@gmail.com<mailto:erickerick...@gmail.com>><mailto:erickerick...@gmail.com<mailto:erickerick...@gmail.com><mailto:erickerick...@gmail.com<mailto:erickerick...@gmail.com>>>>
> >>  wrote:
> >> 8 nodes, 4 shards apiece? All in the same JVM? People have gotten by
> >> the GC pain by running in separate JVMs with less Java memory each on
> >> big beefy machines.... That's not a recommendation as much as an
> >> observation.
> >>
> >> That aside, unless you have some very strange stuff going on this is
> >> totally weird. Are you hitting OOM errors at any time you have this
> >> problem? Once you hit an OOM error, all bets are off about how Java
> >> behaves. If you are hitting those, you can't hope for stability until
> >> you fix that issue. In your writeup there's some evidence for this
> >> when you say that if you index multiple docs at a time you get
> >> failures.
> >>
> >> Do your Solr logs show any anomalies? My guess is that you'll see
> >> exceptions in your Solr logs that will shed light on the issue.
> >>
> >> Best,
> >> Erick
> >>
> >> On Mon, May 16, 2016 at 8:03 AM, Stephen Weiss 
> >> <steve.we...@wgsn.com<mailto:steve.we...@wgsn.com><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com><mailto:steve.we...@wgsn.com<mailto:steve.we...@wgsn.com>>>>
> >>  wrote:
> >>> Hi everyone,
> >>>
> >>> I'm running into a problem with SolrCloud replicas and thought I would 
> >>> ask the list to see if anyone else has seen this / gotten past it.
> >>>
> >>> Right now, we are running with only one replica per shard.  This is 
> >>> obviously a problem because if one node goes down anywhere, the whole 
> >>> collection goes offline, and due to garbage collection issues, this 
> >>> happens about once or twice a week, causing a great deal of instability.  
> >>> If we try to increase to 2 replicas per shard, once we index new 
> >>> documents and the shards autocommit, the shards all get out of sync with 
> >>> each other, with different numbers of documents, different numbers of 
> >>> documents deleted, different facet counts - pretty much totally divergent 
> >>> indexes.  Shards always show green and available, and never go into 
> >>> recovery or any other state as to indicate there's a mismatch.  There are 
> >>> also no errors in the logs to indicate anything is going wrong.  Even 
> >>> long after indexing has finished, the replicas never come back into sync. 
> >>>  The only way to get consistency again is to delete one set of replicas 
> >>> and then add them back in.  Unfortunately, when we do this, we invariabl
 y discover that many documents (2-3%) are missing from the index.
> >>>
> >>> We have tried setting the min_rf parameter, and have found that when 
> >>> setting min_rf=2, we almost never get back rf=2.  We almost always get 
> >>> rf=1, resend the request, and it basically just goes into an infinite 
> >>> loop.  The only way to get rf=2 to come back is to only index one 
> >>> document at a time.  Unfortunately, we have to update millions of 
> >>> documents a day and it isn't really feasible to index this way, and even 
> >>> when indexing one document at a time, we still occasionally find 
> >>> ourselves in an infinite loop.  This doesn't appear to be related to the 
> >>> documents we are indexing - if we stop the index process and bounce solr, 
> >>> the exact same document will go through fine the next time until indexing 
> >>> stops up on another random document.
> >>>
> >>> We have 8 nodes, with 4 shards a piece, all running one collection with 
> >>> about 900M documents.  An important note is that we have a block join 
> >>> system with 3 tiers of documents (products -> skus -> sku_history).  
> >>> During indexing, we are forced to delete all documents for a product 
> >>> prior to adding the product back into the index, in order to avoid 
> >>> orphaned children / grandchildren.  All documents are consistently 
> >>> indexed with the top-level product ID so that we can delete all 
> >>> child/grandchild documents prior to updating the document.  So, for each 
> >>> updated document, we are sending through a delete call followed by an add 
> >>> call.  We have tried putting both the delete and add in the same update 
> >>> request with the same results.
> >>>
> >>> All we see out there on Google is that none of what we're seeing should 
> >>> be happening.
> >>>
> >>> We are currently running Solr 6.0 with Zookeeper 3.4.6.  We experienced 
> >>> the same behavior on 5.4 as well.
> >>>
> >>> --
> >>> Steve
> >>>
> >>> ________________________________
> >>>
> >>> WGSN is a global foresight business. Our experts provide deep insight and 
> >>> analysis of consumer, fashion and design trends. We inspire our clients 
> >>> to plan and trade their range with unparalleled confidence and accuracy. 
> >>> Together, we Create Tomorrow.
> >>>
> >>> WGSN<http://www.wgsn.com/> is part of WGSN Limited, comprising of 
> >>> market-leading products including WGSN.com<http://www.wgsn.com>, WGSN 
> >>> Lifestyle & Interiors<http://www.wgsn.com/en/lifestyle-interiors>, WGSN 
> >>> INstock<http://www.wgsninstock.com/>, WGSN 
> >>> StyleTrial<http://www.wgsn.com/en/styletrial/> and WGSN 
> >>> Mindset<http://www.wgsn.com/en/services/consultancy/>, our bespoke 
> >>> consultancy services.
> >>>
> >>> The information in or attached to this email is confidential and may be 
> >>> legally privileged. If you are not the intended recipient of this 
> >>> message, any use, disclosure, copying, distribution or any action taken 
> >>> in reliance on it is prohibited and may be unlawful. If you have received 
> >>> this message in error, please notify the sender immediately by return 
> >>> email and delete this message and any copies from your computer and 
> >>> network. WGSN does not warrant that this email and any attachments are 
> >>> free from viruses and accepts no liability for any loss resulting from 
> >>> infected email transmissions.
> >>>
> >>> WGSN reserves the right to monitor all email through its networks. Any 
> >>> views expressed may be those of the originator and not necessarily of 
> >>> WGSN. WGSN is powered by Ascential plc<http://www.ascential.com>, which 
> >>> transforms knowledge businesses to deliver exceptional performance.
> >>>
> >>> Please be advised all phone calls may be recorded for training and 
> >>> quality purposes and by accepting and/or making calls from and/or to us 
> >>> you acknowledge and agree to calls being recorded.
> >>>
> >>> WGSN Limited, Company number 4858491
> >>>
> >>> registered address:
> >>>
> >>> Ascential plc, The Prow, 1 Wilder Walk, London W1B 5AP
> >>>
> >>> WGSN Inc., tax ID 04-3851246, registered office c/o National Registered 
> >>> Agents, Inc., 160 Greentree Drive, Suite 101, Dover DE 19904, United 
> >>> States
> >>>
> >>> 4C Serviços de Informação Ltda., CNPJ/MF (Taxpayer's Register): 
> >>> 15.536.968/0001-04, Address: Avenida Cidade Jardim, 377, 7˚ andar CEP 
> >>> 01453-000, Itaim Bibi, São Paulo
> >>>
> >>> 4C Business Information Consulting (Shanghai) Co., Ltd, 富新商务信息咨询(上海)有限公司, 
> >>> registered address Unit 4810/4811, 48/F Tower 1, Grand Gateway, 1 Hong 
> >>> Qiao Road, Xuhui District, Shanghai
> >>
> >>
> >>
> >> ________________________________
> >>
> >> WGSN is a global foresight business. Our experts provide deep insight and 
> >> analysis of consumer, fashion and design trends. We inspire our clients to 
> >> plan and trade their range with unparalleled confidence and accuracy. 
> >> Together, we Create Tomorrow.
> >>
> >> WGSN<http://www.wgsn.com/> is part of WGSN Limited, comprising of 
> >> market-leading products including WGSN.com<http://www.wgsn.com>, WGSN 
> >> Lifestyle & Interiors<http://www.wgsn.com/en/lifestyle-interiors>, WGSN 
> >> INstock<http://www.wgsninstock.com/>, WGSN 
> >> StyleTrial<http://www.wgsn.com/en/styletrial/> and WGSN 
> >> Mindset<http://www.wgsn.com/en/services/consultancy/>, our bespoke 
> >> consultancy services.
> >>
> >> The information in or attached to this email is confidential and may be 
> >> legally privileged. If you are not the intended recipient of this message, 
> >> any use, disclosure, copying, distribution or any action taken in reliance 
> >> on it is prohibited and may be unlawful. If you have received this message 
> >> in error, please notify the sender immediately by return email and delete 
> >> this message and any copies from your computer and network. WGSN does not 
> >> warrant that this email and any attachments are free from viruses and 
> >> accepts no liability for any loss resulting from infected email 
> >> transmissions.
> >>
> >> WGSN reserves the right to monitor all email through its networks. Any 
> >> views expressed may be those of the originator and not necessarily of 
> >> WGSN. WGSN is powered by Ascential plc<http://www.ascential.com>, which 
> >> transforms knowledge businesses to deliver exceptional performance.
> >>
> >> Please be advised all phone calls may be recorded for training and quality 
> >> purposes and by accepting and/or making calls from and/or to us you 
> >> acknowledge and agree to calls being recorded.
> >>
> >> WGSN Limited, Company number 4858491
> >>
> >> registered address:
> >>
> >> Ascential plc, The Prow, 1 Wilder Walk, London W1B 5AP
> >>
> >> WGSN Inc., tax ID 04-3851246, registered office c/o National Registered 
> >> Agents, Inc., 160 Greentree Drive, Suite 101, Dover DE 19904, United States
> >>
> >> 4C Serviços de Informação Ltda., CNPJ/MF (Taxpayer's Register): 
> >> 15.536.968/0001-04, Address: Avenida Cidade Jardim, 377, 7˚ andar CEP 
> >> 01453-000, Itaim Bibi, São Paulo
> >>
> >> 4C Business Information Consulting (Shanghai) Co., Ltd, 富新商务信息咨询(上海)有限公司, 
> >> registered address Unit 4810/4811, 48/F Tower 1, Grand Gateway, 1 Hong 
> >> Qiao Road, Xuhui District, Shanghai
> >
> >
> > ________________________________
> >
> > WGSN is a global foresight business. Our experts provide deep insight and 
> > analysis of consumer, fashion and design trends. We inspire our clients to 
> > plan and trade their range with unparalleled confidence and accuracy. 
> > Together, we Create Tomorrow.
> >
> > WGSN<http://www.wgsn.com/> is part of WGSN Limited, comprising of 
> > market-leading products including WGSN.com<http://www.wgsn.com>, WGSN 
> > Lifestyle & Interiors<http://www.wgsn.com/en/lifestyle-interiors>, WGSN 
> > INstock<http://www.wgsninstock.com/>, WGSN 
> > StyleTrial<http://www.wgsn.com/en/styletrial/> and WGSN 
> > Mindset<http://www.wgsn.com/en/services/consultancy/>, our bespoke 
> > consultancy services.
> >
> > The information in or attached to this email is confidential and may be 
> > legally privileged. If you are not the intended recipient of this message, 
> > any use, disclosure, copying, distribution or any action taken in reliance 
> > on it is prohibited and may be unlawful. If you have received this message 
> > in error, please notify the sender immediately by return email and delete 
> > this message and any copies from your computer and network. WGSN does not 
> > warrant that this email and any attachments are free from viruses and 
> > accepts no liability for any loss resulting from infected email 
> > transmissions.
> >
> > WGSN reserves the right to monitor all email through its networks. Any 
> > views expressed may be those of the originator and not necessarily of WGSN. 
> > WGSN is powered by Ascential plc<http://www.ascential.com>, which 
> > transforms knowledge businesses to deliver exceptional performance.
> >
> > Please be advised all phone calls may be recorded for training and quality 
> > purposes and by accepting and/or making calls from and/or to us you 
> > acknowledge and agree to calls being recorded.
> >
> > WGSN Limited, Company number 4858491
> >
> > registered address:
> >
> > Ascential plc, The Prow, 1 Wilder Walk, London W1B 5AP
> >
> > WGSN Inc., tax ID 04-3851246, registered office c/o National Registered 
> > Agents, Inc., 160 Greentree Drive, Suite 101, Dover DE 19904, United States
> >
> > 4C Serviços de Informação Ltda., CNPJ/MF (Taxpayer's Register): 
> > 15.536.968/0001-04, Address: Avenida Cidade Jardim, 377, 7˚ andar CEP 
> > 01453-000, Itaim Bibi, São Paulo
> >
> > 4C Business Information Consulting (Shanghai) Co., Ltd, 富新商务信息咨询(上海)有限公司, 
> > registered address Unit 4810/4811, 48/F Tower 1, Grand Gateway, 1 Hong Qiao 
> > Road, Xuhui District, Shanghai
> 
> 
> 
> ________________________________
> 
> WGSN is a global foresight business. Our experts provide deep insight and 
> analysis of consumer, fashion and design trends. We inspire our clients to 
> plan and trade their range with unparalleled confidence and accuracy. 
> Together, we Create Tomorrow.
> 
> WGSN<http://www.wgsn.com/> is part of WGSN Limited, comprising of 
> market-leading products including WGSN.com<http://www.wgsn.com>, WGSN 
> Lifestyle & Interiors<http://www.wgsn.com/en/lifestyle-interiors>, WGSN 
> INstock<http://www.wgsninstock.com/>, WGSN 
> StyleTrial<http://www.wgsn.com/en/styletrial/> and WGSN 
> Mindset<http://www.wgsn.com/en/services/consultancy/>, our bespoke 
> consultancy services.
> 
> The information in or attached to this email is confidential and may be 
> legally privileged. If you are not the intended recipient of this message, 
> any use, disclosure, copying, distribution or any action taken in reliance on 
> it is prohibited and may be unlawful. If you have received this message in 
> error, please notify the sender immediately by return email and delete this 
> message and any copies from your computer and network. WGSN does not warrant 
> that this email and any attachments are free from viruses and accepts no 
> liability for any loss resulting from infected email transmissions.
> 
> WGSN reserves the right to monitor all email through its networks. Any views 
> expressed may be those of the originator and not necessarily of WGSN. WGSN is 
> powered by Ascential plc<http://www.ascential.com>, which transforms 
> knowledge businesses to deliver exceptional performance.
> 
> Please be advised all phone calls may be recorded for training and quality 
> purposes and by accepting and/or making calls from and/or to us you 
> acknowledge and agree to calls being recorded.
> 
> WGSN Limited, Company number 4858491
> 
> registered address:
> 
> Ascential plc, The Prow, 1 Wilder Walk, London W1B 5AP
> 
> WGSN Inc., tax ID 04-3851246, registered office c/o National Registered 
> Agents, Inc., 160 Greentree Drive, Suite 101, Dover DE 19904, United States
> 
> 4C Serviços de Informação Ltda., CNPJ/MF (Taxpayer's Register): 
> 15.536.968/0001-04, Address: Avenida Cidade Jardim, 377, 7˚ andar CEP 
> 01453-000, Itaim Bibi, São Paulo
> 
> 4C Business Information Consulting (Shanghai) Co., Ltd, 富新商务信息咨询(上海)有限公司, 
> registered address Unit 4810/4811, 48/F Tower 1, Grand Gateway, 1 Hong Qiao 
> Road, Xuhui District, Shanghai
> 

Reply via email to