> -----Original Message----- > From: Thorsten Schöning [mailto:tschoen...@am-soft.de] > Sent: 09 October 2012 10:17 > To: 'users@subversion.apache.org' > Subject: Re: how large data will affect performance? > > Guten Tag wang.yu, > am Dienstag, 9. Oktober 2012 um 03:35 schrieben Sie: > > > I have a SVN server on windows2003. > > Now the developer want check in about 10G date to it, and in the > > last few months, they will check in about 100G data. > > Will so many data affect on server’s performance? > > This depends heavily on what the data is used for after the commit. If > it is only committed once and after that never ever read or updated or > else, the commit itself will consume resources while processing it, > but afterwards the version only blocks the space it needs on your > hard drive and maybe is never accessed again. Things like dump and > load cycles and such maintenance work on the repo itself is affected, > of course.
Just a thought but... depending on what the data is, you could consider creating another repository for that data (and use an svn:external to pull whatever is required into a working copy). That might make maintenance of your exist repository(s) easier in the future. ~ mark c