I think this is something that can be easily madded with some shell scripting, i suppose you are using Linux or a Unix derivate.
In this case tar + bzip2 is your friend, altough for the volume of files i sugest looking at something like rsync or unison and coupling it with cron for automating the online backup. Otherwise i suggest looking at glob module for listing directories and files and the bz2 module, there is also ziplib module, check the docs: http://docs.python.org/lib/someos.html As for ftp, there is ftplib: http://docs.python.org/lib/module-ftplib.html Good luck, regards Carlos Daniel Ruvalcaba On 6/27/06, Magnus Wirström <[EMAIL PROTECTED]> wrote: > Hi Everyone > > I'm starting to python and i need to write a program that are able to > compress a high amount of files and directories into a single file that > is later going to be transfered with ftp to a backup storage server. > The data is a quite big amount (over 1.5 gb in 40000 files and 1300 > directories) so i would like to have high compression like RAR or bz2. > What way is the best to approach this? I have looked a bit on zlib but i > didn't find a good example that uses it to compress files and > directories in one file from python. or perhaps it's simpler to execute > a external program, doing the work. I would prefer to use python as much > as i can without external apps. > > Does anyone have a idea how this can be made with python. > > Thanks > Magnus Wirström > _______________________________________________ > Tutor maillist - Tutor@python.org > http://mail.python.org/mailman/listinfo/tutor > _______________________________________________ Tutor maillist - Tutor@python.org http://mail.python.org/mailman/listinfo/tutor