On 9/4/2015 9:33 AM, Ana Emília M. Arruda wrote:
Hello Stefan,
Sorry for my late reply.
If you want to run jobs concurrently using the same pool and only one
drive (your FileStorage device), your jobs should be configured to use
this only pool. They will use only one volume in this pool since you
have only one drive.
From your first post, the problem was not about bacula needing to
create another volume, but that you had only one drive and your jobs
were asking for volumes in two different pools.
If you want to run concurrent jobs writing in different volumes, I
would recommend you to configure a virtual autochanger
(http://blog.bacula.org/whitepapers/CommunityDiskBackup.pdf).
I think in many cases there is little point in using concurrent jobs.
Unless all clients are on a 10G network, they are bandwidth limited at
125 MB/s, well within the capability of a single storage device. Unless
concurrent jobs are going to reduce the total backup time for N clients,
it is simpler to back them up sequentially to a single storage device.
The exception would be if you have many slow clients that can only
sustain around 60 MB/s.
About the job that you had problems to cancel, I suppose, because of
my delay in replying to you, that you had already solved it. But if
you would like to confirm if this is a bug or not you can run your
daemons in debug mode: start your daemon(s) in debug mode (./bacula
start -d100, for example. The higher this value, more output will be
produced) or configure debug mode using the "setdebug" bconsole command.
Best regards,
Ana
On Tue, Sep 1, 2015 at 3:54 AM, Stefan Michael Guenther
<[email protected] <mailto:[email protected]>> wrote:
Hi Ana,
> Both of your jobs (BackupCatalog and Backup job AB01) are trying
to use device
> FileStorage. Is this your only device? If so, you can have
concurrent jobs
> writing in the same mounted volume in this device. You cannot
have more than
> one volume mounted in one specific drive. Also, you no not have
"Label Format"
> configured for your pools. This way bacula will not be able to
automatic label
> your volumes. You will need to do this manually.
>
well now the pool and storage definitions look like this:
Pool {
Name = File
Pool Type = Backup
Recycle = yes
AutoPrune = yes
Volume Retention = 365 days
Maximum Volume Bytes = 50G
Maximum Volumes = 10
LabelFormat = File
}
Device {
Name = FileStorage
Media Type = File
Archive Device = /media/cifs/bacula
LabelMedia = yes
Random Access = Yes;
AutomaticMount = yes;
RemovableMedia = no;
AlwaysOpen = yes;
}
But the question is, why does Bacula need to create another
volume? There is an appendable and free volume for this pool
available?
Pool: File
+---------+------------+-----------+---------+----------+----------+--------------+---------+------+-----------+-----------+---------------------+
| MediaId | VolumeName | VolStatus | Enabled | VolBytes | VolFiles
| VolRetention | Recycle | Slot | InChanger | MediaType |
LastWritten |
+---------+------------+-----------+---------+----------+----------+--------------+---------+------+-----------+-----------+---------------------+
| 1 | File001 | Append | 1 | 50,844 | 0
| 31,536,000 | 1 | 0 | 0 | File |
2015-08-30 23:10:02 |
| 7 | File002 | Append | 1 | 191 | 0
| 31,536,000 | 1 | 0 | 0 | File |
0000-00-00 00:00:00 |
+---------+------------+-----------+---------+----------+----------+--------------+---------+------+-----------+-----------+---------------------+
Another strange thing:
According to "status storage", SD is blocked by a job:
Running Jobs:
Writing: Full Backup job Server JobId=13 Volume=""
pool="File" device="FileStorage" (/media/cifs/bacula)
spooling=0 despooling=0 despool_wait=0
Files=0 Bytes=0 Bytes/sec=0
FDReadSeqNo=6 in_msg=6 out_msg=4 fd=6
====
But when I try to cancel that job, Bacula tells me, that it
doesn't exist:
*cancel
Keine Aufgaben laufen.
There has been a job with that ID, but it has already failed:
list jobs
.....
| 13 | IP_PLS | 2015-09-01 03:00:01 | B | I |
0 | 0 | f |
....
Thanks again for your help,
Stefan
------------------------------------------------------------------------------
_______________________________________________
Bacula-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bacula-users
------------------------------------------------------------------------------
_______________________________________________
Bacula-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bacula-users