(Illustration by Gaich Muramatsu)
Hello , I solved my problem by restarting from the scratch again. The stops are which may be important for others: - stop coda on all servers - erase all partitions needed for replications - vice-setup on all servers - Edit file '/vice/db/servers' to include my coda servers with the corresponding ID: (example: tarzan1 1 tarzan2 2 ) - Edit file '/usr/local/etc/coda/server.conf' to define hostname and ipaddress as in /etc/hosts - Make sure that directory '/var/lock/subsys' is existing - Activate 'vice-setup-srvdir' to set up additional partitions (/vicepb, /vicepc in my case, beside /vicepa) File '/vice/db/vicetab' should have entries like: tarzan1 /vicepa ftree width=32,depth=4 tarzan1 /vicepb ftree width=32,depth=4 tarzan1 /vicepc ftree width=256,depth=3 on first coda server - Restart coda server in all nodes. Please check /vice/srv/SrvLog!!!!! - Setup replicating partitions by activing the following scripts on SCM: createvol_rep iersdc tarzan1/vicepb tarzan2/vicepb createvol_rep iers2 tarzan1/vicepc tarzan2/vicepc Now it did work. Regards, Reiner > Hello Jan, > > thanks for your response. > > Well, I made sure that > 'vicepb' partition is setup (using /usr/local/sbin/vice-setup-srvdir) > > I restartet the server and now I found in > /vice/srv/SrvLog thet the codaserver is not running: > > Date: Fri 01/26/2007 > > 09:36:39 Coda Vice, version 6.9.0 log started at Fri Jan 26 > 09:36:39 2007 > > 09:36:39 RvmType is Rvm > 09:36:39 Main process doing a LWP_Init() > 09:36:39 Main thread just did a RVM_SET_THREAD_DATA > > 09:36:39 Setting Rvm Truncate threshhold to 5. > > log_recover failed. > do_rvm_options failed > 09:36:39 rvm_init failed RVM_EIO > > What does this mean? > well codaserver was running after first setup, i.e. where only partition > 'vicepa' > was defined. > > Best regards, Reiner > > Harkes wrote: > >> On Thu, Jan 25, 2007 at 05:51:04PM +0100, Reiner Dassing wrote: >> >>> a new setup of coda-6.9.0 on Debian Linux did allow to createvol_rep >>> the default /vicepa but the command >>> >>> "createvol_rep iersdc tarzan1/vicepb tarzan2/vicepb" >>> shows: "Failed to dump the current VRDB into /vice/db/VRList.new" >> >> >> >> Did you set up /vice/db/vicetab to have a valid entry for /vicepb? >> Also, you may have to restart the servers before the new partition is >> picked up from that file. >> >> >>> + volutil -h tarzan1 dumpvrdb /vice/db/VRList.new >>> + '[' 255 -ne 0 ']' >>> + echo 'Failed to dump the current VRDB into /vice/db/VRList.new' >>> Failed to dump the current VRDB into /vice/db/VRList.new >>> + exit 1 >> >> >> >> If you run that volutil command from the prompt does it give an >> explanation for the error? In the script we redirected stderr to >> /dev/null to avoid too much clutter. >> >> >>> # more /vice/vol/BigVolumeList >>> P/vicepa Htarzan1 T8b3f8c F8b3f78 >>> W/.0 I1000001 H1 P/vicepa m0 M0 U2 W1000001 C45b88f7a D45b88f7a B0 A0 >> >> >> >> This does look like we don't know about any vicepb 'partitions' on >> either tarzan1 or tarzan2, as there would be a P/vicepb line in the >> volume list. But maybe the script never got far enough to update this >> information. >> >> Jan >> >>Received on 2007-01-26 10:29:50