Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Code Block
guc is globus-url-copy - acronym invented by Levente
srm is a software layer over that - it makes sure transfers complete

There is a –r recursive copy option;

Or else you can do a loop in shell:
set daq   = (1380001.daq 1380003.daq 1380004.daq);
foreach f ($daq)
globus-url-copy -p 25 file:/star/bla/bla/$f
gsiftp://pdsfgrid.nersc.gov/bla/bla/bla/$f
end


 If you specify a source ending in "XXX/" , it will treat XXX as a
directory and transfer all files in XXX from the XXX directory

Use -p 25 option for having more streams, but isn't 25 overdoing it? You got 1 file
cut into 25 buffer chunks and all this has to be re-assembled on arrival.
I'd try +-2 around 8.

Use different end-point machine:  dtn01.nersc.gov  to avoid STAR,ATLAS conflicts

there is also carvergrid.nersc.gov gatekeeper I have access too.

 pull from carver, you need to load the osg module ,and then you have the globus commands available

globus-job-run works like remote ssh command
Look at this page:
http://rcsg-gsir.imsb-dsgi.nrc-cnrc.gc.ca/globus_tutorial/#running
for globus-job-run, globus-job-submit and globus-job-get-output
(job is a command or script, not necessarily something that goes into
the batch system)

dtn01 is a data transfer node. It has a gsiftp server but no gatekeeper.
carvergrid has both but dtn01 has a 10Gb erthernet and carvergrid only 1Gb.
You can use dtn01 for globus-url-copy and carvergrid for globus-job... type commands.

 You can test a gridftp from stargrid node (01,02,03,04) right now to either gsiftp://dtn01.nersc.gov/global/scratch/sd/balewski/...  or to 
gsiftp://pdsfdtn1.nersc.gov/project/projectdirs/star/...

...

Code Block
Ok.  Let's do a transfer.

1st check for endpoints (assuming your globus online user name is same as on your laptop, otherwise add yourname@ in front of the cli...)

ssh -t cli.globusonline.org endpoint-list

should show nothing.  But add the -p,

ssh -t cli.globusonline.org endpoint-list -p

This should produce a public list that includes nersc#dtn & nersc#hpss.
(always just use the -t to hide any passwords you might need ...)

Activate both those endpoints,

ssh -t cli.globusonline.org endpoint-activate nersc#*

This will prompt for a pass phrase - use your nim password.

Now list again & this should show the endpoints with a time associated with it
( we can increase the time to 11 days as needed).

ssh -t cli.globusonline.org endpoint-list


Now copy 1 file from HPSS to disk.  Here's my example.

echo "nersc#hpss/home/p/porter/auk.tar nersc#dtn/project/projectdirs/star/porter/auk.tar"
| ssh -t cli.globusonline.org transfer

It will return a task id and you can query that specifically via

ssh -t cli.globusonline.org details your-task-id

If this all works, then try to sync a directory tree from a source to a destination
- here is my example from disk to hpss

echo "nersc#dtn/global/scratch/sd/porter/reco/2007ProductionMinBias/
nersc#hpss/home/p/porter/ -r -s 1 " | ssh -t cli.globusonline.org transfer

You need the slashes at the end of the directory names.  The 's 1' means use filesize to
determine success.

Let me know how that goes.   The same syntax could be used to copy files/directories between
 rcf and nersc once the rcf endpoint is configured.

Jeff

Jeff2:

Code Block

Let's do a transfer.

1st check for endpoints (assuming your globus online user name is same as on your laptop, otherwise add yourname@ in front of the cli...)

ssh -t cli.globusonline.org endpoint-list

should show nothing.  But add the -p,

ssh -t cli.globusonline.org endpoint-list -p

This should produce a public list that includes nersc#dtn & nersc#hpss. (always just use the -t to hide any passwords you might need ...)

Activate both those endpoints,

ssh -t cli.globusonline.org endpoint-activate nersc#*

This will prompt for a pass phrase - use your nim password.

Now list again & this should show the endpoints with a time associated with it 
( we can increase the time to 11 days as needed).

ssh -t cli.globusonline.org endpoint-list


Now copy 1 file from HPSS to disk.  Here's my example.

echo "nersc#hpss/home/p/porter/auk.tar nersc#dtn/project/projectdirs/star/porter/auk.tar" 
| ssh -t cli.globusonline.org transfer

It will return a task id and you can query that specifically via 

ssh -t cli.globusonline.org details your-task-id

If this all works, then try to sync a directory tree from a source to a destination 
- here is my example from disk to hpss

echo "nersc#dtn/global/scratch/sd/porter/reco/2007ProductionMinBias/ 
nersc#hpss/home/p/porter/ -r -s 1 " | ssh -t cli.globusonline.org transfer

You need the slashes at the end of the directory names.  The 's 1' means use filesize to 
determine success.  

Let me know how that goes.   The same syntax could be used to copy files/directories between
 rcf and nersc once the rcf endpoint is configured.

Jeff