Skip to content

Commit 24c5900

Browse files
committed
Merge Split Branch
2 parents 0c35a9b + bfa20cc commit 24c5900

File tree

11 files changed

+75
-60
lines changed

11 files changed

+75
-60
lines changed

preview-staging-transition/assets/search/index.json

Lines changed: 1 addition & 1 deletion
Large diffs are not rendered by default.

preview-staging-transition/assets/search/metadata.json

Lines changed: 1 addition & 1 deletion
Large diffs are not rendered by default.

preview-staging-transition/fellowships/index.html

Lines changed: 18 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -209,14 +209,14 @@ <h1 class="uw-mini-bar mb-4">
209209

210210
<h2 id="featured-fellow">Featured Fellow</h2>
211211

212-
<div id="wil-cram" class="bg-white team-card justify-content-center mb-3 flex-grow-1">
212+
<div id="william-jiang" class="bg-white team-card justify-content-center mb-3 flex-grow-1">
213213
<div class="uw-card-content shadow">
214214
<div class="row">
215215
<div class="col-3">
216216
<div class="overflow-hidden d-flex justify-content-center">
217217

218218

219-
<img src="/web-preview/preview-staging-transition/staff-list/images/wil_cram.jpg" alt="Headshot for Wil Cram" />
219+
<img src="/web-preview/preview-staging-transition/staff-list/images/william_jiang.png" alt="Headshot for William Jiang" />
220220

221221
</div>
222222
</div>
@@ -225,27 +225,32 @@ <h2 id="featured-fellow">Featured Fellow</h2>
225225
<div>
226226
<h2 class="uw-mini-bar mb-0 mt-3">
227227

228-
Wil Cram
228+
William Jiang
229229

230230
</h2>
231-
<div><h6 class="mt-2">Mentor(s):</h6> Greg Thain</div>
231+
<div><h6 class="mt-2">Mentor(s):</h6> Emma Turetsky and Justin Hiemstra</div>
232232
</div>
233233
</div>
234234
</div>
235235
<div class="col-12">
236-
<h4 class="mt-3">Schedd performance analysis for human</h4>
236+
<h4 class="mt-3">Integrating Pelican and Globus</h4>
237237
<div class="pt-2">
238-
<p>The condor_schedd is a single threaded program, and when it is overloaded,
239-
it is difficult for administrators to understand why. There are some
240-
statistics about what it is doing, but there is no clear way to present
241-
this information in a useful way to an administrator. Students working
242-
on this project would build visualizations of complex data, and work
243-
with end users and facilitators to tune output for real world human
244-
consumption.</p>
238+
<p>The Pelican Platform is an open-source software platform designed to create
239+
data federations by unifying diverse data repositories (e.g. POSIX
240+
filesystems, S3) under a single architecture. It differs from the Globus
241+
project, whose main goal is to facilitate peer-to-peer data transfer, and it
242+
thus lacks the means to plug into cluster scheduling frameworks like HTCondor.</p>
243+
244+
<p>This project focused on bridging the gap between Pelican and Globus by
245+
integrating Globus transfer capabilities into Pelican. This will allow users
246+
to leverage Globus’s robust data transfer features while benefiting from
247+
Pelican’s unified data access model. The integration involved developing a
248+
Globus backend for Pelican and enabling seamless data transfers between
249+
Pelican-managed repositories and external systems via Globus.</p>
245250

246251
</div>
247252
<div>
248-
<a class="btn btn-primary" href="./reports/2024/wil-cram">Read Final Report</a>
253+
<a class="btn btn-primary" href="./reports/2024/william-jiang">Read Final Report</a>
249254
</div>
250255
</div>
251256
</div>

preview-staging-transition/uw-research-computing/apptainer-htc.html

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -360,7 +360,7 @@ <h4 id="5-move-the-container-sif-file-to-staging-">5. Move the container .sif fi
360360

361361
<p>Once you are satisfied that your container is built correctly, copy your <code>.sif</code> file to your staging directory.</p>
362362

363-
<pre class="term"><code>mv my-container.sif /staging/$USER
363+
<pre class="term"><code>mv my-container.sif /staging/u/username
364364
</code></pre>
365365

366366
<p>Once the file has transferred, exit the interactive job with</p>
@@ -530,7 +530,7 @@ <h2 id="move-the-container-sif-file-to-staging">Move the Container .sif File to
530530
<p>Since Apptainer <code>.sif</code> files are routinely more than 1GB in size, we recommend that you transfer <code>my-container.sif</code> to your <code>/staging</code> directory.
531531
It is usually easiest to move the container file directly to staging while still in the interactive build job:</p>
532532

533-
<pre class="term"><code>mv my-container.sif /staging/$USER
533+
<pre class="term"><code>mv my-container.sif /staging/u/username
534534
</code></pre>
535535

536536
<p>If you do not have a <code>/staging</code> directory, you can skip this step and the <code>.sif</code> file will be automatically transferred back to the login server when you exit the interactive job.

preview-staging-transition/uw-research-computing/check-quota.html

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -310,9 +310,9 @@ <h2 id="check-your-quota">Check your quota</h2>
310310
<p>This will print a table with your <code>/home</code> and <code>/staging</code> quotas. An example output is shown below.</p>
311311

312312
<pre class="term"><code>[user@ap2002 ~]$ get_quotas
313-
Path Disk_Used(GB) Disk_Limit(GB) Files_Used File_Limit
314-
/home/user 16.0711 40 8039 N/A
315-
/staging/user 13.4731 100 12 1000
313+
Path Disk_Used(GB) Disk_Limit(GB) Files_Used File_Limit
314+
/home/username 16.0711 40 8039 N/A
315+
/staging/u/username 13.4731 100 12 1000
316316
</code></pre>
317317

318318
<p>To print the quota for any path in <code>/staging</code>, <code>/software</code>, and <code>/projects</code> that you have access to, (i.e., a group directory), use the <code>-p</code> option:</p>
@@ -368,10 +368,10 @@ <h3 id="option-2-check-your-message-of-the-day-motd">Option 2: Check your Messag
368368
== NOTICE: THIS NODE IS ON PUPPET ENVIRONMENT "puppet8" ==
369369

370370
Filesystem quota report (last updated 10:33 AM)
371-
Storage Used (GB) Limit (GB) Files (#) File Cap (#) Quota (%)
372-
------------------ ----------- ------------ ----------- -------------- -----------
373-
/home/user 29.38 40 94 0 73.46
374-
/staging/user 50.23 1000 110 10000 5.02
371+
Storage Used (GB) Limit (GB) Files (#) File Cap (#) Quota (%)
372+
------------------ ----------- ------------ ----------- -------------- -----------
373+
/home/username 29.38 40 94 0 73.46
374+
/staging/u/username 50.23 1000 110 10000 5.02
375375
</code></pre>
376376

377377
<p>However, we recommend <a href="#check-your-quota">using the <code>get_quotas</code> command</a> for the most up-to-date information.</p>

preview-staging-transition/uw-research-computing/file-avail-largedata.html

Lines changed: 31 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -338,7 +338,7 @@ <h2>Table of Contents</h2>
338338
</li>
339339
<li><a href="#stage-large-data">Stage large data</a>
340340
<ul>
341-
<li><a href="#request-a-staging-directory">Request a <code>/staging</code> directory</a></li>
341+
<li><a href="#your-personal-staging-directory">Your personal <code>/staging</code> directory</a></li>
342342
<li><a href="#reduce-file-counts">Reduce file counts</a></li>
343343
<li><a href="#use-the-transfer-server">Use the transfer server</a></li>
344344
<li><a href="#remove-files-after-jobs-complete">Remove files after jobs complete</a></li>
@@ -391,19 +391,39 @@ <h2 id="stage-large-data">Stage large data</h2>
391391
<p>In order to stage large data for use on CHTC’s HTC system, users must:</p>
392392

393393
<ol>
394-
<li><strong>Request a <code>/staging</code> directory</strong>: Use our quota request form.</li>
395394
<li><strong>Reduce file counts</strong>: Combine and compress files that are used together.</li>
396395
<li><strong>Transfer files to the HTC system via the transfer server</strong>: Upload your data via our dedicated file transfer server.</li>
397396
<li><strong>Remove files after jobs complete</strong>: Our data staging space is quota-controlled and not backed up.</li>
398397
</ol>
399398

400-
<h3 id="request-a-staging-directory">Request a <code>/staging</code> directory</h3>
399+
<h3 id="your-personal-staging-directory">Your personal <code>/staging</code> directory</h3>
401400

402-
<p>Any one with a CHTC account whose data meets the intended use above can request space in our large data staging area by filling out a quota request form. The default quota is 100 GB / 1000 items; if a larger quota is needed, request a higher quota. The created directory will exist at this path: <code>/staging/username</code></p>
401+
<p>Each user should have a personal <code>/staging</code> directory. The created directory will exist in an alphabetized subdirectory based on the <strong>first letter</strong> of your NetID. For example:</p>
402+
403+
<table>
404+
<thead>
405+
<tr>
406+
<th>NetID</th>
407+
<th>Path to your personal <code>/staging</code> directory</th>
408+
</tr>
409+
</thead>
410+
<tbody>
411+
<tr>
412+
<td><code>alice</code></td>
413+
<td><code>/staging/a/alice</code></td>
414+
</tr>
415+
<tr>
416+
<td><code>bucky</code></td>
417+
<td><code>/staging/b/bucky</code></td>
418+
</tr>
419+
</tbody>
420+
</table>
403421

404422
<p>We can also create group or shared spaces by request.</p>
405423

406-
<p style="text-align: center; margin-bottom: 0; font-weight: bold;">Need a <code>/staging</code> directory or higher quota?</p>
424+
<p>The default quota is 100 GB / 1000 items; if a larger quota is needed, request a higher quota.</p>
425+
426+
<p style="text-align: center; margin-bottom: 0; font-weight: bold;">Need a group <code>/staging</code> directory or higher quota?</p>
407427
<div class="d-flex mb-3">
408428
<div class="p-3 m-auto">
409429
<a class="btn btn-primary" style="text-align: center" href="quota-request">Quota request form</a>
@@ -412,7 +432,7 @@ <h3 id="request-a-staging-directory">Request a <code>/staging</code> directory</
412432

413433
<h3 id="reduce-file-counts">Reduce file counts</h3>
414434

415-
<p>The file system backing our <code>/staging</code>space is optimized to handle small numbers of large files. If your job requires many small files, we recommend placing these files in the <code>/home</code> directory or compressing multiple files into a single zip file or tarball. See <a href="htc-job-file-transfer#data-storage-locations">this table</a> for more information on the differences between <code>/staging</code> and <code>/home</code>.</p>
435+
<p>The file system backing our <code>/staging</code> space is optimized to handle small numbers of large files. If your job requires many small files, we recommend placing these files in the <code>/home</code> directory or compressing multiple files into a single zip file or tarball. See <a href="htc-job-file-transfer#data-storage-locations">this table</a> for more information on the differences between <code>/staging</code> and <code>/home</code>.</p>
416436

417437
<p>Data placed in our large data <code>/staging</code> location should be stored in as few files as possible (ideally, one file per job), and will be used by a job only after being copied from <code>/staging</code> into the job working directory. Similarly, large output should first be written to the job’s working directory then compressed in to a single file before being copied to <code>/staging</code> at the end of the job.</p>
418438

@@ -426,7 +446,7 @@ <h3 id="use-the-transfer-server">Use the transfer server</h3>
426446
<p>Uploading or downloading data to <code>/staging</code> should only be performed via CHTC’s transfer server, <code>transfer.chtc.wisc.edu</code>, and <strong>not via a CHTC Access Point.</strong></p>
427447

428448
<p>For example, you can use <code>scp</code> to transfer files into your <code>/staging</code> directory:</p>
429-
<pre class="term"><code>$ scp large.file netid@transfer.chtc.wisc.edu:/staging/netid/
449+
<pre class="term"><code>$ scp large.file username@transfer.chtc.wisc.edu:/staging/u/username/
430450
</code></pre>
431451

432452
<h3 id="remove-files-after-jobs-complete">Remove files after jobs complete</h3>
@@ -442,7 +462,7 @@ <h3 id="transfer-large-input-files">Transfer large input files</h3>
442462
<p>Staged files should be specified in the job submit file using the <code>osdf:///</code> or <code>file:///</code> syntax,
443463
depending on the size of the files to be transferred. <a href="htc-job-file-transfer#transfer-input-data-to-jobs-with-transfer_input_files">See this table for more information</a>.</p>
444464

445-
<pre class="sub"><code>transfer_input_files = osdf:///chtc/staging/username/file1, file:///staging/username/file2, file3
465+
<pre class="sub"><code>transfer_input_files = osdf:///chtc/staging/u/username/file1, file:///staging/u/username/file2, file3
446466
</code></pre>
447467

448468
<h3 id="transfer-large-output-files">Transfer large output files</h3>
@@ -452,7 +472,7 @@ <h3 id="transfer-large-output-files">Transfer large output files</h3>
452472
<p>Large outputs should be transferred to staging using the same file transfer protocols in HTCondor’s <code>transfer_output_remaps</code> or <code>output_destination</code> option:</p>
453473

454474
<pre class="sub"><code>transfer_output_files = file1, file2, file3
455-
transfer_output_remaps = "file1 = osdf:///chtc/staging/username/file1; file2 = file:///staging/username/file2"
475+
transfer_output_remaps = "file1 = osdf:///chtc/staging/u/username/file1; file2 = file:///staging/u/username/file2"
456476
</code></pre>
457477

458478
<p>In the example above, <code>file1</code> and <code>file2</code> are transferred to <code>/staging</code> via two different file transfer protocols (which depends on their file size), and <code>file3</code> is transferred to the submit directory on <code>/home</code>.</p>
@@ -484,14 +504,14 @@ <h2 id="submit-jobs-using-staged-data">Submit jobs using staged data</h2>
484504

485505
<pre><code class="language-{.sub}">### Example submit file for a single job that stages large data
486506
# Files for the below lines MUST all be somewhere within /home/username,
487-
# and not within /staging/username
507+
# and not within /staging/u/username
488508

489509
executable = run_myprogram.sh
490510
log = myprogram.log
491511
output = $(Cluster).out
492512
error = $(Cluster).err
493513

494-
transfer_input_files = osdf:///chtc/staging/username/myprogram, file:///staging/username/largedata.tar.gz
514+
transfer_input_files = osdf:///chtc/staging/u/username/myprogram, file:///staging/u/username/largedata.tar.gz
495515

496516
# IMPORTANT! Require execute servers that can access /staging
497517
Requirements = (Target.HasCHTCStaging == true)

preview-staging-transition/uw-research-computing/high-memory-jobs.html

Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -436,16 +436,13 @@ <h2 id="a-submit-file">A. Submit File</h2>
436436

437437
<pre><code class="language-{.sub}">### Example submit file for a single staging-dependent job
438438

439-
universe = vanilla
440-
441439
# Files for the below lines will all be somewhere within /home/username,
442-
# and not within /staging/username
440+
# and not within /staging/u/username
443441
log = run_myprogram.log
444442
executable = run_Trinity.sh
445443
output = $(Cluster).out
446444
error = $(Cluster).err
447445
transfer_input_files = trinityrnaseq-2.0.1.tar.gz
448-
should_transfer_files = YES
449446

450447
# Require execute servers that have large data staging
451448
Requirements = (Target.HasCHTCStaging == true)
@@ -495,7 +492,7 @@ <h2 id="c-executable-script">C. "Executable" script</h2>
495492
<pre class="file"><code>#!/bin/bash
496493
# Copy input data from /staging to the present directory of the job
497494
# and un-tar/un-zip them.
498-
cp /staging/username/reads.tar.gz ./
495+
cp /staging/u/username/reads.tar.gz ./
499496
tar -xzvf reads.tar.gz
500497
rm reads.tar.gz
501498

@@ -514,7 +511,7 @@ <h2 id="c-executable-script">C. "Executable" script</h2>
514511
# Trinity will write output to the working directory by default,
515512
# so when the job finishes, it needs to be moved back to /staging
516513
tar -czvf trinity_out_dir.tar.gz trinity_out_dir
517-
cp trinity_out_dir.tar.gz trinity_stdout.txt /staging/username/
514+
cp trinity_out_dir.tar.gz trinity_stdout.txt /staging/u/username/
518515
rm reads_*.fq trinity_out_dir.tar.gz trinity_stdout.txt
519516

520517
### END

preview-staging-transition/uw-research-computing/htc-docker-to-apptainer.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -305,7 +305,7 @@ <h3 id="move-the-image-to-your-staging-directory">Move the image to your staging
305305

306306
<p>Because container images are generally large, we require users to move these images into their staging directories. While you are still in your interactive job, move the image to your staging directory.</p>
307307

308-
<pre class="term"><code>mv container.sif /staging/username/
308+
<pre class="term"><code>mv container.sif /staging/u/username/
309309
</code></pre>
310310

311311
<h2 id="use-the-apptainer-image-in-a-job">Use the Apptainer image in a job</h2>

0 commit comments

Comments
 (0)