Category Archives: grid

Protected: Grayson – Science Workflow on the Hybrid Grid

There is no excerpt because this is a protected post.

Posted in Compute Grids, condor, Engage VO, GPGPU, grid, High Throughput Computing (HTC), High Throughput Parallel Computing (HTPC), multicore, OSG, RENCI, Uncategorized | Tagged

On the Open Science Grid Trail

The open science grid is a distributed heterogeneous network of computing clusters. Its infrastructure and protocols allow members to submit high throughput compute jobs for remote execution. All use is authenticated and authorized via a PKI infrastructure which associates jobs … Continue reading

Aside | Posted on by | Leave a comment

Engage Submit Host Architecture

Engage users log in to the submit host to run jobs on the OSG. It needs to move to a platform with better administrative support. It’s time to do some archaeology to find out what’s there then build a new … Continue reading

Posted in Compute Grids, condor, Engage VO, Globus, grid, High Throughput Computing (HTC), High Throughput Parallel Computing (HTPC), OSG, RENCI | Leave a comment

High Throughput Parallel Molecular Dynamics on OSG

The Goal RENCI’s working with researchers interested in running high throughput parallel molecular dynamics simulations on OSG. Amber9 PMEMD The program we’d like to execute is called PMEMD (Particle Mesh Ewald Molecular Dynamics). PMEMD is a high-performance, parallel component of … Continue reading

Posted in Amber9, Compute Grids, condor, Continuous Integration (CI), Engage VO, grid, High Throughput Computing (HTC), High Throughput Parallel Computing (HTPC), multicore, OSG, pmemd, RENCI | Leave a comment

The OSG Compute Element – A Trip Report

In this post I’ll cover a few details, in hindsight, of setting up the Blueridge CE. There is plenty of documentation on how to set up an OSG compute element. Some of it is referenced from other posts on this … Continue reading

Posted in Compute Grids, Globus, grid, High Throughput Computing (HTC), multicore, OSG | Leave a comment

PMEMD on Blueridge

Baby steps. This is pmemd compiled against native MPI libraries executing on the RENCI Blueridge cluster. The job submission workflow uses the RENCI-CI script library. The job_run script uses Globus tools to transfer the pmemd application and input files to Blueridge. … Continue reading

Posted in Amber9, Compute Grids, Continuous Integration (CI), Engage VO, Globus, grid, High Throughput Computing (HTC), High Throughput Parallel Computing (HTPC), multicore, OSG, pmemd, RENCI | Leave a comment

An OSG Map

Finding OSG Stuff The Open Science Grid is a large global collaboration. Thousands of computing resources are shared through its protocols and conventions. While this is good news, it’s not always a simple matter to understand how the resources are … Continue reading

Posted in Compute Grids, Engage VO, grid, High Throughput Computing (HTC), OSG, RENCI | 2 Comments