SlideShare a Scribd company logo
THE OPEN SCIENCE GRID (OSG)
And how it relates to PRAGMA
by Igor Sfiligoi, UCSD/SDSC
(isfiligoi@sdsc.edu)
PRAGMA 37 - SEPT 13TH, 2019
Premise – Not the Grid you were used to
• It is my understanding Grid has a negative connotation in PRAGMA
• While OSG has “Grid” in its name,
I believe it is far from the “Grid” you experienced 15 years ago
• Please give me a chance to convince you of that
PRAGMA 37 - SEPT 13TH, 2019 2
OSG evolution
• 10 years ago, OSG provided an integrated software stack that organizations
had to deploy in order for their researchers to benefit from OSG.
•We learned the hard way that this is too high a bar
to jump across for most researchers and institutions.
• Today, we offer to operate services to the groups we serve in order
to minimize the threshold to entry for everybody.
• Leveraging work performed by the likes of PRP/NRP-P/TNRP (e.g. Kubernetes)
PRAGMA 37 - SEPT 13TH, 2019 3
OSG Aspirational Goal
A campus Research IT Organization
should not have to learn anything ”non-standard”
in order to have their researchers benefit from OSG,
or have their resources be available via OSG
PRAGMA 37 - SEPT 13TH, 2019 4
Now …. Not there yet, but getting closer by the day…
OSG Mission
Advancing Open Science
through distributed High
Throughput Computing
PRAGMA 37 - SEPT 13TH, 2019 5
Open Science
•All of open science irrespective of discipline
•Include the maximum possible dynamic range
of science, groups, and institutions
• From individual undergraduates
to international collaborations with thousands of members.
• From small colleges, museums, zoos,
to national scale centers of open science.
PRAGMA 37 - SEPT 13TH, 2019 6
What Sciences use OSG?
• Order here is from most active to least active.
• Category is based on self-identification … i.e. people say what they do.
PRAGMA 37 - SEPT 13TH, 2019 7
Not all users are the same
• The individual researchers and small groups
• The campus Research Support Organizations
• Multi-institutional Science Teams
• XENON, GlueX, SPT, Simons, and many many more
• Collaborations between multiple campuses
• The 4 “big science” projects:
• US-ATLAS, US-CMS, LIGO, IceCube
Used services
change between
different categories
PRAGMA 37 - SEPT 13TH, 2019 8
OSG Mission
Advancing Open Science
through distributed High
Throughput Computing
PRAGMA 37 - SEPT 13TH, 2019 9
distributed High Throughput Computing
DRIVING PRINCIPLES:
• Maximize total compute, do not
optimize for any single application
• Allow for resources coming from
anywhere in the world
• Embrace heterogeneity
• Federated policy control
PRAGMA 37 - SEPT 13TH, 2019 10
(as opposed to e.g. HPC)
distributed High Throughput Computing
DRIVING PRINCIPLES:
• Maximize total compute, do not
optimize for any single application
• Allow for resources coming from
anywhere in the world
• Embrace heterogeneity
• Federated policy control
IMPLICATIONS:
• Expect high network latencies
• Expect varying walltime
between identical runs
• Job restarts are the norm
PRAGMA 37 - SEPT 13TH, 2019 11
(Although throughput
should be high)
Thriving in dHTC
• The path to successful dHTC involves two aspects:
• Separate a big computing problem in
many individually schedulable small problems.
• Minimize your requirements in order to
maximize the raw capacity that you can effectively use.
• The infrastructure will transparently take care of job restarts
Ingenious
Parallelism
PRAGMA 37 - SEPT 13TH, 2019 12
dHTC is special
•dHTC scales by definition perfectly.
• When a researcher understands how to partition their workflow into
many individually schedulable compute problems they can scale out
with ease to seemingly arbitrary scales of computing.
• When we integrate all IT resources at Universities, National Labs,
and the commercial cloud, we arrive at a near infinite resource pool.
PRAGMA 37 - SEPT 13TH, 2019 13
Stronger together
• For OSG VO, more than half of the compute comes from smaller sites
(and this does not include any of the big sciences, that have their own pools)
15k cores
PRAGMA 37 - SEPT 13TH, 2019 14
OSG IN DETAIL
PRAGMA 37 - SEPT 13TH, 2019 15
OSG Compute Services
• OSG provides the means for operating a compute federation
• Mostly in the form of Metascheduling Service
• But also operates one access point itself for the smaller players
PRAGMA 37 - SEPT 13TH, 2019 16
Federation i.e. Distributed control
• OSG works on three simple principles:
• Resource Owners determine policy of use
• This means that all policy of use is set locally by the clusters that join the federation.
• Resource Consumers specify the types of resources they are willing to use.
• How much RAM? How many cores per node? …
• OSG submits HTCondor batch system as (pilot) payload into all local batch
systems that match requirements, creating an overlay dHTC system
• Jobs are submitted locally, queue centrally, and execute anywhere that
matches requirements after resource becomes available.
PRAGMA 37 - SEPT 13TH, 2019 17
OSG Data Federation
• OSG operates a content delivery network (CDN)
• Based on XRootD caching technology
• Data origins operated by Science communities
• Data transparently published in our federated namespace.
• Caches hide data access latencies and reduce network traffic.
PRAGMA 37 - SEPT 13TH, 2019 18
OSG Software Distribution
•OSG can host and distribute community software on CVMFS
• A CDN at heart, but based on Apache Squid technology
• CVMFS provides a Read-Only POSIX interface
•OSG encourages the use of containerization
• Encourages participating sites to provide Singularity
• Which is optimized for use by unprivileged users (unlike Docker)
• Containers can be distributed through CVMFS (for performance)
PRAGMA 37 - SEPT 13TH, 2019 19
Facilitation Services
• Proactive, personalized guidance
and support for:
• Institution-specific submit points
• Sharing institutional resources via OSG
• Data federation across OSG sites
• Individual researchers using the
OSG submit infrastructure
• We also offer
• Local workshops
• OSG-hosted education and training
• Learning from the OSG Facilitation
Community
PRAGMA 37 - SEPT 13TH, 2019 20
OSG AND PRAGMA
PRAGMA 37 - SEPT 13TH, 2019 21
OSG AND PRAGMA
• OSG would be happy to facilitate
any Open Science endeavor
that has at least some of its
members in the US
• We are a NSF-funded project
• All OSG services are at your disposal
• Compute federation
• Data federation
• Software distribution
• Facilitation
PRAGMA 37 - SEPT 13TH, 2019 22
OSG and International Setups
•All our major users are international
• CMS, ATLAS, LIGO, IceCube, XENON, GlueX
•But, alas, we have been mostly collaborating with Europe
•PRAGMA could help us engage more
Science communities in the Pacific Rim
PRAGMA 37 - SEPT 13TH, 2019 23
Summary and Conclusion
• OSG’s objective is to “Advance Open Science
through distributed High Throughput Computing”
• We are committed to help any Open Science community
that has at least some members in the USA – Nice overlap with PRAGMA
• OSG can offer a wide variety of services
• Both in terms of technology and facilitation
• We could use some help getting more involved in the Pacific Rim
• Contact us at: help@opensciencegrid.org
PRAGMA 37 - SEPT 13TH, 2019 24

More Related Content

PPTX
Efficient and effective: can we combine both to realize high-value, open, sca...
PPTX
Thoughts on Cybersecurity
PPTX
A Reference Architecture for a Federated Cloud for Research
PDF
OGF standards for cloud computing
PDF
Post IPv6 Implementation and Security: Now What?
PDF
Distributed deep learning reference architecture v3.2l
PDF
OGF Standards Overview - Globus World 2013
PDF
OGF Introductory Overview - FAS* 2014
Efficient and effective: can we combine both to realize high-value, open, sca...
Thoughts on Cybersecurity
A Reference Architecture for a Federated Cloud for Research
OGF standards for cloud computing
Post IPv6 Implementation and Security: Now What?
Distributed deep learning reference architecture v3.2l
OGF Standards Overview - Globus World 2013
OGF Introductory Overview - FAS* 2014

Similar to The Open Science Grid and how it relates to PRAGMA (20)

PDF
GRP 19 - Nautilus, IceCube and LIGO
PPTX
GridIIT Open Science Grid
PDF
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
PDF
Bringing OSG users to the PRP Kubernetes Cluster
PPTX
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
PDF
The Open Science Grid
PDF
QuSandbox+NVIDIA Rapids
PPTX
Cloud Busting: Understanding Cloud-based Digital Forensics
PDF
Edge Computing and OGC SensorThings API
PDF
THE OPEN SCIENCE GRID Ruth Pordes
PDF
OGF Introductory Overview - OGF 44 at EGI Conference 2015
PDF
Calgary-Splunk-User-Group-March-2023.pdf
PPTX
Orchestration, Automation and Virtualisation (OAV) in GÉANT
PDF
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
PDF
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
PDF
Frank Würthwein - NRP and the Path forward
PPTX
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
PPTX
Internet2 - US Ignite Application Summit 2013
PPTX
Science Gateways: one portal, many e-Infrastructures and related services
PPTX
Big Data on Cloud Native Platform
GRP 19 - Nautilus, IceCube and LIGO
GridIIT Open Science Grid
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
Bringing OSG users to the PRP Kubernetes Cluster
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
The Open Science Grid
QuSandbox+NVIDIA Rapids
Cloud Busting: Understanding Cloud-based Digital Forensics
Edge Computing and OGC SensorThings API
THE OPEN SCIENCE GRID Ruth Pordes
OGF Introductory Overview - OGF 44 at EGI Conference 2015
Calgary-Splunk-User-Group-March-2023.pdf
Orchestration, Automation and Virtualisation (OAV) in GÉANT
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
Frank Würthwein - NRP and the Path forward
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
Internet2 - US Ignite Application Summit 2013
Science Gateways: one portal, many e-Infrastructures and related services
Big Data on Cloud Native Platform
Ad

More from Igor Sfiligoi (20)

PDF
Preparing Fusion codes for Perlmutter - CGYRO
PDF
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
PDF
Comparing single-node and multi-node performance of an important fusion HPC c...
PDF
The anachronism of whole-GPU accounting
PDF
Auto-scaling HTCondor pools using Kubernetes compute resources
PDF
Speeding up bowtie2 by improving cache-hit rate
PDF
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
PDF
Comparing GPU effectiveness for Unifrac distance compute
PDF
Managing Cloud networking costs for data-intensive applications by provisioni...
PDF
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
PDF
Using A100 MIG to Scale Astronomy Scientific Output
PDF
Using commercial Clouds to process IceCube jobs
PDF
Modest scale HPC on Azure using CGYRO
PDF
Data-intensive IceCube Cloud Burst
PDF
Scheduling a Kubernetes Federation with Admiralty
PDF
Accelerating microbiome research with OpenACC
PDF
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
PDF
Porting and optimizing UniFrac for GPUs
PDF
Demonstrating 100 Gbps in and out of the public Clouds
PDF
TransAtlantic Networking using Cloud links
Preparing Fusion codes for Perlmutter - CGYRO
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
Comparing single-node and multi-node performance of an important fusion HPC c...
The anachronism of whole-GPU accounting
Auto-scaling HTCondor pools using Kubernetes compute resources
Speeding up bowtie2 by improving cache-hit rate
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
Comparing GPU effectiveness for Unifrac distance compute
Managing Cloud networking costs for data-intensive applications by provisioni...
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Using A100 MIG to Scale Astronomy Scientific Output
Using commercial Clouds to process IceCube jobs
Modest scale HPC on Azure using CGYRO
Data-intensive IceCube Cloud Burst
Scheduling a Kubernetes Federation with Admiralty
Accelerating microbiome research with OpenACC
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Porting and optimizing UniFrac for GPUs
Demonstrating 100 Gbps in and out of the public Clouds
TransAtlantic Networking using Cloud links
Ad

Recently uploaded (20)

PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PPTX
sap open course for s4hana steps from ECC to s4
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PDF
Spectral efficient network and resource selection model in 5G networks
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PPTX
MYSQL Presentation for SQL database connectivity
PPTX
ACSFv1EN-58255 AWS Academy Cloud Security Foundations.pptx
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PPTX
Big Data Technologies - Introduction.pptx
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
Dropbox Q2 2025 Financial Results & Investor Presentation
sap open course for s4hana steps from ECC to s4
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
Spectral efficient network and resource selection model in 5G networks
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
The Rise and Fall of 3GPP – Time for a Sabbatical?
MYSQL Presentation for SQL database connectivity
ACSFv1EN-58255 AWS Academy Cloud Security Foundations.pptx
Building Integrated photovoltaic BIPV_UPV.pdf
Big Data Technologies - Introduction.pptx
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Mobile App Security Testing_ A Comprehensive Guide.pdf
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
Advanced methodologies resolving dimensionality complications for autism neur...
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Chapter 3 Spatial Domain Image Processing.pdf
NewMind AI Weekly Chronicles - August'25 Week I
MIND Revenue Release Quarter 2 2025 Press Release
Reach Out and Touch Someone: Haptics and Empathic Computing
“AI and Expert System Decision Support & Business Intelligence Systems”

The Open Science Grid and how it relates to PRAGMA

  • 1. THE OPEN SCIENCE GRID (OSG) And how it relates to PRAGMA by Igor Sfiligoi, UCSD/SDSC (isfiligoi@sdsc.edu) PRAGMA 37 - SEPT 13TH, 2019
  • 2. Premise – Not the Grid you were used to • It is my understanding Grid has a negative connotation in PRAGMA • While OSG has “Grid” in its name, I believe it is far from the “Grid” you experienced 15 years ago • Please give me a chance to convince you of that PRAGMA 37 - SEPT 13TH, 2019 2
  • 3. OSG evolution • 10 years ago, OSG provided an integrated software stack that organizations had to deploy in order for their researchers to benefit from OSG. •We learned the hard way that this is too high a bar to jump across for most researchers and institutions. • Today, we offer to operate services to the groups we serve in order to minimize the threshold to entry for everybody. • Leveraging work performed by the likes of PRP/NRP-P/TNRP (e.g. Kubernetes) PRAGMA 37 - SEPT 13TH, 2019 3
  • 4. OSG Aspirational Goal A campus Research IT Organization should not have to learn anything ”non-standard” in order to have their researchers benefit from OSG, or have their resources be available via OSG PRAGMA 37 - SEPT 13TH, 2019 4 Now …. Not there yet, but getting closer by the day…
  • 5. OSG Mission Advancing Open Science through distributed High Throughput Computing PRAGMA 37 - SEPT 13TH, 2019 5
  • 6. Open Science •All of open science irrespective of discipline •Include the maximum possible dynamic range of science, groups, and institutions • From individual undergraduates to international collaborations with thousands of members. • From small colleges, museums, zoos, to national scale centers of open science. PRAGMA 37 - SEPT 13TH, 2019 6
  • 7. What Sciences use OSG? • Order here is from most active to least active. • Category is based on self-identification … i.e. people say what they do. PRAGMA 37 - SEPT 13TH, 2019 7
  • 8. Not all users are the same • The individual researchers and small groups • The campus Research Support Organizations • Multi-institutional Science Teams • XENON, GlueX, SPT, Simons, and many many more • Collaborations between multiple campuses • The 4 “big science” projects: • US-ATLAS, US-CMS, LIGO, IceCube Used services change between different categories PRAGMA 37 - SEPT 13TH, 2019 8
  • 9. OSG Mission Advancing Open Science through distributed High Throughput Computing PRAGMA 37 - SEPT 13TH, 2019 9
  • 10. distributed High Throughput Computing DRIVING PRINCIPLES: • Maximize total compute, do not optimize for any single application • Allow for resources coming from anywhere in the world • Embrace heterogeneity • Federated policy control PRAGMA 37 - SEPT 13TH, 2019 10 (as opposed to e.g. HPC)
  • 11. distributed High Throughput Computing DRIVING PRINCIPLES: • Maximize total compute, do not optimize for any single application • Allow for resources coming from anywhere in the world • Embrace heterogeneity • Federated policy control IMPLICATIONS: • Expect high network latencies • Expect varying walltime between identical runs • Job restarts are the norm PRAGMA 37 - SEPT 13TH, 2019 11 (Although throughput should be high)
  • 12. Thriving in dHTC • The path to successful dHTC involves two aspects: • Separate a big computing problem in many individually schedulable small problems. • Minimize your requirements in order to maximize the raw capacity that you can effectively use. • The infrastructure will transparently take care of job restarts Ingenious Parallelism PRAGMA 37 - SEPT 13TH, 2019 12
  • 13. dHTC is special •dHTC scales by definition perfectly. • When a researcher understands how to partition their workflow into many individually schedulable compute problems they can scale out with ease to seemingly arbitrary scales of computing. • When we integrate all IT resources at Universities, National Labs, and the commercial cloud, we arrive at a near infinite resource pool. PRAGMA 37 - SEPT 13TH, 2019 13
  • 14. Stronger together • For OSG VO, more than half of the compute comes from smaller sites (and this does not include any of the big sciences, that have their own pools) 15k cores PRAGMA 37 - SEPT 13TH, 2019 14
  • 15. OSG IN DETAIL PRAGMA 37 - SEPT 13TH, 2019 15
  • 16. OSG Compute Services • OSG provides the means for operating a compute federation • Mostly in the form of Metascheduling Service • But also operates one access point itself for the smaller players PRAGMA 37 - SEPT 13TH, 2019 16
  • 17. Federation i.e. Distributed control • OSG works on three simple principles: • Resource Owners determine policy of use • This means that all policy of use is set locally by the clusters that join the federation. • Resource Consumers specify the types of resources they are willing to use. • How much RAM? How many cores per node? … • OSG submits HTCondor batch system as (pilot) payload into all local batch systems that match requirements, creating an overlay dHTC system • Jobs are submitted locally, queue centrally, and execute anywhere that matches requirements after resource becomes available. PRAGMA 37 - SEPT 13TH, 2019 17
  • 18. OSG Data Federation • OSG operates a content delivery network (CDN) • Based on XRootD caching technology • Data origins operated by Science communities • Data transparently published in our federated namespace. • Caches hide data access latencies and reduce network traffic. PRAGMA 37 - SEPT 13TH, 2019 18
  • 19. OSG Software Distribution •OSG can host and distribute community software on CVMFS • A CDN at heart, but based on Apache Squid technology • CVMFS provides a Read-Only POSIX interface •OSG encourages the use of containerization • Encourages participating sites to provide Singularity • Which is optimized for use by unprivileged users (unlike Docker) • Containers can be distributed through CVMFS (for performance) PRAGMA 37 - SEPT 13TH, 2019 19
  • 20. Facilitation Services • Proactive, personalized guidance and support for: • Institution-specific submit points • Sharing institutional resources via OSG • Data federation across OSG sites • Individual researchers using the OSG submit infrastructure • We also offer • Local workshops • OSG-hosted education and training • Learning from the OSG Facilitation Community PRAGMA 37 - SEPT 13TH, 2019 20
  • 21. OSG AND PRAGMA PRAGMA 37 - SEPT 13TH, 2019 21
  • 22. OSG AND PRAGMA • OSG would be happy to facilitate any Open Science endeavor that has at least some of its members in the US • We are a NSF-funded project • All OSG services are at your disposal • Compute federation • Data federation • Software distribution • Facilitation PRAGMA 37 - SEPT 13TH, 2019 22
  • 23. OSG and International Setups •All our major users are international • CMS, ATLAS, LIGO, IceCube, XENON, GlueX •But, alas, we have been mostly collaborating with Europe •PRAGMA could help us engage more Science communities in the Pacific Rim PRAGMA 37 - SEPT 13TH, 2019 23
  • 24. Summary and Conclusion • OSG’s objective is to “Advance Open Science through distributed High Throughput Computing” • We are committed to help any Open Science community that has at least some members in the USA – Nice overlap with PRAGMA • OSG can offer a wide variety of services • Both in terms of technology and facilitation • We could use some help getting more involved in the Pacific Rim • Contact us at: help@opensciencegrid.org PRAGMA 37 - SEPT 13TH, 2019 24