Loading…
Welcome to the 2025 RMACC HPC Symposium!
Venue: Room 206 clear filter
arrow_back View All Dates
Wednesday, May 21
 

10:30am MDT

AI enabled protein interaction modeling
Wednesday May 21, 2025 10:30am - 11:00am MDT
AI assisted protein interaction modeling, pioneered by AlphaFold and RosettaFold, has become more diverse both with respect to the programs that do it, and how users run these programs. In this talk, we will cover the programs that are supported at the University of Utah, namely Alphafold2, Alphafold3, Colabfold, Boltz1, RFDiffusion and other tools from the Baker lab, the choices we have made with their deployment, and our experiences with using them. With respect to the ways to run, we will go over the standard SLURM scripts to run Alphafold in two stages (CPU only MSA search, GPU accelerated inference), use Colabfold server for faster MSA search, and using Google Colab running on compute nodes for interactive modeling in a notebook interface. Attendees should leave this talk with ideas how to set up and support these tools and contacts to UofU staff for further questions.
Speakers
Wednesday May 21, 2025 10:30am - 11:00am MDT
Room 206

11:15am MDT

Collaborative Cloud Science: Deploying The Littlest JupyterHub on Jetstream2
Wednesday May 21, 2025 11:15am - 12:15pm MDT
Many research teams and educators experience technical and resource challenges when setting up multi-user systems for data analysis and repeatable research. Instead of managing complex, on-premise systems, or paying for commercial cloud offerings, this presentation will show how to quickly start a simple JupyterHub (using TLJH) on a public research cloud like Jetstream2. This reduces setup effort for teams with limited IT support and improves teamwork and research repeatability in data-intensive projects.
This session is appropriate for researchers, educators, and research software engineers with intermediate skills who want to improve cloud access and teamwork, especially from institutions with limited research computing resources.
TLJH (The Littlest JupyterHub) is a simple, lightweight Jupyter Notebook server for small to medium-sized groups. It helps educators and researchers set up a shared Jupyter environment on a single server with minimal setup (no Kubernetes required!). 
Jetstream2 is a flexible, user-friendly cloud computing environment built on OpenStack. It is available to US-based researchers and educators at no cost through support from the National Science Foundation's Advanced Cyberinfrastructure Coordination Ecosystem: Services & Support (ACCESS) program. 
Learning Objectives
  • Learn Cloud Deployment with Jetstream2: Understand how to use Jetstream2, create instances, and benefit from cloud computing for research teamwork.
  • Install TLJH Step-by-Step: Follow the setup process for TLJH and adjust a basic JupyterHub to fit research needs
  • Set Up User Management and Security: Configure login settings, control user access, and adjust network settings to create a secure and easy-to-use research system.
  • Solve Problems Together: Work in small groups to fix common setup issues, share ways to expand the system, and discuss real-world uses of TLJH in research and teaching.
Speakers
Wednesday May 21, 2025 11:15am - 12:15pm MDT
Room 206

1:15pm MDT

Deploy & Manage Kubernetes on Jetstream2 using OpenStack Magnum
Wednesday May 21, 2025 1:15pm - 2:45pm MDT
Many modern research software systems run on Kubernetes for scale and resilience (e.g. JupyterHub, Dask, RStudio, etc.). Deploying Kubernetes in a reliable and robust way has historically been difficult. This tutorial offers a simple way to deploy Kubernetes clusters on Jetstream2 using OpenStack Magnum. By making cluster setup and management easier, this session helps teams with limited IT support to run powerful and scalable computing tools.

Participants will learn how to use OpenStack Magnum to create and manage Kubernetes clusters on the Jetstream2 research cloud. Designed for research software engineers and IT support staff with intermediate Linux skills and a basic understanding of containers and container orchestration, this session provides a repeatable process to build a scalable, container-based research system for their institutions.

Jetstream2 is a flexible, user-friendly cloud computing environment built on OpenStack. It is available to US-based researchers and educators at no cost through support from the National Science Foundation's Advanced Cyberinfrastructure Coordination Ecosystem: Services & Support (ACCESS) program. 

OpenStack is a free cloud computing platform that provides Infrastructure as a Service (IaaS). It helps organizations set up and manage public and private clouds. OpenStack includes tools for computing, networking, storage, and identity management, making it easy to build flexible and scalable cloud systems on different hardware. 

Kubernetes is an open-source container orchestration platform that automates deployment, scaling, and management of containerized applications. Kubernetes helps developers run complex applications reliably and efficiently.

Magnum is an OpenStack service that helps users set up and manage Kubernetes. Magnum offers native integration with OpenStack services, simplified cluster lifecycle management, and enhanced security and resource allocation for containers.
Learning Objectives
  • Understand how OpenStack Magnum automates Kubernetes cluster setup.
  • Understand the advantages of Magnum compared to alternatives.
  • Follow a step-by-step process to create and configure a Kubernetes cluster using Magnum.
  • Deploy test containerized applications and adjust cluster scaling.
Speakers
Wednesday May 21, 2025 1:15pm - 2:45pm MDT
Room 206

3:00pm MDT

Advancing Research with AWS: Compliance, Computing, and Connectivity
Wednesday May 21, 2025 3:00pm - 3:30pm MDT
This presentation explores AWS comprehensive solutions for the research community, addressing three critical challenges: meeting evolving compliance requirements like the new NIH NIST 800-171 standards, delivering scalable high-performance computing, and simplifying data management. We'll examine how the AWS Secure Research Environment (SRE) addresses complex security needs, compare managed AWS ParallelCluster Service with self-managed options, and showcase tools like the Globus S3 Connector for streamlined data handling. Join us to discover how AWS empowers researchers to focus on innovation while maintaining security, performance, and efficiency at scale.
Speakers
Wednesday May 21, 2025 3:00pm - 3:30pm MDT
Room 206

3:45pm MDT

Data Lifecycle Management in HPC – Automating Tiered Storage & Archival Strategies- Arcitecta
Wednesday May 21, 2025 3:45pm - 4:15pm MDT
Managing data at scale in high-performance computing (HPC) environments requires efficient storage and retrieval strategies. Automated tiered storage solutions enable seamless migration of aged data to lower-cost archival tiers while maintaining accessibility. Enriched metadata—spanning tagging, search, discovery, and data provenance—enhances data usability and long-term value. This approach not only optimizes storage costs but also empowers researchers with better data discovery and reuse. Real-world HPC use cases demonstrate how metadata-driven workflows streamline research, ensuring that critical datasets remain accessible and actionable over time.
Speakers
Wednesday May 21, 2025 3:45pm - 4:15pm MDT
Room 206

4:30pm MDT

The Metadata Revolution: Accelerating HPC Workflows Through Intelligent Data Management
Wednesday May 21, 2025 4:30pm - 5:00pm MDT
The next frontier in AI advancement isn’t just about algorithms—it’s about unlocking the wealth of hidden insights trapped within millions of files in HPC environments. While organizations focus on model architectures, the true bottleneck often lies in discovering and preparing relevant data buried in vast storage systems.
This presentation, featuring MetadataHub and a live demonstration, will reveal how intelligent metadata extraction and management transforms unstructured data into AI-ready assets by:
  • Uncovering Hidden Context: Live metadata extraction demonstrating how MetadataHubcaptures content and contextual value, revealing unexpected connections between research datasets and enabling new AI training opportunities that would otherwise remain hidden.
  • Automating Data Discovery: Demonstrating how MetadataHub automates metadata tagging to identify valuable training data across petabyte-scale storage, reducing data preparation time by up to 90%.
  • Enhancing Model Quality: Exploring how rich metadata captured by MetadataHub improves AI model performance by providing better context and enabling more relevant training data selection.
  • Scaling Efficiently: Showcasing metadata-driven automation with MetadataHub that optimizes data pipeline efficiency and resource utilization, including GPU/CPU performance, across HPC environments.
The session will highlight a real-world success story from the Zuse Institute Berlin, where MetadataHub unlocked 200 PB of previously underutilized research data for cutting-edge Generative AI applications. A 15-minute live demonstration will guide attendees through their journey—from data discovery to AI-ready datasets—highlighting practical challenges and solutions.
Attendees will leave with actionable strategies for implementing metadata-driven approaches in their own HPC workflows. By showcasing MetadataHub’s ability to extract content and contextual value, this session will demonstrate how metadata transforms unstructured data into a strategic advantage, accelerating AI initiatives and driving HPC innovation.
Speakers
Wednesday May 21, 2025 4:30pm - 5:00pm MDT
Room 206
 
Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.
Filtered by Date -