Skip to content Skip to navigation
See our Campus Ready site for most up to date information about the fall semester.Campus ReadyCOVID Help

Research Computing Services

Schedule a Consultation

If you need personalized expertise to improve your lab workflow, design one of your project, or fill in compute or data storage requirement for a grant, you can schedule a consultation with our team of experts for a personalized recommendation.


Research Facilitation

It is our goal to make your research efficient, reproducible, and accessible. Our expert are here to provide you with a range of services to help your research succeed. Our research facilitation service includes providing you with resources and knowledge of IT capabilities, as well as helping you with technical writing for grants or publishing.

We take the time to understand your work in order to ensure the scalability and sustainability of your project. We can help you with Version Control, Code testing and optimization, collaboration, software citation, and more. 


HPC Zoom Clinics

Bring your laptop, your code, and your questions to the HPC walk-in clinic and get expert help, right on the spot. Experienced graduate students are encouraged to come help your peers by mentoring them in HPC tips and tricks. Faculty are also welcome to join. Among others, Sarvani Chadalapaka (HPC Manager) and Robert Romero (Sr. Sys Admin) will be available to meet with and help members of the campus research computing community at these sessions.

WHERE 
Zoom - Located after you log into the MERCED Cluster
WHEN 
Every Friday @ 10:30 am – 12:00 noon 


Facilities Statement

The  MERCED Cluster (Multi-Environment Research Computer for Exploration and Discovery Cluster) is a Linux cluster from Advanced Clustering Technologies, funded by an NSF Major Research Instrumentation grant.  It has:
  - 100+ nodes with between 20 and 42 cores, from 128G to 256G per machine for a total of 2000+ core an 18T of RAM.
  - 6 GPU Nodes with 12 NVIDIA Tesla P100
  - 5 storage nodes for a total of 350TB
All above nodes are interconnected via infiniband w/ RDMA for fast (25Gbits/s) and for low latency (sub ms) data transfer.

The Science DMZ is a campus-wide dedicated 1-10G network with direct connections to CENIC’s High-performance Research Network; the Pacific Research Platform (PRP); and other regional, national and even international networks. The Science DMZ also hosts three dedicated Fast Input/Output Network Appliances (FIONAs) data transfer nodes. These FIONAs provide over 250TBs of data disks that can be used for the staging, moving and sharing of large data sets. The PRP provides a uniform and configuration-managed suite of FIONAs across all 10 UC campuses, allowing for seamless and nearly effortless data transfers. The Science DMZ also hosts two (2) additional FIONA8s, which combine the abilities of the traditional data transfer nodes with at least 8 high-end GPUs, which allow researchers to host the data required for machine learning techniques. This world-wide system, known as Nautilus, is available to researchers in any discipline by request. More information on location and other details is available here. The Science DMZ is support by NSF Award #1659210. The PRP is supported by NSF Grant #1541349 under the CC*DNI DIBBs program.

The Wide Area Visualization Environment (WAVE) is constructed of 20 4K (Ultra HD), stereoscopic Organic Light Emitting Diode (OLED) displays tiled together in a 5 x 4 half-pipe matrix. These OLED displays are capable of displaying over 1 billion distinct colors (compared to standard displays of around 1.7 million) and a contrast ratio of 200,000:1. The system is driven by 10 Dual Tesla GPU compute nodes with several TB ultra-fast SSD Drives to work with multi TB data sets. In addition to the stunning 2D/3D imagery, the system is connected to the ScienceDMZ at 10Gbps (and to the rest of campus at 40 Gbps), so multi-site collaborations are possible. While each individual part of the WAVE (e.g. the displays, the GPUs, the network) is actually commodity hardware, the sophisticated engineering and open source software it runs make the WAVE and its kind a unique research tool.

The WAVE Lab have a system of Space and compute reservations.

All compute systems are usually running jobs using an automatic scheduler, but it is possible to place exceptional reservation in place to ensure the compute, network and storage resources will be available during a specific time slot.
 


Data Management

Managing research data can be challenging, especially when the data are large or sensitive (e.g. HIPAA). Through a partnership with the University Library, Research IT can help! We offer services in developing a Data Management Plan through the use of the DMPTool (see DMPTool.org), and we can help in specifying application specific storage solutions for purchase. Request a consultation here. If you require letters of compliance in order to gain access to sensitive data, or if your research requires you to safeguard sensitive data, we can help you navigate the complex regulatory environment with the goal of issuing you letters of compliance so you can get started. See Research Data Management Consultation for more information.