Facilities, Equipment, and Other Resources (FER) Statement—Long Version

This page is for a longer version of FER. For some submissions, you may be space-limited and a more concise version of FER may be desirable. The shorter version is available at this page.

A longer version of the FER is provided here in various forms:

  1. Download PDF

  2. Download a Latex document that will build the above PDF

  3. Download an MS Word Document

The contents of the documents above are identical. One form may be preferable to another in order to match document styles. The document contents also appear below.

1. Introduction

Advanced Research Computing (ARC) advances the use of computing and visualization in Virginia Tech (VT) research by providing a centralized organization that identifies, acquires, maintains, and upgrades/extends compute systems; and educates, supports, problem-solves, and collaborates with the VT community. As one example, ARC offers generous levels of no-cost resources to students and faculty, which supports the great majority of user needs. The ARC website is https://arc.vt.edu. Our vision and mission statements are available at https://arc.vt.edu/content/arc_vt_edu/en/about.html. ARC resides in the Division of Information Technology (DoIT). Cybersecurity is implemented through the Information Technology Security Office (ITSO).

2. Compute Systems

2.1 Cluster Systems

VT’s main compute clusters are listed in Table 1. All clusters serve the entire VT community. TinkerCliffs is the flagship cluster, consisting of CPU and GPU (A100 and H200) nodes. Owl is a water-cooled CPU-based cluster that includes three large memory nodes with up to 8 TB of memory each. Falcon is a GPU cluster comprising L40s, A40, V100, and T4 GPUs. The CUI (Controlled Unclassified Information) cluster is restricted for using ITAR or Export Controlled software/data. The biomed cluster is restricted for biomedical research in need of NIST 800-171 compliance. The five clusters serving all of VT provide 570 nodes, 58,784 CPU cores, 497 GPUs, and 272 TB of memory. The unrestricted clusters share the same GPFS mount so that codes and data can be accessed from any cluster. All clusters use Infiniband (IB) interconnect.

Table 1. Summary of ARC compute clusters.

Cluster

TinkerCliffs

Owl

Falcon

CUI

Biomed

Node type

CPU / GPU

CPU

GPU

CPU/GPU

CPU / GPU

Nodes

308 / 16 / 8 / 14 / 7

84 / 4 / 3

32 / 20 / 40 / 19

6 / 2

6 / 1

Cores per node

128 / 96 / 128 / 128 / 64

96 / 64 / 128

64 / 64 / 24 / 32

64 / 64

64 / 64

GPUs per node

– / – / – / 8x A100 / 8x H200

– / – / –

4x A30 / 4x L40s / 2x V100 / 1x T4

– / 8x A100

– / 8x A100

Memory (GB) per node

256 / 384 / 1024 / 2048 / 2048

768 / 512 / 8192

512 / 512 / 384 / 196

512 / 2048

512 / 2048

2.2 Visionarium

ARC’s Visionarium Lab provides an array of visualization resources, including the VisCube, an immersive three-dimensional visualization environment, with head and device tracking. The ‘HyperCube’ remains the highest resolution and the most accessible immersive visualization lab on campus. The lab is available to graduates and undergraduates to work with their data and visualization solutions in a hardware and software environment with interactive 3D graphics and High-Performance Visualization (HPV). A new system became operational in 2025, expanded to a 14.7-foot by 14.7-foot by 9.2-foot three-dimensional visualization environment, with 47.9 million pixels and 700 ft2 of display surface, returning the Visionarium to a world-class immersive visualization facility.

2.3 Access To Resources Beyond VT

ARC resources are able to leverage Virginia Tech’s excellent network connectivity, and network. Virginia offers access to advanced national networks, including ESnet, Internet2, and Mid Atlantic Crossroads.

3. Upgrades and Maintenance

Approximately every four months, the clusters are taken offline for two-to-four days for major upgrades. To the extent possible, hardware and software maintenance and expansion—which are continual efforts—are completed while the systems are offline to minimize disruptions to users.

4. Leadership and Technical Support

4.1 ARC Leadership

Alberto Cano, AVP for Research Computing, leads ARC. Other leadership team members are: Matthew Brown, Director of Research Computing Services; Mark Gardner, Network Research Manager; Jeremy Johnson, IT Operations Manager; and Nicholas Polys, Director of Visualization.

4.2 Technical Systems Support

The Systems Engineering Team architects, installs, maintains, and upgrades research system network, storage, and compute resources, and workload management. They maintain and implement system security practices and respond to alerts from monitoring/logging systems. These efforts include implementing and maintaining supporting infrastructure systems (e.g., cooling and power systems) and services. The team researches new/emerging technologies, including research on compute hardware and networks, and interacts with vendors to integrate hardware into ARC systems. They maintain user accounts and operate user-facing systems such as ColdFront and Open OnDemand, respectively.

4.3 Technical User Support

Technical user support occurs on two fronts. First, support is provided for general use of the clusters, user and system problems, and individual user or group needs, including proposal preparation (e.g., consultations). Second, system software and applications are maintained and new software is added based on user requirements. Both fronts are served by five Computational Scientists and four Graduate Research Assistants, who also conduct workshops for users. The group also hosts a bimonthly meeting for all VT students and faculty to describe new developments, bring up new issues for input from the user base, and answer questions. These meetings, among other outlets, guide ARC’s future equipment purchases, operations, and services.

5. Cybersecurity

The Information Technology Security Office (ITSO) oversees the security of the cyber infrastructure on campus under the direction of the Chief Information Security Officer.

5.1 InCommon Federation

Virginia Tech participates in the InCommon Federation as an Identity Provider, allowing individuals to use their Virginia Tech credentials to securely authenticate to services provided by InCommon Service Providers. Since InCommon is a federal Trust Framework Provider, Virginia Tech is an approved Credential Service Provider under the FICAM TFS Program.

5.2 Intrusion Detection

The ITSO deploys freeware and commercial intrusion detection sensors to monitor attacks against University computers. The data collected by these sensors is analyzed to identify and respond to cyber-attacks and to analyze malware. Virginia Tech takes a Continuous Monitoring security strategy that allows “hunting” for compromised machines and facilitates remediating events. Continuous Monitoring is a component of the Center for Internet Security’s (CIS) Controls architecture. The CIS Controls are a subset of the NIST 800-53 Priority 1 controls, NIST CSF, and NIST 800-171. One very useful capability is the ability to merge IDS data with GIS maps and building floor plans on campus to facilitate a quick and timely way to determine the scope of cyber-attacks against University computers.

5.3 Cyber Security Operations Center

Virginia Tech has implemented a Cyber Security Operations Center (SOC). This convergence of tools, data, and personnel into the SOC allows IT Security personnel to unite data from network sensors, provide analysis, and coordinate needed responses to further protect assets.

5.4 Centralized Logging

The Division of Information Technology provides a Central Log Service (CLS) to university departments for storing and analyzing logs, both for security and operational purposes.