The CCS Research Network provides a high-performance, dedicated infrastructure designed to support data-intensive research workflows across the University of Kentucky. The network connects compute clusters, Data Transfer Nodes (DTNs), storage systems, and external research networks through a secure, high-bandwidth architecture optimized for scientific data movement.

Network Architecture

The Research Network forms a separate, performance-oriented network path distinct from campus IT traffic. It is optimized for predictable throughput, low-latency communication, and large-scale data movement between local compute/storage systems and external collaborators.

The network core is built with 100 Gbps high-bandwidth switching equipment. These switches interconnect CCS systems, including compute clusters, GPFS storage systems, Ceph storage, and dedicated DTNs.

Science DMZ

The CCS Research Network implements a Science DMZ architecture, which provides:

  • A high-throughput, bypass path for research data
  • Performance-tuned network segments optimized for large scientific datasets
  • Dedicated monitoring and performance-analysis tools

This design enables end-to-end data transfers at speeds that would not be achievable on the general-purpose campus network.

Data Transfer Nodes (DTNs)

CCS operates multiple DTN nodes that serve as optimized endpoints for large-scale data movement. These nodes:

  • Provide high-performance access between GPFS, compute nodes, and external destinations
  • Are tuned for tools such as Globus, Rclone, and other parallel transfer protocols
  • Act as the primary interface between CCS storage and external research networks or institutional collaborators

DTNs are part of the Science DMZ and benefit from its performance-oriented design.

External Research Network Connectivity

The Research Network connects to the outside world through multiple high-speed paths to Internet2, enabling fast, reliable transfers between UK and other national and international research institutions.

These external pathways support:

  • Collaboration with national labs and supercomputing centers
  • Movement of instrument and experimental data to and from CCS systems
  • Integration with NSF-funded testbeds and research cyberinfrastructure

NSF FABRIC Connectivity

The CCS Research Network provides Layer-2 connectivity to two on-campus NSF FABRIC racks, enabling:

  • Direct access to the FABRIC national backbone
  • Experimentation with advanced networking, distributed systems, and programmable infrastructure
  • High-bandwidth research workflows requiring L2 adjacency to remote FABRIC sites

This supports UK researchers working with next-generation networking architectures and federated national testbeds.

VIPLanes

The network continues to support VIPLanes, a software-defined networking (SDN) service that provides programmable L2/L3 connectivity for research workflows. VIPLane allows researchers to bypass bottlenecks in the campus network, enabling significantly more efficient and secure large-scale data transfers.