Texas A&M University is establishing and exploiting advanced 100-Gb/s connectivity, focused on accessing Internet2's OS3E network for its researchers. It is furthering data-intensive discipline research and network research, and is contributing to national best practices and to advanced connectivity in support of data-intensive research throughout the LEARN community in Texas.

The chief objective is to enhance the ability of A&M researchers to engage in data-intensive science, including combined use of campus and national cyberinfrastructure and large data sets. Similarly, routinely using large VLAN-based, wide-area data flows will challenge Texas A&M to grow technically. One example is the integration of these techniques into the LHC "data moving" area.

A second example is exploring federated wide-area file systems as an alternative to the classic "data moving" paradigm, including performance and federated identity issues crucial to the success of this approach. This is a contribution to realizing the evolving Campus Bridging vision. In support of computational scientists at Texas A&M, the project leverages connectivity with the Immersive Visualization Center to enable effective forms of local and remote visualization. Finally, the project enables network researchers to use the OpenFlow-based services of Internet2's OS3E as it evolves.

The advanced connectivity, combined with Texas A&M's ScienceDMZ, enables emerging sets of data-intensive scientists in areas that have not yet become prominent, destroying physical remoteness as a barrier to engaging in collaborative data-intensive science.

Project Report

Texas A&M University has built advanced 100 Gigabit per second (Gbps) computer network infrastructure to support data-intensive science and engineering research as well as computer networking research. We have connected the 100 Gbps network to a separate high-speed network on campus, known as a ScienceDMZ, which provides a "fast-lane" to the 100 Gbps off-campus network with 10 to 40 Gbps connections to researcher laboratories and campus cyberinfrastructure (e.g., High-Performance Computing (HPC) clusters, immersive visualization center, and specialized instruments). The ScienceDMZ also supports Software Defined Networking (SDN) protocols. Texas A&M researchers in many fields of science and engineering utilize national high-performance computing resources, unique research infrastructure on- and off-campus, or sensor networks that require moving large amounts of data to and from campus. For example, our experimental particle physics group participates in the Compact Muon Solenoid (CMS) experiment, which is located at the Large Hadron Collider at CERN in Switzerland, and they are expecting to transfer 100 Terabytes (TB)/month of CMS processed data from CERN to campus in the coming years (1 Terabyte = 1,000 Gigabytes). Similarly, the Texas Center for Climate Studies has a research project in climate modeling that generated 60 TB of data on the NSF-funded Stampede supercomputer, located at the Texas Advanced Computer Center (TACC) at the University of Texas at Austin, which needed to be moved back to the A&M campus for further analysis. Prior to the advanced networking infrastructure built as part of this grant, it took 30 days to transfer this data over the network. When high-performance computer interfaces and high-speed file transfer software are installed in the next few months, we expect this same transfer to take just a few hours. There are also instruments on campus such as the new Cryo Electron Microscope that generate large images that need to transferred to systems on- and off-campus, and the Synchrophasor Test Bed that is under development by the TEES SmartGrid Center will need to access real-time synchrophasor electric power grid data both nationally and internationally. The Synchrophasor Test Bed will also need to access A&M supercomputers for real-time computations. Texas A&M has computer networking research that will utilize the new networking infrastructure as well. Some of the networking research is data-intensive, such as the collection of data from millions of websites or the development of "big data" tools for extracting, mining, and visualizing data. Other networking research in high-performance networking protocols, SDN applications, cloud computing architectures that support large file transfers, or utilize the ExoGeni network will also make use of the new infrastructure. Utilizing the Lonestar Education and Research Network (LEARN) regional optical network, a 100 Gbps circuit was provisioned from the main campus in College Station, Texas to the LEARN/Internet2 Point-of-Presence in Houston, Texas, where it connects to Internet2’s Advanced Layer 2/3 networks. As the first recipient of a NSF CC-NIE grant in Texas, we worked to build consensus among LEARN members for a strategy on connecting to Internet2 at 100 Gbps in both Houston and Dallas, which allows all participating LEARN members to benefit from the 10X increase in bandwidth, shares the cost among the users, and provides redundancy and performance benefits. The 2 X 100 Gbps Internet2 connections are also important to the other four NSF CC-NIE/CC*IIE grant recipients in Texas. Internet2 operates a 100 Gbps Research and Education (R&E) network that connects to more than 250 U.S. universities as well as corporate and government R&D laboratories. Internet2 also connects globally to more than 65 international R&E partners that connect over 100 countries. Outside of Texas, Internet2 will provide the high-speed connections to support the data-intensive and networking research described above. Within Texas, LEARN operates a Texas R&E Network to which all of the universities and much of K-12 connect, and the University of Texas and Texas A&M Systems operate a joint 20 Gbps network backbone on the LEARN optical infrastructure with 100 Gbps-capable routers that connect to LEARN and Internet2 in Houston and Dallas. TACC benefits from the shared UT/A&M infrastructure to connect to Internet2 in Houston at 100 Gbps, which supports national access to the TACC supercomputers funded by NSF. The original ScienceDMZ on the A&M campus connected three HPC clusters and a network measurement node at 10 Gbps. The upgraded ScienceDMZ now connects off-campus at 100 Gbps and has added four 10 Gbps connections to faculty research labs and campus cyberinfrastructure. By the end of 2015, we expect to have a total of 15 research labs, HPC clusters, and specialized instruments connected to the ScienceDMZ at speeds between 10 and 40 Gbps.

Agency
National Science Foundation (NSF)
Institute
Division of Advanced CyberInfrastructure (ACI)
Type
Standard Grant (Standard)
Application #
1246097
Program Officer
Kevin Thompson
Project Start
Project End
Budget Start
2012-10-01
Budget End
2014-09-30
Support Year
Fiscal Year
2012
Total Cost
$499,609
Indirect Cost
Name
Texas A&M University
Department
Type
DUNS #
City
College Station
State
TX
Country
United States
Zip Code
77845