GENI to FABRIC Transition

Having served the NSF research community for more than a decade, the GENI research network shut down its operations on August 1, 2023. Although GENI is no longer operational, there are several alternative NSF-funded cyberinfrastructure facilities for researchers to use. We expect most users will switch to FABRIC, which, like GENI, offers a programmable network platform with advanced, future-looking, “routing” infrastructure connected via high speed links with the potential for QoS guarantees.

The following instructions have been created to help users move from GENI to FABRIC. Please consult the appropriate instructions based on your role:
Researchers
Instructors
Students

IT Staff at GENI sites can find information needed to shutdown their GENI rack in this GENI Rack Decommission Instructions document.

NSF Funds Seven New Campus Cyberinfrastructure Projects Using FABRIC

The NSF Campus Cyberinfrastructure (CC*) program invests in coordinated campus-level networking and cyberinfrastructure improvements, innovation, integration, and engineering for science applications and distributed research projects. We are excited to be working with seven awardees from this CC* solicitation (NSF 20-507) as they use FABRIC to perform experimental deployment, protocol testing, and evaluation. Learn more about each project below:

CC* Integration-Large: Q-Factor: A Framework to Enable Ultra High-Speed Data Transfer Optimization based on Real-Time Network State Information provided by Programmable Data Planes

PI: Jeronimo Bezerra | Florida International University | Award Abstract | Project Overview Slides

Communication networks are critical components of today’s scientific workflows. Researchers require long-distance, ultra high-speed networks to transfer huge data from acquisition sites (such as Vera C. Rubin Observatory, also known as the Large Synoptic Survey Telescope in Chile) to processing sites, and to share measurements with scientists worldwide. However, while network bandwidth is continuously increasing, the majority of data transfers are unable to efficiently utilize the added capacity due to inherent limitations of parameter settings of the network transport protocols and the lack of network state information at the end hosts. To address these challenges, Q-Factor plans to use temporal network state data to dynamically configure current transport protocol parameters to reach higher network utilization and, as a result, to improve scientific workflows.

CC* Integration-Small: Science Traffic as a Service (STAAS)

PI: Jack Brassil | Princeton University | Award Abstract | Project Overview Slides

Science Traffic as a Service (STAAS) is a decentralized system to collect, create and distribute a diverse collection of real and synthetic science network traffic to the experimental research infrastructure user community. Available on-demand to experimenters through a publish-subscribe dashboard, the tool will elevate traffic selection and distribution to a first class experimental instrumentation resource.

The STAAS team is developing a cooperative, scalable, infrastructure-agnostic end system traffic generation service. They will deploy the system first on the Network Programming Initiative testbed reaching from Princeton to Cornell, and then ready STAAS for later deployment on the FABRIC research infrastructure. Their key insight is that plenty of science flows are already in transit at any moment, either on campus or crossing the campus border. They plan to tap these available active science flows and process each for forwarding onto the experimental testbed, while preserving both the timing integrity of the flows and the data privacy of their headers and payloads. To do this safely, science traffic will flow only when offered by an originating campus and requested by the experimenter. By introducing a service model, STAAS hopes to help advance the networking research community’s understanding of emerging science traffic and to help the operators of scientific instruments improve the efficiency of global data distribution.

CC* Integration-Small: Integrating Application Agnostic Learning with FABRIC for Enabling Realistic High-Fidelity Traffic Generation and Modeling

PI: Deniz Gurkan | University of Houston | Award Abstract | Project Overview Slides

Novel approaches to networking and application development require high-fidelity testing and evaluation supported by realistic network usage scenarios. This project will provide researchers the means to easily utilize the capabilities of the FABRIC testbed through a suite of new tools – smoothing the transition of existing experiments to the testbed and enabling exciting new areas of research through three systems facilitating end to end traffic modeling and generation in the FABRIC environment. A model repository will be created for the storage and access of custom models by experimenters and will be seeded with stock models of some popular applications for immediate use. The use of the models within FABRIC-hosted experiments will be advanced through a bespoke matching system that will align experiment resources with model requirements. Finally, for experiments developing novel applications, a tool will be provided for creating new models using data captured with FABRIC infrastructure components.

CC* Integration-Large: SciStream: Architecture and Toolkit for Data Streaming between Federated Science Instruments

PI: Rajkumar Kettimuthu | University of Chicago | Award Abstract | Project Overview Slides

Scientific instruments are capable of generating data at very high speeds. However, with traditional file-based data movement and analysis methods, data are often processed at a much lower speed, leading to either operating the instruments at a lower speed or discarding a portion of the data without processing it. To address this issue, the SciStream project will develop software tools to stream data at very high speeds from scientific instruments to supercomputers at a distant location. Through a set of gateway nodes optimized for wide area memory-to-memory transfers at the perimeter of the campus network, SciStream establishes necessary bridging between source and destination. SciStream hides the complexities in network connections from the end user and provides a high level of security for all the network connections. Key design choices such as application-agnostic streaming and support for best-effort streaming will make SciStream appealing to a broader science community.

CC* Integration-Small: Error Free File Transfer for Big Science

PI: Craig Patridge | Colorado State University | Award Abstract | Project Overview Slides

This project will improve the Internet’s ability to support big data transfers, both for science and commerce, for decades to come. Scientific data transfers have gotten so large that previously rare transmission errors in the Internet are causing some scientific data transfers to be corrupted. The Internet’s error checking mechanisms were designed at a time when a megabyte was a large file. Now files can contain terabytes. The old error checking mechanisms are in danger of being overwhelmed. This project seeks to find new error checking mechanisms for the Internet to safely move tomorrow’s scientific data efficiently and without errors by addressing two fundamental issues.

First, the Internet’s checksums and message digests are too small (32-bits) and probably are poorly tuned to today’s error patterns. The first step in this project is to collect information about the kinds of transmission errors currently happening in the Internet for a comprehensive study. Second, today’s file transfer protocols, if they find a file has been corrupted in transit, simply discard the file and transfer it again. Rather, the file transfer protocol should seek to repair the corrupted parts of the file. As the project collects data about errors, it will also design a new file transfer protocol that can incrementally verify and repair files.

CC* Integration-Large: N-DISE: NDN for Data Intensive Science Experiments

PI: Edmund Yeh | Northeastern University | Award Abstract | Project Overview Slides

The project, N-DISE (Named Data Networking for Data Intensive Science Experiments), aims to accelerate the pace of breakthroughs and innovations in data-intensive science fields such as the Large Hadron Collider (LHC) high energy physics program and the BioGenome and human genome projects. Based on Named Data Networking (NDN), a data-centric architecture, N-DISE will deploy and commission a highly efficient and field-tested petascale data distribution, caching, access and analysis system serving major science programs.

The N-DISE project will design and develop high-throughput caching and forwarding methods, containerization techniques, hierarchical memory management subsystems, congestion control mechanisms, integrated with Field Programmable Gate Arrays (FPGA) acceleration subsystems, to produce a system capable of delivering LHC and genomic data over a wide area network at throughputs approaching 100 Gbits per second, while significantly decreasing download time. In addition, N-DISE will utilize NDN’s built-in data security support to ensure data integrity and provenance tracing. N-DISE will leverage existing infrastructure and build an enhanced testbed with four additional high performance NDN data cache servers at participating institutions.

CC* Integration-Large: An ‘On-the-fly’ Deeply Programmable End-to-End Network-Centric Platform for Edge-to-Core Workflows

PI: Michael Zink | University of Massachusetts Amherst | Award Abstract | Project Overview Slides

Unmanned Aerial Vehicles (also known as drones) are becoming popular in the sky. Their application reaches from hobby drones for leisurely activities to life-critical drones for organ transport to commercial applications such as air taxis. The safe, efficient, and economic operation of such drones poses a variety of challenges that have to be addressed by the science community. For example, drones need very detailed, close to the ground weather information for safe operations, and data processing and energy consumption of drones need to be intelligently handled. This project will provide tools that will allow researchers and drone application developers to address operational drone challenges by using advanced computer and network technologies.

This project will provide an architecture and tools that will enable scientists to include edge computing devices in their computational workflows. This capability is critical for low latency and ultra-low latency applications like drone video analytics and route planning for drones. The proposed work will include four major tasks. First, cutting edge network and compute infrastructure will be integrated into the overall architecture to make them available as part of scientific workflows. Second, in-network processing at the network edge and core will be made available through new programming abstractions. Third, enhanced end-to-end monitoring capabilities will be offered. Finally, the architecture will leverage the Pegasus Workflow Management System to integrate in-network and edge processing capabilities.

Do you have a project idea that would benefit from using FABRIC? We are interested in working with you. Apply for a Campus Cyberinfrastructure (CC*) grant – the solicitation’s next deadline is March 1, 2021.

NSF announces $3 million award to expand FABRIC cyberinfrastructure globally

Advanced network offers platform to reimagine the Internet and speed scientific discovery

A new $3 million grant from the National Science Foundation (NSF) will expand FABRIC, a project to build the nation’s largest cyberinfrastructure testbed, to four preeminent scientific institutions in Asia and Europe. The expansion represents an ambitious effort to accelerate scientific discovery by creating the networks needed to move vast amounts of data across oceans and time zones seamlessly and securely.

Science is fast outgrowing the capabilities of today’s Internet infrastructure. To fully capitalize on big data, artificial intelligence, advanced computation and the Internet of Things requires robust, interconnected computers, storage, networks and software. Uneven progress in science cyberinfrastructure has led to bottlenecks that stymie collaboration and slow the process of discovery.

FABRIC, launched in 2019 with a $20 million grant from NSF, is building a cyberinfrastructure platform where computer scientists can reimagine the Internet and test new ways to store, compute, and move data. With the new NSF award, a sister project called FABRIC Across Borders (FAB) will link FABRIC’s nationwide infrastructure to nodes in Japan, Switzerland, the U.K. and the Netherlands.

“FAB allows collaborative international science projects to experiment with ways to do their science more efficiently,” said FAB Principal Investigator Anita Nikolich, Director of Technology Innovation at the University of Illinois School of Information Sciences and Cyber Policy Fellow at the Harris School of Public Policy at University of Chicago. “Sending large quantities of data long distances—across borders and oceans—is complicated when your science depends on real-time processing so you don’t miss once in a lifetime events. Being able to put FABRIC nodes in physically distant places allows us to experiment with the infrastructure to support new capabilities and also bring disparate communities together.”

FAB will be led by the University of Illinois along with core team members from RENCI at the University of North Carolina at Chapel Hill; the University of Kentucky; the Department of Energy’s Energy Sciences Network (ESnet); Clemson University; and the University of Chicago. Over three years, the team will work with international partners to place FABRIC nodes at the University of Tokyo; CERN, the European Organization for Nuclear Research in Geneva, Switzerland; the University of Bristol in the U.K.; and the University of Amsterdam.

The project is driven by science needs in fields that are pushing the limits of what today’s Internet can support. As new scientific instruments are due to come online in the next few years—generating ever larger data sets and demanding ever more powerful computation—FAB gives researchers a testbed to explore and anticipate how all that data will be handled and shared among collaborators spanning continents.

“FAB will offer a rich set of network-resident capabilities to develop new models for data delivery from the Large Hadron Collider (LHC) at CERN to physicists worldwide,” said Rob Gardner, Deputy Dean for Computing and research professor in the Physical Sciences Division at the University of Chicago and member of FAB’s core team. “As we prepare for the high luminosity LHC, the FAB international testbed will provide a network R&D infrastructure we’ve never had before, allowing us to consider novel analysis systems that will propel discoveries at the high energy frontier of particle physics.”

“FABRIC will tremendously help the ATLAS experiment in prototyping and testing at scale some of the innovative ideas we have to meet the high throughput and big data challenges ATLAS will face during the high luminosity LHC era,” said ATLAS computing coordinators Alessandro Di Girolamo, a staff scientist in CERN’s IT department, and Zach Marshall, an ATLAS physicist from Lawrence Berkeley National Laboratory. “The ATLAS physics community will be excited to test new ways of doing analysis, better exploiting the distributed computing infrastructure we run all around the world.”

To ensure the project meets the needs of the scientists it aims to serve, FAB will be built around use cases led by scientific partners in five areas:

Physics (high energy physics use cases at CERN’s Large Hadron Collider)
Space (astronomy and cosmology use cases in the Legacy Survey of Space and Time and the Cosmic Microwave Background-Stage 4 project)
Smart cities (sensing and computing use cases to advance smart, connected communities for the NSF SAGE project and work at the University of Antwerp and the University of Bristol)
Weather (use cases to improve weather and climate prediction at the University of Miami and Brazil’s Center for Weather Forecast and Climatic Studies)
Computer science (use cases in private 5G networks at the University of Tokyo; censorship evasion at Clemson University; network competition and sharing at the University of Kentucky; and software-defined networking and P4 programming at South Korea’s national research and engineering network, KREONET)
FAB will connect with existing U.S. and international cyberinfrastructure testbeds and bring programmable networking hardware, storage, computers, and software into one interconnected system. All software associated with FAB will be open source and posted in a publicly available repository: https://github.com/fabric-testbed/.

FABRIC Experimenters Workshop

Date: April 8-9 2021
Location: Virtual
Registration: https://t.co/fwIniOlxNs

The goal of this workshop is to introduce FABRIC to experimenters and those considering experimenting. We will share information about deployed sites, recent decisions that could affect your experiments, and demos of FABRIC’s current capabilities. Additionally, we will share the planned timeline, including future site deployment and testbed access. More information can be found on our website.

This workshop is open to everyone – industry, academic, government, and international researchers. Register here!

OmniSOC Year 3 Virtual Event

Date: February 26, 2021
Location: Virtual
Event Website: https://omnisoc.iu.edu/seminar-series/

To celebrate OmniSOC’s third year of operations and its team’s initiative to open OmniSOC membership to all members of the higher education and research community, OmniSOC will conduct a three-day virtual event February 24–26, 2021. This event features internal discussions with founding and current members, its lead technology partner Elastic, and content for potential new members and the larger higher education and research community. On February 26 at 1 PM ET, FABRIC Co-PI Anita Nikolich will speak on a panel discussing the potential OmniSOC and ResearchSOC data holds for cybersecurity researchers.

GENI Survey

In June of 2020, the GENI ENTeR project team (consisting of UNC RENCI and University of Kentucky) sent a survey to over 4000 experimenters who were known to have used GENI within the past 2 years.

We received 183 complete responses (251 total), of which 22% were from faculty, 68% from students, 10% from professional researchers, postdocs and technical staff.

• Across those groups, 40% used GENI for research, 80% used it for education (many used GENI for both).
• Over 40% used it periodically during the semester, and 30% reported using it weekly.
• Their activities spanned a wide variety of research topics, from network architectures and protocols, to security, virtualization and software-defined networking.
• Overwhelmingly 83% of respondents found GENI to be beneficial or very beneficial to their research or educational activities.

Faculty using GENI for teaching reported using GENI in classes that had a total over 4000 students. Many taught regularly recurring classes on Computer Communications, Networking, Cybersecurity, and Cloud Computing.

Respondents collectively reported publishing 245 papers that benefitted from GENI. Respondents provided 31 specific citations of these papers published between 2011 and 2020; 26 of these 31 were not previously listed in the GENI Bibliography.

For more details please download the full survey report, GENI 2020 Survey.

GENI Regional Workshop at the University of Oregon

Jun Li, professor and director of the Center for Cyber Security and Privacy at the University of Oregon, hosted a GENI Regional Workshop (GRW) on November 3-4, 2017. The workshop attracted over 40 faculty, graduate and undergraduate students, a majority of whom were from universities and colleges in Oregon, Washington and Northern California.

Participants at the GRW at the University of Oregon

Participants at the GRW at the University of Oregon

The theme of the workshop was security experiments, in keeping with the interests of Prof. Li and his colleagues.  It included tutorials that used GENI resources to experiment with man-in-the-middle attacks on Wi-Fi hotspots, denial of service attack mitigation using software defined networking, using virtual network function concepts to implement a dynamically scalable intrusion detection system, and creating and detecting covert storage channels.

The workshop opened the morning of November 3 with talks by Chip Elliot of Raytheon BBN Technologies, Eric Keller of the University of Colorado and Prasad Calyam of the University of Missouri at Columbia.  Chip talked about moving beyond today’s Internet with its physical routers, switches and exchange points to a software defined infrastructure that is sliced and virtualized.  Eric Keller talked about stateless network functions that make networks more resilient and easier to manage.  Prasad spoke on the importance of edge computing and the division of tasks between the edge and the cloud for real-time applications and for the Internet of Things.

Jun Li, Prasad Calyam and Eric Keller make presentations at the GRW.

Jun Li, Prasad Calyam and Eric Keller make presentations at the GRW.

The morning also included presentations by the GRW host Jun Li and his colleague Ram Durairajan on their research projects including a network security research project called Drawbridge that uses GENI for experimentation and evaluation.

Two tracks of tutorials were offered the afternoon of November 3.  A beginners track consisted of tutorials on getting started with GENI and an introduction to running software defined networking (SDN) experiments on GENI.  A parallel track for those with some prior experience with GENI consisted of two security tutorials that used GENI wired and wireless resources.

Three tutorials were offered on Saturday November 4: Building a load adaptive intrusion detection system on GENI using Network Function Virtualization, covert channel creation and detection, and edge-cloud experiments using GENI and the Chameleon cloud testbed.

Xenia Mountrouidou, Joon Yee Chuah and Violet Syrotiuk lead tutorials at the GRW.

Xenia Mountrouidou, Joon Yee Chuah and Violet Syrotiuk lead tutorials at the GRW.

Instructors at the GRW were Nabeel Akhtar (Boston University), Joon Yee Chuah (Texas Advanced Computing Center), Fraida Fund (New York University), Xenia Mountrouidou (College of Charleston), Paul Ruth (RENCI), Violet Syrotiuk (Arizona State University) and Vic Thomas (BBN Technologies).

For details of this workshop please see http://geni.uoregon.edu.

If you are interested in hosting a GRW, please contact Violet Syrotiuk or Abraham Matta.  These 1–2 day workshops deliver a focused introduction to GENI, using lectures and hands-on tutorials, that new users find most helpful to begin using GENI effectively in their work.