The CFDE Workbench - Abstract The NIH Common Fund (CF) programs have produced transformative datasets, databases, methods, bioinformatics tools and workflows that are significantly advancing biomedical research in the United States and worldwide. Currently, CF programs are mostly isolated. However, integrating data from across CF programs has the potential for synergistic discoveries. In addition, since CF programs have a time limit of 10 years, sustainability of the widely used CF digital resources after the programs expire is critical. To address these challenges, the NIH established the Common Fund Data Ecosystem (CFDE) program which has been recently approved to continue to its second new phase. For the second phase of the CFDE, this project will establish the Data Resource Center (DRC) and the Knowledge Center (KC). Our efforts will culminate in producing The CFDE Workbench which will be composed of three main products: the CFDE information portal, the CFDE data resource portal, and the CFDE knowledge portal. These three web portals will be full-stack web-based applications with a backend database and will be integrated into one public site. The CFDE information portal will be the entry point to the other two portals. It will contain information about the CFDE in a dedicated About page, information about each participating and non-participating CF program, information about each data coordination center (DCC), a link to a catalog of CF datasets, and a link to a catalog of CF tools and workflows, news, events, funding opportunities, standards and protocols, educational programs and opportunities, social media feeds, and publications. The CFDE data resource portal will contain metadata, data, workflows, and tools which are the products of the CF programs, and their data coordination centers (DDCs). We will adopt the C2M2 data model for storing information about metadata describing DCC datasets. We will also archive relatively small omics datasets that do not have a home in widely established repositories and do not require PHI protection. In addition, we will expand the cataloging to CF tools, APIs, and workflows. Importantly, we will develop a search engine that will index and present results from all these assembled digital assets. In addition, continuing the work established in the CFDE pilot phase, users of the data portal will be able to fetch identified datasets through links provided by the DCCs via the DRS protocol. This will include links to raw and processed data. The CFDE knowledge portal will provide access to CF programs processed data in various formats including: 1) knowledge graph assertions; 2) gene, drug, metabolite, and other set libraries; 3) data matrices ready for machine learning and other AI applications; 4) signatures; and 5) bipartite graphs. In addition, the extract, transform, and load (ETL) scripts to process the data into these formats will be provided. Since such processed data is relatively small, we will archive and serve this processed data, mint it with unique IDs, and serve it via APIs. In addition, we will develop workflows that will demonstrate how the processed data can be harmonized. At the same time, we will document APIs from all CF DCCs and provide example Jupyter Notebooks that demonstrate how these datasets can be accessed, processed, and combined for integrative omics analysis. For the knowledge portal we will also develop a library of tools that utilize these processed datasets. These tools will have some uniform requirements enabling a plug-and-play architecture. To achieve these goals, we will work collaboratively with the other CFDE newly established centers, the participating CFDE DCCs, the CFDE NIH team, and relevant external entities and potential consumers of these three software products. These interactions will be achieved via face-to-face meetings, virtual working groups meeting, one-on-one meetings, Slack, GitHub, project management software, and e-mail exchange. Via these interactions,