2020
DOI: 10.1051/epjconf/202024504042
|View full text |Cite
|
Sign up to set email alerts
|

Moving the California distributed CMS XCache from bare metal into containers using Kubernetes

Abstract: The University of California system maintains excellent networking between its campuses and a number of other Universities in California, including Caltech, most of them being connected at 100 Gbps. UCSD and Caltech Tier2 centers have joined their disk systems into a single logical caching system, with worker nodes from both sites accessing data from disks at either site. This successful setup has been in place for the last two years. However, coherently managing nodes at multiple physical locations is not tri… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

3
3

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 8 publications
(10 reference statements)
0
5
0
Order By: Relevance
“…By 2028, the community expects the data volume to increase by thirty fold [1]. We observe that a significant portion of the popular datasets are shared among users in the same geographical region [1], which suggests that regional data storage caches could reduce data access latency by holding popular datasets closer to user analyses [2][3][4][5][6][7][8][9]. In-network cache or regional data caching mechanism [6][7][8][9] has been deployed in Southern California for the US CMS, one of the LHC experiment.…”
Section: Introductionmentioning
confidence: 93%
See 1 more Smart Citation
“…By 2028, the community expects the data volume to increase by thirty fold [1]. We observe that a significant portion of the popular datasets are shared among users in the same geographical region [1], which suggests that regional data storage caches could reduce data access latency by holding popular datasets closer to user analyses [2][3][4][5][6][7][8][9]. In-network cache or regional data caching mechanism [6][7][8][9] has been deployed in Southern California for the US CMS, one of the LHC experiment.…”
Section: Introductionmentioning
confidence: 93%
“…We observe that a significant portion of the popular datasets are shared among users in the same geographical region [1], which suggests that regional data storage caches could reduce data access latency by holding popular datasets closer to user analyses [2][3][4][5][6][7][8][9]. In-network cache or regional data caching mechanism [6][7][8][9] has been deployed in Southern California for the US CMS, one of the LHC experiment. The caching approach improves overall application performance by decreasing data access latency and increasing data access throughput.…”
Section: Introductionmentioning
confidence: 96%
“…• RedirCache -Reading data over SoCal Cache [20] Redirector distributed between 2 Sites: Caltech and UCSD. Tests (RedirCache1) were repeated to highlight identified issues below.…”
Section: Testing Environmentmentioning
confidence: 99%
“…When a user's computing job needs a file from SoCal Repo, the system first looks up the location of the file using the "Trivial File Catalogue" (TFC) [8,9]. Following the established convention for the tiered storage system, the data files are grouped into the namespace for the local cache nodes and the TFC points to a "local redirector" in XRootD where the "local redirector" knows all regional caches.…”
Section: Introductionmentioning
confidence: 99%