SC06: Difference between revisions

From ReddNet
Jump to navigation Jump to search
 
Line 22: Line 22:
of disk servers with an aggregate transfer rate of 3-4 GBytes/sec.  The plot below shows  
of disk servers with an aggregate transfer rate of 3-4 GBytes/sec.  The plot below shows  
that we were able to sustain traffic at nearly 3.3 GBytes/sec for a two hour period (and we could
that we were able to sustain traffic at nearly 3.3 GBytes/sec for a two hour period (and we could
have gone over for several hours more):
have continued several hours more):


[[Image:35Gbs.png]]
[[Image:35Gbs.png]]

Latest revision as of 16:17, 30 November 2006

Meetings

  • Discussion with ORNL and Ultralight to set up REDDnet nodes at ORNL
  • People we should think about meeting with
    • Jackie Chen and/or her group
    • Tony Messacappa

Talks and Demonstrations

L-Store: Storage Virtualization for the Open Science Grid

Alan Tackett, Vanderbilt University/ACCRE

L-Store allows for distributed storage devices (which can be distributed in a wide geographic region) to function as a single file system. Specifically, L-Store provides a distributed namespace for storing arbitrary sized data objects. It is agnostic to the data transfer and storage mechanism, although currently only IBP is supported. L-Store utilizes a chord based DHT implementation to provide metadata scalability. Multiple metadata servers can be used to increase reliability and fault tolerance, and real time addition and deletion of metadata server nodes is allowed. L-Store supports Weaver Erasure encoding of data, so that files can be stored across multiple storage devices and can be recovered even if some of these devices fail.

L-Store/IBP is highly performant, in our booth we will be demonstrating sustained writes to a rack of disk servers with an aggregate transfer rate of 3-4 GBytes/sec. The plot below shows that we were able to sustain traffic at nearly 3.3 GBytes/sec for a two hour period (and we could have continued several hours more):

Error creating thumbnail: Unable to save thumbnail to destination


At SC06, we successfully filled a 10 GigE pipe with traffic from/to Caltech (Pasadena) to our depot on the floor at Tampa:

Error creating thumbnail: Unable to save thumbnail to destination

Nevoa Storage System: Managing scalable storage virtualization infrastructure

Hunter Hagewood, Nevoa Networks The deployment of distributed systems both in wide and local area networks requires a more extensive set of management tools in order to attain the same levels of realibility typically found in self-contained systems. Nevoa Networks currently offers such tools through its Nevoa StorCore product line. Nevoa StorCore allows you to easily configure, monitor, share, and manage the storage resources of your logistical network. Partitioning pooled storage capacity, defining realiability policies, automating fault detection decisions, tiering or peering with other logistical networks are just some of the functionalities Nevoa StorCore offers. This demonstration will include a walk through of Nevoa StorCore's features, examples of national and international deployments and an introduction to Nevoa Explorer Personal Edition, the client component of the Nevoa Storage System for end-users.

How to create a multi-user, wide area facility for data intensive visualization

Contact: Jian Huang, University to Tennessee <huang@cs.utk.edu> Participants: VU, ORNL

Contrary to conventional wisdom, we show that a distributed collection of heterogeneous computing resources can be used effectively to create a scalable infrastructure for high performance, data intensive visualization. In this demonstration, we will show how a large scale visualization of cutting edge simulation datasets can be shared among multiple, distributed users working on the same pool of distributed resources. The demonstration will use IBP depots, equipped with our Visualization Cookbook Library (VCL), at the ACCRE booth, at other booths on the SC06 show floor and on REDDnet's wide area infrastructure. As shown, good performance can be attained with only a standard Internet connection and in spite of the presence of several other simultaneous users.

SRM

Contact: Alex Sim of LBL <ASim@lbl.gov>

This is only for SRM v2.2 implementations with the updated WSDL. We'll use srm-tester to make put/put_done/bring-onling/get/release/remove,and the same with space reservation (with space token). In addition, we'll have srmLs. Then, we'll have srmCopy for a remote copy from a gridftp server and from another SRM server. Lastly, we'll have two SRMs involved in a coordinated remote copy by get_1/put_2/gridftp_between_1_2/release_1/put_done_2/srmLs_2/remove_2.

All tests will be repeated, and will not leave any left-over files by removing the previous test. We'll prepare some gui to show the in-between progresses, and the result on the GUI itself as well as on the web.

All current 6 implementations on 7 sites are following:

  • CERN - DPM
  • FNAL - dCache
  • INFN - StoRM
  • LBNL - SRM
  • RAL, CERN - CASTOR
  • VU - LStore

Any non-implemented and un-supported methods in a site will be screened in advance and will be skipped for the demo. The servers are expected to be up and running during the SC06. (11/12-11/16/2006)

I'll need to ask you your institutional logo and project title, etc when we prepare the slides, etc. If you do not want to participate in the SC06 demo, please let me know.

Schedule of Talks and Demonstrations

Monday

Tuesday

  • 1:00-2:00 -- Multi-user, Data Intensive Visualization

Wednesday

  • 1:00-2:00 -- Multi-user, Data Intensive Visualization

Thursday

  • 1:00-2:00 -- Multi-user, Data Intensive Visualization

Computing resources used in demonstrations

Vanderbilt/ACCRE Booth

Thankyous.jpg

100 TBytes of online disk storage, 40 Gbs Networking, and a 96 CPU compute cluster

  • 40 Capricorn Technologies 3 TB disk servers with dual-core Athlon processors.
  • Foundry sx 800 Switch with 192 1-GigE ports and 4 10-GigE ports.
  • Four 10-GigE Foundry ports connected to the showroom floor (SCinet) and the external network.
  • A 96 CPU "client" cluster consisting of:
    • 2 dual-dual-core opteron servers
    • 9 dual-CPU operton servers
    • 14 IBM JS20 dual-CPU PowerPC blades
    • 7 IBM x336 servers (dual-CPU Intel)
    • 14 IBM x335 servers (dual-CPU Intel)

Collaborating booths at SC06

Current REDDnet nodes, national and international

  • ANSP - Academic Network at São Paulo (Brazil)
  • California Institute of Technology
  • Fermi National Accelerator Laboratory
  • Universidade do Estado do Rio de Janeiro (Brazil)
  • University of Florida
  • University of Michigan
  • Vanderbilt University