San Diego Supercomputer Center
   HOME
*



picture info

San Diego Supercomputer Center
The San Diego Supercomputer Center (SDSC) is an organized research unit of the University of California, San Diego (UCSD). SDSC is located at the UCSD campus' Eleanor Roosevelt College east end, immediately north the Hopkins Parking Structure. The current SDSC director is Michael L. Norman, UCSD physics professor, who succeeded noted grid computing pioneer Francine Berman. He was named director September 2010, having been interim director more than a year. SDSC was founded in 1985 and describes its mission as "developing and using technology to advance science." The SDSC was one of the five original NSF Supercomputing Centers and the National Science Foundation (NSF) continues to be a primary funder of the SDSC. Its research pursuits are high performance computing, grid computing, computational biology, geoinformatics, computational physics, computational chemistry, data management, scientific visualization, and computer networking. SDSC computational biosciences contribu ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




SDSC Logo
SDSC may refer to: * San Diego Supercomputer Center * Satish Dhawan Space Centre * Strategic and Defence Studies Centre * Secure Digital Standard Capacity Secure Digital, officially abbreviated as SD, is a proprietary non-volatile flash memory card format developed by the SD Association (SDA) for use in portable devices. The standard was introduced in August 1999 by joint efforts between SanDis ... card * São Carlos Airport (ICAO-Code) {{disambig ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Protein Data Bank
The Protein Data Bank (PDB) is a database for the three-dimensional structural data of large biological molecules, such as proteins and nucleic acids. The data, typically obtained by X-ray crystallography, NMR spectroscopy, or, increasingly, cryo-electron microscopy, and submitted by biologists and biochemists from around the world, are freely accessible on the Internet via the websites of its member organisations (PDBe, PDBj, RCSB, and BMRB). The PDB is overseen by an organization called the Worldwide Protein Data Bank, wwPDB. The PDB is a key in areas of structural biology, such as structural genomics. Most major scientific journals and some funding agencies now require scientists to submit their structure data to the PDB. Many other databases use protein structures deposited in the PDB. For example, SCOP and CATH classify protein structures, while PDBsum provides a graphic overview of PDB entries using information from other sources, such as Gene ontology. History Two force ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Cyberinfrastructure
United States federal research funders use the term cyberinfrastructure to describe research environments that support advanced data acquisition, data storage, data management, data integration, data mining, data visualization and other computing and information processing services distributed over the Internet beyond the scope of a single institution. In scientific usage, cyberinfrastructure is a technological and sociological solution to the problem of efficiently connecting laboratories, data, computers, and people with the goal of enabling derivation of novel scientific theories and knowledge. Origin The term National Information Infrastructure had been popularized by Al Gore in the 1990s. This use of the term "cyberinfrastructure" evolved from the same thinking that produced Presidential Decision Directive NSC-63 on Protecting America's Critical Infrastructures (PDD-63). PDD-63 focuses on the security and vulnerability of the nation's "cyber-based information systems" as well ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


E-Science
E-Science or eScience is computationally intensive science that is carried out in highly distributed network environments, or science that uses immense data sets that require grid computing; the term sometimes includes technologies that enable distributed collaboration, such as the Access Grid. The term was created by John Taylor, the Director General of the United Kingdom's Office of Science and Technology in 1999 and was used to describe a large funding initiative starting in November 2000. E-science has been more broadly interpreted since then, as "the application of computer technology to the undertaking of modern scientific investigation, including the preparation, experimentation, data collection, results dissemination, and long-term storage and accessibility of all materials generated through the scientific process. These may include data modeling and analysis, electronic/digitized laboratory notebooks, raw and fitted data sets, manuscript production and draft versions, pre-p ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Supercomputer Sites
A supercomputer is a computer with a high level of performance as compared to a general-purpose computer. The performance of a supercomputer is commonly measured in floating-point operations per second (FLOPS) instead of million instructions per second (MIPS). Since 2017, there have existed supercomputers which can perform over 1017 FLOPS (a hundred quadrillion FLOPS, 100 petaFLOPS or 100 PFLOPS). For comparison, a desktop computer has performance in the range of hundreds of gigaFLOPS (1011) to tens of teraFLOPS (1013). Since November 2017, all of the world's fastest 500 supercomputers run on Linux-based operating systems. Additional research is being conducted in the United States, the European Union, Taiwan, Japan, and China to build faster, more powerful and technologically superior Exascale computing, exascale supercomputers. Supercomputers play an important role in the field of computational science, and are used for a wide range of computationally intens ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

National Digital Information Infrastructure And Preservation Program
The National Digital Information Infrastructure and Preservation Program (NDIIPP) of the United States was an archival program led by the Library of Congress to archive and provide access to digital resources. The program convened several working groups, administered grant projects, and disseminated information about digital preservation issues. The U.S. Congress established the program in 2000, and official activity specific to NDIIPP itself wound down between 2016 and 2018. The Library was chosen because of its role as one of the leading providers of high-quality content on the Internet. The Library of Congress has formed a national network of partners dedicated to preserving specific types of digital content that is at risk of loss. In July 2010, the Library launched a National Digital Stewardship Alliance (NDSA) to extend the work of NDIIPP to more institutions. The organization, which has been hosted by the Digital Library Federation since January 2016, focuses on several g ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

National Digital Library Program
The Library of Congress National Digital Library Program (NDLP) is assembling a digital library of reproductions of primary source materials to support the study of the history and culture of the United States. Begun in 1995 after a five-year pilot project, the program began digitizing selected collections of Library of Congress archival materials that chronicle the nation's rich cultural heritage. In order to reproduce collections of books, pamphlets, motion pictures, manuscripts and sound recordings, the Library has created a wide array of digital entities: bitonal document images, grayscale and color pictorial images, digital video and audio, and searchable e-texts. To provide access to the reproductions, the project developed a range of descriptive elements: bibliographic records, finding aids, and introductory texts and programs, as well as indexing the full texts for certain types of content. The reproductions were produced with a variety of tools: image scanners, digit ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


RIPE Atlas
Réseaux IP Européens (RIPE, French for "European IP Networks") is a forum open to all parties with an interest in the technical development of the Internet. The RIPE community's objective is to ensure that the administrative and technical coordination necessary to maintain and develop the Internet continues. It is not a standards body like the Internet Engineering Task Force (IETF) and does not deal with domain names like ICANN. RIPE is not a legal entity and has no formal membership. This means that anybody who is interested in the work of RIPE can participate through mailing lists and by attending meetings. RIPE has a chair to keep an eye on work between RIPE meetings and to act as its external liaison. Rob Blokzijl, who was instrumental in the formation of RIPE, was the initial chair and remained in that position until 2014, when he appointed Hans Petter Holen as his successor. The RIPE community interacts via RIPE Mailing Lists, RIPE Working Groups, and RIPE Meetings. Al ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




PlanetLab
PlanetLab was a group of computers available as a testbed for computer networking and distributed systems research. It was established in 2002 by Prof. Larry L. Peterson and Prof. David Culler, and as of June 2010, it was composed of 1090 nodes at 507 sites worldwide. Each research project had a "slice", or virtual machine access to a subset of the nodes. Accounts were limited to persons affiliated with corporations and universities that hosted PlanetLab nodes. However, a number of free, public services have been deployed on PlanetLab, including CoDeeN CoDeeN is a proxy server system created at Princeton University in 2003 and deployed for general use on PlanetLab. It operates as per the following: # Users set their internet caches to a nearby high bandwidth proxy that participates in the syst ..., the Coral Content Distribution Network, and Open DHT. Open DHT was taken down on 1 July 2009. PlanetLab was officially shut down in May 2020 but continues in Europe. References ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Rocks Cluster Distribution
Rocks Cluster Distribution (originally NPACI Rocks) is a Linux distribution intended for high-performance computing (HPC) clusters. It was started by National Partnership for Advanced Computational Infrastructure and the San Diego Supercomputer Center (SDSC) in 2000. It was initially funded in part by an NSF grant (2000–07), but was funded by the follow-up NSF grant through 2011. Distribution Rocks was initially based on the Red Hat Linux (RHL) distribution, however modern versions of Rocks were based on CentOS, with a modified Anaconda installer that simplifies mass installation onto many computers. Rocks includes many tools (such as Message Passing Interface (MPI)) which are not part of CentOS but are integral components that make a group of computers into a cluster. Installations can be customized with additional software packages at install-time by using special user-supplied CDs (called "Roll CDs"). The "Rolls" extend the system by integrating seamlessly and automat ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Argonne National Laboratory
Argonne National Laboratory is a science and engineering research United States Department of Energy National Labs, national laboratory operated by University of Chicago, UChicago Argonne LLC for the United States Department of Energy. The facility is located in Lemont, Illinois, outside of Chicago, and is the largest national laboratory by size and scope in the Midwest. Argonne had its beginnings in the Metallurgical Laboratory of the University of Chicago, formed in part to carry out Enrico Fermi's work on nuclear reactors for the Manhattan Project during World War II. After the war, it was designated as the first national laboratory in the United States on July 1, 1946. In the post-war era the lab focused primarily on non-weapon related nuclear physics, designing and building the first power-producing nuclear reactors, helping design the reactors used by the United States' nuclear navy, and a wide variety of similar projects. In 1994, the lab's nuclear mission ended, and today ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]