architect

Results 1 - 25 of 1907Sort Results By: Published Date | Title | Company Name
By: Cray     Published Date: Jul 02, 2015
The Cray XC series is a distributed memory system developed as part of Cray’s participation in the Defense Advanced Research Projects Agency’s (DARPA) High Productivity Computing System (HPCS) program. Previously codenamed “Cascade,” the Cray XC system is capable of sustained multi-petaflops performance and features a hybrid architecture combining multiple processor technologies, a high performance network and a high performance operating system and programming environment.
Tags : 
     Cray
By: Intel     Published Date: Aug 06, 2014
Designing a large-scale, high-performance data storage system presents significant challenges. This paper describes a step-by-step approach to designing such a system and presents an iterative methodology that applies at both the component level and the system level. A detailed case study using the methodology described to design a Lustre storage system is presented.
Tags : intel, high performance storage
     Intel
By: InsideHPC Special Report     Published Date: Aug 17, 2016
A single issue has always defined the history of HPC systems: performance. While offloading and co-design may seem like new approaches to computing, they actually have been used, to a lesser degree, in the past as a way to enhance performance. Current co-design methods are now going deeper into cluster components than was previously possible. These new capabilities extend from the local cluster nodes into the “computing network.”
Tags : 
     InsideHPC Special Report
By: IBM     Published Date: Jun 05, 2014
As new research and engineering environments are expanded to include more powerful computers to run increasingly complex computer simulations, the management of these heterogeneous computing environments continues to increase in complexity as well. Integrated solutions that include the Intel® Many Integrated Cores (MIC) architecture can dramatically boost aggregate performance for highly-parallel applications.
Tags : 
     IBM
By: SGI     Published Date: Mar 03, 2015
The SGI UV system is uniquely suited for bioinformatics and genomics by providing the computational capabilities and global shared memory architecture needed for even the most demanding sequencing and analytic tasks, including post sequencing and other data intensive workflows. Because of the systems outstanding speed and throughput, genomics researchers can perform very large jobs in less time, realizing a dramatically accelerated time-to-solution. And best of all, they can explore avenues of research that were computationally beyond the reach of HPC systems lacking the power and in-memory capabilities of the SGI UV.
Tags : 
     SGI
By: Penguin Computing     Published Date: Mar 23, 2015
The Open Compute Project, initiated by Facebook as a way to increase computing power while lowering associated costs with hyper-scale computing, has gained a significant industry following. While the initial specifications were created for a Web 2.0 environment, Penguin Computing has adapted these concepts to create a complete hardware ecosystem solution that addresses these needs and more. The Tundra OpenHPC system is applicable to a wide range of HPC challenges and delivers the most requested features for data center architects.
Tags : penguin computing, open computing, computing power, hyper-scale computing, tundra openhpc
     Penguin Computing
By: IBM     Published Date: May 20, 2015
The latest generation of highly scalable HPC clusters is a game changer for design optimization challenges. HPC clusters, built on a modular, multi-core x86 architecture, provide a cost effective and accessible platform on which to conduct realistic simulation compared with the “big iron” HPC systems of the past or with the latest workstation models. This paper provides 6 steps to making clusters a reality for any business.
Tags : 
     IBM
By: Seagate     Published Date: Sep 30, 2015
In all things, but especially in high-performance computing (HPC), the details matter. In HPC, the most important detail is drive efficiency, especially when analyzing storage performance. This fundamental metric is easy to calculate and can have a huge impact on HPC storage performance. In this guide, we’ll look at why conventional HPC storage architectures do so poorly at drive-level throughput and what that means to an organization’s ability to realize the full potential of its HPC investments. We’ll also explain why ClusterStor™ from Seagate® excels in this arena.
Tags : 
     Seagate
By: Seagate     Published Date: Sep 30, 2015
In all things, but especially in high-performance computing (HPC), the details matter. In HPC, the most important detail is drive efficiency, especially when analyzing storage performance. This fundamental metric is easy to calculate and can have a huge impact on HPC storage performance. In this guide, we’ll look at why conventional HPC storage architectures do so poorly at drive-level throughput and what that means to an organization’s ability to realize the full potential of its HPC investments. We’ll also explain why ClusterStor™ from Seagate® excels in this arena.
Tags : 
     Seagate
By: InsideHPC Special Report     Published Date: Sep 04, 2013
The recently launched Mellanox Connect-IB™ InfiniBand adapter introduced a novel high-performance and scalable architecture for high-performance clusters. The architecture was designed from the ground up to provide high performance and scalability for the largest supercomputers in the world, today and in the future.
Tags : 
     InsideHPC Special Report
By: IBM     Published Date: Nov 14, 2014
The latest generation of highly scalable HPC clusters is a game changer for design optimization challenges. HPC clusters, built on a modular, multi-core x86 architecture, provide a cost effective and accessible platform on which to conduct realistic simulation compared with the “big iron” HPC systems of the past or with the latest workstation models. This paper provides 6 steps to making clusters a reality for any business.
Tags : 
     IBM
By: IBM     Published Date: Feb 13, 2015
Value is migrating throughout the IT industry from hardware to software and services. High Performance Computing (HPC) is no exception. IT solution providers must position themselves to maximize their delivery of business value to their clients – particularly industrial customers who often use several applications that must be integrated in a business workflow. This requires systems and hardware vendors to invest in making their infrastructure “application ready”. With its Application Ready solutions, IBM is outflanking competitors in Technical Computing and fast-tracking the delivery of client business value by providing an expertly designed, tightly integrated and performance optimized architecture for several key industrial applications. These Application Ready solutions come with a complete high-performance cluster including servers, network, storage, operating system, management software, parallel file systems and other run time libraries, all with commercial-level solution s
Tags : 
     IBM
By: Panasas     Published Date: Oct 02, 2014
HPC and technical computing environments require the collection, storage,and transmission of large-scale datasets. To meet these demands, datacenter architects must consider how increasing storage capacity over time will affect HPC workloads, performance, and system availability. While many enterprises have looked to scale-up NAS to meet their storage needs, this approach can lead to data islands that make it difficult to share data. Distributed, scale-out storage was developed to get around the technology limitations of scale-up NAS architectures.
Tags : 
     Panasas
By: Dell and Intel®     Published Date: Mar 30, 2015
Dell has teamed with Intel® to create innovative solutions that can accelerate the research, diagnosis and treatment of diseases through personalized medicine. The combination of leading-edge Intel® Xeon® processors and the systems and storage expertise from Dell create a state-of-the-art data center solution that is easy to install, manage and expand as required. Labelled the Dell Genomic Data Analysis Platform (GDAP), this solution is designed to achieve fast results with maximum efficiency. The solution is architected to solve a number of customer challenges, including the perception that implementation must be large-scale in nature, compliance, security and clinician uses.
Tags : 
     Dell and Intel®
By: Data Direct Networks     Published Date: Dec 31, 2015
Using high performance parallel storage solutions, geologists and researchers can now incorporate larger data sets and execute more seismic and reservoir simulations faster than ever before, enabling higher fidelity geological analysis and significantly reduced exploration risk. With high costs of exploration, oil and gas companies are increasingly turning to high performance DDN storage solutions to eliminate I/O bottlenecks, minimize risk and costs, while delivering a larger number of higher fidelity simulations in same time as traditional storage architectures.
Tags : 
     Data Direct Networks
By: Intel     Published Date: Aug 26, 2019
Distributed cloud architectures will play an important role in Communication Service Providers’ network transformations. The distributed cloud will be made of numerous edge locations, from the customer premises to mobile base stations to central offices, each of which has its benefits and constraints. CSPs will need to consider attributes such as performance, availability, security, integration, and data and traffic volume to optimize the infrastructure for a myriad of workloads. Learn more about Heavy Reading’s survey results on CSPs adopting a distributed cloud strategy and delivering new services in this white paper.
Tags : 
     Intel
By: Intel     Published Date: Aug 26, 2019
5G services will place greater demands on the network in terms of peak data rates, connection density, and more. Learn how CoSPs are getting ready for 5G and how Intel® architecture is helping accelerate network transformation.
Tags : 
     Intel
By: Intel     Published Date: Sep 25, 2019
Distributed cloud architectures will play an important role in Communication Service Providers’ network transformations. The distributed cloud will be made of numerous edge locations, from the customer premises to mobile base stations to central offices, each of which has its benefits and constraints. CSPs will need to consider attributes such as performance, availability, security, integration, and data and traffic volume to optimize the infrastructure for a myriad of workloads. Learn more about Heavy Reading’s survey results on CSPs adopting a distributed cloud strategy and delivering new services in this white paper.
Tags : 
     Intel
By: Intel     Published Date: Sep 25, 2019
5G services will place greater demands on the network in terms of peak data rates, connection density, and more. Learn how CoSPs are getting ready for 5G and how Intel® architecture is helping accelerate network transformation.
Tags : 
     Intel
By: KPMG     Published Date: Jun 06, 2019
Businesses increasingly see cloud as the best route to enhanced agility and digital transformation. But this survey reveals that keeping the lid on cloud costs is far from straightforward. Read this report to discover: • whether businesses are realising the benefits they anticipated from cloud migration • how effective companies are at managing cloud investments • the risks and consequences of getting locked into inefficient architectures.
Tags : 
     KPMG
By: Cisco Umbrella EMEA     Published Date: Aug 08, 2019
“We ruled out Zscaler due to its overly complex architecture and PAC file requirements. The ongoing management would have been a nightmare.” - Marc Tournier, Information Security and Compliance Manager (CISO), Avril Group
Tags : 
     Cisco Umbrella EMEA
By: Black Duck Software     Published Date: Jul 16, 2010
This paper is for IT development executives looking to gain control of open source software as part of a multi-source development process. You can gain significant management control over open source software use in your development organization. Today, many IT executives, enterprise architects, and development managers in leading companies have gained management control over the externally-sourced software used by their application development groups. Download this free paper to discover how.
Tags : open source, development, architects, application development, software development
     Black Duck Software
By: Intel     Published Date: Sep 27, 2019
The stakes are high in today's data centers. Organisations have access to massive quantities of data promising valuable insights and new opportunities for business. But data center architects need to rethink and redesign their system architectures to ingest, store and process all that information. Similarly, application owners need to assess how they can process data more effectively. Those who don't re-architect might find themselves scrambling just to keep from being drowned in a data deluge.
Tags : 
     Intel
By: Intel     Published Date: Sep 30, 2019
Mountains of data promise valuable insights and innovation for businesses that rethink and redesign their system architectures. But companies that don’t re-architect might find themselves scrambling just to keep from being buried in the avalanche of data. The problem is not just in storing raw data, though. For businesses to stay competitive, they need to quickly and cost-effectively access and process all that data for business insights, research, artificial intelligence (AI), and other uses. Both memory and storage are required to enable this level of processing, and companies struggle to balance high costs against limited capacities and performance constraints. The challenge is even more daunting because different types of memory and storage are required for different workloads. Furthermore, multiple technologies might be used together to achieve the optimal tradeoff in cost versus performance. Intel is addressing these challenges with new memory and storage technologies that emp
Tags : 
     Intel
By: Group M_IBM Q4'19     Published Date: Sep 25, 2019
Agile Integration Methodology: Container-Based and Microservices-Aligned Lightweight Integration Runtimes - This paper explores the merits of agile integration architecture--a container-based, decentralized, and microservices-aligned approach for integration solutions that meets the demands of agility, scalability, and resilience required by digital transformation.
Tags : 
     Group M_IBM Q4'19
Start   Previous   1 2 3 4 5 6 7 8 9 10 11 12 13 14 15    Next    End
Search White Papers      

Add White Papers

Get your white papers featured in the insideHPC White Paper Library contact: Kevin@insideHPC.com