Skip to main contentdfsdf

Wong Iseng's List: HPC in the Cloud

      • Lha kok bisa walaupun tidak komunikasi tapi tetep ada degradasi performance ? Virtual Machinenya sendiri bermasalah ?

    • and this is true even when the nodes don’t communicate

    4 more annotations...

    • One of the major questions the study hopes to answer is how well the DOE’s mid-range scientific workloads match up with various cloud architectures and how those architectures could be optimized for HPC applications. Today most public clouds lack the network performance, as well as CPU and memory capacities to handle many HPC codes. The software environment in public clouds also can be at odds with HPC, since little effort has been made to optimize computational performance at the application level. Purpose-built HPC clouds may be the answer, and much of the Magellan effort will be focused on developing these private “science clouds.”
  • Nov 17, 10

    This is the problem and the question. MPI is not good for cloud, and then what ?

    • For the more traditional MPI applications there were significant slowdowns, over a factor of 10,”
    • Since there has been some followup discussion, I wanted to clarify and add some context. The factor of 10 was a comparison between our unvirtualized Magellan hardware and Amazon’s Elastic Compute Cloud (EC2) using m1.xlarge instances. We ran the NERSC6 benchmarks to perform the comparison. For the seven applications we tested, the mean slowdown factor for EC2 relative to Magellan was 10.8. The best application, GAMESS, was 2.7 times slower, while the worst performance was with PARATEC, which was 51.8 times slower. Again, the Magellan results were on unvirtualized hardware with an Infiniband interconnect.
    • How do people actually use multiple machines to solve a problem? – This is really the root question behind all of this work. The first scenario is high-end shared-memory machines (ala Cray supercomputers) and I’m going to eliminate that type of compute from the conversation due to the fact that it simply can’t be well-replicated in the cloud as we currently know it. The far opposite end of the spectrum is “manual” clustering or map reduce – someone figures out a problem they want to solve, divvy’s it up amongst N nodes, and then individually runs a program on each node with the appropriate settings and then manually aggregates the results. This extreme is most likely done by ad-hoc projects or those not familiar with traditional HPC technologies and approaches. Between the two extremes listed, there are Map/Reduce implementations and traditional MPI programs targeted at distributed memory systems.
    • The first scenario is high-end shared-memory machines

    7 more annotations...

    • Microsoft is also providing an Azure resource for scientists that will not require an installation of Windows HPC Server. The service makes the National Center for Biotechnology Information’s BLAST technology, which lets scientists search the human genome, available on Azure. At SC10, Microsoft said it will demonstrate the NCBI BLAST application on Windows Azure performing 100 billion comparisons of protein sequences.
  • Nov 17, 10

    Latest benchmark, seems that EC2 has improved their performance. But does the performance of previous benchmarks as used by Ed Walker and Jeffrey Napper also improved ?

    • The new EC2 cluster compute instance type is an excellent performing cloud server. Performance exceeded that of most of the "bare metal" cloud servers we benchmarked previously. Combined with 10 Gbps non-blocking clustering capabilities, and on-demand deployment & hourly billing, this new instance type provides exceptional value and capabilities for HPC applications.
  • Nov 17, 10

    These are very thorough collection of peformance analysis of cloud computing infrastructure. See especially highlighted papers

    • Evaluated the performance of resources from four production, commercial clouds. We have added to GrenchMark the C-Meter tool for evaluating the performance of cloud resources [3]. We have studied [2,6,9] the performance of resources from four production, commercial clouds: Amazon Elastic Compute Cloud (EC2), Mosso, Elastic Hosts, and GoGrid.
    • Evaluated the variability of the performance delivered by production cloud services. We have collected year-long traces and, based on them, studied [8] the performance of over fifteen operations provided by nine services in two clouds, Amazon Web Services and Google App Engine.

    5 more annotations...

    • During the past few years it has been no secret that EC2 has been best cloud provider for massive scale, but loosely connected scientific computing environments.
    • pleasantly parallel, high-throughput computing workflows.

    1 more annotation...

  • Nov 18, 10

    So they even already made commercial companies ? And I what I can do was only makes a clipping of these HPC in the Cloud stories ? :))

  • Nov 18, 10

    The offer free setup for developer ? Hmmmmm :)

  • Nov 18, 10

    Sudah bukan masalah lagi ? Sudah tidak ada degradasi performa dari MPI ? :)

    • The Magellan Cloud research team at the National Energy Research Scientific Computing Center (NERSC) was one of those beta customers and got a chance to test drive the new EC2 offering prior to this week's official launch. They reported that a series of HPC application benchmarks "ran 8.5 times faster on Cluster Compute Instances for Amazon EC2 than the previous EC2 instance types."
  • Nov 18, 10

    Sudah dijawab dari bulan Juli sama amazon pertanyaannya. Gw aja yang gak mengikuti perkembangan. Belum lagi yang baru di release seakrang pasti lebih bagus hasilnya.

    • “For perspective, in one of our pre-production tests, an 880 server sub-cluster achieved 41.82 TFlops on a LINPACK test run – we’re very excited that Amazon EC2 customers now have access to this type of HPC performance with the low per-hour pricing, elasticity, and functionality they have come to expect from Amazon EC2.” (Peter De Santis, General Manager of Amazon EC2)
1 - 13 of 13
20 items/page
List Comments (0)