Tag Archives: supercomputing

Live-Blog: Overview of High Performance Computing by Dr. Vipin Chaudhary

(This is a live-blog of Dr. Vipin Chaudhary talk on Trends in High Performance Computing, organized by the IEEE Pune sub-section. Since this is being typed while the talk is going on, it might not be as well organized, or as coherent as other PuneTech articles. Also, links will usually be missing.)

Dr. Vipin Chaudhary, CEO of CRL
Live-blog of a talk by Dr. Vipin Chaudhary, CEO of CRL, on High Performance Computing at Institute of Engineers, Pune. CRL are the makers of Eka, one of the world's fastest privately funded supercomputers. For more information about HPC and CRL, click on the photo above.
Myths about High Performance Computing:

  • Commonly associated with scientific computing
  • Only used for large problems
  • Expensive
  • Applicable to niche areas
  • Understood by only a few people
  • Lots of servers and storage
  • Difficult to use
  • Not scalable and reliable

This is not the reality. HPC is:

  • Backbone for national development
  • Will enable economic growth. Everything from toilets to potato chips are designed using HPC
  • Lots of supercomputing is throughput computing – i.e. used to solve lots of small problems
  • “Mainstream” businesses like Walmart, and entertainment companies like Dreamworks Studioes use HPC.
  • _(and a bunch of other reasons that I did not catch)

China is really catching up in the area of HPC. And Vipin correlates China’s GDP with the development of supercomputers in China. Point: technology is a driver for economic growth.  We need to also invest in this.

Problems solved using HPC:

  • Movie making (like avatar)
  • Real time data analysis
    • weather forecasting
    • oil spill impact analysis
    • forest fire tracking and monitoring
    • biological contamination prediction
  • Drug discover
    • reduce experimental costs through simulations
  • Terrain modeling for wind-farms
    • e.g. optimized site selection, maintenance scheduling
    • and other alternate energy sources
  • Geophysical imaging
    • oil industry
    • earthquake analysis
  • Designing airplanes (Virtual wind tunnel)

Trends in HPC.

The Manycore trend.

Putting many CPUs inside a single chip. Multi-core is when you have a few cores, manycore is when you have many, many cores. This has challenges. Programming manycore processors is very cumbersome. Debugging is much harder. e.g. if you need to get good performance out of these chips then you need to do parallel, assembly programming. Parallel programming is hard. Assembly programming is hard. Both together will kill you.

This will be one of the biggest challenges in computer science in the near future. A typical laptop might have 8 to 10 processses running concurrently. So there is automatic parallelism, as long as number of cores is less than 10. But as chips get 30, 40 cores or more, individual processes will need to be parallel. This will be very challenging.

Oceans of Data but the Pipes are Skinny

Data is growing fast. In sciences, humanities, commerce, medicine, entertainment. The amount of information being created in the world is huge. Emails, photos, audio, documents etc. Genomic data (bio-informatics) data is also huge.

Note: data is growing way, way faster than Moore’s law!

Storing things is not a problem – we have lots of disk space. Fetching and finding stuff is a pain.

Challenges in data-intensive systems:

  • Amount of data to be accessed by the application is huge
  • This requires huge amounts of disk, and very fat interconnects
  • And fast processors to process that data

Conventional supercomputing was CPU bound. Now, we are in the age of data-intensive supercomputing. Difference: old supercomputing had storage elsewhere (away from the processor farm). Now the disks have to be much closer.

Conventional supercomputing was batch processed. Now, we want everything in real-time. Need interactive access. To be able to run analytic and ad hoc queries. This is a new, and difficult challenge.

While Vipin was faculty in SUNY Buffalo, they started an initiative for data-intensive discovery initiative (Di2). Now, CRL is participating. Large, ever-changing data sets. Collecting and maintaining data is of course major problem, but primary focus of Di2 is to search in this data. e.g. security (find patterns in huge logs user actions). This requires a new, different architecture from traditional supercomputing, and the resulting Di2 system significantly outperforms the traditional system.

This also has applications in marketing analysis, financial services, web analytics, genetics, aerospace, and healthcare.

High Performance Cloud Services at CRL

Cloud computing makes sense. It is here to stay. But energy consumption of clouds is a problem.

Hence, CRL is focusing on a green cloud. What does that mean?

Data center optimization:

  • Power consumption optimization on hardware
  • Optimization of the power system itself
  • Optimized cooling subsystem
  • CFD modeling of the power consumption
  • Power dashboards

Workflow optimization (reduce computing resource consumption via efficiencies):

  • Cloud offerings
  • Virtualizations
  • Workload based power management
  • Temperature aware distribution
  • Compute cycle optimization

Green applications being run in CRL

  • Terrain modeling
  • Wind farm design and simulation
  • Geophysical imaging
  • Virtual wind tunnel

Summary of talk

  • Manycore processors are here to stay
    • Programmability have to improve
    • Must match application requirements to processor architecture (one size does not fit all)
  • Computation has to move to where the data is, and not vice versa
  • Data scale is the biggest issue
    • must co-locate data with computing
  • Cloud computing will continue to grow rapidly
    • Bandwidth is an issue
    • Security is an issue
    • These issues need to be solved

Session on High Performance Computing – Dr. Vipin Chaudhari, CEO CRL

What: IEEE Pune presents a session on High Performance Computing, by Dr. Vipin Chaudhary, CEO of the Computation Research Laboratories (CRL) the makers of the Eka supercomputer
When: Saturday, 14 August, 5pm-7pm
Where: Institution of Engineers, Shivajinagar, JM Road, Opposite Modern Cafe
Registration and Fees: This event is free for all. Register by sending mail to IEEE125.Pune.Symposium@gmail.com.
Details: Contact Amey Asodekar 020-6606-8494

Dr. Vipin Chaudhary, CEO of CRL
Dr. Vipin Chaudhary, CEO of CRL, will give a talk on High Performance Computing at Institute of Engineers, Pune. CRL are the makers of Eka, one of the world's fastest privately funded supercomputers. For more information about HPC and CRL, click on the photo above.

Computation Research Laboratories (CRL) is the Pune-based company from the Tatas which built the Eka supercomputer. Eka was the 4th fastest when it launched a few years back, but has now dropped to 33rd; nevertheless, it remains one of the fastest private (i.e. not funded by any government) supercomputers in the world.

Earlier this year, Dr. Vipin Chaudhary took over as the CEO of CRL. I assume this marks a change in direction for CRL. Earlier, the focus was on building Eka, which required lots of cutting edge research in hardware, software, and facilities among other things. During that phase it was developed and run by academics (CRL was started with the help of Dr. Narendra Karmarkar, and most of the senior executives in CRL were ex-IIT-Bombay professors. Now, however, it is likely that they’re looking for a return on the investment, and would like to start marketing high performance computing services using Eka. They have a team working on high performance infrastructure and applications using the Eka hardware, and being a purely private company, are in a unique position to offer their hardware, software and services to companies who might be interested in supercomputing applications (think airplane design and modeling (e.g. somebody like Boeing), or car design (e.g. for in-house use like Tata Motors)). Dr. Chaudhary, who relocated from the US for this role, has been earlier involved with two startups (Corio, acquired by IBM in 2005) and Cradle Technologies (2000-2003), in addition to being an associate professor at SUNY Buffalo. Thus the credentials he brings, specifically his strong technical and business background in this area, are impressive.

CRL, is working on some of the most complex technologies in Pune. For that reason alone, any techie in Pune should be interested in this talk.

Narendra Karmarkar’s recent research

After working on the Interior Point Method, Karmarkar worked on a new architecture for supercomputing, based on concepts from projective geometry. Currently, he is synthesizing these concepts with some new ideas he calls sculpturing free space (a non-linear analogue of what has has popularly been described as folding the perfect corner). This approach allows him to extend this work to the physical design of machines. He is now publishing updates on his recent work on PuneTech.

Upcoming Presentations:

Narendra will be presenting new paradigm as follows:

16th July 2008: “Vacuum nanoelectronics devices from the perspective of optimization theory“, The 21st International Vacuum Nanoelectronics Conference in Wroc?aw, Poland, on July 13-17, 2008.

25th July 2008: “Seminar on Massively Parallel Systems and Global Optimization“, Computation Research in Boston, at  Massachusetts Institute of Technology.