Supercomputing in Plain English: Overview

Supercomputing in Plain English: Overview

Supercomputing in Plain English High Throughput Computing Henry Neeman, University of Oklahoma Director, OU Supercomputing Center for Education & Research (OSCER) Assistant Vice President, Information Technology Research Strategy Advisor Associate Professor, Gallogly College of Engineering Adjunct Associate Professor, School of Computer Science Tuesday April 17 2018 This is an experiment! Its the nature of these kinds of videoconferences that FAILURES ARE GUARANTEED TO HAPPEN! NO PROMISES! So, please bear with us. Hopefully everything will work out well enough.

If you lose your connection, you can retry the same kind of connection, or try connecting another way. Remember, if all else fails, you always have the phone bridge to fall back on. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 2 PLEASE MUTE YOURSELF No matter how you connect, PLEASE MUTE YOURSELF, so that we cannot hear you. At OU, we will turn off the sound on all conferencing technologies. That way, we wont have problems with echo cancellation. Of course, that means we cannot hear questions.

So for questions, youll need to send e-mail: [email protected] PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 3 Download the Slides Beforehand Before the start of the session, please download the slides from the Supercomputing in Plain English website: http://www.oscer.ou.edu/education/ That way, if anything goes wrong, you can still follow along with just audio. PLEASE MUTE YOURSELF.

PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 4 Zoom Go to: http://zoom.us/j/979158478 Many thanks Eddie Huebsch, OU CIO, for providing this. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 5

YouTube You can watch from a Windows, MacOS or Linux laptop or an Android or iOS handheld using YouTube. Go to YouTube via your preferred web browser or app, and then search for: Supercomputing InPlainEnglish (InPlainEnglish is all one word.) Many thanks to Skyler Donahue of OneNet for providing this. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 6 Twitch You can watch from a Windows, MacOS or Linux laptop or an

Android or iOS handheld using Twitch. Go to: http://www.twitch.tv/sipe2018 Many thanks to Skyler Donahue of OneNet for providing this. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 7 Wowza #1 You can watch from a Windows, MacOS or Linux laptop using Wowza from the following URL: http://jwplayer.onenet.net/streams/sipe.html If that URL fails, then go to:

http://jwplayer.onenet.net/streams/sipebackup.html Many thanks to Skyler Donahue of OneNet for providing this. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 8 Wowza #2 Wowza has been tested on multiple browsers on each of: Windows 10: IE, Firefox, Chrome, Opera, Safari MacOS: Safari, Firefox Linux: Firefox, Opera Weve also successfully tested it via apps on devices with: Android iOS

Many thanks to Skyler Donahue of OneNet for providing this. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 9 Toll Free Phone Bridge IF ALL ELSE FAILS, you can use our US TOLL phone bridge: 405-325-6688 684 684 # NOTE: This is for US call-ins ONLY. PLEASE MUTE YOURSELF and use the phone to listen. Dont worry, well call out slide numbers as we go. Please use the phone bridge ONLY IF you cannot connect any other way: the phone bridge can handle only 100 simultaneous connections, and we have over 1000 participants.

Many thanks to OU CIO Eddie Huebsch for providing the phone bridge.. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 10 Please Mute Yourself No matter how you connect, PLEASE MUTE YOURSELF, so that we cannot hear you. (For YouTube, Twitch and Wowza, you dont need to do that, because the information only goes from us to you, not from you to us.) At OU, we will turn off the sound on all conferencing technologies. That way, we wont have problems with echo cancellation. Of course, that means we cannot hear questions. So for questions, youll need to send e-mail. PLEASE MUTE YOURSELF.

Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 11 Questions via E-mail Only Ask questions by sending e-mail to: [email protected] All questions will be read out loud and then answered out loud. DONT USE CHAT OR VOICE FOR QUESTIONS! No one will be monitoring any of the chats, and if we can hear your question, youre creating an echo cancellation problem. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 12

Onsite: Talent Release Form If youre attending onsite, you MUST do one of the following: complete and sign the Talent Release Form, OR sit behind the cameras (where you cant be seen) and dont talk at all. If you arent onsite, then PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 13 TENTATIVE Schedule Tue Jan 23: Storage: What the Heck is Supercomputing? Tue Jan 30: The Tyranny of the Storage Hierarchy Part I Tue Feb 6: The Tyranny of the Storage Hierarchy Part II Tue Feb 13: Instruction Level Parallelism Tue Feb 20: Stupid Compiler Tricks

Tue Feb 27: Multicore Multithreading Tue March 6: Distributed Multiprocessing Tue March 13: NO SESSION (Henry business travel) Tue March 20: NO SESSION (OU's Spring Break) Tue March 27: Applications and Types of Parallelism Tue Apr 3: Multicore Madness Tue Apr 10: NO SESSION (Henry business travel) Tue Apr 17: High Throughput Computing Tue Apr 24: GPGPU: Number Crunching in Your Graphics Card Tue May 1: Grab Bag: Scientific Libraries, I/O Libraries, Visualization Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 14 Thanks for helping! OU IT

OSCER operations staff (Dave Akin, Patrick Calhoun, Kali McLennan, Jason Speckman, Brett Zimmerman) OSCER Research Computing Facilitators (Jim Ferguson, Horst Severini) Debi Gentis, OSCER Coordinator Kyle Dudgeon, OSCER Manager of Operations

Ashish Pai, Managing Director for Research IT Services The OU IT network team OU CIO Eddie Huebsch OneNet: Skyler Donahue Oklahoma State U: Dana Brunson Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 15 This is an experiment! Its the nature of these kinds of videoconferences that FAILURES ARE GUARANTEED TO HAPPEN! NO PROMISES! So, please bear with us. Hopefully everything will work out well enough. If you lose your connection, you can retry the same kind of

connection, or try connecting another way. Remember, if all else fails, you always have the phone bridge to fall back on. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 16 Coming in 2018! Coalition for Advancing Digital Research & Education (CADRE) Conference: 17-18 2018 @ Oklahoma State U, Stillwater OK USA Apr https://hpcc.okstate.edu/cadre-conference

Linux Clusters Institute workshops http://www.linuxclustersinstitute.org/workshops/ Introductory HPC Cluster System Administration: May 14-18 2018 @ U Nebraska, Lincoln NE USA Intermediate HPC Cluster System Administration: Aug 13-17 2018 @ Yale U, New Haven CT USA Great Plains Network Annual Meeting: details coming soon Advanced Cyberinfrastructure Research & Education Facilitators (ACI-REF) Virtual Residency Aug 5-10 2018, U Oklahoma, Norman OK USA PEARC 2018, July 22-27, Pittsburgh PA USA https://www.pearc18.pearc.org/ IEEE Cluster 2018, Sep 10-13, Belfast UK https://cluster2018.github.io OKLAHOMA SUPERCOMPUTING SYMPOSIUM 2018, Sep 25-26 2018 @ OU SC18 supercomputing conference, Nov 11-16 2018, Dallas TX USA http://sc18.supercomputing.org/ Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018

17 Outline What is High Throughput Computing? Tightly Coupled vs Loosely Coupled What is Opportunistic Computing? HTCondor Grid Computing Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018

18 What is High Throughput Computing? High Throughput Computing High Throughput Computing (HTC) means getting lots of work done per large time unit (for example, jobs per month). This is different from High Performance Computing (HPC), which means getting a particular job done in less time (for example, calculations per second). Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 20

Throughput vs Performance Throughput is a side effect of how much time your job takes from when you first submit it until it completes. Performance is the factor that controls how much time your jobs takes from when it first starts running until it completes. Example:

You submit a very big job at 1:00am on January 1. It sits in the queue for a while. It starts running at 5:00pm on January 2. It finishes running at 6:00pm on January 2. Its performance is fast; its throughput is slow. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 21 High Throughput on a Cluster? Is it possible to get high throughput on a cluster? Sure it just has to be a cluster that no one else is trying to use! Normally, a cluster that is shared by many users is fully loaded with jobs all the time. So your throughput depends on when you submit your jobs, how big your jobs are, and even how many jobs you submit at a time. Depending on a variety of factors, a job you submit may

wait in the queue for anywhere from seconds to days. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 22 Tightly Coupled vs Loosely Coupled Tightly Coupled vs Loosely Coupled Tightly coupled means that all of the parallel tasks have to advance forward in lockstep, so they have to communicate frequently. Loosely coupled means that the parallel tasks can largely or

completely ignore each other (little or no communication), and they can advance at different rates. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 24 Tightly Coupled Example Consider weather forecasting. You take your simulation domain for example, the continental United States split it up into chunks, and give each chunk to an MPI process. But, the weather in northern Oklahoma affects the weather in southern Kansas. So, every single timestep, the process that contains northern Oklahoma has to communicate with the process that contains southern Kansas, so that the interface between the processes has the same weather at the same time.

Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 25 Tightly Coupled Example OK/KS boundary http://www.caps.ou.edu/wx/p/r/conus/fcst/ Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 26 Loosely Coupled Example

An application is known as embarrassingly parallel, or loosely coupled, if its parallel implementation: 1. can straightforwardly be broken up into roughly equal amounts of work per processor, AND 2. has minimal parallel overhead (for example, communication among processors). We love embarrassingly parallel applications, because they get near-perfect parallel speedup, sometimes with only modest programming effort. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 27 Monte Carlo Methods Monte Carlo is a city in the tiny European country Monaco. People gamble there; that is, they play games of chance, which involve randomness.

Monte Carlo methods are ways of simulating (or otherwise calculating) physical phenomena based on randomness. Monte Carlo simulations typically are embarrassingly parallel. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 28 Monte Carlo Methods: Example Suppose you have some physical phenomenon. For example, consider High Energy Physics, in which we bang tiny particles together at incredibly high speeds. BANG! We want to know, say, the average properties of this phenomenon. There are infinitely many ways that two particles can be banged together. So, we cant possibly simulate all of them.

Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 29 Monte Carlo Methods: Example Suppose you have some physical phenomenon. For example, consider High Energy Physics, in which we bang tiny particles together at incredibly high speeds. BANG! We want to know, say, the average properties of this phenomenon. There are infinitely many ways that two particles can be banged together. So, we cant possibly simulate all of them. Instead, we can randomly choose a finite subset of these infinitely many ways and simulate only the subset. Supercomputing in Plain English: Hi Thruput

Tue Apr 17 2018 30 Monte Carlo Methods: Example Suppose you have some physical phenomenon. For example, consider High Energy Physics, in which we bang tiny particles together at incredibly high speeds. BANG! We want to know, say, the average properties of this phenomenon. There are infinitely many ways that two particles can be banged together. So, we cant possibly simulate all of them. The average of this subset will be close to the actual average. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018

31 Monte Carlo Methods In a Monte Carlo method, you randomly generate a large number of example cases (realizations) of a phenomenon, and then take the average of the properties of these realizations. When the realizations average converges (that is, doesnt change substantially if new realizations are generated), then the Monte Carlo simulation stops. This can also be implemented by picking a high enough number of realizations to be sure, mathematically, of convergence. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 32 MC: Embarrassingly Parallel Monte Carlo simulations are embarrassingly parallel, because

each realization is completely independent of all of the other realizations. That is, if youre going to run a million realizations, then: 1. you can straightforwardly break up into roughly 1M / Np chunks of realizations, one chunk for each of the Np processes, AND 2. the only parallel overhead (for example, communication) comes from tracking the average properties, which doesnt have to happen very often. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 33 Serial Monte Carlo Suppose you have an existing serial Monte Carlo simulation: PROGRAM monte_carlo CALL read_input()

DO realization = 1, number_of_realizations CALL generate_random_realization() CALL calculate_properties() END DO CALL calculate_average() END PROGRAM monte_carlo How would you parallelize this? Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 34 Parallel Monte Carlo: MPI PROGRAM monte_carlo_mpi [MPI startup] IF (my_rank == server_rank) THEN CALL read_input()

END IF CALL MPI_Bcast() number_of_realizations_per_process = & & number_of_realizations / number_of_processes DO realization = 1, number_of_realizations_per_process CALL generate_random_realization() CALL calculate_realization_properties() CALL calculate_local_running_average(...) END DO IF (my_rank == server_rank) THEN [collect properties] ELSE [send properties] END IF CALL calculate_global_average_from_local_averages() CALL output_overall_average(...) [MPI shutdown] END PROGRAM monte_carlo_mpi Supercomputing in Plain English: Hi Thruput

Tue Apr 17 2018 35 Parallel Monte Carlo: HTC Suppose you have an existing serial Monte Carlo simulation: PROGRAM monte_carlo CALL read_input() number_of_realizations_per_job = & & number_of_realizations / number_of_jobs DO realization = 1, number_of_realizations_per_job CALL generate_random_realization() CALL calculate_properties() END DO CALL calculate_average_for_this_job() CALL output_average_for_this_job() END PROGRAM monte_carlo

To parallelize this for HTC, simply submit number_of_jobs jobs, and then at the very end run a little program to calculate the overall average. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 36 What is Opportunistic Computing? Desktop PCs Are Idle Half the Day Desktop PCs tend to be active during the workday. But at night, during most of the year, theyre idle. So were

only getting half their value (or less). Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 38 Supercomputing at Night Many institutions have lots of desktop PCs that are idle during the evening and during intersessions. Wouldnt it be great to put them to work on something useful to the institution? That is: What if they could pretend to be a big supercomputer at night, when theyd otherwise be idle anyway? This is sometimes known as opportunistic computing: When a desktop PC is otherwise idle, you have an opportunity to do number crunching on it.

Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 39 Supercomputing at Night Example SETI the Search for Extra-Terrestrial Intelligence is looking for evidence of green bug-eyed monsters on other planets, by mining radio telescope data. [email protected] runs number crunching software as a screensaver on idle PCs around the world (2+ million PCs in 252 countries): http://setiathome.berkeley.edu/ There are many similar projects:

WorldCommunityGrid, [email protected] (curing human diseases) [email protected] (Laser Interferometer Gravitational wave Observatory) [email protected] (gravitational potential) [email protected] (protein folding) PrimeGrid ClimatePrediction Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 40 BOINC The projects listed on the previous page use a software package named BOINC (Berkeley Open Infrastructure for

Network Computing), developed at the University of California, Berkeley: http://boinc.berkeley.edu/ To use BOINC, you have to insert calls to various BOINC routines into your code. It looks a bit similar to MPI: int main () { /* main */ boinc_init(); boinc_finish(); } /* main */ Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 41 BOINC: a Big Supercomputer

As of yesterday (Mon Apr 16 2018), according to the BOINC website, the following were active: 168,847 volunteers 834,604 computers 24-hour average: 10.115 PetaFLOPS If this were a single machine, on the most recent Top 500 list, itd be the 11th fastest supercomputer in the world. Note to self: If your science isnt easy to fund, do amazing computing, so someone will fund that .

Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 42 HTCondor https://research.cs.wisc.edu/htcondor/ HTCondor is Like BOINC HTCondor steals computing time on existing desktop PCs when theyre idle.

HTCondor runs in background when no one is sitting at the desk. HTCondor allows an institution to get much more value out of the hardware thats already purchased, because theres little or no idle time on that hardware all of the idle time is used for number crunching. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 44 HTCondor is Different from BOINC

To use HTCondor, you dont need to rewrite your software to add calls to special routines; in BOINC, you do. HTCondor works great under Unix/Linux, but less well under Windows or MacOS (more on this presently); BOINC works well under all of them. Its non-trivial to install HTCondor on your own personal desktop PC; its straightforward to install a BOINC application such as [email protected] Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 45 Useful Features of HTCondor

Opportunistic computing: HTCondor steals time on existing desktop PCs when theyre otherwise not in use. HTCondor doesnt require any changes to the software. HTCondor can automatically checkpoint a running job: Every so often, HTCondor saves to disk the state of the job (the values of all the jobs variables, plus where the job is in the program). Therefore, HTCondor can preempt running jobs if more important jobs come along, or if someone sits down at the desktop PC. Likewise, HTCondor can migrate running jobs to other PCs, if someone sits at the PC or if the PC crashes. And, HTCondor can do all of its I/O over the network, so that the job on the desktop PC doesnt consume the desktop PCs local disk.

Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 46 HTCondor Pool @ Your Institution You can deploy an HTCondor pool at your institution! This can provide a significant amount of computing capacity. The hardware and software cost is zero, and the labor cost is modest. When we did this at OU, we had been seeing empirically that lab PCs were available for HTCondor jobs about 80% of the time. So doing this increased bang-for-the-buck by 5x. We stopped doing this when we realized that the lab team was overwhelmed and didnt have time to maintain HTCondor. Supercomputing in Plain English: Hi Thruput

Tue Apr 17 2018 47 HTCondor Limitations The Unix/Linux version has more features than Windows, which is referred to as clipped.

The MacOS version runs natively on BSD Unix, but isnt mature. Under Windows, you cant checkpoint your jobs (though you can pause them). Your code shouldnt be parallel to do opportunistic computing (MPI requires a fixed set of resources throughout the entire run), and it shouldnt try to do any funky communication (for example, opening sockets). For a Red Hat/CentOS Linux HTCondor pool, you have to be able to compile your code with gcc, g++, g77 or gfortran. Also, depending on the PCs that have HTCondor on them, you may have limitations on, for example, how big your jobsSupercomputing RAM footprint canHibe. in Plain English: Thruput Tue Apr 17 2018

48 Running a HTCondor Job Running a job on HTCondor pool is a lot like running a job on a cluster: 1. You compile your code using the compilers appropriate for that resource. 2. You submit a batch script to the HTCondor system, which decides when and where your job runs, magically and invisibly. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 49 Sample HTCondor Batch Script Universe Executable

Notification Notify_User Arguments Input Output Error Log InitialDir Queue = = = = = = = = =

= standard /home/hneeman/NBody/nbody_compiled_for_HTCondor Error [email protected] 1000 100 /home/hneeman/NBody/nbody_input.txt nbody_$(Cluster)_$(Process)_out.txt nbody_$(Cluster)_$(Process)_err.txt nbody_$(Cluster)_$(Process)_log.txt /home/hneeman/NBody/Run001 The batch submission command is condor_submit, used like so: condor_submit nbody.condor Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018

50 Grid Computing What is Grid Computing? The term grid computing is poorly defined, but the best definition Ive seen so far is: a distributed, heterogeneous operating system. A grid can consist of: compute resources; storage resources; networks; data collections; shared instruments; sensor networks; and so much more .... Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018

52 Grid Computing is Like and Unlike ... IBMs website has a very good description of grid computing:

Like the Web, grid computing keeps complexity hidden: multiple users enjoy a single, unified experience. Unlike the Web, which mainly enables communication, grid computing enables full collaboration toward common ... goals. Like peer-to-peer, grid computing allows users to share files. Unlike peer-to-peer, grid computing allows many-to-many sharing not only files but other resources as well. Like clusters and distributed computing, grids bring computing resources together. Unlike clusters and distributed computing, which need physical proximity and operating homogeneity, grids can be geographically distributed and heterogeneous. Like virtualization technologies, grid computing enables the virtualization of IT resources. Unlike virtualization technologies, which virtualize a single system, grid computing enables the virtualization of vast and disparate IT resources. http://www.thocp.net/hardware/grid_computers.htm Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018

53 HTCondor is Grid Computing HTCondor creates a grid out of disparate desktop PCs. (Actually, they dont have to be desktop PCs; they dont even have to be PCs. You can use HTCondor to schedule a cluster, or a big iron supercomputer. But most dont.) From a users perspective, all of the PCs are essentially invisible; the user just knows how to submit a job, and everything happens magically and invisibly, and at some point the job is done and a result appears. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 54 Cloud Computing

The term cloud computing is often used interchangeably with grid computing. Mostly, its come to mean commercial clouds: https://www.sdxcentral.com/articles/ news/microsoft-azure-revenues-gre w-98-in-latest-quarter/2018/02/ Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 55 Cloud Computing is Big Business

AWS had $17.459 billion in sales in 2017, up 43% from $12.219 billion in sales in 2016. Operating income for AWS was up 39% from $3.1 billion in 2016 to $4.3 billion in 2017 [+24.6%]. http://www.businessinsider.com/amazon-web-services-2017-revenue-2018-2 AWS accounted for 73% of [Amazons] net income [in 2017Q4]. http://www.businessinsider.com/amazon-fourth-quarter-2017-earnings-2018-2 Microsoft: Azure up 98%, Surface up 1%, and Windows up 4% https://venturebeat.com/2018/01/31/microsoft-reports-28-9-billion-in-q2-2018-revenue-azur e-up-98-surface-up-1-and-windows-up-4/ IBM posted robust cloud-revenue growth for calendar 2017 of

$17 billion, up 24%, with even stronger cloud growth in the fourth quarter with $5.5 billion in cloud revenue, up 30%. https://www.forbes.com/sites/bobevans1/2018/01/26/amazon-to-become-1-in -cloud-computing-revenue-by-beating-ibms-17-billion/#31eea59a6b3e Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 56 TENTATIVE Schedule Tue Jan 23: Storage: What the Heck is Supercomputing? Tue Jan 30: The Tyranny of the Storage Hierarchy Part I Tue Feb 6: The Tyranny of the Storage Hierarchy Part II Tue Feb 13: Instruction Level Parallelism Tue Feb 20: Stupid Compiler Tricks Tue Feb 27: Multicore Multithreading Tue March 6: Distributed Multiprocessing Tue March 13: NO SESSION (Henry business travel) Tue March 20: NO SESSION (OU's Spring Break)

Tue March 27: Applications and Types of Parallelism Tue Apr 3: Multicore Madness Tue Apr 10: NO SESSION (Henry business travel) Tue Apr 17: High Throughput Computing Tue Apr 24: GPGPU: Number Crunching in Your Graphics Card Tue May 1: Grab Bag: Scientific Libraries, I/O Libraries, Visualization Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 57 Thanks for helping! OU IT

OSCER operations staff (Dave Akin, Patrick Calhoun, Kali McLennan, Jason Speckman, Brett Zimmerman) OSCER Research Computing Facilitators (Jim Ferguson, Horst Severini) Debi Gentis, OSCER Coordinator Kyle Dudgeon, OSCER Manager of Operations Ashish Pai, Managing Director for Research IT Services The OU IT network team OU CIO Eddie Huebsch

OneNet: Skyler Donahue Oklahoma State U: Dana Brunson Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 58 This is an experiment! Its the nature of these kinds of videoconferences that FAILURES ARE GUARANTEED TO HAPPEN! NO PROMISES! So, please bear with us. Hopefully everything will work out well enough. If you lose your connection, you can retry the same kind of connection, or try connecting another way. Remember, if all else fails, you always have the phone bridge to fall back on. PLEASE MUTE YOURSELF.

PLEASE MUTE YOURSELF. PLEASE MUTE YOURSELF. Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 59 Coming in 2018! Coalition for Advancing Digital Research & Education (CADRE) Conference: 17-18 2018 @ Oklahoma State U, Stillwater OK USA Apr https://hpcc.okstate.edu/cadre-conference Linux Clusters Institute workshops http://www.linuxclustersinstitute.org/workshops/ Introductory HPC Cluster System Administration: May 14-18 2018 @ U Nebraska, Lincoln NE USA Intermediate HPC Cluster System Administration: Aug 13-17 2018 @ Yale U, New Haven CT USA

Great Plains Network Annual Meeting: details coming soon Advanced Cyberinfrastructure Research & Education Facilitators (ACI-REF) Virtual Residency Aug 5-10 2018, U Oklahoma, Norman OK USA PEARC 2018, July 22-27, Pittsburgh PA USA https://www.pearc18.pearc.org/ IEEE Cluster 2018, Sep 10-13, Belfast UK https://cluster2018.github.io OKLAHOMA SUPERCOMPUTING SYMPOSIUM 2018, Sep 25-26 2018 @ OU SC18 supercomputing conference, Nov 11-16 2018, Dallas TX USA http://sc18.supercomputing.org/ Supercomputing in Plain English: Hi Thruput Tue Apr 17 2018 60 Thanks for your

attention! Questions? www.oscer.ou.edu

Recently Viewed Presentations

  • Cell Organelle chart summary - Central Bucks School District

    Cell Organelle chart summary - Central Bucks School District

    From the diagram above, there is an organelle in Cell A that surrounds the cell that gives it shape and extra support. What is the name of the organelle? Explain. Ribosomes, the organelles that make proteins, can be found in...
  • A New Kind of Flour Baby PPT

    A New Kind of Flour Baby PPT

    A New Kind of "Flour Baby" ... place the entire bag of flour into the same pantyhose leg used to created the head. Push the flour all the way through until it is touching the polyester fiberfill. ... The body...
  • Ancient Egypt - Home - Thomas County Schools

    Ancient Egypt - Home - Thomas County Schools

    The pharaoh was the political and religious leader of the Egyptian people, Had two title: 'Lord of the Two Lands' and 'High Priest of Every Temple'. 'Lord of the Two Lands' meant he was the ruler of Upper and Lower...
  • Literary Analysis Strategy - featuredartistnetwork.com

    Literary Analysis Strategy - featuredartistnetwork.com

    Example Thesis Statements. Use a Debatable Thesis: Pride . and Prejudice is about Elizabeth . Bennet's. effort to overcome her own . proud . behavior and discrimination towards Mr. Darcy, as well as how her family is affected by the...
  • Beam Elements - Matlab

    Beam Elements - Matlab

    Analysis of Plastic Behavior. Plastic deformation in metals is an inherently nonlinear process. ... Multilinear Kinematic Hardening - series of straight lines after yielding. Nonlinear Kinematic Hardening. Similar models exist for isotropic hardening.
  • DANGER CONFINED SPACE Special Training & Precautions required

    DANGER CONFINED SPACE Special Training & Precautions required

    Risk Assessment of Oxygen Depletion and Hazardous Gas Emissions at Mine Reclamation Sites Ladan Mohammadi and John Meech NORMAN B. KEEVIL INSTITUTE OF MINING ENGINEERING, UBC, Vancouver, B.C.
  • Wounded Warrior Family Support Debit Card Program Maj

    Wounded Warrior Family Support Debit Card Program Maj

    Wounded Warrior Family Support Debit Card Program Maj L. Gaines HQMC P&R, RFF 15 August 2011 Background Marine Corps issues approximately 1500 bedside Invitation Travel Orders (ITOs) annually Approximately $2.6M war related and $.8M non-war related bedside ITOs issued in...
  • Welcome to Class St Margaret Year 6 - st-francis.kent.sch.uk

    Welcome to Class St Margaret Year 6 - st-francis.kent.sch.uk

    Welcome to St Paul & St Peter Year 4 Class Teachers: Mrs. Hawkes & Mrs Zokra Teaching Assistants: Mrs Duran & Mrs Martin What you need to know… Equipment and water bottles PE kits and uniform Homework/home learning Support for...