Wednesday, May 20
8:00 am | – | 9:00 am | Breakfast and Registration
vegetarian quiche, whole wheat toast, muffins, fruit platter, orange juice, coffee, tea
|
|
Session Moderators: John "TJ" Knoeller, Brian Lin | ||||
9:00 am | – | 9:30 am | Welcome |
Miron Livny Center for High Throughput Computing
|
9:40 am | – | 10:10 am | What's new in HTCondor? What's coming up?
Overview of upcoming enhancements in HTCondor v8.4.x and beyond. Note the opening video from the talk can also be found on YouTube at https://youtu.be/D-FnZJ8avsU
|
Todd Tannenbaum Center for High Throughput Computing
|
10:15 am | – | 10:45 am | Break
coffee, tea, soda, cookies, fresh fruit, yogurt, granola
|
|
Session Moderators: Aaron Moate, Mátyás Selmeci | ||||
10:45 am | – | 11:05 am | Batch Services at CERN - Present Status and Future Evolution |
Helge Meinhard CERN
|
11:10 am | – | 11:30 am | Implementing an HTCondor service at CERN
At CERN our batch system runs job submissions from physicists around the world, looking to analyze the LHC data. Currently, this represents around 65,000 cores running 400,000 jobs a day, at any one time over 60,000 jobs are executing. This has now become a matter of scaling, and when compared to other systems, HTCondor looks to be our best option to take us to 50,000 worker nodes and beyond. Therefore, CERN IT plans to move to HTCondor as their primary batch management software in the next few years. In this presentation we report about the status of the project and our future plans.
|
Iain Steers CERN
|
11:35 am | – | 11:55 am | CMS Experience Provisioning Cloud Resources with GlideinWMS |
Anthony Tiradani Fermi National Accelerator Laboratory
|
12:00 pm | – | 1:10 pm | Lunch
prime rib, chicken roulade, green salad with vinagrette, eggplant and mozzeralla napoleon with roasted tomato sauce, bread, brownies, bars
|
|
Data Management | ||||
Session Moderators: Tim Theisen, Lauren Michael | ||||
1:10 pm | – | 1:30 pm | Efficiently Sharing Common Data
We examine the pros and cons of combining two technologies: Using BitTorrent for data transfer, and caching data on execute nodes using a pluggable cache replacement policy.
|
Zach Miller Center for High Throughput Computing
|
1:35 pm | – | 1:55 pm | Data caching with HTCondor: The HTCondor CacheD
We present the HTCondor CacheD, a caching daemon that optimizes stage-in time for large shared input data through through novel transfer methods and proactive replication.
|
Derek Weitzel University of Nebraska – Lincoln
|
2:00 pm | – | 2:20 pm | HTCondor Input File Transfer via Squid Caching Proxy |
Carl Vuosalo Center for High Throughput Computing
|
2:25 pm | – | 2:45 pm | Scientific Workflows - How Pegasus can enhance your DAGMan experience |
Karan Vahi USC Information Sciences Institute
|
2:50 pm | – | 3:25 pm | Break
coffee, tea, soda, snack mix, peanut butter filled pretzels, tortilla chips, salsa, guacamole, vegetable and dip platter
|
|
Session Moderators: Jaime Frey, Kent Wenger | ||||
3:25 pm | – | 3:45 pm | "No Idle Cores": Automatic Load Balancing in our ATLAS Pool
In our ATLAS pool, multicore jobs coexist with single core jobs of many different varieties, and ensuring that all types of jobs have access to the resources they require is a challenging problem. I will discuss how we partially solved the problem a combination of Hierarchical Group Quotas and Partitionable Slots, and then present work we did to allow automatic balancing between groups based on demand from PANDA, the ATLAS workload-management system -- balancing the effects of single vs. multicore jobs and queues with differing priorities.
|
William Strecker-Kellogg Brookhaven National Laboratory
|
3:50 pm | – | 4:10 pm | Scientific Computing on emerging infrastructures using HTCondor |
Sanjay Padhi University of California, San Diego
|
4:15 pm | – | 4:35 pm | One Pool To Rule Them All: The CMS HTCondor/glideinWMS Global Pool
The CMS experiment at CERN runs all distributed computing now through a single global HTCondor/glideInWMS pool. I will discuss the reasons for doing this, how it is constructed and our experience running with it.
|
David Mason Fermi National Accelerator Laboratory
|
4:40 pm | – | 5:00 pm | Adopting HTCondor at Raytheon
Raytheon Company, with 2014 sales of $23 billion and 61,000 employees worldwide, is a technology and innovation leader specializing in defense, security, and civil markets throughout the world. Today's extensive High Performance Computing needs require Raytheon's Information Technology organization to continuously push the envelope to create leading-edge technical processing solutions for complex systems and program challenges. Beginning in late 2012, Michael Pelletier took on such a challenge, a 750+-thread compute cluster which was straining the limits of the existing infrastructure, as well as the capabilities of the IT support team and its engineering partners. By adopting HTCondor, and with the help of the CHTC development team, he was able to transform the performance of the existing hardware, and leverage HTCondor's unique capabilities simplify workflow and multiply the productivity of the modeling, simulation, and analysis tasks across multiple Raytheon programs. He will discuss Raytheon's adoption of HTCondor and the innovations and improvements which flowed from it.
|
Michael V. Pelletier Raytheon Integrated Defense Systems - Information Technology
|
Reception sponsored by Cycle Computing
Courtesy of Cycle Computing we will have a reception following the Wednesday session. It will be at Lucky's Bar and Grille, 1421 Regent Street from 6pm to 7pm. Drinks and appetizers will be served. It is a short walk from the meeting location. Walking directions
Specific talks and times are subject to change.