U-M plans participation at SC17 conference in Denver

By | General Interest, Happenings, HPC, News

Several University of Michigan researchers and professional IT staff will attend the Supecomputing 17 (SC17) conference in Colorado from Nov. 12-17, and participate in a number of different ways, including demonstrations, presentations and tutorials.

In addition to the events and presentations listed below, Amy Liebowitz, a network architect at ITS, is working on SCINet, a high-capacity network created every year for the conference. Liebowitz is on the routing team, which is responsible for installing, configuring and supporting the high performance conference network. The Routing Team also coordinates external connectivity with commodity Internet and R&E WAN service providers.

Please visit us at Booth 471 on the exhibit floor, or at one of the following events:

Sunday, Nov. 12

8:30 a.m. – 5 p.m.: Quentin Stout (EECS) and Christiane Jablonowski (CLASP) will teach the “Parallel Computing 101” tutorial.

Tuesday, Nov. 14

10:30 a.m.:  Research Networking at the University of Michigan
Eric Boyd, Director of Research Networks, U-M
U-M Booth #471

11 a.m.: The OSiRIS Project: Open Storage Research Infrastructure
Ben Meekhof, HPC Storage Administrator, Advanced Research Computing – Technology Services (ARC-TS)
U-M Booth #471

12:30 p.m.: GPU-Accelerated Predictive Material Design
Simon Adorf, Ph.D. Candidate, Chemical Engineering Department, U-M
U-M Booth #471

1:30 p.m.: Simple Data and Workflow Management with Signac
Simon Adorf, Ph.D. Candidate, Chemical Engineering Department, U-M
U-M Booth #471

1:30 – 3 p.m.: Matt McLean, a Big Data systems administrator with ARC-TS, will serve as a panelist at a presentation titled “The ARM Software Ecosystem: Are We There Yet?

2 p.m.: The Michigan Institute for Computational Discovery and Engineering
Mariana Carrasco-Teja, Assistant Director, MICDE
U-M Booth #471

5:15 – 7 p.m.: Jeff Sica, a research database administrator with ARC-TS, will help lead a Birds of a Feather session titled “Containers in HPC.”

6:15 – 8:30 p.m.: ARC at U-M is a sponsor of a networking and career networking reception put on by Women in HPC. ARC Director Sharon Broude Geva will speak at the event.

Wednesday, November 15

10 a.m.: The OSiRIS Project: Open Storage Research Infrastructure
Shawn McKee, U-M Department of Physics, OSiRIS Principal Investigator
U-M Booth #471

11 a.m.: Research Networking at the University of Michigan
Eric Boyd, Director of Research Networks, U-M
U-M Booth #471

3 p.m.: The New Cavium ThunderX Big Data Cluster at U-M
Matt McLean, Big Data Systems Administrator, ARC-TS
U-M Booth #471

Thursday, November 16

11 a.m.: The New Cavium ThunderX Big Data Cluster at U-M
Matt McLean, Big Data Systems Administrator, ARC-TS
U-M Booth #471

 

 

U-M partners with Cavium on Big Data computing platform

By | Feature, General Interest, Happenings, HPC, News

A new partnership between the University of Michigan and Cavium Inc., a San Jose-based provider of semiconductor products, will create a powerful new Big Data computing cluster available to all U-M researchers.

The $3.5 million ThunderX computing cluster will enable U-M researchers to, for example, process massive amounts of data generated by remote sensors in distributed manufacturing environments, or by test fleets of automated and connected vehicles.

The cluster will run the Hortonworks Data Platform providing Spark, Hadoop MapReduce and other tools for large-scale data processing.

“U-M scientists are conducting groundbreaking research in Big Data already, in areas like connected and automated transportation, learning analytics, precision medicine and social science. This partnership with Cavium will accelerate the pace of data-driven research and opening up new avenues of inquiry,” said Eric Michielssen, U-M associate vice president for advanced research computing and the Louise Ganiard Johnson Professor of Engineering in the Department of Electrical Engineering and Computer Science.

“I know from experience that U-M researchers are capable of amazing discoveries. Cavium is honored to help break new ground in Big Data research at one of the top universities in the world,” said Cavium founder and CEO Syed Ali, who received a master of science in electrical engineering from U-M in 1981.

Cavium Inc. is a leading provider of semiconductor products that enable secure and intelligent processing for enterprise, data center, wired and wireless networking. The new U-M system will use dual socket servers powered by Cavium’s ThunderX ARMv8-A workload optimized processors.

The ThunderX product family is Cavium’s 64-bit ARMv8-A server processor for next generation Data Center and Cloud applications, and features high performance custom cores, single and dual socket configurations, high memory bandwidth and large memory capacity.

Alec Gallimore, the Robert J. Vlasic Dean of Engineering at U-M, said the Cavium partnership represents a milestone in the development of the College of Engineering and the university.

“It is clear that the ability to rapidly gain insights into vast amounts of data is key to the next wave of engineering and science breakthroughs. Without a doubt, the Cavium platform will allow our faculty and researchers to harness the power of Big Data, both in the classroom and in their research,” said Gallimore, who is also the Richard F. and Eleanor A. Towner Professor, an Arthur F. Thurnau Professor, and a professor both of aerospace engineering and of applied physics.

Along with applications in fields like manufacturing and transportation, the platform will enable researchers in the social, health and information sciences to more easily mine large, structured and unstructured datasets. This will eventually allow, for example, researchers to discover correlations between health outcomes and disease outbreaks with information derived from socioeconomic, geospatial and environmental data streams.

U-M and Cavium chose to run the cluster on Hortonworks Data Platform, which is based on open source Apache Hadoop. The ThunderX cluster will deliver high performance computer services for the Hadoop analytics and, ultimately, a total of three petabytes of storage space.

“Hortonworks is excited to be a part of forward-leading research at the University of Michigan exploring low-powered, high-performance computing,” said Nadeem Asghar, vice president and global head of technical alliances at Hortonworks. “We see this as a great opportunity to further expand the platform and segment enablement for Hortonworks and the ARM community.”

CSCAR provides walk-in support for new Flux users

By | Data, Educational, Flux, General Interest, HPC, News

CSCAR now provides walk-in support during business hours for students, faculty, and staff seeking assistance in getting started with the Flux computing environment.  CSCAR consultants can walk a researcher through the steps of applying for a Flux account, installing and configuring a terminal client, connecting to Flux, basic SSH and Unix command line, and obtaining or accessing allocations.  

In addition to walk-in support, CSCAR has several staff consultants with expertise in advanced and high performance computing who can work with clients on a variety of topics such as installing, optimizing, and profiling code.  

Support via email is also provided via hpc-support@umich.edu.  

CSCAR is located in room 3550 of the Rackham Building (915 E. Washington St.). Walk-in hours are from 9 a.m. – 5 p.m., Monday through Friday, except for noon – 1 p.m. on Tuesdays.

See the CSCAR web site (cscar.research.umich.edu) for more information.

University of Michigan researcher contributes to NASA findings on carbon in the atmosphere showcased in the journal Science

By | General Interest, Happenings, News

 

High-resolution satellite data from NASA’s Orbiting Carbon Observatory-2 are revealing the subtle ways that carbon links everything on Earth – the ocean, land, atmosphere, terrestrial ecosystems and human activities. Scientists using the first 2 1/2 years of OCO-2 data have published a special collection of five papers today in the journal Science that demonstrates the breadth of this research. In addition to showing how drought and heat in tropical forests affected global carbon dioxide levels during the 2015-16 El Niño, other results from these papers focus on ocean carbon release and absorption, urban emissions and a new way to study photosynthesis. A final paper by OCO-2 Deputy Project Scientist Annmarie Eldering of NASA’s Jet Propulsion Laboratory in Pasadena, California, and colleagues gives an overview of the state of OCO-2 science.

Manish Verma, a Geospatial/Data Science Consultant at the University of Michigan’s Consulting for Statistics, Computing and Analytics Research (CSCAR) unit, contributed as a coauthor to an article on a new way to measure photosynthesis over time and space.

Using data from the OCO-2, Verma’s analysis helped expand the utility of measurements of solar induced fluorescence (SIF), which indicates active photosynthesis in plants. Verma’s work showed that SIF data collected from the OCO-2 satellite provides reliable information on the variability of photosynthesis at a much smaller scale — down to individual ecosystems.

This can, in turn, “lead to more reliable estimates of carbon sources — that is, when, where, why and how carbon is exchanged between land and atmosphere — as well as a deeper understanding of carbon-climate feedbacks,” according to the Science article.

For more, see the NASA press release (https://www.nasa.gov/feature/jpl/new-insights-from-oco-2-showcased-in-science) and the Science article (http://science.sciencemag.org/content/358/6360/eaam5747.full)

The University of Michigan is live on IBM.com

By | News, Research

IBM is showcasing the current research developed with ConFlux, our ground-breaking cluster that uses IBM’s HPC and storage technology to enable scientists to draw on huge volumes of bid data and use machine learning to create reliable models for compute-intensive research.

“ With IBM hardware boosting our HPC environment, we can offer scientists the tools to conduct research that could revolutionize entire industries. ”
Todd Raeker, Research Technology Consultant for the University of Michigan.

To learn more please visit http://www-03.ibm.com/software/businesscasestudies/us/en/corp?synkey=A323848E50678F66

 

U-M joins NSF-funded SLATE project to simplify scientific collaboration on a massive scale

By | Feature, General Interest, Happenings, News, Research

From the Cosmic Frontier to CERN, New Platform Stitches Together Global Science Efforts

SLATE will enable creation of new platforms for collaborative science

Today’s most ambitious scientific quests — from the cosmic radiation measurements by the South Pole Telescope to the particle physics of CERN — are multi-institutional research collaborations requiring computing environments that connect instrumentation, data, and computational resources. Because of the scale of the data and the complexity of this science,  these resources are often distributed among university research computing centers, national high performance computing centers, or commercial cloud providers.  This can cause scientists to spend more time on the technical aspects of computation than on discoveries and knowledge creation, while computing support staff are required to invest more effort integrating domain specific software with limited applicability beyond the community served.  

With Services Layer At The Edge (SLATE), a $4 million project funded by the National Science Foundation, the University of Michigan joins a team led by the Enrico Fermi and Computation Institutes at University of Chicago to provide technology that simplifies connecting university and laboratory data center capabilities to the national cyberinfrastructure ecosystem. The University of Utah is also participating. Once installed, SLATE connects local research groups with their far-flung collaborators, allowing central research teams to automate the exchange of data, software and computing tasks among institutions without burdening local system administrators with installation and operation of highly customized scientific computing services. By stitching together these resources, SLATE will also expand the reach of domain-specific “science gateways” and multi-site research platforms.  

“Science, ultimately, is a collective endeavor. Most scientists don’t work in a vacuum, they work in collaboration with their peers at other institutions,” said Shawn McKee, a co-PI on the project and director of the Center for Network and Storage-Enabled Collaborative Computational Science at the University of Michigan. “They often need to share not only data, but systems that allow execution of workflows across multiple institutions. Today, it is a very labor-intensive, manual process to stitch together data centers into platforms that provide the research computing environment required by forefront scientific discoveries.”

SLATE works by implementing “cyberinfrastructure as code”, augmenting high bandwidth science networks with a programmable “underlayment” edge platform. This platform hosts advanced services needed for higher-level capabilities such as data and software delivery, workflow services and science gateway components.  

U-M  has numerous roles in the project including:

  • defining, procuring and configuring much of the SLATE hardware platform
  • working on the advanced networking aspects (along with Utah) which includes Software Defined Networking (SDN) and Network Function Virtualization (NFV),
  • developing the SLATE user interface and contributing to the core project design and implementation.

The project is similar to the OSiRIS project led by McKee, which also aims to remove bottlenecks to discovery posed by networking and data transfer infrastructure.

SLATE uses best-of-breed data center virtualization components, and where available, software defined networking, to enable automation of lifecycle management tasks by domain experts. As such, it simplifies the creation of scalable platforms that connect research teams, institutions and resources, accelerating science while reducing operational costs and development time. Since SLATE needs only commodity components, it can be used for distributed systems across all data center types and scales, thus enabling creation of ubiquitous, science-driven cyberinfrastructure.

slateAt UChicago, the SLATE team will partner with the Research Computing Center and Information Technology Services to help the ATLAS experiment at CERN, the South Pole Telescope and the XENON dark matter search collaborations create the advanced cyberinfrastructure necessary for rapidly sharing data, computer cycles and software between partner institutions.  The resulting systems will provide blueprints for national and international research platforms supporting a variety of science domains.  

For example, the SLATE team will work with researchers from the Computation Institute’s Knowledge Lab to develop a hybrid platform that elastically scales computational social science applications between commercial cloud and campus HPC resources. The platform will allow researchers to use their local computational resources with the analytical tools and sensitive data shared through Knowledge Lab’s Cloud Kotta infrastructure, reducing cost and preserving data security.

“SLATE is about creating a ubiquitous cyberinfrastructure substrate for hosting, orchestrating and managing the entire lifecycle of higher level services that power scientific applications that span multiple institutions,” said Rob Gardner, a Research Professor in the Enrico Fermi Institute and Senior Fellow in the Computation Institute. “It clears a pathway for rapidly delivering capabilities to an institution, maximizing the science impact of local research IT investments.”

Many universities and research laboratories use a “Science DMZ” architecture to balance security with the ability to rapidly move large amounts of data in and out of the local network. As sciences from physics to biology to astronomy become more data-heavy, the complexity and need for these subnetworks grows rapidly, placing additional strain on local IT teams.

That stress is further compounded when local scientists join multi-institutional collaborations, often requiring the installation of specialized, domain-specific services for the sharing of compute and data resources.

With SLATE, research groups will be able to fully participate in multi-institutional collaborations and contribute resources to their collective platforms with minimal hands-on effort from their local IT team. When joining a project, the researchers and admins can select a package of software from a cloud-based service — a kind of “app store” — that allows them to connect and work with the other partners.

“Software and data can then be updated automatically by experts from the platform operations and research teams, with little to no assistance required from local IT personnel,” said Joe Breen, Senior IT Architect for Advanced Networking Initiatives at the University of Utah’s Center for High Performance Computing. “While the SLATE platform is designed to work in any data center environment, it will utilize advanced network capabilities, such as software defined overlay networks, when the devices support it.”

By reducing the technical expertise and time demands for participating in multi-institution collaborations, the SLATE platform will be especially helpful to smaller universities that lack the resources and staff of larger institutions and computing centers. The SLATE functionality can also support the development of “science gateways” which make it easier for individual researchers to connect to HPC resources such as the Open Science Grid and XSEDE.

“A central goal of SLATE is to lower the threshold for campuses and researchers to create research platforms within the national cyberinfrastructure,” Gardner said.

Initial partner sites for testing the SLATE platform and developing its architecture include New Mexico State University and Clemson University, where the focus will be creating distributed  cyberinfrastructure in support of large scale bioinformatics and genomics workflows. The project will also work with the Science Gateways Community Institute, an NSF funded Scientific Software Innovation Institute, on SLATE integration to make gateways more powerful and reach more researchers and resources.

###

The Computation Institute (CI), a joint initiative of the University of Chicago and Argonne National Laboratory, is an intellectual nexus for scientists and scholars pursuing multi-disciplinary research and a resource center for developing and applying innovative computational approaches. Founded in 1999, it is home to over 100 faculty, fellows, and staff researching complex, system-level problems in such areas as biomedicine, energy and climate, astronomy and astrophysics, computational economics, social sciences and molecular engineering. CI is home to diverse projects including the Center for Robust Decision Making on Climate and Energy Policy, Knowledge Lab, The Urban Center for Computation and Data and the Center for Data Science and Public Policy.

For more information, contact Dan Meisler, Communications Manager, Advanced Research Computing at U-M: dmeisler@umich.edu, 734-764-7414

Info sessions on graduate studies in computational and data sciences — Sept. 21 and 25

By | Educational, Events, General Interest, News, Research

Learn about graduate programs that will prepare you for success in computationally intensive fields — pizza and pop provided

  • The Ph.D. in Scientific Computing is open to all Ph.D. students who will make extensive use of large-scale computation, computational methods, or algorithms for advanced computer architectures in their studies. It is a joint degree program, with students earning a Ph.D. from their current departments, “… and Scientific Computing” — for example, “Ph.D. in Aerospace Engineering and Scientific Computing.”
  • The Graduate Certificate in Computational Discovery and Engineering trains graduate students in computationally intensive research so they can excel in interdisciplinary HPC-focused research and product development environments. The certificate is open to all students currently pursuing Master’s or Ph.D. degrees at the University of Michigan.
  • The Graduate Certificate in Data Science is focused on developing core proficiencies in data analytics:
    1) Modeling — Understanding of core data science principles, assumptions and applications;
    2) Technology — Knowledge of basic protocols for data management, processing, computation, information extraction, and visualization;
    3) Practice — Hands-on experience with real data, modeling tools, and technology resources.

Times / Locations:

Siqian Shen (IOE) to receive an Early Career Award from the Department of Energy

By | News, Research

siqian-shen-featuredMICDE Associate Director Siqian Shen has been selected to receive an Early Career Award for the Department of Energy Office of Science by the DoE Office of Advanced Scientific Computing Research. The objective of her proposal titled “Extreme‐Scale Stochastic Optimization and Simulation via Learning‐Enhanced Decomposition and Parallelization” is to develop an efficient and unified framework that integrates machine learning with discrete optimization and risk‐averse modeling. The models considered represent a broad class of complex decision‐making problems, where 0‐1 or continuous decisions are made before and/or after knowing multiple and potentially correlated sources of uncertainties. This research will shed new light on the traditional decomposition algorithms for high‐performance computing.

Prof. Shen was recently promoted to Associate Professor of Industrial and Operations Engineering. To learn more about her research please visit http://micde.umich.edu/faculty-member/siqian-shen/.

The Early Career Award program from the US Department of Energy is a funding opportunity for researchers in universities and DOE national laboratories to support the development of individual research programs of outstanding scientists early in their careers. For the past 8 years this program has helped stimulate research careers in the disciplines supported by the DOE Office of Science. These include Advanced Scientific Computing Research (ASCR); Biological and Environmental Research (BER); Basic Energy Sciences (BES), Fusion Energy Sciences (FES); High Energy Physics (HEP), and Nuclear Physics (NP).

U-M, SJTU research teams share $1 million for data science projects

By | Data, General Interest, Happenings, News, Research

Five research teams from the University of Michigan and Shanghai Jiao Tong University in China are sharing $1 million to study data science and its impact on air quality, galaxy clusters, lightweight metals, financial trading and renewable energy.

Since 2009, the two universities have collaborated on a number of research projects that address challenges and opportunities in energy, biomedicine, nanotechnology and data science.

In the latest round of annual grants, the winning projects focus on data science and how it can be applied to chemistry and physics of the universe, as well as finance and economics.

For more, read the University Record article.

For descriptions of the research projects, see the MIDAS/SJTU partnership page.

SAVE THE DATE: MIDAS Annual Symposium, Oct. 11

By | Events, General Interest, News

Please join us for the 2017 Michigan Institute for Data Science Symposium.

The keynote speaker will be Cathy O’Neil, mathematician and best-selling author of “Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy.”

Other speakers include:

  • Nadya Bliss, Director of the Global Security Initiative, Arizona State University
  • Francesca Dominici, Co-Director of the Data Science Initiative and Professor of Biostatistics, Harvard T.H. Chan School of Public Health
  • Daniela Whitten, Associate Professor of Statistics and Biostatistics, University of Washington
  • James Pennebaker, Professor of Psychology, University of Texas

More details, including how to register, will be available soon.