|By Business Wire||
|November 5, 2012 03:20 PM EST||
This year, SC12 will not only showcase the next generation of high performance computing applications but it will also be home to seven innovative network research projects through a special program called the SCinet Research Sandbox (SRS).
SCinet is the primary high performance network infrastructure built each year for SC exhibitors to highlight their cutting edge computing applications and collaborations. As a key component of SCinet, the SRS is designed to enable researchers to experiment and demonstrate their ideas on innovative network architectures, applications and protocols in the unique live environment of the SCinet network. This year, the SRS will provide researchers with dedicated access to multiple 100 Gigabits per second wide area network links as well as a 10 Gigabit per second (Gbps) OpenFlow network testbed.
“In addition to supporting the extreme demands of the HPC-based demonstrations that have become the trademark of the conference, SCinet also seeks to foster and highlight developments in network research that will be necessary to support the next-generation of science applications,” said Brian Tierney, SRS co-chair for SC12 and head of ESnet’s Advanced Network Technologies Group. “Both 100 Gbps networking and OpenFlow have become some of the most influential networking technologies of this decade. SRS allows the community to showcase innovations on these platforms while in their infancy to demonstrate the impact they may have on the entire HPC community in the future.”
“Openflow has the potential to greatly improve applications that are necessary for advancements in HPC such as GridFTP and others. By allowing these applications to access the network in a highly intelligent and programmable manner, OpenFlow can significantly improve end-to-end network performance especially for bulk data transfers, which will be an increasingly difficult challenge in the era of data intensive science,” said Andrew Lee, SRS co-chair for SC12 and Principal Network Systems Engineer for the Global Research Network Operations Center at Indiana University. “Demonstrations like those being supported by the SCinet Research Sandbox are laying the groundwork for these critical advancements as well as demonstrating to the community - in a tangible way - the possibilities that Openflow provides."
Seven projects have been selected as part of the SRS program, all of which will be showcased as part of the Technical Program and will be demonstrating their research in several exhibit booths during the conference. For detailed information on the projects and their presentations visit: http://sc12.supercomputing.org/content/scinet-research-sandbox
2012 SRS projects include:
Efficient LHC Data Distribution across 100Gbps Networks
The analysis of data leading to the recent discoveries at the Large Hadron Collider (LHC) produces data flows of more than 100 Petabytes per year, and increasingly relies on the efficient movement of data sets between the globally distributed computing sites. The team will demonstrate the state-of-the-art data movement tools, as enabling technology for high-throughput data distribution over 100Gbps WAN circuits. The demo will interconnect three major LHC Tier-2 computing sites and the SC12 show floor (booth 809) using 100Gbps technology.
Collaborating organizations: University of Victoria,
University of Michigan, California Institute of Technology, Vanderbilt
University, Internet2, ESnet, CENIC, Starlight, PacWave/GLORIAD,
KNU/KISTI as well as vendors: Alcatel Lucent, Ciena, Cisco, Juniper
Networks, Mellanox. Dell-Force10, SuperMicro, ASA Micro Systems, Data
Direct Network, Fusion-IO, PADTECH
Demonstration booth: 809
Presentation: November 15, 2012, 8:50am MT in Room 155-F2
Exploiting Network Parallelism for Improving Data Transfer Performance
The task of scientific bulk data movement, e.g. migrating collected results from the instrumentation to the processing and storage facilities, is hampered by a lack of available network resources. Traditional R&E connectivity can be congested on portions of an end-to-end path causing degradation of overall performance. This SRS project will explore dynamic network control to facilitate efficient bulk data movement, combining opportunistic use of "traditional" networks with dedicated reservations over virtual circuits and OpenFlow enabled resources. The GridFTP application has been instrumented with the eXtensible Session Protocol (XSP), an intelligent system capable of controlling programmable networks. The project intends to show end-to-end performance improvement between the SC12 conference and campuses involved in the DYNES project, through a combination of regular connectivity, dynamic bandwidth allocations, TCP acceleration, and operations using multiple paths.
Collaborating organizations: Indiana University, Lawrence
Berkeley National Laboratory, Argonne National Laboratory and Internet2
Demonstration booths: 1042, 1343
Presentation: Thursday, November 15, 2012, 9:10am MT in Room 155-F2
Multipathing with MPTCP and OpenFlow
This demo shows several emerging network technologies and how these can be used in big data transfers between data centers. In this demonstration traffic is sent simultaneously across multiple OpenFlow controlled paths between Geneva and Salt Lake City. The congestion control mechanism of Multipath TCP (MPTCP) favours the least congested paths and ensures that the load balancing across the paths is always optimal.
Collaborating organizations: SURFnet, SARA, iCAIR and the
California Institute of Technology
Demonstration booths: 2333, 809, 501
Presentation: November 15, 2012, 10:30am MT in Room 155-F2
The emerging era of “Big Science” demands the highest possible network performance. End-to-end circuit automation and workflow-driven customization are two essential capabilities needed for networks to scale to meet this challenge. This demonstration showcases how combining software-defined networking techniques with virtual circuits capabilities can transform the network into a dynamic, customer-configurable virtual switch. In doing so, users are able to rapidly customize network capabilities to meet their unique workflows with little to no configuration effort. The demo also highlights how the network can be automated to support multiple collaborations in parallel.
Collaborating organizations: ESnet, Ciena
Demonstration booth: 2437
Presentation: November 15, 2012, 11:10am MT in Room 155-F2
OpenFlow Enabled Hadoop over Local and Wide Area Clusters
The Hadoop Distributed File Systems and Hadoop's implementation of MapReduce is one of the most widely used platforms for data intensive computing. The shuffle and sort phases of a MapReduce computation often saturate network links to nodes and the reduce phase of the computation must wait for data. This study explores the use of OpenFlow to the control network configuration for different flows to thereby provide different network characteristics for different categories of Hadoop traffic.
Collaborating organizations: Infoblox, Laboratory for
Advanced Computing, University of Chicago, Open Cloud Consortium,
Livermore Valley Open Campus
Demonstration booth: 501
Presentation: November 15, 2012, 10:50am MT in Room 155-F2
Large-scale data intensive science requires global collaboration and sophisticated high capacity data management. The emergence of more flexible networking, for example, using techniques based on OpenFlow, provides opportunities to address these issues because these techniques enable a high degree of network customization and dynamic provisioning. These techniques enable large-scale facilities to be created that can be used to prototype new architecture, services, protocols, and technologies. A number of research organizations from several countries have designed and implemented a persistent international experimental research facility that can be used to prototype, investigate, and test network innovations for large-scale global science. For SC12, this international experimental network facility will be extended to from sites across the world to the conference showfloor, and it will be used to support several testbeds and to showcase a series of complementary demonstrations.
Collaborating organizations: iCAIR,
Northwestern University; National Center for High-Performance Computing,
Taiwan; University of Applied Sciences, Taiwan; National Cheng-Kung
University, Taiwan; SARA, The Netherlands, California Institute of
Technology/CERN; SURFnet, The Netherlands.
Demonstration booths: 2333, 501, 843, 809
Presentation: November 15, 2012, 8:30am MT in Room 155-F2
Scalable Cyber-Security for Terabit Cloud Computing
Reservoir Labs will demonstrate R-Scope®, a scalable, high-performance network packet inspection technology that forms the core of a new generation of Intrusion Detection Systems enabling the construction and deployment of cyber security infrastructures scaling to terabit per second ingest bandwidths. This scalability is enabled by the use of low- power and high-performance manycore network processors combined with Reservoir’s enhancements to Bro. The innovative R-Scope PACE-T appliance, implemented on a 1U Tilera TILExtreme-Gx platform, will demonstrate the capacity to perform cyber-security analysis at 80Gbps, by combining cyber-security aware front-end network traffic load balancing tightly coupled with the full back-end analytic power of Bro. This fully-programmable platform incorporates the full Bro semantics into the appliance’s load-balancing front-end and the back-end analytic nodes.
Collaborating organizations: Reservoir Labs, SCinet
Presentation: November 15, 2012, 9:30am MT in Room 155-F2
This press release is available online at http://sc12.supercomputing.org/content/scinet-research-sandbox-shows-groundbreaking-network-research.
SC12, sponsored by the ACM (Association for Computing Machinery) and the IEEE Computer Society, offers a complete technical education program and exhibition to showcase the many ways high performance computing, networking, storage and analysis lead to advances in scientific discovery, research, education and commerce. This premier international conference includes a globally attended technical program, workshops, tutorials, a world class exhibit area, demonstrations and opportunities for hands-on learning.
Photos/Multimedia Gallery Available: http://www.businesswire.com/cgi-bin/mmg.cgi?eid=50466050&lang=en
Identity is in everything and customers are looking to their providers to ensure the security of their identities, transactions and data. With the increased reliance on cloud-based services, service providers must build security and trust into their offerings, adding value to customers and improving the user experience. Making identity, security and privacy easy for customers provides a unique advantage over the competition.
Jul. 28, 2016 11:45 PM EDT Reads: 1,204
In his session at @DevOpsSummit at 19th Cloud Expo, Yoseph Reuveni, Director of Software Engineering at Jet.com, will discuss Jet.com's journey into containerizing Microsoft-based technologies like C# and F# into Docker. He will talk about lessons learned and challenges faced, the Mono framework tryout and how they deployed everything into Azure cloud. Yoseph Reuveni is a technology leader with unique experience developing and running high throughput (over 1M tps) distributed systems with extre...
Jul. 28, 2016 10:15 PM EDT Reads: 2,216
"We've discovered that after shows 80% if leads that people get, 80% of the conversations end up on the show floor, meaning people forget about it, people forget who they talk to, people forget that there are actual business opportunities to be had here so we try to help out and keep the conversations going," explained Jeff Mesnik, Founder and President of ContentMX, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 28, 2016 10:15 PM EDT Reads: 1,436
SYS-CON Events announced today that Isomorphic Software will exhibit at DevOps Summit at 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Isomorphic Software provides the SmartClient HTML5/AJAX platform, the most advanced technology for building rich, cutting-edge enterprise web applications for desktop and mobile. SmartClient combines the productivity and performance of traditional desktop software with the simp...
Jul. 28, 2016 10:15 PM EDT Reads: 1,213
As companies gain momentum, the need to maintain high quality products can outstrip their development team’s bandwidth for QA. Building out a large QA team (whether in-house or outsourced) can slow down development and significantly increases costs. This eBook takes QA profiles from 5 companies who successfully scaled up production without building a large QA team and includes: What to consider when choosing CI/CD tools How culture and communication can make or break implementation
Jul. 28, 2016 10:00 PM EDT Reads: 1,719
"When you think about the data center today, there's constant evolution, The evolution of the data center and the needs of the consumer of technology change, and they change constantly," stated Matt Kalmenson, VP of Sales, Service and Cloud Providers at Veeam Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 28, 2016 10:00 PM EDT Reads: 1,443
"There's a growing demand from users for things to be faster. When you think about all the transactions or interactions users will have with your product and everything that is between those transactions and interactions - what drives us at Catchpoint Systems is the idea to measure that and to analyze it," explained Leo Vasiliou, Director of Web Performance Engineering at Catchpoint Systems, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York Ci...
Jul. 28, 2016 10:00 PM EDT Reads: 2,096
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
Jul. 28, 2016 09:00 PM EDT Reads: 2,706
Extreme Computing is the ability to leverage highly performant infrastructure and software to accelerate Big Data, machine learning, HPC, and Enterprise applications. High IOPS Storage, low-latency networks, in-memory databases, GPUs and other parallel accelerators are being used to achieve faster results and help businesses make better decisions. In his session at 18th Cloud Expo, Michael O'Neill, Strategic Business Development at NVIDIA, focused on some of the unique ways extreme computing is...
Jul. 28, 2016 07:15 PM EDT Reads: 383
"We view the cloud not really as a specific technology but as a way of doing business and that way of doing business is transforming the way software, infrastructure and services are being delivered to business," explained Matthew Rosen, CEO and Director at Fusion, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 28, 2016 06:45 PM EDT Reads: 1,640
Redis is not only the fastest database, but it is the most popular among the new wave of databases running in containers. Redis speeds up just about every data interaction between your users or operational systems. In his session at 19th Cloud Expo, Dave Nielsen, Developer Advocate, Redis Labs, will share the functions and data structures used to solve everyday use cases that are driving Redis' popularity.
Jul. 28, 2016 06:30 PM EDT Reads: 1,663
Aspose.Total for .NET is the most complete package of all file format APIs for .NET as offered by Aspose. It empowers developers to create, edit, render, print and convert between a wide range of popular document formats within any .NET, C#, ASP.NET and VB.NET applications. Aspose compiles all .NET APIs on a daily basis to ensure that it contains the most up to date versions of each of Aspose .NET APIs. If a new .NET API or a new version of existing APIs is released during the subscription peri...
Jul. 28, 2016 06:00 PM EDT Reads: 965
Organizations planning enterprise data center consolidation and modernization projects are faced with a challenging, costly reality. Requirements to deploy modern, cloud-native applications simultaneously with traditional client/server applications are almost impossible to achieve with hardware-centric enterprise infrastructure. Compute and network infrastructure are fast moving down a software-defined path, but storage has been a laggard. Until now.
Jul. 28, 2016 05:30 PM EDT Reads: 1,872
"My role is working with customers, helping them go through this digital transformation. I spend a lot of time talking to banks, big industries, manufacturers working through how they are integrating and transforming their IT platforms and moving them forward," explained William Morrish, General Manager Product Sales at Interoute, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 28, 2016 05:30 PM EDT Reads: 2,217
[webcast] Continuous Delivery in the Enterprise | @DevOpsSummit @IBMDevOps #IBM #DevOps #ContinuousDelivery
To leverage Continuous Delivery, enterprises must consider impacts that span functional silos, as well as applications that touch older, slower moving components. Managing the many dependencies can cause slowdowns. See how to achieve continuous delivery in the enterprise.
Jul. 28, 2016 05:00 PM EDT Reads: 440
You think you know what’s in your data. But do you? Most organizations are now aware of the business intelligence represented by their data. Data science stands to take this to a level you never thought of – literally. The techniques of data science, when used with the capabilities of Big Data technologies, can make connections you had not yet imagined, helping you discover new insights and ask new questions of your data. In his session at @ThingsExpo, Sarbjit Sarkaria, data science team lead ...
Jul. 28, 2016 04:30 PM EDT Reads: 1,196
SYS-CON Events announced today the Kubernetes and Google Container Engine Workshop, being held November 3, 2016, in conjunction with @DevOpsSummit at 19th Cloud Expo at the Santa Clara Convention Center in Santa Clara, CA. This workshop led by Sebastian Scheele introduces participants to Kubernetes and Google Container Engine (GKE). Through a combination of instructor-led presentations, demonstrations, and hands-on labs, students learn the key concepts and practices for deploying and maintainin...
Jul. 28, 2016 04:30 PM EDT Reads: 918
Security, data privacy, reliability, and regulatory compliance are critical factors when evaluating whether to move business applications from in-house, client-hosted environments to a cloud platform. Quality assurance plays a vital role in ensuring that the appropriate level of risk assessment, verification, and validation takes place to ensure business continuity during the migration to a new cloud platform.
Jul. 28, 2016 04:15 PM EDT Reads: 386
Extracting business value from Internet of Things (IoT) data doesn’t happen overnight. There are several requirements that must be satisfied, including IoT device enablement, data analysis, real-time detection of complex events and automated orchestration of actions. Unfortunately, too many companies fall short in achieving their business goals by implementing incomplete solutions or not focusing on tangible use cases. In his general session at @ThingsExpo, Dave McCarthy, Director of Products...
Jul. 28, 2016 04:15 PM EDT Reads: 1,781
Security, data privacy, reliability and regulatory compliance are critical factors when evaluating whether to move business applications from in-house client hosted environments to a cloud platform. In her session at 18th Cloud Expo, Vandana Viswanathan, Associate Director at Cognizant, In this session, will provide an orientation to the five stages required to implement a cloud hosted solution validation strategy.
Jul. 28, 2016 03:45 PM EDT Reads: 1,127