Welcome!

@BigDataExpo Authors: Ed Featherston, Liz McMillan, Elizabeth White, Scott Allen, Dana Gardner

Related Topics: @BigDataExpo

@BigDataExpo: Blog Feed Post

EMC's Greenplum Unit Helps with Large Publicly Accessible Hadoop Cluster

Need to test your Hadoop app on a thousand nodes? Here’s how.

It isn’t often that you can get access to a thousand-node network to test your latest app, but thanks to the efforts of EMC’s Greenplum unit and some additional computing vendors, you can, and more amazingly, it is free of charge too.

The network was announced last fall at Strata and connects 1,000 specialized servers from Supermicro running dual Intel Xeon processors with 48 GB of RAM apiece along with Mellanox 10 GB Ethernet adapters and switches, and a total of 12,000 Seagate 2 TB drives. It is all contained within Greenplum’s Las Vegas data center, with the goal of having the largest publicly accessible Hadoop cluster around. While Yahoo and eBay and others have some fairly large Hadoop clusters, they generally don’t let anyone else come in and try out their apps. The cluster goes under the name of Analytics Workbench. On this page, you can click on the “learn more” button and submit your name if you are interested in using the cluster.

The goal, according to Greenplum staffers, is to have a community and collaborative big data platform that can be applied to a set of analytical problems that have wide appeal. When the Strata announcement was made last fall, Greenplum stated that they wanted to eventually publish any results from the cluster, but they haven’t yet. Intel was one of the first clients to use the workbench (and running a thousand-node job too), but they are still reviewing their results.

Other clients that are running tests on the cluster include Mellanox and VMware, who both donated gear to power it, and a research team from the University of Central Florida. A group from NASA Goddard is using it to perform an analysis of historical weather patterns. The cluster formally opened up in July, and yes, it is really is free of charge. Applicants need to be vetted and work closely with the Greenplum engineers to get their apps uploaded and configured to the cluster.

“We accept bids based on any submitted application and developers can request specific time and resources,” says William Davis, one of the Greenplum product marketers involved with the cluster’s creation. Applications are reviewed by an internal group of Hadoop experts called the Jedi Council, and they try to select who will have the best fit for the next test run on the cluster.

Greenplum intends to use the cluster in a variety of ways besides public testing. Sometime next quarter they will launch a training program for Hadoop. A unique aspect of the program is that each member of the course will be granted access to the cluster to use as a sandbox environment for their own project. They are still working out the details on how this will work. The company has other fee-based programs to leverage its experience with this cluster, including what it calls its Analytics Lab packages. This uses their team of data scientists on specific vertical markets or particular custom applications.

There are several other tools that are offered on the cluster in addition to Hadoop including MapReduce, the parallel job processing software; VMware’s Rubicon system management team; and standard Hadoop add-ons such as Hive, Pig, and Mahout.

Greenplum isn’t the first to have such a large test bed assembled, but probably the first to use this level of gear for Hadoop and other data science activities. In the late 1980s, a group of Novell engineers in Utah created the “SuperLab” which eventually grew to 1,700 PCs connected together. The lab was used to prove the features and scalability of Novell’s Netware network operating system, a piece of software that at one time could be found in most enterprises but now is largely a historical curiosity. Just to give you some perspective, in 1999 the PCs in Novell’s lab had a whopping 256 MB of RAM and 8 GB of storage (try buying that on today’s PCs). How times have changed.

Anyway, the SuperLab team left Novell a few years later and built their own private test lab for a startup called Keylabs. I was one of their early customers, using the facility to publish some of the test results in cNet and other IT publications of the first Web server comparison tests.

The Keylabs engineers very quickly discovered that automating the sequencing and actions of the individual PCs was tedious, and they wrote software that eventually spawned Altiris. Part of the assets of this company was later purchased by Symantec and is still used for their desktop imaging and management tool line.

Speaking of scaling up to a thousand machines automatically, running tests on this scale can be tricky. Greenplum has already seen several hardware failures that take down particular nodes as they have begun using their cluster. And like Keylabs, understanding how to sequence all this gear to come online quickly can be vexing: imagine if each machine takes just ten minutes to boot up and launch an app: times ten or twenty nodes that isn’t much of a big deal, but when you are trying to bring up hundreds it could tie up the cluster for the better part of a week in just starting up the tests. “It is a bit of a challenge in educating our customers on how to use and manage something of this size and how to deploy their software across the entire cluster. You can’t deploy software serially, and we have to make sure that our customers understand these issues,” says Davis.

So get your application in now for testing your app. You could be making computing history.

More Stories By David Strom

David Strom is an international authority on network and Internet technologies. He has written extensively on the topic for 20 years for a wide variety of print publications and websites, such as The New York Times, TechTarget.com, PC Week/eWeek, Internet.com, Network World, Infoworld, Computerworld, Small Business Computing, Communications Week, Windows Sources, c|net and news.com, Web Review, Tom's Hardware, EETimes, and many others.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@BigDataExpo Stories
When it comes to cloud computing, the ability to turn massive amounts of compute cores on and off on demand sounds attractive to IT staff, who need to manage peaks and valleys in user activity. With cloud bursting, the majority of the data can stay on premises while tapping into compute from public cloud providers, reducing risk and minimizing need to move large files. In his session at 18th Cloud Expo, Scott Jeschonek, Director of Product Management at Avere Systems, discussed the IT and busin...
There will be new vendors providing applications, middleware, and connected devices to support the thriving IoT ecosystem. This essentially means that electronic device manufacturers will also be in the software business. Many will be new to building embedded software or robust software. This creates an increased importance on software quality, particularly within the Industrial Internet of Things where business-critical applications are becoming dependent on products controlled by software. Qua...
In addition to all the benefits, IoT is also bringing new kind of customer experience challenges - cars that unlock themselves, thermostats turning houses into saunas and baby video monitors broadcasting over the internet. This list can only increase because while IoT services should be intuitive and simple to use, the delivery ecosystem is a myriad of potential problems as IoT explodes complexity. So finding a performance issue is like finding the proverbial needle in the haystack.
"We host and fully manage cloud data services, whether we store, the data, move the data, or run analytics on the data," stated Kamal Shannak, Senior Development Manager, Cloud Data Services, IBM, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to imp...
Large scale deployments present unique planning challenges, system commissioning hurdles between IT and OT and demand careful system hand-off orchestration. In his session at @ThingsExpo, Jeff Smith, Senior Director and a founding member of Incenergy, will discuss some of the key tactics to ensure delivery success based on his experience of the last two years deploying Industrial IoT systems across four continents.
With the proliferation of both SQL and NoSQL databases, organizations can now target specific fit-for-purpose database tools for their different application needs regarding scalability, ease of use, ACID support, etc. Platform as a Service offerings make this even easier now, enabling developers to roll out their own database infrastructure in minutes with minimal management overhead. However, this same amount of flexibility also comes with the challenges of picking the right tool, on the right ...
Organizations planning enterprise data center consolidation and modernization projects are faced with a challenging, costly reality. Requirements to deploy modern, cloud-native applications simultaneously with traditional client/server applications are almost impossible to achieve with hardware-centric enterprise infrastructure. Compute and network infrastructure are fast moving down a software-defined path, but storage has been a laggard. Until now.
DevOps at Cloud Expo – being held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA – announces that its Call for Papers is open. Born out of proven success in agile development, cloud computing, and process automation, DevOps is a macro trend you cannot afford to miss. From showcase success stories from early adopters and web-scale businesses, DevOps is expanding to organizations of all sizes, including the world's largest enterprises – and delivering real results. Am...
"This week we're really focusing on scalability, asset preservation and how do you back up to the cloud and in the cloud with object storage, which is really a new way of attacking dealing with your file, your blocked data, where you put it and how you access it," stated Jeff Greenwald, Senior Director of Market Development at HGST, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform. In his session at @ThingsExpo, Craig Sproule, CEO of Metavine, demonstrated how to move beyond today's coding paradigm and shared the must-have mindsets for removing complexity from the develo...
The 19th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Digital Transformation, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportuni...
“We're a global managed hosting provider. Our core customer set is a U.S.-based customer that is looking to go global,” explained Adam Rogers, Managing Director at ANEXIA, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Predictive analytics tools monitor, report, and troubleshoot in order to make proactive decisions about the health, performance, and utilization of storage. Most enterprises combine cloud and on-premise storage, resulting in blended environments of physical, virtual, cloud, and other platforms, which justifies more sophisticated storage analytics. In his session at 18th Cloud Expo, Peter McCallum, Vice President of Datacenter Solutions at FalconStor, discussed using predictive analytics to mon...
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
IoT is rapidly changing the way enterprises are using data to improve business decision-making. In order to derive business value, organizations must unlock insights from the data gathered and then act on these. In their session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, and Peter Shashkin, Head of Development Department at EastBanc Technologies, discussed how one organization leveraged IoT, cloud technology and data analysis to improve customer experiences and effi...
"We've discovered that after shows 80% if leads that people get, 80% of the conversations end up on the show floor, meaning people forget about it, people forget who they talk to, people forget that there are actual business opportunities to be had here so we try to help out and keep the conversations going," explained Jeff Mesnik, Founder and President of ContentMX, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
"When you think about the data center today, there's constant evolution, The evolution of the data center and the needs of the consumer of technology change, and they change constantly," stated Matt Kalmenson, VP of Sales, Service and Cloud Providers at Veeam Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Let’s face it, embracing new storage technologies, capabilities and upgrading to new hardware often adds complexity and increases costs. In his session at 18th Cloud Expo, Seth Oxenhorn, Vice President of Business Development & Alliances at FalconStor, discussed how a truly heterogeneous software-defined storage approach can add value to legacy platforms and heterogeneous environments. The result reduces complexity, significantly lowers cost, and provides IT organizations with improved efficienc...
With 15% of enterprises adopting a hybrid IT strategy, you need to set a plan to integrate hybrid cloud throughout your infrastructure. In his session at 18th Cloud Expo, Steven Dreher, Director of Solutions Architecture at Green House Data, discussed how to plan for shifting resource requirements, overcome challenges, and implement hybrid IT alongside your existing data center assets. Highlights included anticipating workload, cost and resource calculations, integrating services on both sides...