Click here to close now.


@BigDataExpo Authors: Yeshim Deniz, Liz McMillan, Esmeralda Swartz, Carmen Gonzalez, Dana Gardner

Related Topics: @CloudExpo, Java IoT, Microservices Expo, Linux Containers, Containers Expo Blog, @BigDataExpo

@CloudExpo: Article

APIs: A Soup to Nuts Analysis

Your APIs are doorways into your business systems

APIs are transforming businesses to extend the reach of their information systems and data. In technical terms, APIs are rather finite, allowing two software applications to exchange data and execute procedures. Yet their business impact is so great that API management requires a far from finite thought process. A "soup to nuts" approach is needed to help companies determine how to achieve the most business success with APIs.

An API enables applications to access back-end systems using lightweight, standards-based protocols such as REST and JSON. For example, if you were in food distribution, the actual business of soup and nuts, you could develop an API that lets mobile app developers tap directly into your ERP system. They could build consumer apps for recipes that showed ingredients that you had in stock. Grocery stores could build apps that gave their shoppers bulk discounts on large orders and so forth. The API extends your food distribution business into customer realms where you've never played before. This is fantastic. But it also presents a number of challenges to the management of the business.

An API Is a Product
Whether you like it or not, your API is basically a software product. It should be managed like one. As external parties connect with your API, they will reasonably expect the kind of professional communication and seamless updates that they would find with any web-based software business. If you are a mobile app developer who is accustomed to tapping into's API or any number of comparable interfaces, you will expect the same kind of experience when you connect with a corporate API.

Perhaps more significantly, the API represents an investment of resources. As such, it should serve a well-thought-out and thorough business strategy just like any software product. This means managing its lifecycle. The software lifecycle starts with planning, continues through development and operation, and ends with retirement and replacement. Ideally each phase of the lifecycle should be managed and monitored for optimal results. Certainly better API lifecycle management will foster more satisfied developers and partner communities.

Planning and Building Your APIs
Sometimes an API just is. Someone in the IT department created it for some reason that has since been forgotten, and there it is. It may or may not be great, but you have it. Should you use it "as is" or wait? This is more common than you might imagine. If you are approaching your APIs as if they were products, you would wait and work through a plan first. You have to get together with business and technology stakeholders and determine the business purpose of the APIs that you will be building. Now, this begs the question: Who exactly is this "you" we are referring to? That is one of your first planning challenges. Who is going to own the API business products that will extend the business into new spheres? So first you have to figure out who owns the API, and often it will be a collection of people that represent different parts of a business. Then the next planning task is to understand the cost/benefit outcomes for the business and intended users.

Chances are there will be more than one API developed and within each API there will be a host of different features to be developed. You will have to devise a priority for the rollout of new features. For instance, with the recipe app, the first release of the API might include simple searching and browsing. The second release might enable users to order products for delivery. The third release could let users pay for orders with credits cards and so forth. Not all of these features can be perfected at once.

Once the API is up, the organization has to be in place to support it. It is necessary to think through the empty seats that might need to be filled so that users of your API will feel as if they are connected to a responsive, living business. For example, if you are planning to invite developers from Europe to connect with your API, you need to have a technical point of contact for them as well as documentation that will be meaningful to them, perhaps translated into different languages. You will have to structure your business to support and manage the API.

Running the API: Protect, Secure, Manage
Your APIs are doorways into your business systems. That's great, but it is also a bit scary. APIs need strong, coherent security and management. If you have a handful of APIs, you will be able to stay on top of their availability, security and provisioning by hand. As your API program grows and it will if you are doing it right, you will likely find that using an API platform becomes a best practice - simply because it can take care of the serious work of API management. Effectively running an API through a platform involves the following:

  • Support non-functional requirements - These include message protocol handling, security policy, authentication and authorization, etc.
  • Manage provisioning and access control for apps - This means the selective provisioning of API access. Corporate APIs are not like their consumer counterparts where you want millions of users. With a corporate API, controlled access is where the users are the most valuable, not the most numerous. The API platform should enable selective provisioning that is still highly automated and light in terms of administrative load. Unfettered usage can bring a host of problems. Traffic and load management can get strained. Worse is the possible need to add extra server instances of costly business software just to satisfy API-driven demand. If the increased load is not generating revenue, satisfying it with more instances will be a waste of money.
  • Monetization and control - An API might be a profit center itself. For example, a food distribution company API could be monetized from selling access to data about sales trends in the food industry. This scenario requires the ability to license access to the API.
  • Provide API monitoring - APIs are just like any other piece of enterprise IT that is monitored for its system health, response times, and availability. In some cases, uptime may not be a big issue. However when your customers and partners are connecting to an API, you will want to know whether it is up, down, or running slowly. The API platform should provide monitoring functionality as well as failover for APIs that go down.

Sharing Your APIs: Publish, Support, Syndicate
APIs succeed when they are shared. The API platform can help create a marketplace where developers can discover your APIs and request permission to use them. The marketplace can be internal to your company or set up for external relationships. Whether it is a developer portal or something comparable, the platform should provide the below sharing functionality:

  • Enable you to interact with and recognize your API developers. In some cases, this process can be set up on a self-service basis to allow it to scale without a major resource investment on your part.
  • Facilitate the creation of great documentation about your API and how to use it.
  • Make testing against your API as easy as possible.
  • Monetize your API to assist in future cost benefit analysis.

Analyze Your API Program: Measure, Report, Iterate
Measurement of program results is in the DNA of most good IT managers. The API platform can help make this process as simple as possible. You will likely want to measure and report on the success of your program across the whole API lifecycle. Lifecycle and results are usually linked. For instance, if you see adoption of the recipe API rise as new features are added and new versions are introduced, that is an important finding to determine the payoff of earlier investments in the program and guide future direction. The right API measuring tools can help you drive improvement back into the planning stage. It is a never-ending cycle.

The Platform Approach to API Lifecycle Management
The execution of an API program that manages APIs like products across their full lifecycles requires a combination of organization and technology. The technology alone cannot make it happen. The organization cannot do it alone without the proper tools. In our experience working with many large enterprises, the best practice is to match the API owners with a platform approach to lifecycle management. The platform can be a complete offering, such as our own SOA Software API Management Solution, or it can be built in a variety of other ways. What's important is to recognize that the people, process and platform need to work together to effect comprehensive API lifecycle management. This will ensure the success of the API program.

Lifecycle Management
Ideally lifecycle management will be baked into the structure of the platform tooling, allowing for automated approvals and workflow for each stage of the API lifecycle. This means having the ability to align work streams related to API lifecycle, such as costing, product management, documentation and legal. People involved in running the API program can thus work together efficiently. The API development capability itself should have API modeling, templates, versioning and change management, and impact analysis. These features serve the goal of managing the full API lifecycle without an untenable resource investment.

API management typically works using a proxy that can be managed and monitored more easily than the API itself. For multiple APIs, the best approach is a "gateway" or collective proxy with added integration and mediation capabilities including:

  • Process Composition/Orchestration.
  • Security, including AU/AZ, attack prevention, and protecting your systems from abuse.
  • Caching and paging.
  • Supporting multiple mobile app platforms.
  • Managing quality of service (QoS).

Exposing APIs to your systems involves building a community. Developers whom you don't know will be writing code to access your data through APIs. For the process to work, they need to feel as if they are connecting with real people. This can be achieved in part with self-service and automation and actual human involvement. A developer portal is essential for the following functions:

  • Self-service community to promote innovation and lower support costs.
  • Interactive documentation to increase adoption and encourage experimentation with your APIs.
  • App provisioning so you can gain visibility and control over the apps that are hitting your APIs.
  • Integrated testing to speed up learning.
  • Analytics to provide feedback and measure success.

Getting There
You have your soup. You want the nuts. There is a path to dessert. However all this requires a mature and profitable API program. Our recommendation is to think about what you want to accomplish with APIs from the vantage points of productization and lifecycle. Building APIs puts you into the business of creating software products. While you may not see the API as a product, it is one. And, like any software product, it needs to be managed across its entire lifecycle. From planning through running and sharing, the API has to be monitored and secured. The work of managing it needs to be handled efficiently so that the program can be financially and strategically beneficial to the business. The best practice to ensure all of these outcomes is to use an API management platform to automate the handling of all of your APIs as they progress across their lifecycles. That will get you the nuts. The alternative is to go nuts, and that's not what anyone wants.

More Stories By Alistair Farquharson, CTO, SOA Software

Alistair Farquharson is a visionary industry veteran focused on using disruptive technologies to drive business growth and improve efficiency and agility within organizations. As the CTO of SOA Software Alistair is helping to shape and mature the enterprise API and SOA industry. Alistair is responsible for product strategy and development for this leading Enterprise API and SOA Governance company. He spends a great deal of time and energy shaping customer and industry direction with regular conference keynote appearances, discussions with customers, and even the occasional foray into writing code. Alistair has been at the forefront of many technology waves from enterprise web architecture, through web services and SOA, and now into APIs. His expertise spans a wide range of technologies and businesses, and he brings a unique global focus to everything he does.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.

@BigDataExpo Stories
As more intelligent IoT applications shift into gear, they’re merging into the ever-increasing traffic flow of the Internet. It won’t be long before we experience bottlenecks, as IoT traffic peaks during rush hours. Organizations that are unprepared will find themselves by the side of the road unable to cross back into the fast lane. As billions of new devices begin to communicate and exchange data – will your infrastructure be scalable enough to handle this new interconnected world?
SYS-CON Events announced today that Super Micro Computer, Inc., a global leader in high-performance, high-efficiency server, storage technology and green computing, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Supermicro (NASDAQ: SMCI), the leading innovator in high-performance, high-efficiency server technology is a premier provider of advanced server Building Block Solutions® for Data ...
SYS-CON Events announced today the Containers & Microservices Bootcamp, being held November 3-4, 2015, in conjunction with 17th Cloud Expo, @ThingsExpo, and @DevOpsSummit at the Santa Clara Convention Center in Santa Clara, CA. This is your chance to get started with the latest technology in the industry. Combined with real-world scenarios and use cases, the Containers and Microservices Bootcamp, led by Janakiram MSV, a Microsoft Regional Director, will include presentations as well as hands-on...
SYS-CON Events announced today that Dyn, the worldwide leader in Internet Performance, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Dyn is a cloud-based Internet Performance company. Dyn helps companies monitor, control, and optimize online infrastructure for an exceptional end-user experience. Through a world-class network and unrivaled, objective intelligence into Internet condit...
SYS-CON Events announced today that Sandy Carter, IBM General Manager Cloud Ecosystem and Developers, and a Social Business Evangelist, will keynote at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA.
The Internet of Things (IoT) is growing rapidly by extending current technologies, products and networks. By 2020, Cisco estimates there will be 50 billion connected devices. Gartner has forecast revenues of over $300 billion, just to IoT suppliers. Now is the time to figure out how you’ll make money – not just create innovative products. With hundreds of new products and companies jumping into the IoT fray every month, there’s no shortage of innovation. Despite this, McKinsey/VisionMobile data...
The IoT market is on track to hit $7.1 trillion in 2020. The reality is that only a handful of companies are ready for this massive demand. There are a lot of barriers, paint points, traps, and hidden roadblocks. How can we deal with these issues and challenges? The paradigm has changed. Old-style ad-hoc trial-and-error ways will certainly lead you to the dead end. What is mandatory is an overarching and adaptive approach to effectively handle the rapid changes and exponential growth.
With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo in Silicon Valley. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be! Internet of @ThingsExpo, taking place Nov 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 17th Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading in...
There will be 20 billion IoT devices connected to the Internet soon. What if we could control these devices with our voice, mind, or gestures? What if we could teach these devices how to talk to each other? What if these devices could learn how to interact with us (and each other) to make our lives better? What if Jarvis was real? How can I gain these super powers? In his session at 17th Cloud Expo, Chris Matthieu, co-founder and CTO of Octoblu, will show you!
The IoT is upon us, but today’s databases, built on 30-year-old math, require multiple platforms to create a single solution. Data demands of the IoT require Big Data systems that can handle ingest, transactions and analytics concurrently adapting to varied situations as they occur, with speed at scale. In his session at @ThingsExpo, Chad Jones, chief strategy officer at Deep Information Sciences, will look differently at IoT data so enterprises can fully leverage their IoT potential. He’ll sha...
The enterprise is being consumerized, and the consumer is being enterprised. Moore's Law does not matter anymore, the future belongs to business virtualization powered by invisible service architecture, powered by hyperscale and hyperconvergence, and facilitated by vertical streaming and horizontal scaling and consolidation. Both buyers and sellers want instant results, and from paperwork to paperless to mindless is the ultimate goal for any seamless transaction. The sweetest sweet spot in innov...
Today air travel is a minefield of delays, hassles and customer disappointment. Airlines struggle to revitalize the experience. GE and M2Mi will demonstrate practical examples of how IoT solutions are helping airlines bring back personalization, reduce trip time and improve reliability. In their session at @ThingsExpo, Shyam Varan Nath, Principal Architect with GE, and Dr. Sarah Cooper, M2Mi's VP Business Development and Engineering, will explore the IoT cloud-based platform technologies driv...
The Internet of Everything is re-shaping technology trends–moving away from “request/response” architecture to an “always-on” Streaming Web where data is in constant motion and secure, reliable communication is an absolute necessity. As more and more THINGS go online, the challenges that developers will need to address will only increase exponentially. In his session at @ThingsExpo, Todd Greene, Founder & CEO of PubNub, will explore the current state of IoT connectivity and review key trends an...
SYS-CON Events announced today that DataClear Inc. will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. The DataClear ‘BlackBox’ is the only solution that moves your PC, browsing and data out of the United States and away from prying (and spying) eyes. Its solution automatically builds you a clean, on-demand, virus free, new virtual cloud based PC outside of the United States, and wipes it clean...
SYS-CON Events announced today that Machkey International Company will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Machkey provides advanced connectivity solutions for just about everyone. Businesses or individuals, Machkey is dedicated to provide high-quality and cost-effective products to meet all your needs.
Nowadays, a large number of sensors and devices are connected to the network. Leading-edge IoT technologies integrate various types of sensor data to create a new value for several business decision scenarios. The transparent cloud is a model of a new IoT emergence service platform. Many service providers store and access various types of sensor data in order to create and find out new business values by integrating such data.
There are so many tools and techniques for data analytics that even for a data scientist the choices, possible systems, and even the types of data can be daunting. In his session at @ThingsExpo, Chris Harrold, Global CTO for Big Data Solutions for EMC Corporation, will show how to perform a simple, but meaningful analysis of social sentiment data using freely available tools that take only minutes to download and install. Participants will get the download information, scripts, and complete en...
Too often with compelling new technologies market participants become overly enamored with that attractiveness of the technology and neglect underlying business drivers. This tendency, what some call the “newest shiny object syndrome,” is understandable given that virtually all of us are heavily engaged in technology. But it is also mistaken. Without concrete business cases driving its deployment, IoT, like many other technologies before it, will fade into obscurity.
Achim Weiss is Chief Executive Officer and co-founder of ProfitBricks. In 1995, he broke off his studies to co-found the web hosting company "Schlund+Partner." The company "Schlund+Partner" later became the 1&1 web hosting product line. From 1995 to 2008, he was the technical director for several important projects: the largest web hosting platform in the world, the second largest DSL platform, a video on-demand delivery network, the largest eMail backend in Europe, and a universal billing syste...
Cloud computing delivers on-demand resources that provide businesses with flexibility and cost-savings. The challenge in moving workloads to the cloud has been the cost and complexity of ensuring the initial and ongoing security and regulatory (PCI, HIPAA, FFIEC) compliance across private and public clouds. Manual security compliance is slow, prone to human error, and represents over 50% of the cost of managing cloud applications. Determining how to automate cloud security compliance is critical...

Tweets by @BigDataExpo

@BigDataExpo Blogs
This week, the team assembled in NYC for @Cloud Expo 2015 and @ThingsExpo 2015. For the past four years, this has been a must-attend event for MetraTech. We were happy to once again join industry visionaries, colleagues, customers and even competitors to share and explore the ways in which the Internet of Things (IoT) will impact our industry. Over the course of the show, we discussed the types of challenges we will collectively need to solve to capitalize on the opportunity IoT presents.
Today’s connected world is moving from devices towards things, what this means is that by using increasingly low cost sensors embedded in devices we can create many new use cases. These span across use cases in cities, vehicles, home, offices, factories, retail environments, worksites, health, logistics, and health. These use cases rely on ubiquitous connectivity and generate massive amounts of data at scale. These technologies enable new business opportunities, ways to optimize and automate, along with new ways to engage with users.
Dasher Technologies is helping to usher in the democratization of big data value to more players in less time with analytics in a cloud services model
Heat maps offer a unique way to represent data sets in a variety of settings. A common example you see most mornings is through weather reports on the news showing the movement and predictions of pressure and precipitation, intensifying across bold color schemes. Another familiar example is voting representation in specific areas of the country during election times (think red state vs. blue state). Retail stores utilize heat maps to enable managers and executives to better understand the functionality of their space, as well as the habits and preferences of the people entering the store.
Developing software for the Internet of Things (IoT) comes with its own set of challenges. Security, privacy, and unified standards are a few key issues. In addition, each IoT product is comprised of at least three separate application components: the software embedded in the device, the backend big-data service, and the mobile application for the end user's controls. Each component is developed by a different team, using different technologies and practices, and deployed to a different stack/target - this makes the integration of these separate pipelines and the coordination of software upd...
“Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications and services) that can be rapidly provisioned and released with minimal management.” While this definition is broadly accepted and has, in fact, been my adopted standard for years, it only describes technical aspects of cloud computing. The amalgamation of technologies used to deliver cloud services is not even half the story. Above all else, the successful employment requires a tight linkage to the econ...
Today air travel is a minefield of delays, hassles and customer disappointment. Airlines struggle to revitalize the experience. GE and M2Mi will demonstrate practical examples of how IoT solutions are helping airlines bring back personalization, reduce trip time and improve reliability. In their session at @ThingsExpo, Shyam Varan Nath, Principal Architect with GE, and Dr. Sarah Cooper, M2Mi's VP Business Development and Engineering, will explore the IoT cloud-based platform technologies driving this change including privacy controls, data transparency and integration of real time context w...
DevOps Summit at Cloud Expo 2014 Silicon Valley was a terrific event for us. The Qubell booth was crowded on all three days. We ran demos every 30 minutes with folks lining up to get a seat and usually standing around. It was great to meet and talk to over 500 people! My keynote was well received and so was Stan's joint presentation with RingCentral on Devops for BigData. I also participated in two Power Panels – ‘Women in Technology’ and ‘Why DevOps Is Even More Important than You Think,’ both featuring brilliant colleagues and moderators and it was a blast to be a part of.
All we need to do is have our teams self-organize, and behold! Emergent design and/or architecture springs up out of the nothingness! If only it were that easy, right? I follow in the footsteps of so many people who have long wondered at the meanings of such simple words, as though they were dogma from on high. Emerge? Self-organizing? Profound, to be sure. But what do we really make of this sentence?
Today’s modern day industrial revolution is being shaped by ubiquitous connectivity, machine to machine (M2M) communications, the Internet of Things (IoT), open APIs leading to a surge in new applications and services, partnerships and eventual marketplaces. IoT has the potential to transform industry and society much like advances in steam technology, transportation, mass production and communications ushered in the industrial revolution in the 18th and 19th centuries.
The potential of big data is only limited by the creative thinking of your business stakeholders, and that may be the most important concept in the “thinking like a data scientist” process. The “thinking like a data scientist” process guides the business stakeholders into envisioning how big data can optimize their key business processes, create a more compelling customer engagement and uncover new monetization opportunities. But neither the business stakeholders, nor the data scientists, can likely do that envisioning entirely by themselves.
Disaster recovery (DR) has traditionally been a major challenge for IT departments. Even with the advent of server virtualization and other technologies that have simplified DR implementation and some aspects of on-going management, it is still a complex and (often extremely) costly undertaking. For those applications that do not require high availability, but are still mission- and business-critical, the decision as to which [applications] to spend money on for true disaster recovery can be a struggle.
SCOPE is an acronym for Structured Computations Optimized for Parallel Execution, a declarative language for working with large-scale data. It is still under development at Microsoft. If you know SQL then working with SCOPE will be quite easy as SCOPE builds on SQL. The execution environment is different from that RDBMS oriented data. Data is still modeled as rows. Every row has typed columns and eveyr rowset has a well-defined schema. There is a SCOPe compiler that comes up with optimized execution plan and a runtime execution plan.
If you’re running Big Data applications, you’re going to want to look at some kind of distributed processing system. Hadoop is one of the best-known clustering systems, but how are you going to process all your data in a reasonable time frame? MapReduce has become a standard, perhaps the standard, for distributed file systems. While it’s a great system already, it’s really geared toward batch use, with jobs needing to queue for later output. This can severely hamper your flexibility. What if you want to explore some of your data? If it’s going to take all night, forget about it.
Too many multinational corporations delete little, if any, data even though at its creation, more than 70 percent of this data is useless for business, regulatory or legal reasons.[1] The problem is hoarding, and what businesses need is their own “Hoarders” reality show about people whose lives are driven by their stuff[2] (corporations are legally people, after all). The goal of such an intervention (and this article)? Turning hoarders into collectors.

About @BigDataExpo
Big Data focuses on how to use your own enterprise data – processed in the Cloud – most effectively to drive value for your business.