Welcome!

@BigDataExpo Authors: Pat Romanski, Liz McMillan, Elizabeth White, Amy Eager, John Rauser

Related Topics: @BigDataExpo, Cognitive Computing , Machine Learning

@BigDataExpo: Article

Patent Data Quality | @CloudExpo #BigData #Analytics #AI #MachineLearning

Is clean data a pipe dream?

The United States Patent and Trademark Office (USPTO) recently announced an expansion of PatentsView, its visualization tool for US patents. First launched a few years ago, the intent behind the tool was to make 40 years of patent filing data available for free to those interested in examining "the dynamics of inventor patenting activity over time." In spite of being limited to patents (not applications) and with a focus only on the US, it offers some interesting visualizations around locations and citations.

In a blog post last month, USPTO director Michelle Lee said the PatentView tool is based on "the highest-quality patent data available," connecting 40 years' worth of information about inventors, their organizations, and their locations in unprecedented ways. The newly revamped interface presents three user-friendly starting points - relationship, locations, and comparison visualizations - which allow for deeper exploration and detailed views. However, through no fault of their own, the USPTO dataset is rife with spelling errors, doesn't reflect patent reassignments, and doesn't resolve company subsidiaries or acquisitions.

This issue is not unique to the USPTO. Other PTO offices around the world face similar barriers to presenting "clean" data. The first issue, spelling errors, merely reflects the fact that assignee information (among other fields like inventor names) is manually entered and hence prone to error and inconsistency. For example, "International Business Machines" has been spelled 1,200 different ways as a patent assignee over the last two decades in the USPTO data set.

In addition, PTO data doesn't get corrected or updated based on later corrections or patent reassignments. For example, patent US8176440 was originally - and incorrectly - assigned to Silicon Labs. My company, Innography, filed a certificate of correction to update the assignment, yet the USPTO data and PatentsView still don't reflect this. In fact, Innography research shows that nearly 20 percent of US patents are reassigned in their lifetimes, translating into a significant number of company portfolio errors based on this factor alone.

Finally, PTO data also doesn't reflect when companies purchase each other, when there's a spinoff, or when a subsidiary files patents. Microsoft, for example, now owns all LinkedIn's patents, even if the reassignments haven't been processed.

As a result, PTO data falls far short of reflecting reality, where patents and companies are bought and sold every day, and where data-entry errors exist and are corrected. The accuracy of the data is very low when it comes to representing company patent portfolios in the real world.

The Cost of Free Data
The USPTO aims to increase the transparency of patenting and invention processes. But if the quality of data and search results is questionable, what good is it to IP practitioners?

There is rich information available through the patenting process, including economic research, prior-art searching, and discovery of broader trends around filing patterns. However, it was never intended to be used as-is to inform strategic business decisions such as in and out licensing, merger and acquisition activities, or portfolio pruning and maintenance decisions.

It makes sense for PTOs to offer their data for free as a way to engage the community's interest in patenting processes. However, too many lightweight patent analytics tools use this flawed data verbatim to tout their "data quality" to IP professionals.

Many patent analyses start with a company's patent portfolio, such as competitive benchmarking, acquisition analysis, and negotiation preparation. In addition, just about every board-level question about patents requires accurate patent ownership information: "Are we ahead of or behind this competitor?" "What companies should we be worried about in this technology area?"

Poor data quality makes it difficult, if not impossible, to answer those questions accurately. To create the most accurate data set possible, companies must use other sources of information to crosscheck and improve patent data accuracy.

Innography data scientists process more than 2,000 company acquisitions annually, and our user base suggests another 5,000 updates each year. As a result, Innography has created more than 10 million data-correction rules over the last decade, which are continuously updated via machine learning and crowdsourcing.

Company leaders must be able to use patent reports to assess market opportunities and make strategic business decisions. This requires an IP analytics solution that reflects real-world changes, and doesn't rely on poor data quality from outdated PTO assignee information.

More Stories By Tyron Stading

Tyron Stading is president and founder of Innography, and chief data officer for CPA Global. He has been named one of the “World’s Leading IP Strategists" by IAM, and one of National Law Journal's "50 Intellectual Property Trailblazers & Pioneers". Before Innography, Tyron was an IBM worldwide industry solutions manager in the telecommunications and utilities sector, and worked at several start-ups focused on mobile communications and networks security. He has published multiple research papers and filed more than three dozen patents. Tyron has a BS in Computer Science from Stanford University and an MS in Technology Commercialization from The University of Texas.

@BigDataExpo Stories
"We focus on composable infrastructure. Composable infrastructure has been named by companies like Gartner as the evolution of the IT infrastructure where everything is now driven by software," explained Bruno Andrade, CEO and Founder of HTBase, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
IoT solutions exploit operational data generated by Internet-connected smart “things” for the purpose of gaining operational insight and producing “better outcomes” (for example, create new business models, eliminate unscheduled maintenance, etc.). The explosive proliferation of IoT solutions will result in an exponential growth in the volume of IoT data, precipitating significant Information Governance issues: who owns the IoT data, what are the rights/duties of IoT solutions adopters towards t...
With the introduction of IoT and Smart Living in every aspect of our lives, one question has become relevant: What are the security implications? To answer this, first we have to look and explore the security models of the technologies that IoT is founded upon. In his session at @ThingsExpo, Nevi Kaja, a Research Engineer at Ford Motor Company, discussed some of the security challenges of the IoT infrastructure and related how these aspects impact Smart Living. The material was delivered interac...
Cloud applications are seeing a deluge of requests to support the exploding advanced analytics market. “Open analytics” is the emerging strategy to deliver that data through an open data access layer, in the cloud, to be directly consumed by external analytics tools and popular programming languages. An increasing number of data engineers and data scientists use a variety of platforms and advanced analytics languages such as SAS, R, Python and Java, as well as frameworks such as Hadoop and Spark...
No hype cycles or predictions of zillions of things here. IoT is big. You get it. You know your business and have great ideas for a business transformation strategy. What comes next? Time to make it happen. In his session at @ThingsExpo, Jay Mason, Associate Partner at M&S Consulting, presented a step-by-step plan to develop your technology implementation strategy. He discussed the evaluation of communication standards and IoT messaging protocols, data analytics considerations, edge-to-cloud tec...
When growing capacity and power in the data center, the architectural trade-offs between server scale-up vs. scale-out continue to be debated. Both approaches are valid: scale-out adds multiple, smaller servers running in a distributed computing model, while scale-up adds fewer, more powerful servers that are capable of running larger workloads. It’s worth noting that there are additional, unique advantages that scale-up architectures offer. One big advantage is large memory and compute capacity...
New competitors, disruptive technologies, and growing expectations are pushing every business to both adopt and deliver new digital services. This ‘Digital Transformation’ demands rapid delivery and continuous iteration of new competitive services via multiple channels, which in turn demands new service delivery techniques – including DevOps. In this power panel at @DevOpsSummit 20th Cloud Expo, moderated by DevOps Conference Co-Chair Andi Mann, panelists examined how DevOps helps to meet the de...
"When we talk about cloud without compromise what we're talking about is that when people think about 'I need the flexibility of the cloud' - it's the ability to create applications and run them in a cloud environment that's far more flexible,” explained Matthew Finnie, CTO of Interoute, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
The taxi industry never saw Uber coming. Startups are a threat to incumbents like never before, and a major enabler for startups is that they are instantly “cloud ready.” If innovation moves at the pace of IT, then your company is in trouble. Why? Because your data center will not keep up with frenetic pace AWS, Microsoft and Google are rolling out new capabilities. In his session at 20th Cloud Expo, Don Browning, VP of Cloud Architecture at Turner, posited that disruption is inevitable for comp...
SYS-CON Events announced today that Datanami has been named “Media Sponsor” of SYS-CON's 21st International Cloud Expo, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Datanami is a communication channel dedicated to providing insight, analysis and up-to-the-minute information about emerging trends and solutions in Big Data. The publication sheds light on all cutting-edge technologies including networking, storage and applications, and thei...
SYS-CON Events announced today that Silicon India has been named “Media Sponsor” of SYS-CON's 21st International Cloud Expo, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Published in Silicon Valley, Silicon India magazine is the premiere platform for CIOs to discuss their innovative enterprise solutions and allows IT vendors to learn about new solutions that can help grow their business.
Join us at Cloud Expo June 6-8 to find out how to securely connect your cloud app to any cloud or on-premises data source – without complex firewall changes. More users are demanding access to on-premises data from their cloud applications. It’s no longer a “nice-to-have” but an important differentiator that drives competitive advantages. It’s the new “must have” in the hybrid era. Users want capabilities that give them a unified view of the data to get closer to customers and grow business. The...
"We do one of the best file systems in the world. We learned how to deal with Big Data many years ago and we implemented this knowledge into our software," explained Jakub Ratajczak, Business Development Manager at MooseFS, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
In his session at @ThingsExpo, Eric Lachapelle, CEO of the Professional Evaluation and Certification Board (PECB), provided an overview of various initiatives to certify the security of connected devices and future trends in ensuring public trust of IoT. Eric Lachapelle is the Chief Executive Officer of the Professional Evaluation and Certification Board (PECB), an international certification body. His role is to help companies and individuals to achieve professional, accredited and worldwide re...
SYS-CON Events announced today that TechTarget has been named “Media Sponsor” of SYS-CON's 21st International Cloud Expo, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. TechTarget storage websites are the best online information resource for news, tips and expert advice for the storage, backup and disaster recovery markets.
The current age of digital transformation means that IT organizations must adapt their toolset to cover all digital experiences, beyond just the end users’. Today’s businesses can no longer focus solely on the digital interactions they manage with employees or customers; they must now contend with non-traditional factors. Whether it's the power of brand to make or break a company, the need to monitor across all locations 24/7, or the ability to proactively resolve issues, companies must adapt to...
You know you need the cloud, but you’re hesitant to simply dump everything at Amazon since you know that not all workloads are suitable for cloud. You know that you want the kind of ease of use and scalability that you get with public cloud, but your applications are architected in a way that makes the public cloud a non-starter. You’re looking at private cloud solutions based on hyperconverged infrastructure, but you’re concerned with the limits inherent in those technologies.
"Loom is applying artificial intelligence and machine learning into the entire log analysis process, from start to finish and at the end you will get a human touch,” explained Sabo Taylor Diab, Vice President, Marketing at Loom Systems, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
SYS-CON Events announced today that EnterpriseTech has been named “Media Sponsor” of SYS-CON's 21st International Cloud Expo, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. EnterpriseTech is a professional resource for news and intelligence covering the migration of high-end technologies into the enterprise and business-IT industry, with a special focus on high-tech solutions in new product development, workload management, increased effi...
Automation is enabling enterprises to design, deploy, and manage more complex, hybrid cloud environments. Yet the people who manage these environments must be trained in and understanding these environments better than ever before. A new era of analytics and cognitive computing is adding intelligence, but also more complexity, to these cloud environments. How smart is your cloud? How smart should it be? In this power panel at 20th Cloud Expo, moderated by Conference Chair Roger Strukhoff, paneli...