MarkLogicの新しいウェブサイトは www.progress.com/marklogic です。それをご覧ください。
BLOG ARTICLE

Knowledge Sharing Challenges

Back to blog
07.05.2022
所要時間:7分
Back to blog
07.05.2022
所要時間:7分

We’re all familiar with small teams of smart people, working together to understand data and what it might mean. These teams might be working in finance, life sciences, logistics, government – any pursuit that involves interpreting digital facts and their interpretation in context.

But the familiar model of using small teams of smart people doesn’t scale well at all.

What works for three doesn’t work as well when there are ten involved. Get to thirty participants and you’ll see serious problems.

Get to hundreds or thousands of participants, and it is now a formidably daunting task – especially if the smart people are scattered across multiple organizations.

Nor can small teams keep up as data complexity scales: whether that involves more data sources, more complexity in data sources, or just the complexity that arises from sheer volume.

Ad-hoc tools like spreadsheets, data marts, and analytic workbenches can only get you so far. And we have plenty of good tools for managing, sharing and analyzing data.

But we don’t have good tools to do the same with what we know about it.

In a previous article, we’ve discussed the importance of facts and what they mean. Evaluating information – and what it means in context – is vitally important to the organizations that invest in these teams.

We’ve also discussed the notion of data agility. We want our teams of smart people to be agile with complex data: being able to quickly make new interpretations regarding any aspect of information, and acting on them quickly.

Sharing data – relatively easy. Sharing our specialized knowledge about data – somewhat harder.

Language Matters for Knowledge Sharing

When teams of smart people work together, they will typically develop a language of precise terms and definitions to explain ideas and concepts.

To outsiders it appears as dense jargon, but to the people who use the language, it’s a very efficiently encoded way of describing important things.

That unique and specialized language also encodes what is known about the topic: definitions, rules, relationships, and so on.

Specialized knowledge is created and shared through specialized language.

When you listen to any two experts start to get into a discussion, there’s often a “definition of terms” phase. Once you’re past that – it’s off to the races. Friction is removed.

Conversely, if those experts can’t agree on terms and definitions, it often gets stuck there. Friction overcomes progress.

When people formalize their language, they are also formalizing their knowledge. Formalized knowledge scales much better than unformalized versions, and makes knowledge sharing easier.

This is true in both the digital and analog worlds.

Language Concepts in the Digital World

In today’s world, the majority of the facts we evaluate are digital ones: an event, an observation, a data feed, or something similar.

We ingest the data and evaluate it. Is it important? Can it safely be ignored? What should be done with it? What does it mean?

An obscure transaction could be relevant to someone managing risk across a large portfolio. The publication of a new research report could be of immediate importance to a life sciences firm, but literally thousands of such notes are published every month.

In any security or intelligence function, many millions of scraps of new and potentially relevant information become available – what do they mean in context? In complex manufacturing or logistics functions, small problems can quickly morph into serious ones if important connections aren’t understood.

What is the formalized language we use to describe and share our specialized knowledge of these and other important things? The answer is – typically there isn’t a good one.

Why Scaling the Number of Smart People Doesn’t Work Well

Richard Feynman‘s first notable achievement was organizing humans as a parallel computer at a time when computers didn’t exist. The enormously complex math calculations required during the WWII war effort led to the recruitment of hundreds of math PhDs and similar.

But they were taking way too long to get their work done.

He broke the math problem into components, assigned each to different teams to work in parallel, and created a novel way to rejoin their results at the end, greatly accelerating the outcome at a critical juncture in history.

The principle he discovered still holds true today: adding more smart people alone won’t solve the problem any faster or better unless there’s a way to connect their combined knowledge effectively. How does this apply today?

Simply put: without shared, formalized language about digital facts and what they mean, teams of smart people won’t scale their specialized knowledge about data effectively.

In the example above, the specialized language used was mathematical descriptions of known facts about particle physics. With that in hand, effective knowledge sharing became possible.

In our current digital world, if there isn’t a formalized language around data and what it means, specialized knowledge about the data can’t be efficiently shared and scaled.

Small teams working together can do this intuitively, larger teams don’t have that luxury.

Knowledge Sharing: A Systems Architecture View

If we were to share this concept from a systems architecture perspective, perhaps using a whiteboard, we’d draw a picture of each person involved – each with their knowledge about the data, what it means, what to do with it, etc. – as a “node” in an idealized knowledge-sharing network, often across organizational boundaries.

You’d want those people “nodes” to be able to publish and subscribe against a shared pool of data elements along with their full and richly described meaning: precise definitions, terms used and those terms defined, the rules and why they exist, important concepts defined precisely, and so on.

Ideally, the data elements would be fully self-described by metadata that encodes what is known about it, how those concepts and meanings relate to others, and so on – a semantic knowledge graph of complex data and its informed interpretations.

One wouldn’t want to allow data to be separated from its interpretation, for example any data that is subject to strict compliance and/or security measures. Not only is uninterpreted data of very low value, it creates risk in several different ways.

As above, a formalized and codified language about terms and concepts relevant to the data at hand greatly improves the speed and fidelity of knowledge sharing, whether in the digital or analog world.

When everyone is using the same language, friction is removed.

Examples Abound

Once you recognize the pattern, you will see it almost anywhere you care to look in any organization, large or small: teams of smart people trying to keep up with torrent of digital facts in the form of complex data.

Without the right tools, it can be difficult, frustrating work.

Any dynamic risk management function has this challenge. Any research or applied R&D pursuit has this challenge. Any complex coordination function has this challenge. Any organization that takes compliance or security seriously has this challenge. Any organization that wants to differentiate itself with unique know-how and expertise has this challenge.

The underlying reason that this specialized knowledge is so difficult to scale is that we aren’t using the right tools to formalize and share what we know about the data.

The people involved don’t share a machine-interpretable and human-interpretable language describing the digital facts and what they mean. If perhaps they are fortunate enough to have this in some form, it is typically divorced from the data that created it.

As a result, any organization with this challenge will find it very difficult indeed to collectively scale their shared, specialized knowledge about data and what it means.

Our Story So Far

Digital facts – and what they mean – are quite important now, and will become even more important over time.

Data agility is the ability to quickly react to new digital facts and new interpretations.

Specialized knowledge about digital facts – data – is hard to scale as there is no shared, formalized language about the digital facts: their precise definitions and interpretations. If there is one, it is typically divorced from the data that created it in the first place.

Ideally, we would keep data and what we know about it – in the form of encoded interpretations – together at all times. Doing so creates data agility.

What we’ll see next is that – surprisingly – the pieces and parts we need for this task have been around for a while, and are somewhat familiar.

We’re just learning to use the tools in new ways.


Learn More

Download our white paper, Data Agility with a Semantic Knowledge Graph

ジェレミー・ベントレー

ジェレミー・ベントレーは、Smartlogicの創設者でセマンティックAIプラットフォーム「Semaphore」を開発しました。今回のSmartlogicの買収に伴いMarkLogicの一員となっています。ジェレミー自身もエンジニアであり、エンタープライズ情報管理に関する問題を解決することにキャリアの大半を費やしてきました。現在は、Semaphoreのモデリング、自動分類、テキスト分析、視覚化機能を使用して新しい製品を開発するイノベーターたちをクライアントとしています。そうしたクライアントの業種はさまざまで、金融、出版、石油・ガス、政府、ライフサイエンスなど多岐に及んでいます。彼らはみな、情報資産への依存度が高く、それらを収益化、管理、統一化する必要性があるという点で共通しています。Smartlogic入社以前は、ニューヨークのフィンテック企業であるマイクロバンク・ソフトウェア(サンガード・データ・システムズによって買収)でマネージングディレクターを務めていました。エディンバラ大学で機械工学の学士(成績優秀者)を取得しています。

この投稿者の投稿をもっと読む

この記事を共有する

続きを読む

Related Posts

今回お読みいただいた投稿のほかにも、おすすめの記事がいくつかあります。またブログの概要ページにアクセスすると、さらに多くの記事をご覧いただけます。

Business Insights

The Next Critical Step for AI: Eliminate Data Bias

Learn about data bias in AI, ways technology can help overcome it, why AI still needs humans, and how you can achieve transparency.

ブログ投稿すべてを表示
Business Insights

データアジリティを実現する方法

ビジネス環境の変化にうまく対応するには、データアジリティが必須です。ビジョナリー組織が何をしたのか、それと同じことをするにはどうすればよいのかをご紹介します。

ブログ投稿すべてを表示
Business Insights

データのアジリティがビジネスに不可欠な理由

Data agility is the ability to make simple, powerful, and immediate changes to any aspect of how information is interpreted and acted on.

ブログ投稿すべてを表示

Sign up for a Demo

もうコンポーネントをつなぎ合わせて時間を無駄にする必要はありません。MarkLogicは、マルチモデルデータベース、検索、セマンティックAI技術の機能を1つのプラットフォームに一元化し、マスタリング、メタデータ管理、政府レベルのセキュリティなどを提供しています。

デモをリクエストする