Open Framework, Information Management Strategy & Collaborative Governance | Data & Social Methodology - MIKE2.0 Methodology
Members
Collapse Expand Close

To join, please contact us.

Improve MIKE 2.0
Collapse Expand Close
Need somewhere to start? How about the most wanted pages; or the pages we know need more work; or even the stub that somebody else has started, but hasn't been able to finish. Or create a ticket for any issues you have found.

Archive for the ‘Information Development’ Category

by: Bsomich
21  Dec  2013

Weekly IM Update.

Missed what happened in the MIKE2.0 Community this week? Check out our weekly update:

 

 
 logo.jpg

The Five Phases of MIKE2.0

In order to realize results more quickly, the MIKE2.0 Methodology has abandoned the traditional linear or waterfall approach to systems development. In its stead, MIKE2.0 has adopted an iterative, agile approach called continuous implementation. This approach divides the development and rollout of anentire system into a series of implementation cycles. These cycles identify and prioritize the portions of the system that can be constructed and rolled out before the entire system is complete. Each cycle also includes

  • A feedback step to evaluate and prioritize the implementation results
  • Strategy changes
  • Improvement requests on the future implementation cycles.

Following this approach, there are five phases to the MIKE2.0 Methodology:

Feel free to check them out when you have a moment.

  Sincerely,

MIKE2.0 Community

Contribute to MIKE:

Start a new article, help with articles under construction or look for other ways to contribute.

Update your personal profile to advertise yourself to the community and interact with other members.

Useful Links:
Home Page
Login
Content Model
FAQs
MIKE2.0 Governance

Join Us on
42.gif

Follow Us on
43 copy.jpg

Join Us on
images.jpg

Did You Know?
All content on MIKE2.0 and any contributions you make are published under the Creative Commons license. This allows you free re-use of our content as long as you add a brief reference back to us.

This Week’s Food for Thought:

Turning decision making into a game

Organisations are more complex today than ever before, largely because of the ability that technology brings to support scale, centralisation and enterprise-wide integration.  One of the unpleasant side effects of this complexity is that it can take too long to get decisions made.Read more.

Making Sense of Microsemantics

Like many, I’m one who’s been around since the cinder block days, once entranced by shiny Tektronix tubes stationed nearby a dusty card sorter. After years using languages as varied as Assembler through Scheme, I’ve come to believe the shift these represented, from procedural to declarative, has well-improved the flexibility of software organizations produce.Read more.

Does your company need data visualization apps?

Few learned folks dispute the fact that the era of Big Data has arrived. Debate terms if you like, but most of us are bombarded with information these days. The question is turning to, How do we attempt to understand all of this data?

Read more.

 

  Forward this message to a friend

 

Category: Information Development
No Comments »

by: Bsomich
14  Dec  2013

Weekly IM Update.

 logo.jpg

What is an Open Methodology Framework?

An Open Methodology Framework is a collaborative environment for building methods to solve complex issues impacting business, technology, and society.  The best methodologies provide repeatable approaches on how to do things well based on established techniques. MIKE2.0′s Open Methodology Framework goes beyond the standards, techniques and best practices common to most methodologies with three objectives:

  • To Encourage Collaborative User Engagement
  • To Provide a Framework for Innovation
  • To Balance Release Stability with Continuous Improvement

We believe that this approach provides a successful framework accomplishing things in a better and collaborative fashion. What’s more, this approach allows for concurrent focus on both method and detailed technology artifacts. The emphasis is on emerging areas in which current methods and technologies lack maturity.

The Open Methodology Framework will be extended over time to include other projects. Another example of an open methodology, is open-sustainability which applies many of these concepts to the area of sustainable development. Suggestions for other Open Methodology projects can be initiated on this article’s talk page.

We hope you find this of benefit and welcome any suggestions you may have to improve it.

Sincerely,

MIKE2.0 Community

Popular Content

Did you know that the following wiki articles are most popular on Google? Check them out, and feel free to edit or expand them!

What is MIKE2.0?
Deliverable Templates
The 5 Phases of MIKE2.0
Overall Task List
Business Assessment Blueprint
SAFE Architecture
Information Governance Solution

Contribute to MIKE:

Start a new article, help with articles under construction or look for other ways to contribute.

Update your personal profile to advertise yourself to the community and interact with other members.

Useful Links:
Home Page
Login
Content Model
FAQs
MIKE2.0 Governance

Join Us on
42.gif

Follow Us on
43 copy.jpg

Join Us on
images.jpg

 

This Week’s Blogs for Thought:

The Sound of Sound Information Security
I like it when I stumble across examples of information management concepts.  While working on a podcast interview with William McKnight discussing his new book Information Management: Strategies for Gaining a Competitive Advantage with Data, I asked William for a song recommendation to play as background music while I read his bio during the opening segment of the podcast.

After William emailed me an Apple iTunes audio file for the song “Mother North” off of the 1996 album Nemesis Divina by Norwegian black metal band Satyricon, as the following screen capture shows, I ran into an issue when I attempted to play the song on my computer.

Read more.

In Defense of Structured Data

Writing a book is always a rewarding and challenging endeavor. Ask any author. As I pen these words, I am going through the final edits on The Visual Organization. As someone who recognizes that plenty of great work has already been done in the area of data visualization, I dutifully cite a slew of sources.

It’s just smart to recognize the previous contributions of others in any field, especially the influential ones. What’s more, attribution and intellectual property are fundamental to the world of publishing. And I certainly like receiving credit for things I have written and said. I get a little miffed when I don’t.

Read more.

The Gordian Knot of Big Data Integration

According to legend, the Gordian Knot was an intricate knot binding the ox-cart of the peasant-turned-king Gordias of Phrygia (located in the Balkan Peninsula of Southeast Europe) to a pole within the temple of Zeus in Gordium.  It was prophesied that whoever loosed the knot would become ruler of all Asia.

In 333 BC, Alexander the Great sliced the knot in half with a stroke of his sword (the so-called “Alexandrian solution”) and that night there was a violent thunderstorm, which Alexander took as a sign that Zeus was pleased and would grant Alexander many victories.  He then proceeded to conquer Asia, and much of the then known world, apparently fulfilling the prophecy.

Read more.

 

Forward to a Friend!

Know someone who might be interested in joining the Mike2.0 Community? Forward this message to a friendQuestions?

If you have any questions, please email us at mike2@openmethodology.org.

Category: Information Development
No Comments »

by: Phil Simon
09  Dec  2013

In Defense of Structured Data

Writing a book is always a rewarding and challenging endeavor. Ask any author. As I pen these words, I am going through the final edits on The Visual Organization. As someone who recognizes that plenty of great work has already been done in the area of data visualization, I dutifully cite a slew of sources.

It’s just smart to recognize the previous contributions of others in any field, especially the influential ones. What’s more, attribution and intellectual property are fundamental to the world of publishing. And I certainly like receiving credit for things I have written and said. I get a little miffed when I don’t.

While this is often a bit of an administrative burden on the author, I get it. This means that I have to ask for and receive permissions from media sites, bloggers, and individuals if I want to use more than a certain number of their words.

Better Data Collection Mechanisms

Still, there’s a way to collect and track this information that makes it less onerous for all involved. No, we don’t have to build a sophisticated database, but simple spreadsheets are far superior for this type of rudimentary data management. (And, make no mistake, that’s exactly what this exercise invovles.)

Lamentably, though, not everyone sees it this way. For instance, the data-challenged wizards at FastCompany had me follow an inefficient and borderline excruciating “process” of writing several e-mails to several different companies. This included a series of questions collected in a free-response format. A web-based survey would have better for all concerned. Responding to individual e-mails doesn’t lend itself to easy and comprehensive analysis.

At the end of the e-mail chain, someone new sent me what appeared to be an arbitrary number on what I’d have to pay to use 87 words from an article of theirs. That price was in my view excessive and capricious, underscored by the nature the process.

Simon Says: Structure What You Can

You’ll get no argument from me on the importance of unstructured data. Still, structured data remains exceptionally valuable. There’s just no excuse for treating everything as unstructured.

The smartest folks and organizations out there are breaking from old habits and recognizing that one size doesn’t necessarily fit all. Creating structure where it makes sense saves a great deal of time and allows others to see things that others miss. As Henry David Thoreau once said, “It’s not what you look at that matters, it’s what you see.”

Feedback

What say you?

Tags:
Category: Information Development
No Comments »

by: Bsomich
07  Dec  2013

Weekly IM Update.

 
 logo.jpg

Business Drivers for Better Metadata Management

There are a number Business Drivers for Better Metadata Management that have caused metadata management to grow in importance over the past few years at most major organisations. These organisations are focused on more than just a data dictionary across their information – they are building comprehensive solutions for managing business and technical metadata.Our wiki article on the subject explores many factors contributing to the growth of metadata and guidance to better manage it:  

Feel free to check it out when you have a moment.

Sincerely,MIKE2.0 Community

Contribute to MIKE:

Start a new article, help with articles under construction or look for other ways to contribute.

Update your personal profile to advertise yourself to the community and interact with other members.

Useful Links: Home Page Login Content Model FAQs MIKE2.0 Governance

Join Us on
42.gif

Follow Us on 43 copy.jpg

Join Us on
images.jpg

Did You Know? All content on MIKE2.0 and any contributions you make are published under the Creative Commons license. This allows you free re-use of our content as long as you add a brief reference back to us.

 

This Week’s Blogs for Thought:

Twitter: Rubbish, Valuable, or Both?

We reach a certain age and the music gets too loud. We believe that the world is going to hell in a hand basket. Things were just better when we were young.

This is always the case, and today is no exception. The complaints can be deafening. Young people don’t read books anymore. Look at what young folks are wearing. We decry the state of society and the future.

Of course, we’ve seen this movie before; every generation does this. Many people my age and older dismiss Twitter and the very idea tweeting. Somehow, reading books and newspapers were more sophisticated than texting, tweeting, blogging, and friending.

Read more.

The Data of Damocles

While the era of Big Data invokes concerns about privacy and surveillance, we still tender our privacy as currency for Internet/mobile-based services as the geo-location tags, date-time stamps, and other information associated with our phone calls, text messages, emails, and social networking status updates become the bits and bytes of digital bread crumbs we scatter along our daily paths as our self-surveillance avails companies and governments with the data needed to track us, target us with personalized advertising, and terrorize us with the thought of always being watched.Read more.

Evernote’s Three Laws of Data Protection

“It’s all about bucks, kid. The rest is conversation.” –Michael Douglass as Gordon Gekko, Wall Street (1987) Sporting more than 60 million users, Evernote is one of the most popular productivity apps out there these days. You may in fact use the app to store audio notes, video, pics, websites, and perform a whole host of other tasks. Read more.

  Forward this message to a friend

Category: Information Development
No Comments »

by: Bsomich
16  Nov  2013

Weekly IM Update.

Missed what happened in the MIKE2.0 community this week? Here’s a great recap:
 
 logo.jpg

MIKE2.0 Wiki Topic Index

We’ve recently added a new topic index to make searching our wiki more user-friendly. This week we’re featuring the Overview and Key Concepts Section, which includes:

Concept Papers (2)
Implementation Guide (10)
Introduction to MIKE2.0 (20)
MIKE2 Activities (68)
Continous Improvement (6)
Strategic Mobilisation (1)
Information Development Phases (6)
Overview and Key Concepts (6)
Concept Papers (2)
Information Development Concepts (19)
Application Development (4)
Information Strategy, Architecture and Governance
Enterprise Information Assessment (3)
Enterprise Information Integration (1)
Enterprise Information Management Strategy (1)
IT Transformation (8)
Model Driven Architecture (1)
Networked Information Governance (3)
Services Oriented Architecture (14)
Regulatory Reporting (2)
Requirements Gathering (17)
Software Delivery Lifecycle (12)
Testing (9)
Infrastructure Development (16)
Introduction to MIKE2.0 (20)

We hope you’ll browse through when you have a moment! For more information on MIKE2.0 or how to get involved with the online MIKE2.0 community, please contact us.

Sincerely,

MIKE2.0 Community  

Contribute to MIKE:

Start a new article, help with articles under construction or look for other ways to contribute.

Update your personal profile to advertise yourself to the community and interact with other members.

Useful Links: Home Page Login Content Model FAQs MIKE2.0 Governance

Join Us on
42.gif

Follow Us on 43 copy.jpg

Join Us on
images.jpg

 

Did You Know? All content on MIKE2.0 and any contributions you make are published under the Creative Commons license. This allows you free re-use of our content as long as you add a brief reference back to us.

 

This Week’s Blogs for Thought:

Semantic Business Vocabularies and Rules

For many in the traditional applications development community, “semantics” sounds like a perfect candidate for a buzzword tossed at management in an effort to pry fresh funding for what may appear to be academic projects with not much discernible practical payback. Indeed when challenged for examples of “semantic applications” often one hears stumbling litanies about “Linked Open Data”, ubiquitous “URLs” and “statement triples”. Traditional database folks might then retort where’s the beef?

Read more.

We is Smaller Than the Sum of its Me Parts

In Why Does E=mc2? (And Why Should We Care?), Brian Cox and Jeff Forshaw explained “the energy released in chemical reactions has been the primary source of power for our civilization since prehistoric times.  The amount of energy that can be liberated for a given amount of coal, oil, or hydrogen is at the most fundamental level determined by the strength of the electromagnetic force, since it’s this force that determines the strength of the bonds between atoms and molecules that are broken and reformed in chemical reactions.  However, there’s another force of nature that offers the potential to deliver vastly more energy for a given amount of fuel, simply because it’s much stronger.”

Read more.

VanRoekel is Obama’s CIO

I recently had the opportunity to hear Steven VanRoekel, the second person who’s ever occupied a position I am glad to hear about: Federal Chief Information Officer of the United States. His talk was about federal information technology and initiatives he’s now implementing for the White House Office of Management and Budget. Prior to OMB, VanRoekel’s career ran from Microsoft (speechwriter for Bill Gates) through the Federal Communications Commission. During his tenure at the FCC, VanRoekel has been credited with modernizing aging IT infrastructures, using open-source based, cloud powered platforms for both web and VOIP applications; plus launching a Twitter account (with more than 400,000 followers!); plus the first federal “developer” website, crowd-sourcing data for projects like the National Broadband Map; plus the first to accept public comment via social media.

Read more.

Category: Information Development
No Comments »

by: John McClure
14  Nov  2013

VanRoekel is Obama’s CIO

I recently had the opportunity to hear Steven VanRoekel, the second person who’s ever occupied a position I am glad to hear about: Federal Chief Information Officer of the United States. His talk was about federal information technology and initiatives he’s now implementing for the White House Office of Management and Budget.

Prior to OMB, VanRoekel’s career ran from Microsoft (speechwriter for Bill Gates) through the Federal Communications Commission. During his tenure at the FCC, VanRoekel has been credited with modernizing aging IT infrastructures, using open-source based, cloud powered platforms for both web and VOIP applications; plus launching a Twitter account (with more than 400,000 followers!); plus the first federal “developer” website, crowd-sourcing data for projects like the National Broadband Map; plus the first to accept public comment via social media.

Whew! This fellow surely walks the talk!

He spoke for 30 minutes (you can see the video here) about topics that included President Obama’s open data policy, digital government strategy and the upcoming Federal Information Technology Acquisition Reform Act.

Here’s a few highlights.

VanRoekel has typical CIO responsibilities: planning & managing about US$ 80B civilian systems’ budget and its attendant workforce; while meeting responsibilities for governance of what seems an infinite trove of public and classified information, all increasingly exposed as cybersecurity targets. No small job at all, so his views must be experienced and mature.

Open data. VanRoekel voices a deep committment to nurturing both public open-source and private development of new information-based applications which benefit communities and individuals in their daily lives (of course, he’s referring to the semantic web). He’d anticipates many opportunities for data collection and data sharing, even by corporate entities. (However judging by the paucity of details about initiatives in this regard it’s not clear what priority this issue gets in this CIO’s time. This is a pity because surely EPA planners, and maybe common sense, indicate immediate benefits from local-level 2-1-1 semantic triple-stores.)

Private data. VanRoekel has apparently made recent Policy Directives onward applicable which require special reporting regarding protection of private individual and corporate data both during transmission and during storage. He cites that the government tracks web requests only at the regional level, so presumably he’s saying not by IP address.

Service quality. VanRoekel sounds a note about the future, suggesting even something sounding like government-sponsored Google Ads are on the way. He also touted a “Presidential Innovation Fellows” initiative as the latest avenue for coalescing ideas into working prototypes. In this regard he mentions Project MyUSA (My.USA.gov) as a site that “re-imagines the government interface”, offering APIs, personal toolbars, and offering highly useful doorways to governments treasure chest.

Cloud computing. VanRoekel claims to be a principal engineer of the federal governments computing infrastructure, and is an exceptional supporter of cloud computing economies in a federal budgeting context. In response to the impact of Snowden revelations he sees the current disruption in the reputations of American cloud service providers in the global market, is indicative of the need for nations to coordinate cross=boundary cloud operations and other regulatory matters (no doubt such as tarriffs).

A very worthwhile talk.

Category: Information Development
No Comments »

by: John McClure
11  Nov  2013

Semantic Business Vocabularies and Rules

For many in the traditional applications development community, “semantics” sounds like a perfect candidate for a buzzword tossed at management in an effort to pry fresh funding for what may appear to be academic projects with not much discernible practical payback. Indeed when challenged for examples of “semantic applications” often one hears stumbling litanies about “Linked Open Data”, ubiquitous “URLs” and “statement triples”. Traditional database folks might then retort where’s the beef? because URLs in web applications are certainly just as ubiquitous, are stored in database columns which are named just as “semantic properties” are; are in rows with foreign keys as construable as “subjects” in a standard semantic statement; and that’s still not to mention the many, many other SQL tools in which enterprises have heavily, heavily invested over many, many years! So…

It’s a good question – where IS the beef?

The Object Management Group (OMG), a highly respected standards organization with outsized impacts on many enterprises, has recently released a revision of its Semantics of Business Vocabulary and Rules (SBVR) that provides one clear answer to this important question.

Before we go there, let’s stipulate that for relatively straight-forward (but nonetheless closed world) applications, it’s probably not worth the time nor expense to ‘semanticize’ the application and its associated database. These are applications that have few database tables; that change at a low rate; that have fairly simple forms-based user interfaces, if any; and that are often connected through transaction queues with an enterprise’s complex of applications. For these, fine, exclude them from projects migrating an enterprise to a semantic-processing orientation.

Another genre of applications are those said ‘mission critical’. These applications are characterized by a large number of database tables and consequently a large number of columns and datatypes the application needs to juggle. These applications have moderate to high rates of change to accommodate functional requirements due to shifting (and new additions to) of dynamic enterprises — not so much mission creep as it is the normal response to the tempo of the competitive environments in which enterprises exist.

The fact is that the physical schema for a triples-based (or quad-based) semantic database does not change; the physical schema is static. Rather, it’s the ontology, the logical database schema, that changes to meet new requirements of the enterprise. This is an important outcome of a re-engineered applications development process: this eliminates often costly tasks associated with designing, configuring and deploying physical schema.

Traditionalists might view this shift as mere cleverness, one equally accomplished by tools which transform logical database designs into physical database schema. Personally I don’t have the background to debate the effectiveness of these tools. However, let’s take a larger view, one suggested by the OMG specification for Business Vocabularies and Rules.

Business Policies – where it begins, and will end

Classically business management establishes policies which are sent to an Information Technology department for incorporation to new and existing applications. It is then the job of systems analysts to stare at these goats and translate them into coding specifications for development and testing. Agile and other methodologies help speed this process internally to the IT department, however until the fundamental dynamic between management and IT changes, this cycle remains slow, costly and mistake-prone.

Now this is where OMG’s SBVR applies: it is an ontology for capturing rules such as “If A, then thou shalt not do X when Y or Z applies; otherwise thou shalt do B and C” into a machine-processable form (that is, into semantic statements). Initially suitably trained system analysts draft these statements as well as pertinent queries which are to be performed by applications at the proper moment. However at some point tools will appear that permit management themselves to draft and test the impact of new and changed policies against live databases.

This is real business process re-engineering at its brightest. Policy implementation and operational costs are affected as the same language (a somewhat ‘structured English’) is used to state what should and must be as that used to state what is. Without that common language, enterprises can only rely on the skills of systems analysts to adequately communicate business, and regulatory, requirements to others.

Capturing & weaving unstructured lexical information into enterprise applications, has never been possible with traditional databases. This is why ‘semantics’ is such a big deal.

Cheers!

Category: Enterprise Content Management, Information Development, Information Management, Information Strategy, Information Value, Master Data Management, Semantic Web
No Comments »

by: Bsomich
09  Nov  2013

Weekly IM Update.

Missed what happened in the MIKE2.0 community this week? Here’s a great recap:

 

 
 logo.jpg

How to Become an Information Centric Business

The digital age has created an abundance of information, but its sheer volume and complexity threatens to render it useless unless it is properly harnessed. Information-centric organisations can achieve that by linking up their business strategy, operating model and technology with an underlying information strategy and competency.
Information Development using MIKE2.0” is an effective approach to IM strategy development and transformation. MIKE2.0 was created by industry practitioners from hundreds of successful projects in information governance, business intelligence, data management and enterprise content management. We cover key concepts of information governance, delivery methodology, the SAFE reference architecture and example tools like IM QuickScan to assess an organisation’s information maturity. Several case studies are also presented which illustrate successful enterprise information management strategies.

Get Involved

We’ll be discussing our new book, “Information Development using MIKE2.0” at a number of industry events. It has also been published in paperback (available on Amazon.com and Barnes & Noble) as well as all major e-book publishing platforms if you would like a copy. For more information on MIKE2.0 or how to get involved with the online MIKE2.0 community, please contact us.

Sincerely,

MIKE2.0 Community  

Contribute to MIKE:

Start a new article, help with articles under construction or look for other ways to contribute.

Update your personal profile to advertise yourself to the community and interact with other members.

Useful Links: Home Page Login Content Model FAQs MIKE2.0 Governance

Join Us on
42.gif

Follow Us on 43 copy.jpg

Join Us on
images.jpg

 

Did You Know? All content on MIKE2.0 and any contributions you make are published under the Creative Commons license. This allows you free re-use of our content as long as you add a brief reference back to us.

 

This Week’s Blogs for Thought:

Wikidata Shows the Way

Studies consistently rank DBpedia as a crucial repository in the semantic web; its data is extracted from Wikipedia and then structured according to DBpedia’s own ontology. Available under Creative Commons and GNU licenses, the repository can be queried directly on the DBpedia site and it can be downloaded by the public for use within other semantic tool environments. This is a truly AMAZING resource! The English version of the DBpedia knowledge base for instance now has over two billion ‘triples’ to describe 4,000,000+ topics — 20% are persons, 16% places, 5% organizations including 95,000 companies and educational institutions, plus creative works, species, diseases, and so on — with equally impressive statistics concerning their knowledge bases in over one hundred other languages. And the DBpedia Ontology itself has over 3,000,000 classes properties and instances. What a breath-taking undertaking in the public sphere!

Read more.

Onions in India and the Speed of Data

Even in “advanced” societies, the process of collecting and verifying economic data has historically been pretty challenging. Whether its GDP growth, unemployment rates, housing prices, or inflation, there’s typically been a considerable lag between the actual trend and the presentation of that trend’s data. For instance, it’s helpful to know that inflation was 2.5 percent in the fourth quarter of 2012. But it’s less helpful when that information is made public in the second quarter of 2013.

What if that lag could be significantly reduced?

Read more.

Automation and the Danger of Lost Knowledge

In my previous post I pondered the quality of machine-generated data, cautioning that even though it overcomes some of the inherent errors of human-generated data, it’s not immune to data quality issues.  Nicholas Carr, in his recent article All Can Be Lost: The Risk of Putting Our Knowledge in the Hands of Machines, pondered the quality of our increasingly frequent decision to put so many things into the hands of our automatons.

Read more.

Forward this message to a friend

Category: Information Development
No Comments »

by: John McClure
07  Nov  2013

Wikidata shows the way

Studies consistently rank DBpedia as a crucial repository in the semantic web; its data is extracted from Wikipedia and then structured according to DBpedia’s own ontology. Available under Creative Commons and GNU licenses, the repository can be queried directly on the DBpedia site and it can be downloaded by the public for use within other semantic tool environments.

This is a truly AMAZING resource! The English version of the DBpedia knowledge base for instance now has over two billion ‘triples’ to describe 4,000,000+ topics — 20% are persons, 16% places, 5% organizations including 95,000 companies and educational institutions, plus creative works, species, diseases, and so on — with equally impressive statistics concerning their knowledge bases in over one hundred other languages. And the DBpedia Ontology itself has over 3,000,000 classes properties and instances. What a breath-taking undertaking in the public sphere!

Recently I had a wonderful opportunity to hear about DBpedia’s latest projects for their repository, here are the slides. DBpedia is now surely moving towards adoption of an important tool — Wikidata — in order to aggregate DBpedia’s 120 language-specific databases, into one single, multi-lingual repository.

Wikidata‘s own project requirements are interesting to the MIKE2 community as they parallel significant challenges common to most enterprises in areas of data provenance and data governance. Perhaps in response to various public criticisms about the contents of Wikipedia, Wikidata repositories support source citations for every “fact” the repository contains.

The Wikidata perspective is that it is a repository of “claims” as distinguished from “facts”. Say for example that an estimate of a country’s Gross National Product is recorded. This estimate is a claim will often change over time, and will often be confronted by counter-claims from different sources. What Wikidata does is to provide a data model that keeps track of all claimed values asserted about something, with the expectation this kind of detailed information will lead to mechanisms directly relevant to the level of “trust” that may be confidently associated with any particular ‘statement of fact’.

The importance of source citations is not restricted to the credibility of Wikipedia itself and its derivative repositories; rather this is a universal requirement common to all enterprises whether they be semantics-oriented or not. A simple proposition born of science — to distinguish one’s original creative and derived works from those learned from others — is now codified and freighted with intellectual property laws (copyrights, patents), subjects of complex international trade treaties.

Equally faced by most enterprises is a workforce located around the globe, each with varying strengths in the English language. By using a semantic repository deeply respectful of multilingual requirements — such as Wikidata is — enterprises can deploy ontologies and applications that improve worker productivity across-the-board, regardless of language.
Wikidata is a project funded by Wikipedia-Germany. Enterprises might consider helping to fund open-source projects of this nature, as these are most certainly investments whose value cannot be over-estimated.
Visit here for more Semantic MediaWiki conferences and slides. Ciao!

Category: Information Development, Information Governance, Information Management, Information Strategy, Information Value, Open Source, Semantic Web
No Comments »

by: Phil Simon
07  Nov  2013

Onions in India and the Speed of Data

Even in “advanced” societies, the process of collecting and verifying economic data has historically been pretty challenging. Whether it’s GDP growth, unemployment rates, housing prices, or inflation, there’s typically been a considerable lag between the actual trend and the presentation of that trend’s data. For instance, it’s helpful to know that inflation was 2.5 percent in the fourth quarter of 2012. But it’s less helpful when that information is made public in the second quarter of 2013.

What if that lag could be significantly reduced?

Better Technology Yields Better Data

And that’s the premise behind Premise Data. The company wants to change the way decision makers think about macroeconomic indicators by changing the way they consume that data. According to a recent GigaOM piece:

The big-picture problem, according to Premise co-founder David Soloff, is that the economic indicators many investors, business leaders, and decision makers rely on are just too outdated by the time they’re released. People “lurch to the number,” but the world has often moved since it was calculated and they won’t know it until the next report comes out, he explained.

“As these things speed up,” Soloff added, “the indicators bear less and less resemblance to what people experience on a day-to-day basis.”

And that can be a major problem. In fact, it has been for years. The difference today is that the technology finally exists to do something about it. And that’s where Premise comes in. While most macroeconomic indicators are derived from the top down, the Premise model is more bottom-up. Again, from the piece:

Contributors take a picture of the item either on the shelf or in a market stall; it syncs with Premise’s servers in the cloud; and Premise’s system is then able to extract information from the photos. It can determine information such as price, brand, and quality of the items, and even environmental information such as how clean the store is and how stocked the shelves are.

The article provides an example of real-time price collection of onions in India. Think about it. Equipped with a smartphone, it’s easy to collect and report this data instantaneously.

Simon Says

There’s a major lesson here for mature organizations stuck in the 20th century: More than ever, speed matters. Of course, plenty of enterprises understand the cardinal importance of velocity. For instance, hedge funds have recognized for a long time that even milliseconds matter. Still, far too many refuse to embrace new data collection methods.

Slowly but surely, many NGOs, governments, agencies, and other large public sector organizations are finally starting to change their data collection methods. Isn’t it time that you ask yourself it your organization is collecting data as quickly as possible? While speed doesn’t trump accuracy and quality, waiting months for critical enterprise data is no longer acceptable.

Feedback

What say you?

Tags:
Category: Information Development, Information Value
No Comments »

Calendar
Collapse Expand Close
TODAY: Thu, July 24, 2014
July2014
SMTWTFS
293012345
6789101112
13141516171819
20212223242526
272829303112
Recent Comments
Collapse Expand Close