Docs | NatSec.io
  • About
    • Project Overview
    • πŸ“–Story
    • 🎯Goals
      • πŸ‘₯People
      • πŸŽ“Education
      • πŸ”Iteration
    • 🧩Elements
      • πŸ’»Data
        • Entities
        • Categories
        • Tags
        • Connections
        • Verification
      • πŸ“ˆGraph
      • πŸ—ΊοΈMap
      • ⏳Timeline
    • πŸ’‘Use Cases
      • πŸ›οΈGovernment
      • ❀️Non-Profit
      • 🏫Academic
      • πŸ’°Commercial
    • πŸ’ͺSupport
      • βœ…User Agreement
      • πŸ₯ΈPrivacy
      • πŸ›‘οΈSecurity
      • πŸ™‹Contributors
      • 🀝Partners
      • πŸ“License
      • βž•Additional Licenses
      • βš–οΈEthics
    • πŸ“£Feedback
    • πŸ“±Contact Us
  • Frequently Asked Questions
    • FAQ Overview
    • πŸ’»Data FAQ
    • πŸ“ˆGraph FAQ
    • πŸ—ΊοΈMap FAQ
    • ⏳Timeline FAQ
    • πŸ“‘Documentation FAQ
    • How can I support or contribute to this project?
    • How is this project sustained?
    • What criteria do you use to collect project data?
    • How do I report an error or missing data?
    • How do I access raw project data?
    • What is the difference between the graph and the map?
    • How can I update project documentation?
  • Release Notes
    • Release Notes Overview
    • ⌚Feature Timeline
    • 🐣Beta
    • πŸ₯šAlpha
      • v0.1.0
  • Workflow
    • Workflow Overview
    • πŸ§‘β€πŸ’»Contribution
      • πŸ’»Adding or Updating Data
      • πŸ“ˆModifying the Graph
      • ⏳Modifying the Timeline
      • πŸ—ΊοΈModifying the Map
      • πŸ“’Updating Documentation
    • 🧐Evaluation
      • 🧹Cleaning Data
      • 🚫Prohibited Content
      • πŸ€”Adjudicating Changes
    • 🧠Review
      • πŸ‘‚User Feedback
      • ➰Process Improvement
      • ✨Outcomes
  • Learning
    • Learning Overview
    • πŸ› οΈTools
      • GitBook
      • Google Forms
      • Google Sheets
      • Kumu
      • Slack
    • πŸ’‘Concepts
      • Data Structures
      • Data Visualization
      • Dimensionality
    • πŸ§‘β€πŸ«Courses
      • Udemy / DigitalU
      • FreeCodeCamp
    • πŸ—ƒοΈResearch
      • Existing Research
      • Questions for Researchers
Powered by GitBook
On this page
  • Speed vs. Trust
  • More to Follow

Was this helpful?

  1. About
  2. Elements
  3. Data

Verification

How we manage trust in the Data.

PreviousConnectionsNextGraph

Last updated 2 years ago

Was this helpful?

More details on Verification will be available soon. An overview is provided below. We welcome suggestions, examples, and inputs!

Speed vs. Trust

One of the inherent challenges of crowd-sourced data is veracity (aka truthfulness). To address this question, we have to balance two factors: speed and trust.

Speed is a hallmark feature of crowd-sourcing, with many different inputs and updates being possible across a large number of people in a short period of time. Unfortunately, the resulting data may be inaccurate, resulting in the second factor.

Trust is required for users to apply data they come across to their daily lives, which is an objective of this project on numerous levels. With crowd-sourcing not providing an inherent trust in data veracity, we need some method of creating it, at least in some situations.

Human-Mediated Verification

One of the primary means we expect to use for Verification, especially at first or with new Data sets, will be human-in-the-loop mediation.

While specific details and instructions will depend on the tool(s) employed (and be documented in Workflow), the general idea as that at least one (though ideally at least two) individuals will evaluate an input or update for whether it is satisfactory (again, using criteria listed elsewhere). If so, it will be marked as Verified as of that time and be updated later if there are additional changes (meaning that either some Entities or Connections will have their Verification status change over time or updates will not be shown until verified (meaning a lag-time between submission and when an update reflects in the broader Data, though a public holding tank for unverified updates could mediate this in some cases).

The downside of this approach is that it will be somewhat resource-intensive and prone to human failure. That's where the next option comes in.

Automated Verification

Where possible, we would love to be able to automate verification. Then again, easier said than done.

Our current hypothesis is that there are probably certain data sources (e.g. those ending in .gov) that are more likely to be inherently trustworthy and could substitute for human intervention in certain cases.

If you have ideas or examples, !

Unverified Data

While there are certainly cases where Verification will matter, there will be many cases where just exploring may be sufficient. This is similar to how social media works, with all the incumbent risks on the part of those who use the Data (but with some of the speed that can make it still useful).

More to Follow

This is an important topic and we're just starting to scratch the surface.

🧩
πŸ’»
we're all ears