I’m planning to submit this proposal for Squidsway, so posting here to get feedback:
Squidsway overview:
The ‘product’ is:
- an ongoing research project, with the reports and insights that will generate, and
- the open-source tools which will be built to generate those insights.
I am looking to:
index and curate onchain data, combined with offchain data, in order to generate insights which support healthier democracy within OpenGov - see ‘Aims’.
(Onchain data, eg: voting data and timings; delegations; wallet age/ size/ activity
Curated onchain data, eg: tagging proposals and voting blocs, such as ‘treasury conservative’; ‘resubmitted proposal’; ‘DV cohort 5 delegate’
Offchain data, eg: mostly proposal discussion, but also tweet volume, token price, sentiment analysis)
Aims:
Reduce Information Asymmetry:
-
To reduce the information asymmetry between unconnected voters / proposers and those who are well funded / connected.
-
Reduce the time cost of making informed votes
-
Reduce the time cost of making successful proposals.
-
Improve the quality of proposals and reduce the number of unsuccessful proposals rejected for predictable reasons.
-
Reduce mistrust in the ecosystem by providing objective data in place of rumour and supposition.
Increase voting:
-
both by reducing the time cost involved in meaningfully voting
-
and identifying patterns, especially in non-voting, to generate evidence-based insights to inform other initiatives to increase voting (eg via UX, incentives, publicity)
Individual voters (and non-voters) get:
-
Easily organised information on the voting and outlook of each DAO, allowing them to assess where and whether to delegate.
-
Organised data to compare some current proposal against comparable historical ones.
DAOs get:
-
Increased visibility to potential delegating voters / members
-
Objective data to counter FUD
-
Less proposals rejected for reason of proponents failure to understand the guidelines that each DAO may generally recommend
-
and less time spent in OpenGov discussions explaining these to proponents
Proposers get:
- Concrete data on what kind of proposals pass or fail, based on amount, time, type or other qualitative measure but, also (and in particular), regarding technical guidelines (eg, “since 2024, X% of proposals requesting DOT failed, but Y% requesting USDC succeeded”)
OpenGov gets:
-
Concrete data on non-voting versus voting wallets, in order to better identify pain points in voting, and give us the option to target non-voters (with incentives, awareness, etc.) to encourage voting
-
Concrete data on Decentralised Voices, allowing W3F to improve the program without relying solely on subjective and incomplete means of assessing improvements.
-
More delegation by individual voters.
-
The success or failure of proposals being better determined by community sentiment, and less by proponents’ failure to adhere to technicalities.
Background:
I’ve been a developer in the Polkadot ecosystem since soon after launch. Before becoming a developer, though, I had over a decade learning inside groups and organisations continually hitting pitfalls trying to follow the aims of decentralisation. (The promise of being able to use game theory, and to encode logic, to mitigate such pitfalls is a major reason why I moved into web3 - or just ‘crypto’, as it was called then ;). In particular, I’ve seen how, even when all actors involved do approach a collective with goodwill, information asymmetry creates power silos which self-reinforce and are hard to shift (even after it becomes clear to those in power silos, as it inevitably does, that the rot creates worse outcomes for them too).
So my motivation in this project is to add another (non-treasury) dimension to the sterling work done by folks like OpenGov Watch and Anaelle collating and publishing treasury information, and help steer OpenGov, the Polkadot DAO, towards being a healthy informed democracy.
The negative patterns we’ve seen develop in political life worldwide over the last 10 years or so are neither accidental nor some deliberate plan. They are the results playing out of formal and informal structures unfit for the internet age. This is the case for informational structures even more so than organisational and human ones. Though in some cases information is hoarded by elites or misinformation deliberately spread at scale, more often, needed information of sufficient detail is simply not easily available in an accessible enough form, leading voters to fail to identify bullshit, to feel unable to make worthwhile voting decisions, or even to feel that crucial information is hidden from them- leading to toxicity. All problems which face many voters in OpenGov.
My aim in this project is to make Open Gov fluid and easier to engage with. Even if that’s not a convincing enough reason for you, though, dear reader, I hope you’ll be convinced by the benefits to the different Open Gov participants above
Methodology:
Very very agile.
The aim of the research is to tell us something we didn’t know, rather than setting out to prove or disprove a set of hypotheses.
This means that the treasury will, at each stage/ in each proposal, be funding something that it does not know what it will be.
This agile way of working is necessary because:
-
We need to go where the evidence takes us
-
It’s likely that many of each of the small technical steps that make up a milestone can only be identified once a previous step is complete, so identifying and costing out of these each small technical steps in advance would either lead to wasted labour or lead the research down an inflexible path.
The fact that the treasury is funding something unknown should be mitigated by the ongoing nature of the project, and the fact that each funding milestone is a small amount.
Deliverables:
Each sprint will:
generate data,
which I will report and
will contribute to growing the Squidsway tool,
which will be open-source.
MVP phase:
I envision the first sprints to produce only some superficial statistics.
These first couple or so sprints will be to create an MVP to demonstrate progress, with more valuable data and insights coming later on. The reason for not running more sprints within each milestone (ie larger, longer milestones, thus having more by the first milestone) is due to the current conservative and low-trust environment in OpenGov and this being my first funding proposal.
In the first sprints I will deploy the basic indexing infrastructure (likely SQD), collate parent elements (referenda, delegations, etc.) and add components to tag (eg ‘requested_in_DOT’, ‘beneficiary_is_multisig’) chain data.
Validation phase:
The real value of the tool lies in capturing offchain data (when combined with the chain indexing).
A simple example of offchain data is the DOT price.
Slightly more complex would be extracting time and quantitive data on referenda from Polkassembly/ Subsquare.
More complex than that would be to run an LLM over sources like Polkassembly to collect qualitative data, in particular to be able to classify referenda by subject (eg ‘marketing’, ‘ambassador program’, ‘software development’) and other qualities (eg ‘resubmitted ref’, ‘vote nay’, ‘drama’, ‘much detail’).
The research output of the early part of the validation phase is likely to be concrete but boring statistics on uncontentious facts - labour-saving if that’s the info you were looking for, but not especially insightful in itself.
The research output of the later part of the validation phase is likely to be insights based on these concrete stats, which is where I hope the tool will start to demonstrate its future value. My aim is that by the end of this phase we will see both novel insights, and insights which would have been more labour-intensive by other research means.
One of the sprints within this phase will also be to create a module for efficiently indexing time based data (eg DOT price, treasury size).
Future work:
As the tool gets more powerful and hopefully gains acceptance as a potentially valuable part of Polkadot governance, I would like to:
-
Continue adding offchain sources.
-
Routinely produce data viz from insights.
-
Take in suggestions for research directions.
-
Add documentation and helper functions to make it easier for devs to run bespoke governance indexing.
-
Create a UI for LLM-written queries so that non-devs can easily query the data.
Proposal Spacing:
I envision a sprint being between 20 and 80 hours work, and seeking funding (via a new proposal) roughly every two months for between 40 and 160 hours, continuing as long as the community finds it valuable.
Feedback, plz:
This forum post will be added to as I (hopefully) receive feedback.
I hope that some of the folks that read through every proposal can take a moment out to give me some feedback here - I would like to proposal to be close to final form before I submit it as a referendum.