How Should Group Brainstorming Work on the Web?

“I mean ‘storm’ as in storm the Bastilles” — Alex F. Osborn (creator of brainstorming)

The Internet is full of collaborative tools and services, but a popular group brainstorming platform has yet to materialize.

Admittedly, the web in general has replicated many of the features of classic real-world group brainstorming. While none of the popular social media platforms were designed to conduct formal brainstorming sessions, their ease of access and familiarity has made them a popular choice for conducting ad hoc collaboration. Comment fields and voting buttons (points, “Likes”, “Hearts”, up-and-down arrows, etc.) perform many of the functions attributed to brainstorming, if only on a rudimentary level. For anyone who has experienced a formal group brainstorming session in real life, the tradeoffs with using this makeshift method leaves much to be desired.

Lesser known are the group brainstorming applications currently available for private use. None are open online for the general public; access is based on a contracted legal agreement with per-user license fees. The target audience for these offerings is large organizations with complex internal decision-making needs. In cases where the application is hosted as part of a consulting engagement, guidance by a trained facilitator is a central requirement of the service. A compelling argument for why these brainstorming solutions haven’t gained wide acceptance or use is that the licensing and usage model severely limits the potential for ubiquitous ad hoc collaboration.

The problem, therefore, is that no one has succeeded in taking the next step of reimagining group brainstorming to take full advantage of the latest features and functions of the web. Here are but a few potential attributes for such a methodology…

Eliminate the “Tragedy of the Comments”.

The standard methodology for group collaboration today is the use of comment strings. As useful as this is for general interaction on the web, it’s a terrible method for facilitating large-scale collaborative dialogue on the web. In the early days of the Internet, it served as the online representation of one-on-one conversation. Today, comment strings are used even when the number of participants makes coherent dialogue impossible. A better solution would use the natural tendency of fragmentation within group dialogue as a core element of collaboration. Instead of producing one monolithic string of comments, group brainstorming should employ compulsory tangential discussions during a dialogue to maximize the coverage of the conceptual terrain for the topic at hand, regardless of the subject matter or number of users.

Eliminate Prior Reputation of the User from the Evaluation Process.

Even though users formally register to participate, all of the interaction among participants during a project is anonymous. Biases such as user reputation or size of a participant’s social network are simply not allowed to skew the evaluation process. Most of the valuations attributed to content submitted by users of other collaborative systems rely heavily on social networking. The problem with that reliance is what social theorists call reputational or information “cascades,” more commonly referred to as groupthink.

Integrate “Coopetition” as a function of the process itself.

Use gaming theory to loop the motivations of cooperation and competition into each other. For example, giving a glowing review to the proposal that receives the highest score raises the final ranking of the evaluating participant in a project. Even the decision to punish or even purge abusive users is determined by those participants who will benefit the most by passing sound judgment.

Explicit Sampled Comparative Evaluation: No central judges, no one has to read every comment, and all comments are reviewed.

Allowing users to evaluate any piece of submitted content they want — independent of any requirement of comparing it with other submitted content is the standard mode of most systems today. But this creates an inherent fallacy. Since users are not formally comparing one piece of content versus another, what do their scores really mean?

If the collectively perceived value of submitted content is purely relative to the value of other pieces of content, then the real value of content that isn’t based on comparative judgement is unclear at best. At worst, it doesn’t mean anything.

Should a user be able to review only one piece of submitted content (e.g. blindly giving a score based on a request from a related acquaintance) or should they be required to compare it to at least one other piece of content before making an evaluation? The content displayed on systems such as these that have the highest value could very likely be nothing more than accidental (manipulated?) groupthink. Like a feedback loop between a microphone and a nearby speaker, any content could become “the most valued” for no other reason than the momentum of its popularity at the beginning of a collaboration.

The best current example of explicit comparative voting (made famous by the notorious web site “Hot or Not”) is the Pairwise comparison methodology. Here’s how it works: Users are presented with two pieces of content (typically a line of text or a picture) and are asked to decide which one they like the best. It has an impressive amount of analytical grounding. Of the known Pairwise-based systems out there, none have a facility for submitting commentary for consideration by content originators and/or subsequent evaluators. Since there’s almost no feedback loop, it creates a “rank reversal” problem.

Evolve comments through periodic upgrades, not by adding yet another line to the string.

By making the upgrading of comments a feature, user content is never dated. For historical reference, versioning could be employed to maintain the logical continuity for reviewers in the future. This would make different to other online brainstorming systems as’s e-commerce platform is to eBay’s auctioning system.

And here’s where the functionality of the blockchain stretches the possibilities even further…

Replace scores with cryptoassets.

Imagine ad hoc brainstorming for money for any subject matter, any amount of fees and with an unlimited number of participants. And the money/value used in the session is held in escrow and distributed through smart contracts.

Make Group Brainstorming as a Community-operated social utility, not a centralized commercial business.

A brainstorming system could be built as a DAO on Ethereum (and/or Rootstock when available). Participants would be free to co-create outside of the confines of someone else’s business environment.




Remember—all I'm offering you is my insight. Nothing more.

Love podcasts or audiobooks? Learn on the go with our new app.

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Michael Christopher Johnson

Michael Christopher Johnson

Remember—all I'm offering you is my insight. Nothing more.

More from Medium

How to create offers to sell or buy altcoins on DeFi OTC Marsbase

A Very Useful Article About Online Passwords

Introducing Credit Data Alliance (CreDA)

The Human Touch — How Video Greetings Can Transform A Website & Generate Better Leads