Hi @Czar -- thank you for these thoughtful comments. It's really great to hear from a Commons admin. Here are my replies and follow-up questions:
In case you haven't already, have you tried asking WP editor communities...
Yes, this was something we talked about in the original discussion about structured tasks, particularly in this thread. A type of task that several communities mentioned as being particular desirable would be some sort of copyediting/spellchecking/grammar task. We talked a lot about how one of the challenges is to find spellchecking or grammar-checking engines in all languages. This is an idea that I'm still learning more about and figuring out what's possible.
In choosing which structured tasks to build, we're trying to take several things into account, including (a) whether the task is useful, (b) whether newcomers can do it successfully, and (c) whether we have a technical approach to building it. The technical approach part is what helped us decide to prioritize "add a link" and "add an image". For both of those, the WMF Research team already had ideas for algorithms that could get us there. But for a task like copyediting, the path is not as clear.
Relatedly, I would be curious if developing wikis would rather have newcomers add images or instead...
Thanks for bringing this up. We've heard about ideas and concerns around Wikidata infoboxes from @Mike Peel and others. It would be super helpful if you could check out the questions I posted for Mike, and add any of your own thoughts.
"50% of articles have no images"...
Good point -- we already saw in the user tests we've done so far that users are unclear on what makes an image appropriate for an article. They're not sure whether they are strictly judging on relevance, or also on quality, or on other things. We would need the feature to include a little bit of onboarding to give them guidelines around how to decide.
"Will newcomers think this task is interesting?"...
In terms of whether users will want to go through a feed of articles missing images, we think there are some users who will really like it. In the fifteen user tests we've run so far (we'll post the summarized results in the next couple weeks), several users have said things like, "This is fun" or "I could do this all day." And in building this task, we would also allow users to narrow the articles to their topics of interest, which we do for existing newcomer tasks, and we think that would make the work more engaging.
But the idea you're talking about is different. We call it the "entry point in reading experience": if someone is just reading Wikipedia articles, how about something tells them something like, "This article has 4 suggested edits you can make." Then they could discover the link suggestions or image suggestions, or any other kinds of tasks in the future. And presumably, they would be interested in the content, because they navigated to that article on their own interest. Yes, we think that would be a great thing to build, and we definitely want to in the future. This workflow actually exists in the Wikipedia Android app for the "article description" and "image caption" tasks. But we're starting out with the "feed" approach for two reasons: (a) this gives us more control over how many tasks we make available and which users can do them, and (b) it is technically less ambitious to build the feed because we only need to generate and store suggestions on a sufficient number of articles to populate the feed, rather than on all articles to prepare them to be discovered.
I tried your prototype and felt like I couldn't confidently associate...
Regarding the format, I definitely agree that the context of the article is really important to have. Although our prototype shows one image match at a time on one screen with a small amount of information, it may make more sense to adopt a design in which the user actually navigates to the full article in order to complete the image match. Our team calls this a "Concept A" design, as opposed to the "Concept B" design that you saw in the prototype. We're actually building "Concept A" now for "add a link" (in which the user goes to the full article to complete the task. To read more about Concept A vs. Concept B, see this section.
Could you tell me more what you mean when you say, "I don't really trust Commons metadata for these associations"? What goes wrong?
If I were doing this as a new editor, I'd only want high-quality matches.
I think this could be a good idea. Perhaps we could show only the highest confidence matches for a user's first tasks, so that they can quickly grasp the objective of the task and be productive. Then they could advance to lower confidence matches as they gain skills.
For judging quality, if you're able to detect whether an image has artefacting/pixelation...
The WMF researcher who works on the image algorithm does plan to add some computer vision elements in the future to filter out low quality images, or perhaps NSFW images.
I wouldn't trust new users to write captions...
I agree that writing captions might be the most challenging part of the task for newcomers. But I think, at least on English Wikipedia, that images in articles should almost always have captions, unless they are an obvious image in an infobox. Is that right? If so, I'm trying to think of how else the images could be appropriately captioned, if not by the newcomers who make the match.
I went to follow along with the image trials...
I don't think I can share the spreadsheets publicly, but the prototype you were playing with actually does contain 60 actual matches from the algorithm. I hope that can give you a sense of the sorts of matches that it makes. What do you think of them?