Notes on LLMs as Workers in Human-Computational Algorithms? Replicating Crowdsourcing Pipelines with LLMs
This is a summary of an important research paper that provides a 17:1 time savings. It was crafted by humans working with several AI's. The goal is to save time and curate good ideas.

Link to paper: https://arxiv.org/abs/2307.10168
Paper published on: 2023-07-20
Paper's authors: Tongshuang Wu, Haiyi Zhu, Maya Albayrak, Alexis Axon, Amanda Bertsch, Wenxing Deng, Ziqi Ding, Bill Guo, Sireesh Gururaja, Tzu-Sheng Kuo, Jenny T. Liang, Ryan Liu, Ihita Mandal, Jeremiah Milbauer, Xiaolin Ni, Namrata Padmanabhan, Subhashini Ramkumar, Alexis Sudjianto, Jordan Taylor, Ying-Jui Tseng, Patricia Vaidos, Zhijin Wu, Wei Wu, Chenyang Yang
GPT3 API Cost: $0.03
GPT4 API Cost: $0.11
Total Cost To Write This: $0.14
Time Savings: 17:1
The ELI5 TLDR:
A recent research paper looked at how well Large Language Models (LLMs) can replicate tasks that are usually done by humans. The study found that LLMs can simulate some human abilities, but their success varies. LLMs respond differently to instructions compared to humans, with LLMs being more responsive to certain types of instructions. The study also found that replicating crowdsourcing pipelines with LLMs is possible, but there were challenges in translating the pipeline into LLM prompts. Different students had different success in replicating the pipelines, and there are opportunities for improvement in LLM instruction tuning and output quality. The study also suggests that LLMs can be useful in helping study designers, but their limitations need to be understood. Overall, LLMs have potential but their success depends on various factors and improvements can be made to optimize their use.
The Deeper Dive:
Understanding the Capabilities of Large Language Models in Replicating Crowdsourcing Pipelines
The research paper we're discussing today is an exploration into the capabilities of Large Language Models (LLMs) in replicating more complex crowdsourcing pipelines. The authors have focused on understanding how well LLMs can simulate human-like behavior in tasks that are typically crowd-sourced. The paper brings to light some interesting findings about how LLMs respond to instructions, how they compare to humans in these tasks, and the challenges and opportunities that arise when trying to replicate crowdsourcing pipelines with LLMs.
The Capabilities and Limitations of LLMs in Complex Tasks
The study finds that modern LLMs can simulate some of the abilities of crowdworkers in complex tasks, but the level of success is variable. This variability is influenced by several factors, including the requesters' understanding of LLM capabilities, the specific skills required for sub-tasks, and the optimal interaction modality.
Interestingly, the study finds that LLMs and humans respond differently to instructions. LLMs are more responsive to adjectives and comparison-based instructions. On the other hand, humans receive more scaffolds and interface-enforced interactions, which provide guardrails on output quality and structure that are not available to LLMs.
The study also highlights the need to improve LLM instruction tuning and consider non-textual instructions. It suggests that the effectiveness of replicated LLM chains depends on students' perceptions of LLM strengths.
Replicating Crowdsourcing Pipelines with LLMs
The study required students to replicate crowdsourcing pipelines by writing prompts for LLMs to complete different microtasks. Students implemented two solutions: a baseline solution and a replica of the crowdsourcing pipeline. The replication success was measured based on peer grading results and the effectiveness of the replicated chains.
The findings from the study suggest that all the pipelines were replicable with LLMs, and there was at least one correct replication and an effective one for each pipeline. However, prompting challenges were identified as a major factor for replication failure, with students finding it difficult to translate the pipeline into LLM prompts.
Variance in Replication and Opportunities for Improvement
The study observed a replication variance, with different students' replications of the same pipeline differing significantly. This variance was influenced by students' perceptions of LLM capabilities.
The authors identified several opportunities for improvement. These include developing frameworks to adjust prompt granularity, tuning LLM instructions, and exploring the optimal modality of instruction. They also identified output quality scaffolds and output structure scaffolds as areas for improvement in LLM chains.
Implementing Different Versions of the Find-Fix-Verify Pipeline
The students implemented different versions of the Find-Fix-Verify pipeline, with variations in the Find and Verify steps. Some students extended the Find step to include more types of writing issues, while others focused on fixing grammatical errors in the Verify step. This shows the flexibility and potential adaptability of LLMs in different tasks within a pipeline.
LLMs and Human-LLM Complementarity in Task Delegation
The research highlights the limitations of LLMs in understanding and following instructions, and their inability to take advantage of multimodal cues. Adapting existing techniques, such as using stricter templates or transforming generative tasks into multiple-choice tasks, can help align LLMs with human intuition.
The study also emphasizes the need for human-LLM complementarity in task delegation. The findings suggest that LLMs can be useful for helping study designers reflect on their high-level requirements, but the literal instruction may need to be redesigned. The research also discusses the educational value of allowing students to interact with LLMs to gain awareness of their limitations and prevent excessive reliance on them.
Concluding Thoughts
In conclusion, the study presents an in-depth exploration of LLMs' capabilities and limitations in replicating crowdsourcing pipelines. While LLMs show promise, their success is variable and influenced by several factors. The study provides valuable insights into how to optimize the use of LLMs in complex tasks and how to improve their performance by adjusting prompt granularity, tuning instructions, and exploring the optimal modality of instruction. These findings can be instrumental for businesses and product developers looking to leverage the power of LLMs in their operations.



