Authors:

Alessandro BertiMayssa MaatallahUrszula JessenMichal SrokaSonia Ayachi Ghannouchi

Paper:

https://arxiv.org/abs/2408.07720

Introduction

Process Mining (PM) is a data science discipline that extracts process-related insights from event data recorded by information systems. Techniques in PM include process discovery, conformance checking, and predictive analytics. Large Language Models (LLMs) have shown promise as PM assistants, capable of responding to inquiries and generating executable code. However, LLMs struggle with complex tasks requiring advanced reasoning. This paper proposes the AI-Based Agents Workflow (AgWf) paradigm to enhance PM on LLMs by decomposing complex tasks into simpler workflows and integrating deterministic tools with LLMs’ domain knowledge.

Related Work

Process Mining on LLMs

Previous works have explored textual abstractions of PM artifacts for LLMs to respond to inquiries and translate inquiries into executable SQL statements. Various PM tasks, such as semantic anomaly detection and root cause analysis, have been implemented on LLMs. Different implementation paradigms, including direct provision of insights, code generation, and hypotheses generation, have been proposed.

Connection to Traditional Process Mining Workflows

Scientific workflows, such as those based on RapidMiner, Knime, or SLURM, have been used in PM to increase reproducibility and standardize experiments. However, AgWf(s) aim to ensure the feasibility of the overall pipeline by adopting a divide-et-impera approach and using the right tool for each task. Unlike traditional PM workflows, AgWf(s) are non-deterministic by nature.

Methodology

AI-Based Agents Workflows

An AI-Based Agents Workflow (AgWf) is defined as a tuple containing deterministic tools (F) and AI-based tasks (T). The workflow involves selecting tools for tasks and executing tasks to produce the final response. Different tasks can be associated with different agents, each with specific skills. The workflow is static, but tasks and tool selections are non-deterministic.

AgWf Running Example

An example AgWf aims to exploit two different abstractions (directly-follows graph and process variants) to respond to a user’s inquiry. The individual results are merged by an ensemble, potentially leading to better results. The workflow includes clear start and end tasks, with interleaved tasks allowing different sequences of execution. The execution of each task appends the result to the input string, optimizing the inquiry and providing a comprehensive response.

Possible Implementations

Different implementations of AgWf(s) for the same tasks can vary in effectiveness. A single-task implementation may struggle with complex tasks, while a multi-task implementation can provide more comprehensive insights. Decomposing tasks into smaller units, each with a specific tool, leads to better results. For example, bias detection in process mining can be implemented with multiple tasks, each focusing on different aspects of the analysis.




Types of Tasks

Different types of tasks in AgWf workflows include:

  • Prompt Optimizers: Transform the original inquiry into a language tailored to AI agents.
  • Ensembles: Collate insights from different tasks into a coherent text.
  • Routers: Decide which dependent nodes should be executed based on the prompt.
  • Evaluators: Assess the quality of the output of previous tasks.
  • Output Improvers: Enhance the quality of the output of previous tasks.

Implementation Framework

The CrewAI Python framework supports the implementation of AgWf(s) on top of LLMs. Key concepts include defining AI-based agents, tasks, and tools, and executing tasks sequentially or concurrently. The framework supports entity memory and callbacks for advanced features. Examples of AI-based workflows implemented in CrewAI include fairness assessment and root cause analysis.

Next Steps

Automatic Definition of AgWf(s)

Future research should explore automatic decomposition of tasks into AgWf(s) by an orchestrating LLM. This involves understanding the original task and requesting clarifications.

Tasks Keeping the Human-in-the-Loop

Some tasks may benefit from user clarifications. For example, a prompt optimizer may need additional information to optimize a generic inquiry.

Evaluating AgWf(s)

Assessing the quality of outputs from AgWf(s) is challenging. The effectiveness of the overall workflow depends on the performance of individual agents. Collaboration and psychological traits of agents should be considered.

Maturity of the Tool Support

Current frameworks for implementing AgWf(s) need further development to reach high maturity. Solutions like LangGraph and AutoGen offer advanced features but require stability and user-friendly interfaces.

Conclusion

AgWf(s) offer a powerful tool for PM-on-LLMs by decomposing tasks into smaller units and combining AI-based tasks with deterministic tools. This approach aims to maximize the quality of outputs, differing from traditional scientific workflows focused on reproducibility. Future research should focus on automatic workflow definition, human-in-the-loop tasks, evaluation frameworks, and tool support maturity.

Share.

Comments are closed.

Exit mobile version