Imagine you have a very smart, fast-talking assistant who claims they can draw complex blueprints for a factory just by listening to you describe how the factory works. You say, "First, the raw materials come in, then they get washed, then they go to the oven," and the assistant instantly draws a diagram.
This is exactly what the researchers in this paper tried to build: an AI "Copilot" that turns your spoken or written descriptions into BPMN diagrams (which are just fancy, standardized blueprints for business processes).
The team built a tool called KICoPro and asked five real-life experts (people who draw these blueprints for a living) to try it out. They wanted to know: Is this thing actually useful, or is it just a cool toy?
Here is the story of what they found, explained simply:
1. The "Nice Face, Shaky Hands" Problem
The experts liked the look and feel of the tool. It was easy to chat with, the buttons worked, and it felt friendly. On a "usability" test, it scored a decent 67 out of 100.
The Analogy: Imagine a car with a beautiful leather interior, a smooth steering wheel, and a great radio. You love sitting in it. But when you try to drive it, the engine sputters, the brakes are unreliable, and the map is often wrong.
- The Result: The experts said, "I like driving this car, but I wouldn't trust it to get me to the hospital in an emergency." Their trust score was only 48 out of 100. They didn't believe the AI would get the job right every time.
2. The "Mind Reading" Struggle
The experts found a weird problem: They knew what they wanted (a blueprint), but they didn't know how to ask for it.
- The Analogy: It's like ordering a custom cake from a baker. You say, "I want a cake." The baker gives you a plain sponge. You say, "I want a chocolate cake with strawberries." The baker gives you a burnt chocolate cake with no strawberries. You realize you need to be a "cake whisperer" to get what you want, but the baker never asks, "Did you want vanilla or chocolate?"
- The Finding: The AI never asked clarifying questions. If the description was vague, the AI just guessed. If the process was long and complicated, the AI got confused and only drew half the picture.
3. The "Chunking" Hack
To get good results, the experts had to do extra mental work. They couldn't just describe the whole factory at once. They had to break their big idea into tiny, bite-sized pieces and ask the AI to draw them one by one.
- The Analogy: Instead of telling a painter, "Paint me the whole Grand Canyon," you have to say, "Paint the sky," then "Paint the rocks," then "Paint the river."
- The Problem: This made the experts tired. They had to hold the whole picture in their heads and stitch the pieces together themselves. The tool was supposed to save them work, but it actually made them work harder to "fix" the AI's mistakes.
4. The "Silent Partner"
The experts noticed the AI followed the rules of the road (BPMN standards) poorly. Sometimes it drew lines that shouldn't be there or missed important details like "who does this task?"
- The Analogy: It's like a GPS that sometimes tells you to drive into a lake because it didn't check the map carefully.
- The Fear: In a real business, if you trust the AI and it draws a wrong blueprint, you might build a factory that doesn't work. The experts said, "I can't trust this for important decisions yet."
5. What the Experts Dreamed Of
Even with the flaws, the experts had big ideas for how this tool could be amazing in the future:
- The "Sketch-to-Draft" Bot: Imagine drawing a messy picture on a napkin, and the AI turns it into a professional blueprint instantly.
- The "Quality Police": The AI could check your existing blueprints to make sure you didn't break any company rules.
- The "Local Brain": A version of the AI trained specifically on your company's history, so it knows your specific jargon and rules.
The Big Takeaway
The main lesson from this paper is: Just because a tool is easy to use doesn't mean it's trustworthy.
You can have a chatbot that feels great to talk to (high usability), but if it keeps making mistakes or guessing wrong (low reliability), professionals won't use it for serious work.
The Conclusion: We can't just test AI with computers (checking if the code is right). We have to test it with humans to see if they feel safe using it. The future of AI in business isn't just about making it smarter; it's about making it clearer, more honest about its mistakes, and better at asking questions before it starts drawing.
Get papers like this in your inbox
Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.