RoboCon 2026

What If Robot Framework Have a Brain
2026-02-13 , RoboCon

This talk explores AI-driven automation in Robot Framework through an intelligent Agent that enhances testing with capabilities like Agent.Do and Agent.Check. By leveraging large language models and visual understanding, the Agent interprets test intentions, interacts with GUI elements, and performs visual assertions. It also explores how this can lead toward more autonomous test execution, where the Agent can understand and carry out complete testing goals through another keyword dedicated to this purpose.


This session explores the integration of artificial intelligence into test automation through a novel AI Agent built on top of Robot Framework.
The agent introduces intent-level automation, allowing testers to describe what to test instead of how to test it.

With new intent-based keywords such as Agent.Do and Agent.Check, the framework interprets high-level testing goals, transforming them into concrete test actions and assertions in real time.

At its core, the agent combines the reasoning capabilities of Large Language Models (LLMs) with visual understanding models.
It can interpret a tester’s intent, identify and interact with GUI elements, and verify expected outcomes visually without relying on locator-based definitions.
This enables a more resilient, self-adaptive testing approach suitable for rapidly evolving user interfaces.

  • LLM Client Layer:
    A modular interface supporting multiple LLMs to interpret and execute test intents while staying fully compatible with Robot Framework logs.

  • VLM (Visual Language Model):
    Merges a vision model with an OCR to extract visual context, semantics, and element coordinates from screenshots.

The roadmap explores advancing the Agent toward higher levels of autonomy and adaptive decision-making in test execution.

This session includes a live demonstration of the current prototype, focusing on:

  • Reproducibility of the approach
  • Scalability across diverse applications and test environments
  • Quantitative and qualitative improvements in test robustness and maintenance

The session will show concrete results and a live demo of the current prototypes, emphasizing reproducibility, measurable gains, and practical AI-in-testing outcomes.


Lessons Learned:

-How to use this verification inside Robot Framework, with simple, runnable examples.
-In which cases visual/intent checks are most useful ( complicated visual asssertion, time consuming assertion, dynamic app )
-Which LLMs are better for which tasks.
-How to craft good prompts for visual assertions and how to test AI assertions themselves (mock bugs to see whether the LLM detects them, provide references, set confidence thresholds, etc.).
-When delegating actions to the agent: clicking “inaccessible” elements (covered/disabled/hidden), targeting items whose text/IDs change (timestamps, randomized lists), and handling dynamic data.

Categorize / Tags:

AI Testing, Visual Testing, Test Automation

Is this suitable for ..?: Beginner RF User, Intermediate RF User, Advanced RF User Describe your intended audience:

For both web and mobile testing, with a demonstration on mobile, where its unique challenges give AI more room to add value.

In-Person or Online talk/workshop?:

In person

Active Robot Framework community member, Python developer, and QA engineer, currently focused on mobile apps and applying my AI background to QA

Pavlo Ivashchenko

Pavlo Ivashchenko is a Senior Java Software Engineer at Danske Bank in Vilnius, Lithuania. He holds a Master’s degree in Software Engineering from Vilnius University and contributes to open-source projects including Eclipse JKube and assistant-ui. Pavlo is passionate about automation, developer tooling, and test automation frameworks.