publications
publications by categories in reversed chronological order. generated by jekyll-scholar.
2026
- EdgeWisePersona: A Dataset for On-Device User Profiling from Natural Language InteractionsPatryk Bartkowiak and Michal PodstawskiIn UMAP 2026, Jun 2026
This paper introduces a novel dataset and evaluation benchmark designed to assess and improve small language models deployable on edge devices, with a focus on user profiling from multi-session natural language interactions in smart home environments. At the core of the dataset are structured user profiles, each defined by a set of routines - context-triggered, repeatable patterns of behavior that govern how users interact with their home systems. Using these profiles as input, a large language model (LLM) generates corresponding interaction sessions that simulate realistic, diverse, and context-aware dialogues between users and their devices. The primary task supported by this dataset is profile reconstruction: inferring user routines and preferences solely from interaction history. To assess how well current models can perform this task under realistic conditions, we benchmarked several state-of-the-art compact language models and compared their performance against large foundation models. Our results show that while small models demonstrate some capability in reconstructing profiles, they still fall significantly short of large models in accurately capturing user behavior. This performance gap poses a major challenge - particularly because on-device processing offers critical advantages, such as preserving user privacy, minimizing latency, and enabling personalized experiences without reliance on the cloud. By providing a realistic, structured testbed for developing and evaluating behavioral modeling under these constraints, our dataset represents a key step toward enabling intelligent, privacy-respecting AI systems that learn and adapt directly on user-owned devices.
@inproceedings{bartkowiak2026edgewisepersona, title = {EdgeWisePersona: A Dataset for On-Device User Profiling from Natural Language Interactions}, author = {Bartkowiak, Patryk and Podstawski, Michal}, booktitle = {UMAP 2026}, month = jun, year = {2026}, archiveprefix = {arXiv}, primaryclass = {cs.HC}, url = {https://arxiv.org/abs/2505.11417}, } - Exploring the Cognitive Capabilities of Large Language Models in Autonomous and Swarm Navigation SystemsDawid Ewald, Filip Rogowski, Maciej Susniak, and 2 more authorsElectronics, Jan 2026
The rapid evolution of autonomous vehicles necessitates increasingly sophisticated cognitive capabilities to handle complex, unstructured environments. This study explores the cognitive potential of Large Language Models (LLMs) in autonomous navigation and swarm control systems, addressing the limitations of traditional rule-based approaches. The research investigates whether multimodal LLMs, specifically a customized version of LLaVA 7B (Large Language and Vision Assistant), can serve as a central decision-making unit for autonomous vehicles equipped with cameras and distance sensors. The developed prototype integrates a Raspberry Pi module for data acquisition and motor control with a main computational unit running the LLM via the Ollama platform. Communication between modules combines REST API for sensory data transfer and TCP sockets for real-time command exchange. Without fine-tuning, the system relies on advanced prompt engineering and context management to ensure consistent reasoning and structured JSON-based control outputs. Experimental results demonstrate that the model can interpret real-time visual and distance data to generate reliable driving commands and descriptive situational reasoning. These findings suggest that LLMs possess emerging cognitive abilities applicable to real-world robotic navigation and lay the groundwork for future swarm systems capable of cooperative exploration and decision-making in dynamic environments. These insights are particularly valuable for researchers in swarm robotics and developers of edge-AI systems seeking efficient, multimodal navigation solutions.
@article{ewald2026exploring, title = {Exploring the Cognitive Capabilities of Large Language Models in Autonomous and Swarm Navigation Systems}, author = {Ewald, Dawid and Rogowski, Filip and Susniak, Maciej and Bartkowiak, Patryk and Blumensztajn, Piotr}, journal = {Electronics}, volume = {15}, number = {1}, pages = {35}, month = jan, year = {2026}, publisher = {MDPI}, doi = {10.3390/electronics15010035}, url = {https://doi.org/10.3390/electronics15010035}, }
2025
- Seamlessly Integrating Tree-Based Positional Embeddings into Transformer Models for Source Code RepresentationPatryk Bartkowiak and Filip GralinskiIn XLLM at ACL 2025, Aug 2025
Transformer-based models have demonstrated significant success in various source code representation tasks. Nonetheless, traditional positional embeddings employed by these models inadequately capture the hierarchical structure intrinsic to source code, typically represented as Abstract Syntax Trees (ASTs). To address this, we propose a novel tree-based positional embedding approach that explicitly encodes hierarchical relationships derived from ASTs, including node depth and sibling indices. These hierarchical embeddings are integrated into the transformer architecture, specifically enhancing the CodeBERTa model. We thoroughly evaluate our proposed model through masked language modeling (MLM) pretraining and clone detection fine-tuning tasks. Experimental results indicate that our Tree-Enhanced CodeBERTa consistently surpasses the baseline model in terms of loss, accuracy, F1 score, precision, and recall, emphasizing the importance of incorporating explicit structural information into transformer-based representations of source code.
@inproceedings{bartkowiak-gralinski-2025-seamlessly, title = {Seamlessly Integrating Tree-Based Positional Embeddings into Transformer Models for Source Code Representation}, author = {Bartkowiak, Patryk and Gralinski, Filip}, booktitle = {XLLM at ACL 2025}, month = aug, year = {2025}, address = {Vienna, Austria}, publisher = {Association for Computational Linguistics}, url = {https://aclanthology.org/2025.xllm-1.10/}, doi = {10.18653/v1/2025.xllm-1.10}, pages = {91--98}, isbn = {979-8-89176-286-2}, }