✨ Coming Soon: Adversarial RL for LLM, Specultive Decoding, Triton Kernels, and more ...

Aug 1, 2025·
Alessandro Palmas
Alessandro Palmas
· 2 min read

🚧 What’s Next: Current Directions and Upcoming Projects

This page offers a concise roadmap of the key research areas and projects I’m actively working on or preparing to launch. These efforts range from foundational LLM optimization to new frontiers in embodied and agentic AI.


⚔️ Adversarial Reinforcement Learning for LLM Alignment

Exploring how adversarial reinforcement learning can enhance the robustness of LLM alignment. The goal is to move beyond static fine-tuning by simulating challenging interactions that force the model to generalize alignment behavior.


⚡ Speculative Decoding for Faster Inference

Implementing speculative decoding techniques to accelerate LLM inference while maintaining output fidelity. This work supports real-time applications and makes large models more accessible in latency-sensitive environments.


🔧 Experimenting with Triton for LLM Speedup

Using Triton to design custom GPU kernels tailored for LLM workloads. The focus is on optimizing memory access patterns and improving throughput during both inference and training.


🧠 Testing Agentic Frameworks for Multi-Step Coordination

Prototyping multi-agent systems built on top of agentic frameworks to coordinate tool use and multi-step reasoning. Emphasis is on modular orchestration, context sharing, and task decomposition.


🦾 Moving Toward Vision-Language-Action Models

Shifting toward embodied AI by working with Vision-Language-Action models that perceive, reason, and act. Also contributing to the development of new benchmarks that better reflect grounded, sequential decision-making tasks.

Get in touch if you want to build together or even just have a chat!