Testing OpenCode Agents with Devstral 2 and Superpowers
I've been testing OpenCode with Devstral 2 and the Superpowers framework, and I've noticed that while I can get an agent to work fairly autonomously, there are some challenges when direct feedback isn't available to pass to the agent. This makes certain problems difficult to debug.
The RSS Reader Challenge
One specific example I've encountered is implementing smooth scrolling in a small CLI tool. I wanted to achieve dynamic line-by-line scrolling, but the best I've managed so far is jerky, inconsistent scrolling. The program is an RSS reader that lists feeds grouped by source.
The agents are becoming increasingly capable, but fine-grained control over certain UI elements in terminal applications remains challenging. The scrolling behavior I want would provide a much better user experience when browsing through long lists of RSS feed items.
Autonomous Debugging Limitations
Without the ability to provide direct feedback about the scrolling behavior, the agent struggles to:
- Understand the exact nature of the "jerky" scrolling issue
- Determine what constitutes "smooth" scrolling in a terminal environment
- Iterate on solutions without human guidance
- Test different approaches systematically
Current Implementation
The current implementation shows feeds in a basic list format:
Source 1:
- Article 1 title
- Article 2 title
- Article 3 title
Source 2:
- Article 1 title
- Article 2 title
When there are many articles, the scrolling becomes problematic as the terminal tries to handle the output.
Potential Solutions
I'm considering several approaches to improve this:
- Pagination: Show a fixed number of items per "page" with next/previous navigation
- Terminal UI Library: Use a library like
blessedorinkfor better terminal control - Custom Scrolling Logic: Implement manual scroll position tracking
- Performance Optimization: Reduce the rendering overhead that causes jerkiness
The Future of AI Agents
Despite these challenges, I'm impressed with how capable these agents are becoming. They can handle complex tasks, understand context, and even make autonomous decisions about implementation approaches. The scrolling issue is just one example where human feedback could bridge the gap between "good enough" and "polished."
As these systems continue to evolve, I expect we'll see improvements in:
- Understanding subtle UI/UX requirements
- Handling real-time interactive elements
- Providing more granular control over output formatting
- Better debugging capabilities for terminal applications
The agents are getting better every day, and while there are still some rough edges to smooth out, the progress is remarkable. I'm excited to see what the next generation of these tools will be capable of.
DISCLAIMER: I've let AI write down this post