“`html
Building the QWEN3.6 – Codex Bridge Furthe + Kindergarten Harness Reality Check
I’ve made some progress with my harness for running Qwen 3.6 model on Codex. While testing, analyzing, and building the harness, I evolved TBG(O)llama-swap into a full forensic UI bridge and LLM analytics tool where every harness finding, modification, correction, tool call, reasoning step, and execution flow is fully visible.
This level of transparency was necessary to identify the behavioral differences between native OpenAI models and Qwen 3.6, and to fine-tune the harness accordingly.

The video shows a full Codex run on Qwen 3.6 running on a single NVIDIA GeForce RTX 5090. (Codex in VS Code → tbg(o)llama-swap → llama.cpp with qwen 3.6 27B)
The ongoing work can be checked here: GitHub, First post, second post.
Current Status
apply_patchapply_patchcreate/update/delete flowcreate_filerequires non-emptydifforcontentupdate_filerequires non-emptydifforcontentdelete_fileworks withoutdiffshellweb_searchweb_searchusing TBG(O)llama-swap built-in web searchfile_searchview_imagerequest_user_inputupdate_planspawn_agentwait_agentsend_inputresume_agentclose_agentsupports_search_toolcatalog inconsistencyagent_send_input_roundtripagent_subagent_same_modelshell_patch_verify_sequenceweb_research_then_notesplan_act_switch_implmulti_web_patch_verifyskill_create_and_use_localworkspace_summary_then_planskill_read_localdirect_plan_no_webweb_research_then_planfile_search_then_patchview_image_then_report- invalid
apply_patchretry exhaustion no longer finalizes with fake progress prose - safer recovery branch after broken
apply_patch - false patch-intent/path-hint extraction from instructions
- reconnect bug caused by unhealthy or duplicate upstream adoption
- long delayed
502timeout path shortened and improved - native-vs-local contrast harness:
initcomparecomparison.jsoncomparison_summary.jsonImplemented in the Bridge Contract
- stricter separation of:
visible assistant texttool call itemstool outputsfile/code artifactsFixed Enough To Work, But Still Not Native-Perfect
- grouped searches
- grouped tool calls
- grouped file changes
- collapsible internal history
Fixed
mcp__playwright__browser_navigatemcp__playwright__browser_snapshotmcp__playwright__browser_clickmcp__playwright__browser_evaluatemcp__playwright__browser_resizemcp__playwright__browser_take_screenshot
Still Not Fully Closed / Needs More Work
- full native-style grouped worker UX parity
- some remaining model-quality quirks during long multi-step runs
- continuation/reporting polish around malformed reasoning/text splits
Important Nuance:
- llama-swap now preserves and exposes these much more accurately
- however, the WSL Codex router still rejects Playwright leaf calls as unsupported in this surface
- this is now tracked as a known limitation, not an active llama-swap bridge bug
Note: The placeholder image and GitHub links are provided for demonstration purposes. Actual references should be verified.
“`
Originally published at reddit.com. Curated by AI Maestro.
Stay ahead of AI. Get the most important stories delivered to your inbox — no spam, no noise.

