MCP-enabled computer-use agents are proliferating faster than our ability to evaluate them — and most existing benchmarks depend on commercial APIs that get deprecated without notice. OpenMCP is an open-source, fully self-hosted benchmarking harness built by NYU researchers that lets anyone reproducibly evaluate MCP agents across diverse hardware, from H100 datacenter GPUs to a Raspberry Pi 5 at the edge.
Announcing NSDI 2026 Bird-of-Feather (BoF) Session on Reproducibility
Our Reproducibility Ambassador Is Heading to NSDI ’26 — Here’s What to Expect
A PhD student selected through the Reproducibility Ambassador program funded by the NSF REPETO project is heading to NSDI '26 to share how researchers can package and reproduce their experiments using Chameleon and Trovi. Join the BoF sessions on May 4th and May 5th to see it in action.
Running Artifact Evaluations on Chameleon
A practical guide for AE organizers using shared research infrastructure
Chameleon has supported artifact evaluations at more than 30 events across 16 major HPC and systems conferences. This guide distills those lessons into practical advice for AE organizers: how to plan hardware access, structure author and reviewer workflows, and keep reproducible artifacts alive after the evaluation closes.
Baremetal H100 nodes on Chameleon
Chameleon Newsletter & Changelog March 2026
Welcome to the Chameleon March 2026 Newsletter!
This month we're highlighting the last chance to register for the Sixth Chameleon User Meeting, a new webinar recording from UTEP's MINCER team, platform updates including multi-instance GPU support, new cloud traces, switch performance improvements, and several testbed usability enhancements.
Wax: Making Stale Profiles Work for Data Center Optimization
Leveraging source code and debug information to unlock up to 93% of fresh-profile performance gains
Data centers waste enormous performance potential because software evolves faster than profiles can be collected. Wax, developed by Tawhid Bhuiyan and colleagues at Columbia University, uses source code and debug information to recover 65–93% of fresh-profile performance gains from stale profiles—outperforming the state of the art by up to 7.86 percentage points across five real-world data center applications.
Running LLMs on Chameleon GPUs from FABRIC via Stitch Ports
A Trovi artifact for cross-testbed GPU workflows using Chameleon and FABRIC
Chameleon's stitch port feature lets you connect bare-metal GPU servers directly to FABRIC's network fabric — no public IP required. This post walks through a Trovi artifact that provisions a Chameleon GPU node, links it to a FABRIC slice over fabnet, and runs LLM inference across the stitched network. Grab the notebook and adapt it to your own cross-testbed workflows.
Chameleon User Meeting 2026: Valerie Taylor (ANL) to Deliver Keynote
On Using LLMs to Power Down HPC
This April, we're bringing together researchers, educators, and infrastructure builders for two days of honest conversation about AI on research computing infrastructure. We're thrilled to kick things off with a keynote from one of HPC's leading voices.
Chameleon Newsletter & Changelog February 2026
Welcome to the Chameleon February 2026 Newsletter!
This month we're highlighting an exciting keynote announcement for our upcoming User Meeting, reproducibility work at FAST 2026, a new webinar recording, and several platform updates including improvements to Trovi, cc-snapshot, CHI@Edge, and python-chi.
A Holistic Approach to Teaching Cloud Computing
How Chameleon Cloud Helped One Community College Bring Hands-On Cloud Education to Hundreds of Students
Teaching cloud computing without tying students to a single vendor is harder than it sounds. Dr. Michael MacLeod at Austin Community College found a solution in Chameleon, and the results speak for themselves: 14 courses, 420 students, and a curriculum that's still growing. Read more in the blog.