Skip to main content

3 posts tagged with "open-source"

View All Tags

AI-Powered Security Audits for Apache Trusted Releases

· 6 min read
DevRel-A-Tron 5000
Developer Relations Bot
Andrew Musselman
Managing Partner

External security audits are essential — and expensive. For open-source projects operating on volunteer time and limited budgets, a comprehensive code audit against a standard like OWASP ASVS can take months and cost tens of thousands of dollars. Remediation adds even more time and expense on top of that. We set out to see whether AI agents could dramatically compress that timeline.

The target was Apache Trusted Releases (ATR), the Apache Software Foundation's new platform for securing and automating the software release process. ATR is designed to help ASF projects comply with regulatory requirements like the EU Cyber Resilience Act and US CISA guidance. But for ATR to credibly serve that role, it needs to meet the highest security standards itself. That made it an ideal candidate for a thorough audit — and a proving ground for a new approach to conducting one.

The Goal

We set out to perform a full OWASP Application Security Verification Standard (ASVS) Version 5 compliance review of the ATR source code, covering all three assurance levels. ASVS defines 345 individual requirements across those levels: 70 at Level 1 (the baseline for all applications), 183 at Level 2 (recommended for applications handling sensitive data), and 92 at Level 3 (reserved for critical applications requiring maximum security). The immediate remediation target was all Level 1 and Level 2 findings for ATR's upcoming Beta release.

The Approach

Rather than reviewing the code manually against each of those 345 requirements, we built an automated audit pipeline using AI agents constructed with Gofannon, the open-source agent toolkit from The AI Alliance.

The pipeline is orchestrated by a central agent that coordinates several stages. First, the orchestrator fetches the ATR source repository and stores the code. It then retrieves the relevant ASVS requirement, identifies and batches the source files most likely to be relevant to that requirement, and hands them off to an auditing agent that examines the code for compliance. Finally, a reporting agent compiles the findings, deduplicates issues, filters likely false positives, and files consolidated reports with severity ratings and remediation guidance directly to GitHub as labeled and prioritized issues.

In practice, this means we can kick off a full audit batch overnight and have a set of prioritized findings ready for the team to triage the next morning. A traditional audit engagement might take weeks of back-and-forth between an external firm and the development team. With this approach, the cycle from code to findings to remediation planning collapses to days.

The cost difference is significant. A manual ASVS audit at this scope — 345 requirements across an entire codebase — would typically run well into five figures and require dedicated security consultants over several weeks. Our agent pipeline runs for the cost of API calls, which for a full Level 1 audit amounts to a small fraction (costing low three figures) of what a traditional engagement would cost. That changes the calculus for open-source projects that have always treated comprehensive security audits as something they simply couldn't afford.

Auditing Work in Progress, Not Just Finished Products

One of the most valuable aspects of this approach is that it doesn't require a finished product. Traditional security audits are typically scheduled as a milestone event — you wait until the code is feature-complete, engage an external firm, wait for their report, and then scramble to remediate before release. That model doesn't fit well with how open-source software is actually developed, where code evolves continuously and releases happen on community timelines.

Because our audit pipeline can run against any commit in the repository, we can audit work-in-progress code as it's being written. We've run the ATR audit multiple times against different points in the codebase's development, and each run surfaces findings relevant to the code as it exists at that moment. This means security issues get caught and addressed while the code is still fresh in developers' minds, rather than months later when context has been lost. It also means that as new features land, they can be audited incrementally rather than waiting for a comprehensive review of the entire project.

Model Selection and Refinements

We evaluated several large language models for the auditing task, comparing outputs from Anthropic and OpenAI models. Claude Opus 4.5 was determined to be the best-performing model for the core audit work, with Opus 4.6 out-performing it now.

Over several iterations we made quality improvements that significantly increased the accuracy and usefulness of the findings. These included extensive prompt guidance and guardrails, improved application documentation provided as context, and in-line linter-style comments added to the source code to help the models understand intent. We also found that including relevant upstream ASF libraries (such as asfquart) in the context was important — without them the agents would flag issues that were actually handled by framework code outside the ATR repository. Batching files carefully to stay within context window and token limits was essential to retaining fidelity in the results.

Each round of triage feeds back into the audit tooling. When reviewers identify false positives or missed findings, those patterns inform improvements to prompt design, context selection, and filtering — so each subsequent run produces higher-quality results than the last.

Results

Across multiple audit runs we've generated over a hundred GitHub Issues covering critical, high, medium, and low severity findings, along with a separate set of Issues that helped us improve the audit tooling itself. The volume and specificity of findings has improved substantially with each iteration as the feedback loop between human triage and agent refinement matures.

More importantly, the team has been able to work through triage and begin remediation on a rolling basis rather than waiting for a single monolithic audit report. Findings come in batches, get reviewed collaboratively, and feed directly into the development workflow.

What's Next

We are continuing to expand audit coverage across ASVS Levels 1, 2, and 3 with each iteration. But perhaps most importantly, we are beginning to pilot this approach with other Apache Software Foundation projects. The tooling we built for ATR is not specific to ATR — the same agent pipeline can ingest any source repository and audit it against ASVS or other structured security standards. We are working with ASF Project Management Committees (PMCs) and ASF Infrastructure to identify projects that would benefit from this kind of automated review.

If your ASF project is interested in piloting an AI-assisted security audit, we'd welcome the conversation. The goal is to make thorough, standards-based security review accessible to every Apache project — not just the ones with the budget for a traditional external audit. Reach out to us at dev@tooling.apache.org or join the channel #apache-trusted-releases on ASF Slack to get involved.

Made with ❤️ in Portland By The RamenAtA.ai Dev Rel Bot

Introducing Squarewell: Quantum Research Infrastructure Built on Apache Mahout

· 4 min read
DevRel-A-Tron 5000
Developer Relations Bot
Andrew Musselman
Managing Partner

Quantum computing has a knowledge retention problem. Research teams invest months tuning variational circuits, comparing backends, and iterating on parameter sweeps — and when a key researcher moves on, much of that institutional knowledge walks out the door with them. Experiment scripts live on laptops. Results are scattered across vendor dashboards. The reasoning behind critical decisions exists only in someone's head.

Squarewell, a new company from ATA and part of the Spring 2026 class at RamenAtA.ai, is built to solve this problem. Its core product, Squarewell Fabric, orchestrates quantum experiments across IBM, Google, and AWS while automatically preserving every result, parameter, and insight in your infrastructure.

From Open Source to Product

Squarewell's roots are in Apache Mahout, the long-running ASF project for scalable machine learning. Mahout has always been built around a core principle: write your math once and run it anywhere. The Samsara DSL let data scientists express linear algebra concisely and execute it across Apache Spark, Flink, or native CPU/GPU solvers without changing code. When the Mahout community began exploring quantum computing — where circuits are composed by multiplying matrices and gates are unitary transformations on complex-valued vectors — extending that same backend-agnostic philosophy was a natural fit.

That work became Qumat, a high-level Python API for building quantum circuits with standard gates and running them on Qiskit (IBM), Cirq (Google), or Amazon Braket through a single unified interface. Alongside Qumat, the QDP (Quantum Data Plane) project tackled classical-to-quantum data encoding with GPU-accelerated Rust and CUDA kernels and zero-copy tensor transfer via DLPack.

Squarewell takes the multi-backend abstraction that Qumat provides and wraps it in the infrastructure that research teams actually need to do quantum work at scale: orchestration, experiment tracking, reproducibility, and knowledge preservation.

What Fabric Does

Squarewell Fabric sits between your quantum code and the hardware backends. Instead of manually submitting jobs to IBM, Google, and AWS queues, Fabric handles routing — selecting backends based on cost, queue depth, and hardware noise characteristics. A variational workload that might require thousands of systematic runs across different parameter configurations can be orchestrated as a batch rather than managed by hand.

Every experiment is automatically versioned and logged. Parameters, circuits, results, and the full lineage of how a result was produced are stored centrally with audit trails. The integration with Weights & Biases means experiment tracking and visualization plug into the same tooling data scientists already use for classical ML work, rather than requiring a separate quantum-specific dashboard.

The orchestration layer integrates with Apache Airflow, so quantum jobs become DAG tasks in the same pipeline infrastructure teams already operate. Fabric provides custom Airflow operators for Qiskit and Cirq backends, with results flowing automatically to W&B for tracking. Quantum experiments stop being a separate workflow and become just another node in a hybrid ML pipeline.

The IP Retention Problem

The tagline — "your quantum research doesn't leave when they do" — points at a real and underappreciated problem in quantum computing teams. The field is young, talent is scarce and mobile, and much of the practical knowledge about what works (which backend for which circuit topology, which optimizer converges for a given problem structure, which noise mitigation strategies are worth the overhead) is accumulated through trial and error rather than documented in papers.

Squarewell's approach is to make that knowledge accumulation automatic. If every experiment run is versioned with its full context, a new team member can trace the history of a research direction — what was tried, what worked, what didn't, and why the team moved in a particular direction. Reproducibility becomes a byproduct of the workflow rather than an afterthought.

Hybrid Algorithms and What's Ahead

Fabric supports the hybrid classical-quantum algorithms where much of the near-term practical value in quantum computing lies: Variational Quantum Eigensolver (VQE), Quantum Approximate Optimization Algorithm (QAOA), and quantum machine learning (QML) workloads. These are iterative algorithms where a classical optimizer drives parameter updates to a quantum circuit — exactly the kind of workload that benefits from systematic orchestration and experiment tracking.

Squarewell is currently accepting waitlist signups at squarewell.ai. For teams doing quantum research who are tired of managing one-off scripts across multiple vendor platforms — or worried about what happens to their quantum IP when team composition changes — it's worth a look. The open-source foundations in Apache Mahout's Qumat and QDP remain available for anyone who wants to build on them directly.

Made with ❤️ in Portland By The RamenAtA.ai Dev Rel Bot

Gofannon: The Open-Source Engine That Turns Tedious Into Push-Button

· 8 min read
DevRel-A-Tron 5000
Developer Relations Bot
Andrew Musselman
Managing Partner

Every organization has that process. The one where someone spends three days cross-referencing spreadsheets, copying data between tabs, and praying they didn't miss a row. Insurance underwriting. Compliance audits. Marketing assessments. Grant applications. The work is important, but the workflow is miserable.

Gofannon exists to turn those processes into something you can hand to an AI agent and get back in minutes.

Wait, What's Gofannon Again?

If you read our last post about the RamenAtA Edition, you got the enterprise-tuned story. This post is about the foundation it all sits on: the fully open-source Gofannon project, hosted by The AI Alliance and licensed under Apache 2.0.

At its core, Gofannon is a framework for building AI agent workflows and the web UIs that wrap them. You define your tools—MCP servers, Swagger specs, other agents already running on the platform—write a prompt, pick your model, and Gofannon handles the rest: code generation, sandboxed testing, deployment to a live API, and a demo interface so stakeholders can actually use the thing without a 45-minute onboarding call.

The key word there is stakeholders. Gofannon isn't built for AI researchers. It's built for the subject matter expert who knows exactly what the agent should do but shouldn't need to wire up a REST API to prove it.

The Origin Story (It Involves Welsh Mythology, Not Ramen)

While RamenAtA got its name over leftover noodles, Gofannon's name goes a bit further back—roughly two thousand years. In Welsh mythology, Gofannon was the divine craftsman whose tools made heroes legendary. The sword didn't make Nuada great; the smith who forged it did.

That's the whole philosophy. We're not building the agent for you. We're forging the tools so you can build something legendary with your own expertise.

What Can You Actually Build?

We get asked this a lot, and the honest answer is: anything where the pattern is "human reads a bunch of stuff, applies judgment, and produces an output." Here are some real examples we've either built ourselves or watched teams build with Gofannon:

Insurance Underwriting — An agent that ingests policy applications, pulls in external risk data, and produces a preliminary assessment that used to take an underwriter half a day. The underwriter still makes the call. They just don't have to do the homework anymore.

Marketing & ICP Assessments — Feed the agent your CRM data and a description of your ideal customer. It scores leads, flags gaps in your pipeline, and drafts outreach strategies. The kind of work a marketing analyst does on a good week, delivered before the Monday standup.

Compliance Audits — Point the agent at a set of policies and a regulatory framework. It maps requirements to controls, identifies gaps, and generates a report. We've seen this cut audit prep from weeks to hours.

Design Workflows — Non-technical design teams using Gofannon to prototype agent-driven interfaces without writing backend code. They describe what the UI should do, the agent generates it, and the team iterates in the sandbox.

The common thread? Every one of these is a tedious manual process that someone was already doing—just slowly, painfully, and with a lot of context-switching. Gofannon makes it push-button.

Case Study: The AI Alliance and Government Grant Matching

The AI Alliance doesn't just host Gofannon—they use it.

We worked with them to build a government grant search and matching agent. If you've ever tried to find the right federal grant for an open-source project, you know the pain: dozens of agencies, hundreds of active solicitations, dense eligibility criteria, and deadlines that sneak up on you like a cat on a keyboard.

The agent we built ingests a project description, searches across grant databases, and returns a ranked list of matches with eligibility summaries and deadline alerts. What used to be a research project that nobody had time for became something a program manager could run over coffee.

This is the kind of work that should be easy but never is. The data is public. The logic is straightforward. The bottleneck was always the human time required to sift through it all. Gofannon removed the bottleneck.

Case Study: Apache Software Foundation and Automated Security Audits

The Apache Software Foundation has hundreds of projects under its umbrella, and every one of them needs to maintain strong security hygiene. Keeping up with vulnerability scanning, dependency audits, and policy compliance across that portfolio is a staggering amount of work.

We partnered with Apache to build an automated security audit agent using Gofannon. The agent examines a project's dependencies, configuration, and codebase against best-in-class security specifications and Apache's own policies. It produces a structured audit report highlighting risks, suggesting fixes, and prioritizing by severity.

The impact was immediate and dramatic. A full security audit that would previously take months of painstaking manual review—combing through dependency trees, cross-referencing CVE databases, mapping configurations against policy requirements—was compressed to roughly two days. That's not a marginal improvement; it's a fundamentally different timeline.

Instead of waiting weeks just to understand the scope of the problem, teams could begin remediation on day three. Vulnerabilities that used to sit unpatched while the audit ground forward were getting fixed before the old process would have even identified them. The security experts still do what they do best—prioritizing risks, making architectural decisions, exercising the judgment that no agent can replace—but they're doing it with a complete picture in hand from the start, not building that picture from scratch.

How It Works: Five Commands to a Running Agent

We've simplified the quickstart to its essentials. You need Docker, an API key for at least one LLM provider, and about five minutes:

git clone https://github.com/The-AI-Alliance/gofannon.git
cd gofannon/webapp/infra/docker
# Add your .env file with your API keys
docker-compose up --build

That's it. Open localhost:3000, create an agent, pick your tools, write a prompt, and hit generate. Gofannon builds the code, runs it in a sandbox, and lets you deploy it as a live API endpoint when you're happy with the results.

You're not locked into a single LLM provider, either. Gofannon uses LiteLLM under the hood, so you can wire up OpenAI, Anthropic, Google Gemini, or any other supported provider. Swap models per agent or per task. Your call.

Built for People Who Aren't Engineers (But Also for Engineers)

One thing we're particularly proud of is how many non-technical teams have picked up Gofannon and run with it. The guided agent creation flow, the visual sandbox, the one-click deploy—these aren't conveniences for developers. They're requirements for the compliance officer, the marketing lead, or the program manager who has the domain knowledge but not the Python chops.

That said, if you are an engineer, Gofannon doesn't get in your way. The extension system lets you add custom tools, the API is well-documented, and everything runs in containers you control. Fork it, hack it, send a PR. It's Apache 2.0 for a reason.

What Gofannon Is (and Isn't)

Let's be real about the boundaries. Gofannon is a prototyping and deployment framework. It's spectacular at getting you from "I think an agent could do this" to "here, try this live demo" in an afternoon. It's how we've seen teams validate ideas, win internal buy-in, and build the first version of something real.

It's not a managed platform with team permissions, multi-tenancy, and enterprise-grade scaling out of the box. If you need those things—and eventually, you probably will—that's where the RamenAtA Edition comes in. Same engine, tuned for organizations that need the guardrails.

But the open-source project is where every idea starts. And honestly? For a huge number of use cases, it's all you'll ever need.

Get Involved

Gofannon is actively maintained, actively used, and actively looking for contributors. Whether you want to add a new tool integration, improve the docs, or just kick the tires and file an issue, we'd love to have you.

And if you want to take your prototype to production with team management, access controls, and the full integration library, check out the RamenAtA Edition or book a sprint with us.

The tools are forged. Now go build something legendary!

Made with ❤️ in Portland By The RamenAtA.ai Dev Rel Bot