GSoC 2026 Proposal Discussion: AI Chatbot to guide user workflow

Hello everyone! :waving_hand: I’m Daniele Caldarigi, a 2nd year CS Student from Italy. I’m very excited about the AI Chatbot project draft and I’m currently drafting a detailed technical proposal.

I have a strong background in Java, Python (FastAPI), and React, which I believe fits the requirements for this project perfectly. I’m particularly interested in implementing a RAG-based architecture to handle Jenkins documentation.

I have a few questions to better refine my proposal:

  1. Persistence: Should the plugin support multiple persistent chat sessions (like ChatGPT), or is a single ephemeral session preferred for specific workflow assistance?

  2. Tech Stack: Do you have specific preferences or constraints for the AI orchestration layer (e.g., LangChain, LlamaIndex) and the vector database?

  3. Communication: Would you prefer to discuss the proposal details here on the forum or on a shared Google Doc later on?

Regarding contributions, I am already looking into some issue to familiarize myself with the Jenkins ecosystem. Are there any specific areas or plugins related to this project that you would recommend exploring?

Looking forward to your feedback! Best regards, Daniele

1 Like

Hi Daniele

Welcome on board then. Today we really discuss on AI related topic in Jenkins and as you mention there are several inactive at the moment. Some of the people are at fosdem but I think that they will show up :wink:

Hi Daniele,

For the GSoC 2026 proposal we prefer to use a running Google Doc for discussion. As per your few questions:

  1. I think it would be best if we could support multiple persistent chat sessions.
  2. I prefer LangChain for the AI orchestration layer, no preference on the vector database though.
  3. We could follow up here as well as on Gitter/Matrix.

You should try to experiment with Jenkins as a user first, try to use a few of our more popular plugins. For plugin popularity maybe you could refer to https://plugins.jenkins.io/.

Hi Kris :waving_hand:,

Thank you for your feedback! I’ve been experimenting with the Jenkins ecosystem and exploring the core plugins to find a suitable first issue to tackle.

Regarding the AI Chatbot project, I have a couple of questions to better align my proposal with the organization’s vision:

  1. LLM Strategy: Does the Jenkins org have a preference between Open Source/Self-hosted LLMs (e.g., Llama 3 via Ollama/EC2) versus Cloud-based APIs (e.g., OpenAI/Gemini)? I’m considering a provider-agnostic approach using LlamaIndex, but knowing your preference for the final production environment would be very helpful.

  2. Proposal Sharing: I’m currently drafting the Google Doc. What is the preferred way to share it for initial feedback? Should I post the link here in the public channel for community review, or is there a specific process for draft submissions?

Hi Daniele,

Sounds good!

Regarding your quesions:

  1. We do prefer open-source / self-hosted LLMs over cloud-based and proprietary solutions.
  2. We currently do not have a dedicated channel for sharing your draft as a Google doc with us now but later on we will have a Google form you can use to share this with us discretely once Jenkins has been accepted as a GSoC 2026 mentoring org. We will not know until late February though.

Hi Kris :waving_hand:,

Thanks again for your previous feedback.

After getting familiar with Jenkins, I’ve made my first contribution and am looking forward to working on other issues. In the meantime, I’m refining the architecture for the AI Chatbot project proposal. I’m currently evaluating two different approaches for the AI backend and would love to hear the community’s preference:

  1. Local-first (Subprocess/Launcher): Running Python scripts directly on the Jenkins Controller. While it’s more “compact,” it couples the heavy LLM workload with Jenkins’ core processes.

  2. Decoupled (REST API/FastAPI): Communication via HTTP with an external backend. This allows for both self-hosted configurations (e.g., via Docker/Ollama) and cloud-based options (AWS/OpenAI), ensuring Jenkins’ stability and scalability.

Regarding the models, I have one question: is an open-weight model (like Llama or Mistral) acceptable, or does the community strictly prefer a fully open-source model (including training data and code)?