Your own private AI Assistant It’s 2025 after all. You thought we’d leave you stranded with just your own brain to work with? Of course not. Now, you can obviously google things, and you could also ChatGPT things, but these scenarios are mission critical, and we can’t afford to leak any sensitive information about our scenarios! While you may have top-security level clearance, hosted services do not. Therefore, we’ve provisioned for you an AI Assistant, running in your private project, and backed by OUR private version of Granite 3.1! You’re welcome. But be careful: With Granite power, comes Granite responsibilities. Accessing your Assistant From the main dashboard of OpenShift AI (https://rhods-dashboard-redhat-ods-applications.PLACEHOLDER-URL.com/) choose your private project (userX) In that project, click on the link to access the Granite Assistant: When prompted, authenticate. Once inside the AnythingLLM interface, create a new Workspace called Granite: Configure the workspace with a simpler prompt: Answer the question as accurately as you are able to Confirm it’s properly configured by asking it a random question: Intended uses We can’t think of all the things YOU might want to use this for. But we suggest the following might be ways of leveraging the AI assistant: explaining terms and concepts that may not be clear help with code analysis, explanation, modifications error and log messages analysis support with writing and developing notes brainstorming ideas, theories, and tests Getting Support during the event Reach out to the moderators if you need clarifications. 2.1 Getting connected 3.1 Briefing