Technology

GitLab service desk with iRedMail: details

In addition to my previous article about GitLab service desk feature with iRedMail I would like to enhance it a little bit my fixing some missing parts. Starting with connecting to PostgreSQL vmail database: Verifying mail server connectivity and TLS issues. GitLab (and iRedMail also) does not allow expired certificates. So be sure to have both root CA present and your certificate valid. And finally restarting services. Interesting part is that only Nginx do not allow placing R11 root CA before actual certificate. But placing R11 after actual certificate does the job.

Technology

Proxmox 7 pmxcfs failed: recovery without reboot

Failing Proxmox node and unable to access /etc/pve means that you have broken pmxcf. Analyse which services are stuck: Restart services: And in case it was stuck at updating certs (on which it got stuck in my case): In case UI is still unavailable: In my case, at this point it was fine.

Technology

pfSense 2.7.2 with Realtek 2.5GbE ethernet driver

I got 2.5GbE PCI-e card, but no drivers present in pfSense installation. To list devices: There is this Realtek network card: But there is no driver enabled: I installed realtek-re-kmod-198.00_3 package. It was few months ago, but I am pretty sure that is was from pkg and not by downloading it from some repository manually. Finally add to /boot/loader.conf: So you can either download from FreeBSD repository or:

AI/ML

Mattermost AI chatbot with image generation support from Automatic1111

How about AI chatbot integraton in you Mattermost server? With possiblity to generate images using StableDiffusion… So, here is my Indatify’s Mattermost server which I have been playing around for last few nights. It is obvious that interaction with LLM model and generating images is way more playful in Mattermost than using Open WebUI or other TinyChat solution. So here you have an example of such integration. It is regular Mattermost on-premise server: Mattermost First, we need to configure Mattermost to be able to host AI chatbots. Configure Bot account Enable bot account creation, which is disabled by default. Of

AI/ML

Configuring NVIDIA RTX A6000 ADA in Ubuntu 22

I thought that installing NVIDIA RTX A6000 ADA in default Ubuntu 22 server installation would be an easy one. However, installing drivers from the repository made no good. I verified if secure boot is enable and no it was disabled. We need to install few things first: We need to get rig of previously installed drivers: Verify if secure boot is disabled: Get NVIDIA driver, such as NVIDIA-Linux-x86_64-535.216.01.run from their webiste and install it: In case you got rid of previously installed drivers, disabled secure boot and installed build tools, kernel headers… you will be good to go to compile

AI/ML

“You’re trying to frame the request as a documentary photograph”

LLMs contain built-in policies for protecting minors, animals etc. Monkey eating sausage should be against policy. But it can be fooled and finally model stops complaining and describe what we want to. Tried: to generate funny/controversial pictures. Actuall image generate takes place at Stable Diffusion and not at those conversational LLMs. However, once aksed to generate something dubious or funny they tend to reject such requests hiding befind their policies. Refusals from nexusraven and granite3-dense First I asked for Proboscis Monkey holding can of beer and eating sausage. LLM model called nexusraven refused with that request: nexusraven: I cannot fulfill

AI/ML

Code generation and artifacts preview with WebUI and codegemma:7b

Generate WebGL, Three.JS, HTML, CSS, JavaScript, no Python code, single page with rotating cube, ambient lighting. Load libraries from CDN. Let ambient lighting be as such cube edges are visible. add directional lighting also pointing at the cube. Scene needs to be navigable using arrow keys. Ensure browser compability. With codegemma:7b you can generate source code. If asked properly then in WebUI chat a artifacts feature will appear, interpreting your source code immediately, just after source code is generated. This feature is useful for designers, developers and marketers who would like to speed-up scaffolding and migrating from brainstorm into visible

AI/ML

Ollama, WebUI, Automatic1111 – your own, personal, local AI from scratch

My local toolbox was empty, now it’s full. Lately I have been writing about Ollama, WebUI and StableDiffusion on top of Automatic1111. I found myself struggling a little bit to keep up with all those information about how to run it in specific environments. So here you have an extract of step by step installation. Starting with NVIDIA driver and some basic requirements: Next we go for Docker. Ollama, but with binaries instead of Docker container. It will be much easier, and does not require installing Docker extensions for GPU acceleration support: If running Ollama on different server, then need

AI/ML

Custom Gemma AI system prompt to create own chatbot experience

I want to create custom chatbot experience. I want to be based on Google’s Gemma AI Large Language Models. I find Gemma3, especially 27b version very capable while problem solving. It has been trained on such data that I find it interesting. I will use Open WebUI to create custom “model hat” and provide chatbot experience TLDR In order to create your own chatbot, only 3 steps are required: To create own chatbot experience I can use System Prompt feature which is core part of model itself. Running on Ollama, Gemma3:27b is actually a 4-bit quantized version of full 16-bit