<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>DevOps on Mengboy Tech Notes</title>
    <link>https://www.mfun.ink/en/categories/devops/</link>
    <description>Recent content in DevOps on Mengboy Tech Notes</description>
    <generator>Hugo -- 0.156.0</generator>
    <language>en</language>
    <lastBuildDate>Mon, 16 Mar 2026 01:08:00 +0000</lastBuildDate>
    <atom:link href="https://www.mfun.ink/en/categories/devops/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>OpenAI Responses &#43; GitHub Actions PR Risk Gate: Automated Evals, Tiered Blocking, and One-Click Rollback</title>
      <link>https://www.mfun.ink/en/2026/03/16/openai-responses-github-actions-pr-risk-gate/</link>
      <pubDate>Mon, 16 Mar 2026 01:08:00 +0000</pubDate>
      <guid>https://www.mfun.ink/en/2026/03/16/openai-responses-github-actions-pr-risk-gate/</guid>
      <description>&lt;p&gt;You don&amp;rsquo;t need an AI reviewer that “sounds smart.” You need a gate that &lt;strong&gt;stops risky PRs before they hit main&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;This post shows a production-ready minimum setup: OpenAI Responses generates structured risk output, GitHub Actions enforces tiered policies, and critical failures can trigger a one-click rollback.&lt;/p&gt;</description>
    </item>
    <item>
      <title>GitHub Actions &#43; AI Agent Auto-Fix Pipeline: Failure Tiers, Regression Gates, and Security Guardrails</title>
      <link>https://www.mfun.ink/en/2026/02/27/github-actions-ai-agent-auto-fix-pipeline/</link>
      <pubDate>Fri, 27 Feb 2026 01:18:38 +0000</pubDate>
      <guid>https://www.mfun.ink/en/2026/02/27/github-actions-ai-agent-auto-fix-pipeline/</guid>
      <description>&lt;p&gt;When CI keeps failing, the real risk is not “slow fixes” — it is “fast bad fixes.”
This guide gives you a practical &lt;strong&gt;GitHub Actions + AI Agent auto-fix pipeline&lt;/strong&gt; with failure tiering, strict edit boundaries, and merge-time gates.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Redis Distributed Lock Best Practices (with Common Misuse Cases)</title>
      <link>https://www.mfun.ink/en/2026/02/19/redis-distributed-lock-best-practices/</link>
      <pubDate>Thu, 19 Feb 2026 09:55:00 +0800</pubDate>
      <guid>https://www.mfun.ink/en/2026/02/19/redis-distributed-lock-best-practices/</guid>
      <description>&lt;p&gt;In high-concurrency scenarios, distributed locks are essential for ensuring data consistency. However, many developers&amp;rsquo; understanding of Redis distributed locks stops at &amp;ldquo;SETNX&amp;rdquo;, leading to frequent production incidents.&lt;/p&gt;
&lt;p&gt;This article comprehensively covers the correct usage of Redis distributed locks from principles, implementation, common misuse cases to production-grade solutions.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Hugo Auto Deploy with GitHub Actions: Safe Config and Troubleshooting</title>
      <link>https://www.mfun.ink/en/2026/02/16/github-actions-hugo-auto-deploy-safe-config/</link>
      <pubDate>Mon, 16 Feb 2026 11:59:00 +0800</pubDate>
      <guid>https://www.mfun.ink/en/2026/02/16/github-actions-hugo-auto-deploy-safe-config/</guid>
      <description>&lt;p&gt;Your local &lt;code&gt;hugo&lt;/code&gt; build works, but GitHub Actions fails randomly. Classic.
The root cause is usually not the workflow syntax. It is environment drift, missing permissions, and unstable dependencies.&lt;/p&gt;</description>
    </item>
    <item>
      <title>WSL2 &#43; Docker Network Troubleshooting: Fix DNS Timeouts and Image Pull Failures</title>
      <link>https://www.mfun.ink/en/2026/02/11/wsl2-docker-network-troubleshooting/</link>
      <pubDate>Wed, 11 Feb 2026 22:50:00 +0800</pubDate>
      <guid>https://www.mfun.ink/en/2026/02/11/wsl2-docker-network-troubleshooting/</guid>
      <description>&lt;p&gt;If your &lt;strong&gt;WSL2 + Docker&lt;/strong&gt; setup suddenly fails with &lt;code&gt;docker pull&lt;/code&gt; timeouts, &lt;code&gt;Temporary failure in name resolution&lt;/code&gt;, or containers that start but cannot access the internet, don&amp;rsquo;t nuke your environment yet. Most cases are recoverable in 15 minutes.&lt;/p&gt;
&lt;p&gt;This guide gives you a practical sequence: identify whether the fault is DNS, proxy/VPN, virtual NIC, or Docker daemon config—then apply the smallest fix first.&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
