Welcome to The Weekly Five - your curated list of 5
exceptional open source projects I discovered this week.

Build Your Own Stack: Five Tools That Put Control Back on Your Machine

The cloud is convenient until it isn't. Maybe it's a surprise bill. Maybe it's a service sunset. Or maybe it's just that nagging feeling that your data lives on someone else's servers. Whatever the trigger, more developers are reclaiming their infrastructure. This week we're spotlighting five open source projects that let you search, cache, back up, orchestrate, and design without phoning home. Each one runs on hardware you control, scales on your terms, and stays accountable to a community.

Top takeaways

  • Local-first doesn't mean feature-light: these projects match (and sometimes exceed) their hosted counterparts in capability.

  • Self-hosting has never been more approachable, with Docker images, single-binary installs, and batteries-included defaults.

  • Owning your stack future-proofs your work against API changes, price hikes, and vendor lock-in.

Who this issue is for

Developers, homelab tinkerers, and teams who want production-grade tooling without ceding control to third-party services.

1. qmd

Why this made the cut: A local search engine that rivals cloud-based semantic search, all without sending your documents anywhere.

Why it matters

Your notes, meeting transcripts, and internal docs deserve powerful search. But shipping them to an external API? That's a non-starter for many teams. QMD brings state-of-the-art retrieval (BM25 full-text, vector embeddings, LLM re-ranking) to your terminal, running entirely on-device via node-llama-cpp with GGUF models. No network calls. No token fees. No data leakage.

Key features

  • Hybrid retrieval: combines lexical search, vector semantic search, and HyDE (hypothetical document embeddings) via reciprocal rank fusion plus neural re-ranking.

  • Flexible querying: the query command supports typed sub-queries (lex, vec, hyde) so you can tune precision vs. recall on the fly.

  • Document retrieval helpers: get fetches a single doc by path or docid (with fuzzy suggestions), while multi_get handles globs, comma-separated lists, or batch docids.

How to use

  1. Install via npm: npm install -g qmd

  2. Index a folder of markdown files: qmd index ~/notes

  3. Search with natural language: qmd query "action items from last week's standup"

  4. Retrieve a specific doc: qmd get meeting-2024-05-12.md

Because the models run locally, first-run indexing downloads weights (a few GB). After that, queries complete in milliseconds to low seconds depending on corpus size and hardware.

🔗 View on GitHub | GitHub stars: 24,847

2. valkey

Why this made the cut: A community-driven Redis fork that keeps the familiar API while removing licensing uncertainty.

Why it matters

When Redis shifted its license, teams depending on permissive open source faced a choice: accept new terms or migrate. Valkey is the community's answer, a flexible, distributed key-value store optimized for caching and real-time workloads. It's fully BSD-licensed and maintained by nearly 1,000 contributors. If you've built muscle memory around Redis commands, Valkey feels like home.

Key features

  • Drop-in compatibility: existing Redis clients, libraries, and tooling work without modification.

  • Production-ready clustering: supports replication, sharding, and failover out of the box.

  • Active ecosystem: topics like cache, nosql, and key-value-store reflect a broad base of integrations.

How to use

  1. Clone and build: git clone https://github.com/valkey-io/valkey && cd valkey && make

  2. Start the server: ./src/valkey-server

  3. Connect with any Redis-compatible client. In Python, for example:

import redis

r = redis.Redis()

r.set("foo", "bar")

print(r.get("foo"))

For quick Docker setups, run docker run -p 6379:6379 valkey/valkey:latest and you're ready to connect. If you prefer bare-metal deployments, configure a systemd service file pointing to valkey-server with your desired config path.

🔗 View on GitHub | GitHub stars: 25,774

3. databasus

Why this made the cut: Database backups shouldn't require a SaaS subscription; this Go binary handles Postgres, MySQL/MariaDB, and MongoDB with a clean web UI.

Why it matters

Backups are the last line of defense, yet many teams still cobble together cron scripts and hope for the best. Databasus is a self-hosted, open source tool purpose-built for database snapshots. It supports the big three (PostgreSQL, MySQL/MariaDB, MongoDB), ships to S3-compatible storage, and includes a web interface for scheduling and monitoring. Zero open issues at time of writing signals a stable, well-maintained codebase.

Key features

  • Multi-database support: configure Postgres, MySQL/MariaDB, and MongoDB connections in one place.

  • S3 and local storage: push backups to any S3-compatible bucket or keep them on disk.

  • Docker and Kubernetes ready: first-class container support for modern deployment patterns.

How to use

  1. Pull the Docker image: docker pull databasus/databasus

  2. Run with environment variables for your database credentials and S3 keys

  3. Open the web UI at http://localhost:8080 to create backup schedules.

  4. Verify restores periodically: a backup you've never tested is just a hope.

🔗 View on GitHub | GitHub stars: 6,854

4. crewAI

Why this made the cut: Multi-agent AI orchestration that runs wherever your Python environment lives, no vendor lock-in required.

Why it matters

Autonomous AI agents are moving from demo to production, but most orchestration frameworks assume you'll pay per API call to a hosted model. CrewAI lets you define role-playing agents, assign tasks, and coordinate collaboration locally (or with your own hosted LLMs). With over 51k stars and 7k forks, it has become the de facto open framework for agentic workflows.

Key features

  • Role-based agents: define personas (researcher, writer, critic) with goals, backstories, and tools.

  • Task orchestration: chain tasks with dependencies so agents hand off context seamlessly.

  • LLM flexibility: plug in OpenAI, Anthropic, or local models via LangChain integrations.

How to use

  1. Install: pip install crewai

  2. Define agents and tasks in Python:

from crewai import Agent, Task, Crew

researcher = Agent(

    role="Researcher",

    goal="Find accurate data",

    backstory="Senior analyst with 10 years experience"

)

writer = Agent(

    role="Writer",

    goal="Draft clear summaries",

    backstory="Technical writer who simplifies complexity"

)

task = Task(description="Summarize Q2 metrics", agent=researcher)

crew = Crew(agents=[researcher, writer], tasks=[task])

crew.kickoff()
  1. Swap the llm parameter to a local model endpoint for fully offline operation.

Start with two agents and one handoff before scaling to complex crews: debugging is easier when the graph is small. Once you're comfortable, add tools (web search, file access, APIs) to give agents real capabilities.

🔗 View on GitHub | GitHub stars: 51,405

5. editor (Pascal Editor)

Why this made the cut: Browser-based 3D architectural design powered by React Three Fiber and WebGPU, entirely self-hostable.

Why it matters

Cloud CAD tools lock your floor plans behind subscriptions and proprietary formats. Pascal Editor is an open source 3D building editor you can run locally or deploy on your own server. Built on React Three Fiber with WebGPU acceleration, it delivers real-time rendering without sending geometry to a remote service. Architects, game designers, and hobbyists get full creative control.

Key features

  • WebGPU rendering: leverages next-gen browser graphics for smooth, high-fidelity visuals.

  • Monorepo architecture: core, viewer, and editor packages let you embed just the piece you need.

  • Shareable projects: export and share 3D scenes without depending on a proprietary cloud.

How to use

  1. Clone the repo: git clone https://github.com/pascalorg/editor && cd editor

  2. Install dependencies (Turborepo monorepo): pnpm install

  3. Start the dev server: pnpm dev

  4. Open http://localhost:3000 and begin placing walls, floors, and objects.

For embedding in your own app, install the viewer package (pnpm add @pascal-app/core) and render scenes with sensible defaults. The viewer handles camera controls, lighting, and interaction out of the box.

🔗 View on GitHub | GitHub stars: 15,435

If you only try one

Start with qmd. Search is the gateway to every other workflow: you can't back up what you can't find, orchestrate agents over docs you can't query, or design buildings from specs buried in a folder. QMD installs in one command, indexes in minutes, and immediately upgrades how you interact with your own knowledge. Once local semantic search clicks, you'll wonder why you ever tolerated cloud-only alternatives.

If you are doing Open Source I have a good news for you, I work at CodeRabbit which is an AI review tool and its free for Open Source, please reach out to me on X or LinkedIn or just send an email on [email protected] if you need help on adopting CodeRabbit.

You can visit our portal below to create a new account and connect your repository and start reviewing your code.

Keep reading