Monorepo for Tangled tangled.org

knotserver/xrpc: add list repo endpoint #903

closed opened by julien.rbrt.fr targeting master from julien.rbrt.fr/tangled-core: julien/listrepo

Add a list user and repo endpoint on the knot. It is useful when browsing the knot directly and not requiring indexing it via another service.

Labels

None yet.

assignee

None yet.

Participants 2
AT URI
at://did:plc:7kpq3n7brenbgyp2gx36hl6x/sh.tangled.repo.pull/3maelm6pqks22
+1359 -2028
Diff #2
+51
api/tangled/repolistRepos.go
··· 1 + // Code generated by cmd/lexgen (see Makefile's lexgen); DO NOT EDIT. 2 + 3 + package tangled 4 + 5 + // schema: sh.tangled.repo.listRepos 6 + 7 + import ( 8 + "context" 9 + 10 + "github.com/bluesky-social/indigo/lex/util" 11 + ) 12 + 13 + const ( 14 + RepoListReposNSID = "sh.tangled.repo.listRepos" 15 + ) 16 + 17 + // RepoListRepos_Output is the output of a sh.tangled.repo.listRepos call. 18 + type RepoListRepos_Output struct { 19 + Users []*RepoListRepos_User `json:"users" cborgen:"users"` 20 + } 21 + 22 + // RepoListRepos_RepoEntry is a "repoEntry" in the sh.tangled.repo.listRepos schema. 23 + type RepoListRepos_RepoEntry struct { 24 + // defaultBranch: Default branch of the repository 25 + DefaultBranch *string `json:"defaultBranch,omitempty" cborgen:"defaultBranch,omitempty"` 26 + // did: DID of the repository owner 27 + Did string `json:"did" cborgen:"did"` 28 + // fullPath: Full path to the repository 29 + FullPath string `json:"fullPath" cborgen:"fullPath"` 30 + // name: Repository name 31 + Name string `json:"name" cborgen:"name"` 32 + } 33 + 34 + // RepoListRepos_User is a "user" in the sh.tangled.repo.listRepos schema. 35 + type RepoListRepos_User struct { 36 + // did: DID of the user 37 + Did string `json:"did" cborgen:"did"` 38 + Repos []*RepoListRepos_RepoEntry `json:"repos" cborgen:"repos"` 39 + } 40 + 41 + // RepoListRepos calls the XRPC method "sh.tangled.repo.listRepos". 42 + func RepoListRepos(ctx context.Context, c util.LexClient) (*RepoListRepos_Output, error) { 43 + var out RepoListRepos_Output 44 + 45 + params := map[string]interface{}{} 46 + if err := c.LexDo(ctx, util.Query, "", "sh.tangled.repo.listRepos", params, nil, &out); err != nil { 47 + return nil, err 48 + } 49 + 50 + return &out, nil 51 + }
-1
appview/db/repos.go
··· 158 158 from repo_languages 159 159 where repo_at in (%s) 160 160 and is_default_ref = 1 161 - and language <> '' 162 161 ) 163 162 where rn = 1 164 163 `,
+3 -13
appview/pages/markup/extension/atlink.go
··· 35 35 return KindAt 36 36 } 37 37 38 - var atRegexp = regexp.MustCompile(`(^|\s|\()(@)([a-zA-Z0-9]([a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?\.)+[a-zA-Z]([a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(\b)`) 39 - var markdownLinkRegexp = regexp.MustCompile(`(?ms)\[.*\]\(.*\)`) 38 + var atRegexp = regexp.MustCompile(`(^|\s|\()(@)([a-zA-Z0-9.-]+)(\b)`) 40 39 41 40 type atParser struct{} 42 41 ··· 56 55 if m == nil { 57 56 return nil 58 57 } 59 - 60 - // Check for all links in the markdown to see if the handle found is inside one 61 - linksIndexes := markdownLinkRegexp.FindAllIndex(block.Source(), -1) 62 - for _, linkMatch := range linksIndexes { 63 - if linkMatch[0] < segment.Start && segment.Start < linkMatch[1] { 64 - return nil 65 - } 66 - } 67 - 68 58 atSegment := text.NewSegment(segment.Start, segment.Start+m[1]) 69 59 block.Advance(m[1]) 70 60 node := &AtNode{} ··· 97 87 98 88 func (r *atHtmlRenderer) renderAt(w util.BufWriter, source []byte, n ast.Node, entering bool) (ast.WalkStatus, error) { 99 89 if entering { 100 - w.WriteString(`<a href="/`) 90 + w.WriteString(`<a href="/@`) 101 91 w.WriteString(n.(*AtNode).Handle) 102 - w.WriteString(`" class="mention">`) 92 + w.WriteString(`" class="mention font-bold">`) 103 93 } else { 104 94 w.WriteString("</a>") 105 95 }
-121
appview/pages/markup/markdown_test.go
··· 1 - package markup 2 - 3 - import ( 4 - "bytes" 5 - "testing" 6 - ) 7 - 8 - func TestAtExtension_Rendering(t *testing.T) { 9 - tests := []struct { 10 - name string 11 - markdown string 12 - expected string 13 - }{ 14 - { 15 - name: "renders simple at mention", 16 - markdown: "Hello @user.tngl.sh!", 17 - expected: `<p>Hello <a href="/user.tngl.sh" class="mention">@user.tngl.sh</a>!</p>`, 18 - }, 19 - { 20 - name: "renders multiple at mentions", 21 - markdown: "Hi @alice.tngl.sh and @bob.example.com", 22 - expected: `<p>Hi <a href="/alice.tngl.sh" class="mention">@alice.tngl.sh</a> and <a href="/bob.example.com" class="mention">@bob.example.com</a></p>`, 23 - }, 24 - { 25 - name: "renders at mention in parentheses", 26 - markdown: "Check this out (@user.tngl.sh)", 27 - expected: `<p>Check this out (<a href="/user.tngl.sh" class="mention">@user.tngl.sh</a>)</p>`, 28 - }, 29 - { 30 - name: "does not render email", 31 - markdown: "Contact me at test@example.com", 32 - expected: `<p>Contact me at <a href="mailto:test@example.com">test@example.com</a></p>`, 33 - }, 34 - { 35 - name: "renders at mention with hyphen", 36 - markdown: "Follow @user-name.tngl.sh", 37 - expected: `<p>Follow <a href="/user-name.tngl.sh" class="mention">@user-name.tngl.sh</a></p>`, 38 - }, 39 - { 40 - name: "renders at mention with numbers", 41 - markdown: "@user123.test456.social", 42 - expected: `<p><a href="/user123.test456.social" class="mention">@user123.test456.social</a></p>`, 43 - }, 44 - { 45 - name: "at mention at start of line", 46 - markdown: "@user.tngl.sh is cool", 47 - expected: `<p><a href="/user.tngl.sh" class="mention">@user.tngl.sh</a> is cool</p>`, 48 - }, 49 - } 50 - 51 - for _, tt := range tests { 52 - t.Run(tt.name, func(t *testing.T) { 53 - md := NewMarkdown() 54 - 55 - var buf bytes.Buffer 56 - if err := md.Convert([]byte(tt.markdown), &buf); err != nil { 57 - t.Fatalf("failed to convert markdown: %v", err) 58 - } 59 - 60 - result := buf.String() 61 - if result != tt.expected+"\n" { 62 - t.Errorf("expected:\n%s\ngot:\n%s", tt.expected, result) 63 - } 64 - }) 65 - } 66 - } 67 - 68 - func TestAtExtension_WithOtherMarkdown(t *testing.T) { 69 - tests := []struct { 70 - name string 71 - markdown string 72 - contains string 73 - }{ 74 - { 75 - name: "at mention with bold", 76 - markdown: "**Hello @user.tngl.sh**", 77 - contains: `<strong>Hello <a href="/user.tngl.sh" class="mention">@user.tngl.sh</a></strong>`, 78 - }, 79 - { 80 - name: "at mention with italic", 81 - markdown: "*Check @user.tngl.sh*", 82 - contains: `<em>Check <a href="/user.tngl.sh" class="mention">@user.tngl.sh</a></em>`, 83 - }, 84 - { 85 - name: "at mention in list", 86 - markdown: "- Item 1\n- @user.tngl.sh\n- Item 3", 87 - contains: `<a href="/user.tngl.sh" class="mention">@user.tngl.sh</a>`, 88 - }, 89 - { 90 - name: "at mention in link", 91 - markdown: "[@regnault.dev](https://regnault.dev)", 92 - contains: `<a href="https://regnault.dev">@regnault.dev</a>`, 93 - }, 94 - { 95 - name: "at mention in link again", 96 - markdown: "[check out @regnault.dev](https://regnault.dev)", 97 - contains: `<a href="https://regnault.dev">check out @regnault.dev</a>`, 98 - }, 99 - { 100 - name: "at mention in link again, multiline", 101 - markdown: "[\ncheck out @regnault.dev](https://regnault.dev)", 102 - contains: "<a href=\"https://regnault.dev\">\ncheck out @regnault.dev</a>", 103 - }, 104 - } 105 - 106 - for _, tt := range tests { 107 - t.Run(tt.name, func(t *testing.T) { 108 - md := NewMarkdown() 109 - 110 - var buf bytes.Buffer 111 - if err := md.Convert([]byte(tt.markdown), &buf); err != nil { 112 - t.Fatalf("failed to convert markdown: %v", err) 113 - } 114 - 115 - result := buf.String() 116 - if !bytes.Contains([]byte(result), []byte(tt.contains)) { 117 - t.Errorf("expected output to contain:\n%s\ngot:\n%s", tt.contains, result) 118 - } 119 - }) 120 - } 121 - }
-1530
docs/DOCS.md
··· 1 - --- 2 - title: Tangled Documentation 3 - author: The Tangled Contributors 4 - date: 21 Sun, Dec 2025 5 - --- 6 - 7 - # Introduction 8 - 9 - Tangled is a decentralized code hosting and collaboration 10 - platform. Every component of Tangled is open-source and 11 - selfhostable. [tangled.org](https://tangled.org) also 12 - provides hosting and CI services that are free to use. 13 - 14 - There are several models for decentralized code 15 - collaboration platforms, ranging from ActivityPubโ€™s 16 - (Forgejo) federated model, to Radicleโ€™s entirely P2P model. 17 - Our approach attempts to be the best of both worlds by 18 - adopting atprotoโ€”a protocol for building decentralized 19 - social applications with a central identity 20 - 21 - Our approach to this is the idea of โ€œknotsโ€. Knots are 22 - lightweight, headless servers that enable users to host Git 23 - repositories with ease. Knots are designed for either single 24 - or multi-tenant use which is perfect for self-hosting on a 25 - Raspberry Pi at home, or larger โ€œcommunityโ€ servers. By 26 - default, Tangled provides managed knots where you can host 27 - your repositories for free. 28 - 29 - The "appview" at tangled.org acts as a consolidated โ€œviewโ€ 30 - into the whole network, allowing users to access, clone and 31 - contribute to repositories hosted across different knots 32 - seamlessly. 33 - 34 - # Quick Start Guide 35 - 36 - ## Login or Sign up 37 - 38 - You can [login](https://tangled.org) by using your AT 39 - account. If you are unclear on what that means, simply head 40 - to the [signup](https://tangled.org/signup) page and create 41 - an account. By doing so, you will be choosing Tangled as 42 - your account provider (you will be granted a handle of the 43 - form `user.tngl.sh`). 44 - 45 - In the AT network, users are free to choose their account 46 - provider (known as a "Personal Data Service", or PDS), and 47 - login to applications that support AT accounts. 48 - 49 - You can think of it as "one account for all of the 50 - atmosphere"! 51 - 52 - If you already have an AT account (you may have one if you 53 - signed up to Bluesky, for example), you can login with the 54 - same handle on Tangled (so just use `user.bsky.social` on 55 - the login page). 56 - 57 - ## Add an SSH Key 58 - 59 - Once you are logged in, you can start creating repositories 60 - and pushing code. Tangled supports pushing git repositories 61 - over SSH. 62 - 63 - First, you'll need to generate an SSH key if you don't 64 - already have one: 65 - 66 - ```bash 67 - ssh-keygen -t ed25519 -C "foo@bar.com" 68 - ``` 69 - 70 - When prompted, save the key to the default location 71 - (`~/.ssh/id_ed25519`) and optionally set a passphrase. 72 - 73 - Copy your public key to your clipboard: 74 - 75 - ```bash 76 - # on X11 77 - cat ~/.ssh/id_ed25519.pub | xclip -sel c 78 - 79 - # on wayland 80 - cat ~/.ssh/id_ed25519.pub | wl-copy 81 - 82 - # on macos 83 - cat ~/.ssh/id_ed25519.pub | pbcopy 84 - ``` 85 - 86 - Now, navigate to 'Settings' -> 'Keys' and hit 'Add Key', 87 - paste your public key, give it a descriptive name, and hit 88 - save. 89 - 90 - ## Create a Repository 91 - 92 - Once your SSH key is added, create your first repository: 93 - 94 - 1. Hit the green `+` icon on the topbar, and select 95 - repository 96 - 2. Enter a repository name 97 - 3. Add a description 98 - 4. Choose a knotserver to host this repository on 99 - 5. Hit create 100 - 101 - "Knots" are selfhostable, lightweight git servers that can 102 - host your repository. Unlike traditional code forges, your 103 - code can live on any server. Read the [Knots](TODO) section 104 - for more. 105 - 106 - ## Configure SSH 107 - 108 - To ensure Git uses the correct SSH key and connects smoothly 109 - to Tangled, add this configuration to your `~/.ssh/config` 110 - file: 111 - 112 - ``` 113 - Host tangled.org 114 - Hostname tangled.org 115 - User git 116 - IdentityFile ~/.ssh/id_ed25519 117 - AddressFamily inet 118 - ``` 119 - 120 - This tells SSH to use your specific key when connecting to 121 - Tangled and prevents authentication issues if you have 122 - multiple SSH keys. 123 - 124 - Note that this configuration only works for knotservers that 125 - are hosted by tangled.org. If you use a custom knot, refer 126 - to the [Knots](TODO) section. 127 - 128 - ## Push Your First Repository 129 - 130 - Initialize a new git repository: 131 - 132 - ```bash 133 - mkdir my-project 134 - cd my-project 135 - 136 - git init 137 - echo "# My Project" > README.md 138 - ``` 139 - 140 - Add some content and push! 141 - 142 - ```bash 143 - git add README.md 144 - git commit -m "Initial commit" 145 - git remote add origin git@tangled.org:user.tngl.sh/my-project 146 - git push -u origin main 147 - ``` 148 - 149 - That's it! Your code is now hosted on Tangled. 150 - 151 - ## Migrating an existing repository 152 - 153 - Moving your repositories from GitHub, GitLab, Bitbucket, or 154 - any other Git forge to Tangled is straightforward. You'll 155 - simply change your repository's remote URL. At the moment, 156 - Tangled does not have any tooling to migrate data such as 157 - GitHub issues or pull requests. 158 - 159 - First, create a new repository on tangled.org as described 160 - in the [Quick Start Guide](#create-a-repository). 161 - 162 - Navigate to your existing local repository: 163 - 164 - ```bash 165 - cd /path/to/your/existing/repo 166 - ``` 167 - 168 - You can inspect your existing git remote like so: 169 - 170 - ```bash 171 - git remote -v 172 - ``` 173 - 174 - You'll see something like: 175 - 176 - ``` 177 - origin git@github.com:username/my-project (fetch) 178 - origin git@github.com:username/my-project (push) 179 - ``` 180 - 181 - Update the remote URL to point to tangled: 182 - 183 - ```bash 184 - git remote set-url origin git@tangled.org:user.tngl.sh/my-project 185 - ``` 186 - 187 - Verify the change: 188 - 189 - ```bash 190 - git remote -v 191 - ``` 192 - 193 - You should now see: 194 - 195 - ``` 196 - origin git@tangled.org:user.tngl.sh/my-project (fetch) 197 - origin git@tangled.org:user.tngl.sh/my-project (push) 198 - ``` 199 - 200 - Push all your branches and tags to tangled: 201 - 202 - ```bash 203 - git push -u origin --all 204 - git push -u origin --tags 205 - ``` 206 - 207 - Your repository is now migrated to Tangled! All commit 208 - history, branches, and tags have been preserved. 209 - 210 - ## Mirroring a repository to Tangled 211 - 212 - If you want to maintain your repository on multiple forges 213 - simultaneously, for example, keeping your primary repository 214 - on GitHub while mirroring to Tangled for backup or 215 - redundancy, you can do so by adding multiple remotes. 216 - 217 - You can configure your local repository to push to both 218 - Tangled and, say, GitHub. You may already have the following 219 - setup: 220 - 221 - ``` 222 - $ git remote -v 223 - origin git@github.com:username/my-project (fetch) 224 - origin git@github.com:username/my-project (push) 225 - ``` 226 - 227 - Now add Tangled as an additional push URL to the same 228 - remote: 229 - 230 - ```bash 231 - git remote set-url --add --push origin git@tangled.org:user.tngl.sh/my-project 232 - ``` 233 - 234 - You also need to re-add the original URL as a push 235 - destination (git replaces the push URL when you use `--add` 236 - the first time): 237 - 238 - ```bash 239 - git remote set-url --add --push origin git@github.com:username/my-project 240 - ``` 241 - 242 - Verify your configuration: 243 - 244 - ``` 245 - $ git remote -v 246 - origin git@github.com:username/repo (fetch) 247 - origin git@tangled.org:username/my-project (push) 248 - origin git@github.com:username/repo (push) 249 - ``` 250 - 251 - Notice that there's one fetch URL (the primary remote) and 252 - two push URLs. Now, whenever you push, git will 253 - automatically push to both remotes: 254 - 255 - ```bash 256 - git push origin main 257 - ``` 258 - 259 - This single command pushes your `main` branch to both GitHub 260 - and Tangled simultaneously. 261 - 262 - To push all branches and tags: 263 - 264 - ```bash 265 - git push origin --all 266 - git push origin --tags 267 - ``` 268 - 269 - If you prefer more control over which remote you push to, 270 - you can maintain separate remotes: 271 - 272 - ```bash 273 - git remote add github git@github.com:username/my-project 274 - git remote add tangled git@tangled.org:username/my-project 275 - ``` 276 - 277 - Then push to each explicitly: 278 - 279 - ```bash 280 - git push github main 281 - git push tangled main 282 - ``` 283 - 284 - # Knot self-hosting guide 285 - 286 - So you want to run your own knot server? Great! Here are a few prerequisites: 287 - 288 - 1. A server of some kind (a VPS, a Raspberry Pi, etc.). Preferably running a Linux distribution of some kind. 289 - 2. A (sub)domain name. People generally use `knot.example.com`. 290 - 3. A valid SSL certificate for your domain. 291 - 292 - ## NixOS 293 - 294 - Refer to the [knot 295 - module](https://tangled.org/tangled.org/core/blob/master/nix/modules/knot.nix) 296 - for a full list of options. Sample configurations: 297 - 298 - - [The test VM](https://tangled.org/tangled.org/core/blob/master/nix/vm.nix#L85) 299 - - [@pyrox.dev/nix](https://tangled.org/pyrox.dev/nix/blob/d19571cc1b5fe01035e1e6951ec8cf8a476b4dee/hosts/marvin/services/tangled.nix#L15-25) 300 - 301 - ## Docker 302 - 303 - Refer to 304 - [@tangled.org/knot-docker](https://tangled.sh/@tangled.sh/knot-docker). 305 - Note that this is community maintained. 306 - 307 - ## Manual setup 308 - 309 - First, clone this repository: 310 - 311 - ``` 312 - git clone https://tangled.org/@tangled.org/core 313 - ``` 314 - 315 - Then, build the `knot` CLI. This is the knot administration 316 - and operation tool. For the purpose of this guide, we're 317 - only concerned with these subcommands: 318 - 319 - * `knot server`: the main knot server process, typically 320 - run as a supervised service 321 - * `knot guard`: handles role-based access control for git 322 - over SSH (you'll never have to run this yourself) 323 - * `knot keys`: fetches SSH keys associated with your knot; 324 - we'll use this to generate the SSH 325 - `AuthorizedKeysCommand` 326 - 327 - ``` 328 - cd core 329 - export CGO_ENABLED=1 330 - go build -o knot ./cmd/knot 331 - ``` 332 - 333 - Next, move the `knot` binary to a location owned by `root` -- 334 - `/usr/local/bin/` is a good choice. Make sure the binary itself is also owned by `root`: 335 - 336 - ``` 337 - sudo mv knot /usr/local/bin/knot 338 - sudo chown root:root /usr/local/bin/knot 339 - ``` 340 - 341 - This is necessary because SSH `AuthorizedKeysCommand` requires [really 342 - specific permissions](https://stackoverflow.com/a/27638306). The 343 - `AuthorizedKeysCommand` specifies a command that is run by `sshd` to 344 - retrieve a user's public SSH keys dynamically for authentication. Let's 345 - set that up. 346 - 347 - ``` 348 - sudo tee /etc/ssh/sshd_config.d/authorized_keys_command.conf <<EOF 349 - Match User git 350 - AuthorizedKeysCommand /usr/local/bin/knot keys -o authorized-keys 351 - AuthorizedKeysCommandUser nobody 352 - EOF 353 - ``` 354 - 355 - Then, reload `sshd`: 356 - 357 - ``` 358 - sudo systemctl reload ssh 359 - ``` 360 - 361 - Next, create the `git` user. We'll use the `git` user's home directory 362 - to store repositories: 363 - 364 - ``` 365 - sudo adduser git 366 - ``` 367 - 368 - Create `/home/git/.knot.env` with the following, updating the values as 369 - necessary. The `KNOT_SERVER_OWNER` should be set to your 370 - DID, you can find your DID in the [Settings](https://tangled.sh/settings) page. 371 - 372 - ``` 373 - KNOT_REPO_SCAN_PATH=/home/git 374 - KNOT_SERVER_HOSTNAME=knot.example.com 375 - APPVIEW_ENDPOINT=https://tangled.sh 376 - KNOT_SERVER_OWNER=did:plc:foobar 377 - KNOT_SERVER_INTERNAL_LISTEN_ADDR=127.0.0.1:5444 378 - KNOT_SERVER_LISTEN_ADDR=127.0.0.1:5555 379 - ``` 380 - 381 - If you run a Linux distribution that uses systemd, you can use the provided 382 - service file to run the server. Copy 383 - [`knotserver.service`](/systemd/knotserver.service) 384 - to `/etc/systemd/system/`. Then, run: 385 - 386 - ``` 387 - systemctl enable knotserver 388 - systemctl start knotserver 389 - ``` 390 - 391 - The last step is to configure a reverse proxy like Nginx or Caddy to front your 392 - knot. Here's an example configuration for Nginx: 393 - 394 - ``` 395 - server { 396 - listen 80; 397 - listen [::]:80; 398 - server_name knot.example.com; 399 - 400 - location / { 401 - proxy_pass http://localhost:5555; 402 - proxy_set_header Host $host; 403 - proxy_set_header X-Real-IP $remote_addr; 404 - proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; 405 - proxy_set_header X-Forwarded-Proto $scheme; 406 - } 407 - 408 - # wss endpoint for git events 409 - location /events { 410 - proxy_set_header X-Forwarded-For $remote_addr; 411 - proxy_set_header Host $http_host; 412 - proxy_set_header Upgrade websocket; 413 - proxy_set_header Connection Upgrade; 414 - proxy_pass http://localhost:5555; 415 - } 416 - # additional config for SSL/TLS go here. 417 - } 418 - 419 - ``` 420 - 421 - Remember to use Let's Encrypt or similar to procure a certificate for your 422 - knot domain. 423 - 424 - You should now have a running knot server! You can finalize 425 - your registration by hitting the `verify` button on the 426 - [/settings/knots](https://tangled.org/settings/knots) page. This simply creates 427 - a record on your PDS to announce the existence of the knot. 428 - 429 - ### Custom paths 430 - 431 - (This section applies to manual setup only. Docker users should edit the mounts 432 - in `docker-compose.yml` instead.) 433 - 434 - Right now, the database and repositories of your knot lives in `/home/git`. You 435 - can move these paths if you'd like to store them in another folder. Be careful 436 - when adjusting these paths: 437 - 438 - * Stop your knot when moving data (e.g. `systemctl stop knotserver`) to prevent 439 - any possible side effects. Remember to restart it once you're done. 440 - * Make backups before moving in case something goes wrong. 441 - * Make sure the `git` user can read and write from the new paths. 442 - 443 - #### Database 444 - 445 - As an example, let's say the current database is at `/home/git/knotserver.db`, 446 - and we want to move it to `/home/git/database/knotserver.db`. 447 - 448 - Copy the current database to the new location. Make sure to copy the `.db-shm` 449 - and `.db-wal` files if they exist. 450 - 451 - ``` 452 - mkdir /home/git/database 453 - cp /home/git/knotserver.db* /home/git/database 454 - ``` 455 - 456 - In the environment (e.g. `/home/git/.knot.env`), set `KNOT_SERVER_DB_PATH` to 457 - the new file path (_not_ the directory): 458 - 459 - ``` 460 - KNOT_SERVER_DB_PATH=/home/git/database/knotserver.db 461 - ``` 462 - 463 - #### Repositories 464 - 465 - As an example, let's say the repositories are currently in `/home/git`, and we 466 - want to move them into `/home/git/repositories`. 467 - 468 - Create the new folder, then move the existing repositories (if there are any): 469 - 470 - ``` 471 - mkdir /home/git/repositories 472 - # move all DIDs into the new folder; these will vary for you! 473 - mv /home/git/did:plc:wshs7t2adsemcrrd4snkeqli /home/git/repositories 474 - ``` 475 - 476 - In the environment (e.g. `/home/git/.knot.env`), update `KNOT_REPO_SCAN_PATH` 477 - to the new directory: 478 - 479 - ``` 480 - KNOT_REPO_SCAN_PATH=/home/git/repositories 481 - ``` 482 - 483 - Similarly, update your `sshd` `AuthorizedKeysCommand` to use the updated 484 - repository path: 485 - 486 - ``` 487 - sudo tee /etc/ssh/sshd_config.d/authorized_keys_command.conf <<EOF 488 - Match User git 489 - AuthorizedKeysCommand /usr/local/bin/knot keys -o authorized-keys -git-dir /home/git/repositories 490 - AuthorizedKeysCommandUser nobody 491 - EOF 492 - ``` 493 - 494 - Make sure to restart your SSH server! 495 - 496 - #### MOTD (message of the day) 497 - 498 - To configure the MOTD used ("Welcome to this knot!" by default), edit the 499 - `/home/git/motd` file: 500 - 501 - ``` 502 - printf "Hi from this knot!\n" > /home/git/motd 503 - ``` 504 - 505 - Note that you should add a newline at the end if setting a non-empty message 506 - since the knot won't do this for you. 507 - 508 - # Spindles 509 - 510 - ## Pipelines 511 - 512 - Spindle workflows allow you to write CI/CD pipelines in a 513 - simple format. They're located in the `.tangled/workflows` 514 - directory at the root of your repository, and are defined 515 - using YAML. 516 - 517 - The fields are: 518 - 519 - - [Trigger](#trigger): A **required** field that defines 520 - when a workflow should be triggered. 521 - - [Engine](#engine): A **required** field that defines which 522 - engine a workflow should run on. 523 - - [Clone options](#clone-options): An **optional** field 524 - that defines how the repository should be cloned. 525 - - [Dependencies](#dependencies): An **optional** field that 526 - allows you to list dependencies you may need. 527 - - [Environment](#environment): An **optional** field that 528 - allows you to define environment variables. 529 - - [Steps](#steps): An **optional** field that allows you to 530 - define what steps should run in the workflow. 531 - 532 - ### Trigger 533 - 534 - The first thing to add to a workflow is the trigger, which 535 - defines when a workflow runs. This is defined using a `when` 536 - field, which takes in a list of conditions. Each condition 537 - has the following fields: 538 - 539 - - `event`: This is a **required** field that defines when 540 - your workflow should run. It's a list that can take one or 541 - more of the following values: 542 - - `push`: The workflow should run every time a commit is 543 - pushed to the repository. 544 - - `pull_request`: The workflow should run every time a 545 - pull request is made or updated. 546 - - `manual`: The workflow can be triggered manually. 547 - - `branch`: Defines which branches the workflow should run 548 - for. If used with the `push` event, commits to the 549 - branch(es) listed here will trigger the workflow. If used 550 - with the `pull_request` event, updates to pull requests 551 - targeting the branch(es) listed here will trigger the 552 - workflow. This field has no effect with the `manual` 553 - event. Supports glob patterns using `*` and `**` (e.g., 554 - `main`, `develop`, `release-*`). Either `branch` or `tag` 555 - (or both) must be specified for `push` events. 556 - - `tag`: Defines which tags the workflow should run for. 557 - Only used with the `push` event - when tags matching the 558 - pattern(s) listed here are pushed, the workflow will 559 - trigger. This field has no effect with `pull_request` or 560 - `manual` events. Supports glob patterns using `*` and `**` 561 - (e.g., `v*`, `v1.*`, `release-**`). Either `branch` or 562 - `tag` (or both) must be specified for `push` events. 563 - 564 - For example, if you'd like to define a workflow that runs 565 - when commits are pushed to the `main` and `develop` 566 - branches, or when pull requests that target the `main` 567 - branch are updated, or manually, you can do so with: 568 - 569 - ```yaml 570 - when: 571 - - event: ["push", "manual"] 572 - branch: ["main", "develop"] 573 - - event: ["pull_request"] 574 - branch: ["main"] 575 - ``` 576 - 577 - You can also trigger workflows on tag pushes. For instance, 578 - to run a deployment workflow when tags matching `v*` are 579 - pushed: 580 - 581 - ```yaml 582 - when: 583 - - event: ["push"] 584 - tag: ["v*"] 585 - ``` 586 - 587 - You can even combine branch and tag patterns in a single 588 - constraint (the workflow triggers if either matches): 589 - 590 - ```yaml 591 - when: 592 - - event: ["push"] 593 - branch: ["main", "release-*"] 594 - tag: ["v*", "stable"] 595 - ``` 596 - 597 - ### Engine 598 - 599 - Next is the engine on which the workflow should run, defined 600 - using the **required** `engine` field. The currently 601 - supported engines are: 602 - 603 - - `nixery`: This uses an instance of 604 - [Nixery](https://nixery.dev) to run steps, which allows 605 - you to add [dependencies](#dependencies) from 606 - [Nixpkgs](https://github.com/NixOS/nixpkgs). You can 607 - search for packages on https://search.nixos.org, and 608 - there's a pretty good chance the package(s) you're looking 609 - for will be there. 610 - 611 - Example: 612 - 613 - ```yaml 614 - engine: "nixery" 615 - ``` 616 - 617 - ### Clone options 618 - 619 - When a workflow starts, the first step is to clone the 620 - repository. You can customize this behavior using the 621 - **optional** `clone` field. It has the following fields: 622 - 623 - - `skip`: Setting this to `true` will skip cloning the 624 - repository. This can be useful if your workflow is doing 625 - something that doesn't require anything from the 626 - repository itself. This is `false` by default. 627 - - `depth`: This sets the number of commits, or the "clone 628 - depth", to fetch from the repository. For example, if you 629 - set this to 2, the last 2 commits will be fetched. By 630 - default, the depth is set to 1, meaning only the most 631 - recent commit will be fetched, which is the commit that 632 - triggered the workflow. 633 - - `submodules`: If you use [git 634 - submodules](https://git-scm.com/book/en/v2/Git-Tools-Submodules) 635 - in your repository, setting this field to `true` will 636 - recursively fetch all submodules. This is `false` by 637 - default. 638 - 639 - The default settings are: 640 - 641 - ```yaml 642 - clone: 643 - skip: false 644 - depth: 1 645 - submodules: false 646 - ``` 647 - 648 - ### Dependencies 649 - 650 - Usually when you're running a workflow, you'll need 651 - additional dependencies. The `dependencies` field lets you 652 - define which dependencies to get, and from where. It's a 653 - key-value map, with the key being the registry to fetch 654 - dependencies from, and the value being the list of 655 - dependencies to fetch. 656 - 657 - Say you want to fetch Node.js and Go from `nixpkgs`, and a 658 - package called `my_pkg` you've made from your own registry 659 - at your repository at 660 - `https://tangled.sh/@example.com/my_pkg`. You can define 661 - those dependencies like so: 662 - 663 - ```yaml 664 - dependencies: 665 - # nixpkgs 666 - nixpkgs: 667 - - nodejs 668 - - go 669 - # custom registry 670 - git+https://tangled.org/@example.com/my_pkg: 671 - - my_pkg 672 - ``` 673 - 674 - Now these dependencies are available to use in your 675 - workflow! 676 - 677 - ### Environment 678 - 679 - The `environment` field allows you define environment 680 - variables that will be available throughout the entire 681 - workflow. **Do not put secrets here, these environment 682 - variables are visible to anyone viewing the repository. You 683 - can add secrets for pipelines in your repository's 684 - settings.** 685 - 686 - Example: 687 - 688 - ```yaml 689 - environment: 690 - GOOS: "linux" 691 - GOARCH: "arm64" 692 - NODE_ENV: "production" 693 - MY_ENV_VAR: "MY_ENV_VALUE" 694 - ``` 695 - 696 - ### Steps 697 - 698 - The `steps` field allows you to define what steps should run 699 - in the workflow. It's a list of step objects, each with the 700 - following fields: 701 - 702 - - `name`: This field allows you to give your step a name. 703 - This name is visible in your workflow runs, and is used to 704 - describe what the step is doing. 705 - - `command`: This field allows you to define a command to 706 - run in that step. The step is run in a Bash shell, and the 707 - logs from the command will be visible in the pipelines 708 - page on the Tangled website. The 709 - [dependencies](#dependencies) you added will be available 710 - to use here. 711 - - `environment`: Similar to the global 712 - [environment](#environment) config, this **optional** 713 - field is a key-value map that allows you to set 714 - environment variables for the step. **Do not put secrets 715 - here, these environment variables are visible to anyone 716 - viewing the repository. You can add secrets for pipelines 717 - in your repository's settings.** 718 - 719 - Example: 720 - 721 - ```yaml 722 - steps: 723 - - name: "Build backend" 724 - command: "go build" 725 - environment: 726 - GOOS: "darwin" 727 - GOARCH: "arm64" 728 - - name: "Build frontend" 729 - command: "npm run build" 730 - environment: 731 - NODE_ENV: "production" 732 - ``` 733 - 734 - ### Complete workflow 735 - 736 - ```yaml 737 - # .tangled/workflows/build.yml 738 - 739 - when: 740 - - event: ["push", "manual"] 741 - branch: ["main", "develop"] 742 - - event: ["pull_request"] 743 - branch: ["main"] 744 - 745 - engine: "nixery" 746 - 747 - # using the default values 748 - clone: 749 - skip: false 750 - depth: 1 751 - submodules: false 752 - 753 - dependencies: 754 - # nixpkgs 755 - nixpkgs: 756 - - nodejs 757 - - go 758 - # custom registry 759 - git+https://tangled.org/@example.com/my_pkg: 760 - - my_pkg 761 - 762 - environment: 763 - GOOS: "linux" 764 - GOARCH: "arm64" 765 - NODE_ENV: "production" 766 - MY_ENV_VAR: "MY_ENV_VALUE" 767 - 768 - steps: 769 - - name: "Build backend" 770 - command: "go build" 771 - environment: 772 - GOOS: "darwin" 773 - GOARCH: "arm64" 774 - - name: "Build frontend" 775 - command: "npm run build" 776 - environment: 777 - NODE_ENV: "production" 778 - ``` 779 - 780 - If you want another example of a workflow, you can look at 781 - the one [Tangled uses to build the 782 - project](https://tangled.sh/@tangled.sh/core/blob/master/.tangled/workflows/build.yml). 783 - 784 - ## Self-hosting guide 785 - 786 - ### Prerequisites 787 - 788 - * Go 789 - * Docker (the only supported backend currently) 790 - 791 - ### Configuration 792 - 793 - Spindle is configured using environment variables. The following environment variables are available: 794 - 795 - * `SPINDLE_SERVER_LISTEN_ADDR`: The address the server listens on (default: `"0.0.0.0:6555"`). 796 - * `SPINDLE_SERVER_DB_PATH`: The path to the SQLite database file (default: `"spindle.db"`). 797 - * `SPINDLE_SERVER_HOSTNAME`: The hostname of the server (required). 798 - * `SPINDLE_SERVER_JETSTREAM_ENDPOINT`: The endpoint of the Jetstream server (default: `"wss://jetstream1.us-west.bsky.network/subscribe"`). 799 - * `SPINDLE_SERVER_DEV`: A boolean indicating whether the server is running in development mode (default: `false`). 800 - * `SPINDLE_SERVER_OWNER`: The DID of the owner (required). 801 - * `SPINDLE_PIPELINES_NIXERY`: The Nixery URL (default: `"nixery.tangled.sh"`). 802 - * `SPINDLE_PIPELINES_WORKFLOW_TIMEOUT`: The default workflow timeout (default: `"5m"`). 803 - * `SPINDLE_PIPELINES_LOG_DIR`: The directory to store workflow logs (default: `"/var/log/spindle"`). 804 - 805 - ### Running spindle 806 - 807 - 1. **Set the environment variables.** For example: 808 - 809 - ```shell 810 - export SPINDLE_SERVER_HOSTNAME="your-hostname" 811 - export SPINDLE_SERVER_OWNER="your-did" 812 - ``` 813 - 814 - 2. **Build the Spindle binary.** 815 - 816 - ```shell 817 - cd core 818 - go mod download 819 - go build -o cmd/spindle/spindle cmd/spindle/main.go 820 - ``` 821 - 822 - 3. **Create the log directory.** 823 - 824 - ```shell 825 - sudo mkdir -p /var/log/spindle 826 - sudo chown $USER:$USER -R /var/log/spindle 827 - ``` 828 - 829 - 4. **Run the Spindle binary.** 830 - 831 - ```shell 832 - ./cmd/spindle/spindle 833 - ``` 834 - 835 - Spindle will now start, connect to the Jetstream server, and begin processing pipelines. 836 - 837 - ## Architecture 838 - 839 - Spindle is a small CI runner service. Here's a high level overview of how it operates: 840 - 841 - * listens for [`sh.tangled.spindle.member`](/lexicons/spindle/member.json) and 842 - [`sh.tangled.repo`](/lexicons/repo.json) records on the Jetstream. 843 - * when a new repo record comes through (typically when you add a spindle to a 844 - repo from the settings), spindle then resolves the underlying knot and 845 - subscribes to repo events (see: 846 - [`sh.tangled.pipeline`](/lexicons/pipeline.json)). 847 - * the spindle engine then handles execution of the pipeline, with results and 848 - logs beamed on the spindle event stream over wss 849 - 850 - ### The engine 851 - 852 - At present, the only supported backend is Docker (and Podman, if Docker 853 - compatibility is enabled, so that `/run/docker.sock` is created). Spindle 854 - executes each step in the pipeline in a fresh container, with state persisted 855 - across steps within the `/tangled/workspace` directory. 856 - 857 - The base image for the container is constructed on the fly using 858 - [Nixery](https://nixery.dev), which is handy for caching layers for frequently 859 - used packages. 860 - 861 - The pipeline manifest is [specified here](/docs/spindle/pipeline.md). 862 - 863 - ## Secrets with openbao 864 - 865 - This document covers setting up Spindle to use OpenBao for secrets 866 - management via OpenBao Proxy instead of the default SQLite backend. 867 - 868 - ### Overview 869 - 870 - Spindle now uses OpenBao Proxy for secrets management. The proxy handles 871 - authentication automatically using AppRole credentials, while Spindle 872 - connects to the local proxy instead of directly to the OpenBao server. 873 - 874 - This approach provides better security, automatic token renewal, and 875 - simplified application code. 876 - 877 - ### Installation 878 - 879 - Install OpenBao from nixpkgs: 880 - 881 - ```bash 882 - nix shell nixpkgs#openbao # for a local server 883 - ``` 884 - 885 - ### Setup 886 - 887 - The setup process can is documented for both local development and production. 888 - 889 - #### Local development 890 - 891 - Start OpenBao in dev mode: 892 - 893 - ```bash 894 - bao server -dev -dev-root-token-id="root" -dev-listen-address=127.0.0.1:8201 895 - ``` 896 - 897 - This starts OpenBao on `http://localhost:8201` with a root token. 898 - 899 - Set up environment for bao CLI: 900 - 901 - ```bash 902 - export BAO_ADDR=http://localhost:8200 903 - export BAO_TOKEN=root 904 - ``` 905 - 906 - #### Production 907 - 908 - You would typically use a systemd service with a 909 - configuration file. Refer to 910 - [@tangled.org/infra](https://tangled.org/@tangled.org/infra) 911 - for how this can be achieved using Nix. 912 - 913 - Then, initialize the bao server: 914 - 915 - ```bash 916 - bao operator init -key-shares=1 -key-threshold=1 917 - ``` 918 - 919 - This will print out an unseal key and a root key. Save them 920 - somewhere (like a password manager). Then unseal the vault 921 - to begin setting it up: 922 - 923 - ```bash 924 - bao operator unseal <unseal_key> 925 - ``` 926 - 927 - All steps below remain the same across both dev and 928 - production setups. 929 - 930 - #### Configure openbao server 931 - 932 - Create the spindle KV mount: 933 - 934 - ```bash 935 - bao secrets enable -path=spindle -version=2 kv 936 - ``` 937 - 938 - Set up AppRole authentication and policy: 939 - 940 - Create a policy file `spindle-policy.hcl`: 941 - 942 - ```hcl 943 - # Full access to spindle KV v2 data 944 - path "spindle/data/*" { 945 - capabilities = ["create", "read", "update", "delete"] 946 - } 947 - 948 - # Access to metadata for listing and management 949 - path "spindle/metadata/*" { 950 - capabilities = ["list", "read", "delete", "update"] 951 - } 952 - 953 - # Allow listing at root level 954 - path "spindle/" { 955 - capabilities = ["list"] 956 - } 957 - 958 - # Required for connection testing and health checks 959 - path "auth/token/lookup-self" { 960 - capabilities = ["read"] 961 - } 962 - ``` 963 - 964 - Apply the policy and create an AppRole: 965 - 966 - ```bash 967 - bao policy write spindle-policy spindle-policy.hcl 968 - bao auth enable approle 969 - bao write auth/approle/role/spindle \ 970 - token_policies="spindle-policy" \ 971 - token_ttl=1h \ 972 - token_max_ttl=4h \ 973 - bind_secret_id=true \ 974 - secret_id_ttl=0 \ 975 - secret_id_num_uses=0 976 - ``` 977 - 978 - Get the credentials: 979 - 980 - ```bash 981 - # Get role ID (static) 982 - ROLE_ID=$(bao read -field=role_id auth/approle/role/spindle/role-id) 983 - 984 - # Generate secret ID 985 - SECRET_ID=$(bao write -f -field=secret_id auth/approle/role/spindle/secret-id) 986 - 987 - echo "Role ID: $ROLE_ID" 988 - echo "Secret ID: $SECRET_ID" 989 - ``` 990 - 991 - #### Create proxy configuration 992 - 993 - Create the credential files: 994 - 995 - ```bash 996 - # Create directory for OpenBao files 997 - mkdir -p /tmp/openbao 998 - 999 - # Save credentials 1000 - echo "$ROLE_ID" > /tmp/openbao/role-id 1001 - echo "$SECRET_ID" > /tmp/openbao/secret-id 1002 - chmod 600 /tmp/openbao/role-id /tmp/openbao/secret-id 1003 - ``` 1004 - 1005 - Create a proxy configuration file `/tmp/openbao/proxy.hcl`: 1006 - 1007 - ```hcl 1008 - # OpenBao server connection 1009 - vault { 1010 - address = "http://localhost:8200" 1011 - } 1012 - 1013 - # Auto-Auth using AppRole 1014 - auto_auth { 1015 - method "approle" { 1016 - mount_path = "auth/approle" 1017 - config = { 1018 - role_id_file_path = "/tmp/openbao/role-id" 1019 - secret_id_file_path = "/tmp/openbao/secret-id" 1020 - } 1021 - } 1022 - 1023 - # Optional: write token to file for debugging 1024 - sink "file" { 1025 - config = { 1026 - path = "/tmp/openbao/token" 1027 - mode = 0640 1028 - } 1029 - } 1030 - } 1031 - 1032 - # Proxy listener for Spindle 1033 - listener "tcp" { 1034 - address = "127.0.0.1:8201" 1035 - tls_disable = true 1036 - } 1037 - 1038 - # Enable API proxy with auto-auth token 1039 - api_proxy { 1040 - use_auto_auth_token = true 1041 - } 1042 - 1043 - # Enable response caching 1044 - cache { 1045 - use_auto_auth_token = true 1046 - } 1047 - 1048 - # Logging 1049 - log_level = "info" 1050 - ``` 1051 - 1052 - #### Start the proxy 1053 - 1054 - Start OpenBao Proxy: 1055 - 1056 - ```bash 1057 - bao proxy -config=/tmp/openbao/proxy.hcl 1058 - ``` 1059 - 1060 - The proxy will authenticate with OpenBao and start listening on 1061 - `127.0.0.1:8201`. 1062 - 1063 - #### Configure spindle 1064 - 1065 - Set these environment variables for Spindle: 1066 - 1067 - ```bash 1068 - export SPINDLE_SERVER_SECRETS_PROVIDER=openbao 1069 - export SPINDLE_SERVER_SECRETS_OPENBAO_PROXY_ADDR=http://127.0.0.1:8201 1070 - export SPINDLE_SERVER_SECRETS_OPENBAO_MOUNT=spindle 1071 - ``` 1072 - 1073 - On startup, the spindle will now connect to the local proxy, 1074 - which handles all authentication automatically. 1075 - 1076 - ### Production setup for proxy 1077 - 1078 - For production, you'll want to run the proxy as a service: 1079 - 1080 - Place your production configuration in 1081 - `/etc/openbao/proxy.hcl` with proper TLS settings for the 1082 - vault connection. 1083 - 1084 - ### Verifying setup 1085 - 1086 - Test the proxy directly: 1087 - 1088 - ```bash 1089 - # Check proxy health 1090 - curl -H "X-Vault-Request: true" http://127.0.0.1:8201/v1/sys/health 1091 - 1092 - # Test token lookup through proxy 1093 - curl -H "X-Vault-Request: true" http://127.0.0.1:8201/v1/auth/token/lookup-self 1094 - ``` 1095 - 1096 - Test OpenBao operations through the server: 1097 - 1098 - ```bash 1099 - # List all secrets 1100 - bao kv list spindle/ 1101 - 1102 - # Add a test secret via Spindle API, then check it exists 1103 - bao kv list spindle/repos/ 1104 - 1105 - # Get a specific secret 1106 - bao kv get spindle/repos/your_repo_path/SECRET_NAME 1107 - ``` 1108 - 1109 - ### How it works 1110 - 1111 - - Spindle connects to OpenBao Proxy on localhost (typically 1112 - port 8200 or 8201) 1113 - - The proxy authenticates with OpenBao using AppRole 1114 - credentials 1115 - - All Spindle requests go through the proxy, which injects 1116 - authentication tokens 1117 - - Secrets are stored at 1118 - `spindle/repos/{sanitized_repo_path}/{secret_key}` 1119 - - Repository paths like `did:plc:alice/myrepo` become 1120 - `did_plc_alice_myrepo` 1121 - - The proxy handles all token renewal automatically 1122 - - Spindle no longer manages tokens or authentication 1123 - directly 1124 - 1125 - ### Troubleshooting 1126 - 1127 - **Connection refused**: Check that the OpenBao Proxy is 1128 - running and listening on the configured address. 1129 - 1130 - **403 errors**: Verify the AppRole credentials are correct 1131 - and the policy has the necessary permissions. 1132 - 1133 - **404 route errors**: The spindle KV mount probably doesn't 1134 - exist - run the mount creation step again. 1135 - 1136 - **Proxy authentication failures**: Check the proxy logs and 1137 - verify the role-id and secret-id files are readable and 1138 - contain valid credentials. 1139 - 1140 - **Secret not found after writing**: This can indicate policy 1141 - permission issues. Verify the policy includes both 1142 - `spindle/data/*` and `spindle/metadata/*` paths with 1143 - appropriate capabilities. 1144 - 1145 - Check proxy logs: 1146 - 1147 - ```bash 1148 - # If running as systemd service 1149 - journalctl -u openbao-proxy -f 1150 - 1151 - # If running directly, check the console output 1152 - ``` 1153 - 1154 - Test AppRole authentication manually: 1155 - 1156 - ```bash 1157 - bao write auth/approle/login \ 1158 - role_id="$(cat /tmp/openbao/role-id)" \ 1159 - secret_id="$(cat /tmp/openbao/secret-id)" 1160 - ``` 1161 - 1162 - # Migrating knots & spindles 1163 - 1164 - Sometimes, non-backwards compatible changes are made to the 1165 - knot/spindle XRPC APIs. If you host a knot or a spindle, you 1166 - will need to follow this guide to upgrade. Typically, this 1167 - only requires you to deploy the newest version. 1168 - 1169 - This document is laid out in reverse-chronological order. 1170 - Newer migration guides are listed first, and older guides 1171 - are further down the page. 1172 - 1173 - ## Upgrading from v1.8.x 1174 - 1175 - After v1.8.2, the HTTP API for knot and spindles have been 1176 - deprecated and replaced with XRPC. Repositories on outdated 1177 - knots will not be viewable from the appview. Upgrading is 1178 - straightforward however. 1179 - 1180 - For knots: 1181 - 1182 - - Upgrade to latest tag (v1.9.0 or above) 1183 - - Head to the [knot dashboard](https://tangled.org/settings/knots) and 1184 - hit the "retry" button to verify your knot 1185 - 1186 - For spindles: 1187 - 1188 - - Upgrade to latest tag (v1.9.0 or above) 1189 - - Head to the [spindle 1190 - dashboard](https://tangled.org/settings/spindles) and hit the 1191 - "retry" button to verify your spindle 1192 - 1193 - ## Upgrading from v1.7.x 1194 - 1195 - After v1.7.0, knot secrets have been deprecated. You no 1196 - longer need a secret from the appview to run a knot. All 1197 - authorized commands to knots are managed via [Inter-Service 1198 - Authentication](https://atproto.com/specs/xrpc#inter-service-authentication-jwt). 1199 - Knots will be read-only until upgraded. 1200 - 1201 - Upgrading is quite easy, in essence: 1202 - 1203 - - `KNOT_SERVER_SECRET` is no more, you can remove this 1204 - environment variable entirely 1205 - - `KNOT_SERVER_OWNER` is now required on boot, set this to 1206 - your DID. You can find your DID in the 1207 - [settings](https://tangled.org/settings) page. 1208 - - Restart your knot once you have replaced the environment 1209 - variable 1210 - - Head to the [knot dashboard](https://tangled.org/settings/knots) and 1211 - hit the "retry" button to verify your knot. This simply 1212 - writes a `sh.tangled.knot` record to your PDS. 1213 - 1214 - If you use the nix module, simply bump the flake to the 1215 - latest revision, and change your config block like so: 1216 - 1217 - ```diff 1218 - services.tangled.knot = { 1219 - enable = true; 1220 - server = { 1221 - - secretFile = /path/to/secret; 1222 - + owner = "did:plc:foo"; 1223 - }; 1224 - }; 1225 - ``` 1226 - 1227 - # Hacking on Tangled 1228 - 1229 - We highly recommend [installing 1230 - nix](https://nixos.org/download/) (the package manager) 1231 - before working on the codebase. The nix flake provides a lot 1232 - of helpers to get started and most importantly, builds and 1233 - dev shells are entirely deterministic. 1234 - 1235 - To set up your dev environment: 1236 - 1237 - ```bash 1238 - nix develop 1239 - ``` 1240 - 1241 - Non-nix users can look at the `devShell` attribute in the 1242 - `flake.nix` file to determine necessary dependencies. 1243 - 1244 - ## Running the appview 1245 - 1246 - The nix flake also exposes a few `app` attributes (run `nix 1247 - flake show` to see a full list of what the flake provides), 1248 - one of the apps runs the appview with the `air` 1249 - live-reloader: 1250 - 1251 - ```bash 1252 - TANGLED_DEV=true nix run .#watch-appview 1253 - 1254 - # TANGLED_DB_PATH might be of interest to point to 1255 - # different sqlite DBs 1256 - 1257 - # in a separate shell, you can live-reload tailwind 1258 - nix run .#watch-tailwind 1259 - ``` 1260 - 1261 - To authenticate with the appview, you will need redis and 1262 - OAUTH JWKs to be setup: 1263 - 1264 - ``` 1265 - # oauth jwks should already be setup by the nix devshell: 1266 - echo $TANGLED_OAUTH_CLIENT_SECRET 1267 - z42ty4RT1ovnTopY8B8ekz9NuziF2CuMkZ7rbRFpAR9jBqMc 1268 - 1269 - echo $TANGLED_OAUTH_CLIENT_KID 1270 - 1761667908 1271 - 1272 - # if not, you can set it up yourself: 1273 - goat key generate -t P-256 1274 - Key Type: P-256 / secp256r1 / ES256 private key 1275 - Secret Key (Multibase Syntax): save this securely (eg, add to password manager) 1276 - z42tuPDKRfM2mz2Kv953ARen2jmrPA8S9LX9tRq4RVcUMwwL 1277 - Public Key (DID Key Syntax): share or publish this (eg, in DID document) 1278 - did:key:zDnaeUBxtG6Xuv3ATJE4GaWeyXM3jyamJsZw3bSPpxx4bNXDR 1279 - 1280 - # the secret key from above 1281 - export TANGLED_OAUTH_CLIENT_SECRET="z42tuP..." 1282 - 1283 - # run redis in at a new shell to store oauth sessions 1284 - redis-server 1285 - ``` 1286 - 1287 - ## Running knots and spindles 1288 - 1289 - An end-to-end knot setup requires setting up a machine with 1290 - `sshd`, `AuthorizedKeysCommand`, and git user, which is 1291 - quite cumbersome. So the nix flake provides a 1292 - `nixosConfiguration` to do so. 1293 - 1294 - <details> 1295 - <summary><strong>MacOS users will have to setup a Nix Builder first</strong></summary> 1296 - 1297 - In order to build Tangled's dev VM on macOS, you will 1298 - first need to set up a Linux Nix builder. The recommended 1299 - way to do so is to run a [`darwin.linux-builder` 1300 - VM](https://nixos.org/manual/nixpkgs/unstable/#sec-darwin-builder) 1301 - and to register it in `nix.conf` as a builder for Linux 1302 - with the same architecture as your Mac (`linux-aarch64` if 1303 - you are using Apple Silicon). 1304 - 1305 - > IMPORTANT: You must build `darwin.linux-builder` somewhere other than inside 1306 - > the tangled repo so that it doesn't conflict with the other VM. For example, 1307 - > you can do 1308 - > 1309 - > ```shell 1310 - > cd $(mktemp -d buildervm.XXXXX) && nix run nixpkgs#darwin.linux-builder 1311 - > ``` 1312 - > 1313 - > to store the builder VM in a temporary dir. 1314 - > 1315 - > You should read and follow [all the other intructions][darwin builder vm] to 1316 - > avoid subtle problems. 1317 - 1318 - Alternatively, you can use any other method to set up a 1319 - Linux machine with `nix` installed that you can `sudo ssh` 1320 - into (in other words, root user on your Mac has to be able 1321 - to ssh into the Linux machine without entering a password) 1322 - and that has the same architecture as your Mac. See 1323 - [remote builder 1324 - instructions](https://nix.dev/manual/nix/2.28/advanced-topics/distributed-builds.html#requirements) 1325 - for how to register such a builder in `nix.conf`. 1326 - 1327 - > WARNING: If you'd like to use 1328 - > [`nixos-lima`](https://github.com/nixos-lima/nixos-lima) or 1329 - > [Orbstack](https://orbstack.dev/), note that setting them up so that `sudo 1330 - > ssh` works can be tricky. It seems to be [possible with 1331 - > Orbstack](https://github.com/orgs/orbstack/discussions/1669). 1332 - 1333 - </details> 1334 - 1335 - To begin, grab your DID from http://localhost:3000/settings. 1336 - Then, set `TANGLED_VM_KNOT_OWNER` and 1337 - `TANGLED_VM_SPINDLE_OWNER` to your DID. You can now start a 1338 - lightweight NixOS VM like so: 1339 - 1340 - ```bash 1341 - nix run --impure .#vm 1342 - 1343 - # type `poweroff` at the shell to exit the VM 1344 - ``` 1345 - 1346 - This starts a knot on port 6444, a spindle on port 6555 1347 - with `ssh` exposed on port 2222. 1348 - 1349 - Once the services are running, head to 1350 - http://localhost:3000/settings/knots and hit verify. It should 1351 - verify the ownership of the services instantly if everything 1352 - went smoothly. 1353 - 1354 - You can push repositories to this VM with this ssh config 1355 - block on your main machine: 1356 - 1357 - ```bash 1358 - Host nixos-shell 1359 - Hostname localhost 1360 - Port 2222 1361 - User git 1362 - IdentityFile ~/.ssh/my_tangled_key 1363 - ``` 1364 - 1365 - Set up a remote called `local-dev` on a git repo: 1366 - 1367 - ```bash 1368 - git remote add local-dev git@nixos-shell:user/repo 1369 - git push local-dev main 1370 - ``` 1371 - 1372 - The above VM should already be running a spindle on 1373 - `localhost:6555`. Head to http://localhost:3000/settings/spindles and 1374 - hit verify. You can then configure each repository to use 1375 - this spindle and run CI jobs. 1376 - 1377 - Of interest when debugging spindles: 1378 - 1379 - ``` 1380 - # service logs from journald: 1381 - journalctl -xeu spindle 1382 - 1383 - # CI job logs from disk: 1384 - ls /var/log/spindle 1385 - 1386 - # debugging spindle db: 1387 - sqlite3 /var/lib/spindle/spindle.db 1388 - 1389 - # litecli has a nicer REPL interface: 1390 - litecli /var/lib/spindle/spindle.db 1391 - ``` 1392 - 1393 - If for any reason you wish to disable either one of the 1394 - services in the VM, modify [nix/vm.nix](/nix/vm.nix) and set 1395 - `services.tangled.spindle.enable` (or 1396 - `services.tangled.knot.enable`) to `false`. 1397 - 1398 - # Contribution guide 1399 - 1400 - ## Commit guidelines 1401 - 1402 - We follow a commit style similar to the Go project. Please keep commits: 1403 - 1404 - * **atomic**: each commit should represent one logical change 1405 - * **descriptive**: the commit message should clearly describe what the 1406 - change does and why it's needed 1407 - 1408 - ### Message format 1409 - 1410 - ``` 1411 - <service/top-level directory>/<affected package/directory>: <short summary of change> 1412 - 1413 - Optional longer description can go here, if necessary. Explain what the 1414 - change does and why, especially if not obvious. Reference relevant 1415 - issues or PRs when applicable. These can be links for now since we don't 1416 - auto-link issues/PRs yet. 1417 - ``` 1418 - 1419 - Here are some examples: 1420 - 1421 - ``` 1422 - appview/state: fix token expiry check in middleware 1423 - 1424 - The previous check did not account for clock drift, leading to premature 1425 - token invalidation. 1426 - ``` 1427 - 1428 - ``` 1429 - knotserver/git/service: improve error checking in upload-pack 1430 - ``` 1431 - 1432 - 1433 - ### General notes 1434 - 1435 - - PRs get merged "as-is" (fast-forward) -- like applying a patch-series 1436 - using `git am`. At present, there is no squashing -- so please author 1437 - your commits as they would appear on `master`, following the above 1438 - guidelines. 1439 - - If there is a lot of nesting, for example "appview: 1440 - pages/templates/repo/fragments: ...", these can be truncated down to 1441 - just "appview: repo/fragments: ...". If the change affects a lot of 1442 - subdirectories, you may abbreviate to just the top-level names, e.g. 1443 - "appview: ..." or "knotserver: ...". 1444 - - Keep commits lowercased with no trailing period. 1445 - - Use the imperative mood in the summary line (e.g., "fix bug" not 1446 - "fixed bug" or "fixes bug"). 1447 - - Try to keep the summary line under 72 characters, but we aren't too 1448 - fussed about this. 1449 - - Follow the same formatting for PR titles if filled manually. 1450 - - Don't include unrelated changes in the same commit. 1451 - - Avoid noisy commit messages like "wip" or "final fix"โ€”rewrite history 1452 - before submitting if necessary. 1453 - 1454 - ## Code formatting 1455 - 1456 - We use a variety of tools to format our code, and multiplex them with 1457 - [`treefmt`](https://treefmt.com): all you need to do to format your changes 1458 - is run `nix run .#fmt` (or just `treefmt` if you're in the devshell). 1459 - 1460 - ## Proposals for bigger changes 1461 - 1462 - Small fixes like typos, minor bugs, or trivial refactors can be 1463 - submitted directly as PRs. 1464 - 1465 - For larger changesโ€”especially those introducing new features, significant 1466 - refactoring, or altering system behaviorโ€”please open a proposal first. This 1467 - helps us evaluate the scope, design, and potential impact before implementation. 1468 - 1469 - Create a new issue titled: 1470 - 1471 - ``` 1472 - proposal: <affected scope>: <summary of change> 1473 - ``` 1474 - 1475 - In the description, explain: 1476 - 1477 - - What the change is 1478 - - Why it's needed 1479 - - How you plan to implement it (roughly) 1480 - - Any open questions or tradeoffs 1481 - 1482 - We'll use the issue thread to discuss and refine the idea before moving 1483 - forward. 1484 - 1485 - ## Developer certificate of origin (DCO) 1486 - 1487 - We require all contributors to certify that they have the right to 1488 - submit the code they're contributing. To do this, we follow the 1489 - [Developer Certificate of Origin 1490 - (DCO)](https://developercertificate.org/). 1491 - 1492 - By signing your commits, you're stating that the contribution is your 1493 - own work, or that you have the right to submit it under the project's 1494 - license. This helps us keep things clean and legally sound. 1495 - 1496 - To sign your commit, just add the `-s` flag when committing: 1497 - 1498 - ```sh 1499 - git commit -s -m "your commit message" 1500 - ``` 1501 - 1502 - This appends a line like: 1503 - 1504 - ``` 1505 - Signed-off-by: Your Name <your.email@example.com> 1506 - ``` 1507 - 1508 - We won't merge commits if they aren't signed off. If you forget, you can 1509 - amend the last commit like this: 1510 - 1511 - ```sh 1512 - git commit --amend -s 1513 - ``` 1514 - 1515 - If you're submitting a PR with multiple commits, make sure each one is 1516 - signed. 1517 - 1518 - For [jj](https://jj-vcs.github.io/jj/latest/) users, you can run the following command 1519 - to make it sign off commits in the tangled repo: 1520 - 1521 - ```shell 1522 - # Safety check, should say "No matching config key..." 1523 - jj config list templates.commit_trailers 1524 - # The command below may need to be adjusted if the command above returned something. 1525 - jj config set --repo templates.commit_trailers "format_signed_off_by_trailer(self)" 1526 - ``` 1527 - 1528 - Refer to the [jujutsu 1529 - documentation](https://jj-vcs.github.io/jj/latest/config/#commit-trailers) 1530 - for more information.
+136
docs/contributing.md
··· 1 + # tangled contributing guide 2 + 3 + ## commit guidelines 4 + 5 + We follow a commit style similar to the Go project. Please keep commits: 6 + 7 + * **atomic**: each commit should represent one logical change 8 + * **descriptive**: the commit message should clearly describe what the 9 + change does and why it's needed 10 + 11 + ### message format 12 + 13 + ``` 14 + <service/top-level directory>/<affected package/directory>: <short summary of change> 15 + 16 + 17 + Optional longer description can go here, if necessary. Explain what the 18 + change does and why, especially if not obvious. Reference relevant 19 + issues or PRs when applicable. These can be links for now since we don't 20 + auto-link issues/PRs yet. 21 + ``` 22 + 23 + Here are some examples: 24 + 25 + ``` 26 + appview/state: fix token expiry check in middleware 27 + 28 + The previous check did not account for clock drift, leading to premature 29 + token invalidation. 30 + ``` 31 + 32 + ``` 33 + knotserver/git/service: improve error checking in upload-pack 34 + ``` 35 + 36 + 37 + ### general notes 38 + 39 + - PRs get merged "as-is" (fast-forward) -- like applying a patch-series 40 + using `git am`. At present, there is no squashing -- so please author 41 + your commits as they would appear on `master`, following the above 42 + guidelines. 43 + - If there is a lot of nesting, for example "appview: 44 + pages/templates/repo/fragments: ...", these can be truncated down to 45 + just "appview: repo/fragments: ...". If the change affects a lot of 46 + subdirectories, you may abbreviate to just the top-level names, e.g. 47 + "appview: ..." or "knotserver: ...". 48 + - Keep commits lowercased with no trailing period. 49 + - Use the imperative mood in the summary line (e.g., "fix bug" not 50 + "fixed bug" or "fixes bug"). 51 + - Try to keep the summary line under 72 characters, but we aren't too 52 + fussed about this. 53 + - Follow the same formatting for PR titles if filled manually. 54 + - Don't include unrelated changes in the same commit. 55 + - Avoid noisy commit messages like "wip" or "final fix"โ€”rewrite history 56 + before submitting if necessary. 57 + 58 + ## code formatting 59 + 60 + We use a variety of tools to format our code, and multiplex them with 61 + [`treefmt`](https://treefmt.com): all you need to do to format your changes 62 + is run `nix run .#fmt` (or just `treefmt` if you're in the devshell). 63 + 64 + ## proposals for bigger changes 65 + 66 + Small fixes like typos, minor bugs, or trivial refactors can be 67 + submitted directly as PRs. 68 + 69 + For larger changesโ€”especially those introducing new features, significant 70 + refactoring, or altering system behaviorโ€”please open a proposal first. This 71 + helps us evaluate the scope, design, and potential impact before implementation. 72 + 73 + ### proposal format 74 + 75 + Create a new issue titled: 76 + 77 + ``` 78 + proposal: <affected scope>: <summary of change> 79 + ``` 80 + 81 + In the description, explain: 82 + 83 + - What the change is 84 + - Why it's needed 85 + - How you plan to implement it (roughly) 86 + - Any open questions or tradeoffs 87 + 88 + We'll use the issue thread to discuss and refine the idea before moving 89 + forward. 90 + 91 + ## developer certificate of origin (DCO) 92 + 93 + We require all contributors to certify that they have the right to 94 + submit the code they're contributing. To do this, we follow the 95 + [Developer Certificate of Origin 96 + (DCO)](https://developercertificate.org/). 97 + 98 + By signing your commits, you're stating that the contribution is your 99 + own work, or that you have the right to submit it under the project's 100 + license. This helps us keep things clean and legally sound. 101 + 102 + To sign your commit, just add the `-s` flag when committing: 103 + 104 + ```sh 105 + git commit -s -m "your commit message" 106 + ``` 107 + 108 + This appends a line like: 109 + 110 + ``` 111 + Signed-off-by: Your Name <your.email@example.com> 112 + ``` 113 + 114 + We won't merge commits if they aren't signed off. If you forget, you can 115 + amend the last commit like this: 116 + 117 + ```sh 118 + git commit --amend -s 119 + ``` 120 + 121 + If you're submitting a PR with multiple commits, make sure each one is 122 + signed. 123 + 124 + For [jj](https://jj-vcs.github.io/jj/latest/) users, you can run the following command 125 + to make it sign off commits in the tangled repo: 126 + 127 + ```shell 128 + # Safety check, should say "No matching config key..." 129 + jj config list templates.commit_trailers 130 + # The command below may need to be adjusted if the command above returned something. 131 + jj config set --repo templates.commit_trailers "format_signed_off_by_trailer(self)" 132 + ``` 133 + 134 + Refer to the [jj 135 + documentation](https://jj-vcs.github.io/jj/latest/config/#commit-trailers) 136 + for more information.
+172
docs/hacking.md
··· 1 + # hacking on tangled 2 + 3 + We highly recommend [installing 4 + nix](https://nixos.org/download/) (the package manager) 5 + before working on the codebase. The nix flake provides a lot 6 + of helpers to get started and most importantly, builds and 7 + dev shells are entirely deterministic. 8 + 9 + To set up your dev environment: 10 + 11 + ```bash 12 + nix develop 13 + ``` 14 + 15 + Non-nix users can look at the `devShell` attribute in the 16 + `flake.nix` file to determine necessary dependencies. 17 + 18 + ## running the appview 19 + 20 + The nix flake also exposes a few `app` attributes (run `nix 21 + flake show` to see a full list of what the flake provides), 22 + one of the apps runs the appview with the `air` 23 + live-reloader: 24 + 25 + ```bash 26 + TANGLED_DEV=true nix run .#watch-appview 27 + 28 + # TANGLED_DB_PATH might be of interest to point to 29 + # different sqlite DBs 30 + 31 + # in a separate shell, you can live-reload tailwind 32 + nix run .#watch-tailwind 33 + ``` 34 + 35 + To authenticate with the appview, you will need redis and 36 + OAUTH JWKs to be setup: 37 + 38 + ``` 39 + # oauth jwks should already be setup by the nix devshell: 40 + echo $TANGLED_OAUTH_CLIENT_SECRET 41 + z42ty4RT1ovnTopY8B8ekz9NuziF2CuMkZ7rbRFpAR9jBqMc 42 + 43 + echo $TANGLED_OAUTH_CLIENT_KID 44 + 1761667908 45 + 46 + # if not, you can set it up yourself: 47 + goat key generate -t P-256 48 + Key Type: P-256 / secp256r1 / ES256 private key 49 + Secret Key (Multibase Syntax): save this securely (eg, add to password manager) 50 + z42tuPDKRfM2mz2Kv953ARen2jmrPA8S9LX9tRq4RVcUMwwL 51 + Public Key (DID Key Syntax): share or publish this (eg, in DID document) 52 + did:key:zDnaeUBxtG6Xuv3ATJE4GaWeyXM3jyamJsZw3bSPpxx4bNXDR 53 + 54 + # the secret key from above 55 + export TANGLED_OAUTH_CLIENT_SECRET="z42tuP..." 56 + 57 + # run redis in at a new shell to store oauth sessions 58 + redis-server 59 + ``` 60 + 61 + ## running knots and spindles 62 + 63 + An end-to-end knot setup requires setting up a machine with 64 + `sshd`, `AuthorizedKeysCommand`, and git user, which is 65 + quite cumbersome. So the nix flake provides a 66 + `nixosConfiguration` to do so. 67 + 68 + <details> 69 + <summary><strong>MacOS users will have to setup a Nix Builder first</strong></summary> 70 + 71 + In order to build Tangled's dev VM on macOS, you will 72 + first need to set up a Linux Nix builder. The recommended 73 + way to do so is to run a [`darwin.linux-builder` 74 + VM](https://nixos.org/manual/nixpkgs/unstable/#sec-darwin-builder) 75 + and to register it in `nix.conf` as a builder for Linux 76 + with the same architecture as your Mac (`linux-aarch64` if 77 + you are using Apple Silicon). 78 + 79 + > IMPORTANT: You must build `darwin.linux-builder` somewhere other than inside 80 + > the tangled repo so that it doesn't conflict with the other VM. For example, 81 + > you can do 82 + > 83 + > ```shell 84 + > cd $(mktemp -d buildervm.XXXXX) && nix run nixpkgs#darwin.linux-builder 85 + > ``` 86 + > 87 + > to store the builder VM in a temporary dir. 88 + > 89 + > You should read and follow [all the other intructions][darwin builder vm] to 90 + > avoid subtle problems. 91 + 92 + Alternatively, you can use any other method to set up a 93 + Linux machine with `nix` installed that you can `sudo ssh` 94 + into (in other words, root user on your Mac has to be able 95 + to ssh into the Linux machine without entering a password) 96 + and that has the same architecture as your Mac. See 97 + [remote builder 98 + instructions](https://nix.dev/manual/nix/2.28/advanced-topics/distributed-builds.html#requirements) 99 + for how to register such a builder in `nix.conf`. 100 + 101 + > WARNING: If you'd like to use 102 + > [`nixos-lima`](https://github.com/nixos-lima/nixos-lima) or 103 + > [Orbstack](https://orbstack.dev/), note that setting them up so that `sudo 104 + > ssh` works can be tricky. It seems to be [possible with 105 + > Orbstack](https://github.com/orgs/orbstack/discussions/1669). 106 + 107 + </details> 108 + 109 + To begin, grab your DID from http://localhost:3000/settings. 110 + Then, set `TANGLED_VM_KNOT_OWNER` and 111 + `TANGLED_VM_SPINDLE_OWNER` to your DID. You can now start a 112 + lightweight NixOS VM like so: 113 + 114 + ```bash 115 + nix run --impure .#vm 116 + 117 + # type `poweroff` at the shell to exit the VM 118 + ``` 119 + 120 + This starts a knot on port 6444, a spindle on port 6555 121 + with `ssh` exposed on port 2222. 122 + 123 + Once the services are running, head to 124 + http://localhost:3000/settings/knots and hit verify. It should 125 + verify the ownership of the services instantly if everything 126 + went smoothly. 127 + 128 + You can push repositories to this VM with this ssh config 129 + block on your main machine: 130 + 131 + ```bash 132 + Host nixos-shell 133 + Hostname localhost 134 + Port 2222 135 + User git 136 + IdentityFile ~/.ssh/my_tangled_key 137 + ``` 138 + 139 + Set up a remote called `local-dev` on a git repo: 140 + 141 + ```bash 142 + git remote add local-dev git@nixos-shell:user/repo 143 + git push local-dev main 144 + ``` 145 + 146 + ### running a spindle 147 + 148 + The above VM should already be running a spindle on 149 + `localhost:6555`. Head to http://localhost:3000/settings/spindles and 150 + hit verify. You can then configure each repository to use 151 + this spindle and run CI jobs. 152 + 153 + Of interest when debugging spindles: 154 + 155 + ``` 156 + # service logs from journald: 157 + journalctl -xeu spindle 158 + 159 + # CI job logs from disk: 160 + ls /var/log/spindle 161 + 162 + # debugging spindle db: 163 + sqlite3 /var/lib/spindle/spindle.db 164 + 165 + # litecli has a nicer REPL interface: 166 + litecli /var/lib/spindle/spindle.db 167 + ``` 168 + 169 + If for any reason you wish to disable either one of the 170 + services in the VM, modify [nix/vm.nix](/nix/vm.nix) and set 171 + `services.tangled.spindle.enable` (or 172 + `services.tangled.knot.enable`) to `false`.
-93
docs/highlight.theme
··· 1 - { 2 - "text-color": null, 3 - "background-color": null, 4 - "line-number-color": null, 5 - "line-number-background-color": null, 6 - "text-styles": { 7 - "Annotation": { 8 - "text-color": null, 9 - "background-color": null, 10 - "bold": false, 11 - "italic": true, 12 - "underline": false 13 - }, 14 - "ControlFlow": { 15 - "text-color": null, 16 - "background-color": null, 17 - "bold": true, 18 - "italic": false, 19 - "underline": false 20 - }, 21 - "Error": { 22 - "text-color": null, 23 - "background-color": null, 24 - "bold": true, 25 - "italic": false, 26 - "underline": false 27 - }, 28 - "Alert": { 29 - "text-color": null, 30 - "background-color": null, 31 - "bold": true, 32 - "italic": false, 33 - "underline": false 34 - }, 35 - "Preprocessor": { 36 - "text-color": null, 37 - "background-color": null, 38 - "bold": true, 39 - "italic": false, 40 - "underline": false 41 - }, 42 - "Information": { 43 - "text-color": null, 44 - "background-color": null, 45 - "bold": false, 46 - "italic": true, 47 - "underline": false 48 - }, 49 - "Warning": { 50 - "text-color": null, 51 - "background-color": null, 52 - "bold": false, 53 - "italic": true, 54 - "underline": false 55 - }, 56 - "Documentation": { 57 - "text-color": null, 58 - "background-color": null, 59 - "bold": false, 60 - "italic": true, 61 - "underline": false 62 - }, 63 - "DataType": { 64 - "text-color": "#8f4e8b", 65 - "background-color": null, 66 - "bold": false, 67 - "italic": false, 68 - "underline": false 69 - }, 70 - "Comment": { 71 - "text-color": null, 72 - "background-color": null, 73 - "bold": false, 74 - "italic": true, 75 - "underline": false 76 - }, 77 - "CommentVar": { 78 - "text-color": null, 79 - "background-color": null, 80 - "bold": false, 81 - "italic": true, 82 - "underline": false 83 - }, 84 - "Keyword": { 85 - "text-color": null, 86 - "background-color": null, 87 - "bold": true, 88 - "italic": false, 89 - "underline": false 90 - } 91 - } 92 - } 93 -
+214
docs/knot-hosting.md
··· 1 + # knot self-hosting guide 2 + 3 + So you want to run your own knot server? Great! Here are a few prerequisites: 4 + 5 + 1. A server of some kind (a VPS, a Raspberry Pi, etc.). Preferably running a Linux distribution of some kind. 6 + 2. A (sub)domain name. People generally use `knot.example.com`. 7 + 3. A valid SSL certificate for your domain. 8 + 9 + There's a couple of ways to get started: 10 + * NixOS: refer to 11 + [flake.nix](https://tangled.sh/@tangled.sh/core/blob/master/flake.nix) 12 + * Docker: Documented at 13 + [@tangled.sh/knot-docker](https://tangled.sh/@tangled.sh/knot-docker) 14 + (community maintained: support is not guaranteed!) 15 + * Manual: Documented below. 16 + 17 + ## manual setup 18 + 19 + First, clone this repository: 20 + 21 + ``` 22 + git clone https://tangled.org/@tangled.org/core 23 + ``` 24 + 25 + Then, build the `knot` CLI. This is the knot administration and operation tool. 26 + For the purpose of this guide, we're only concerned with these subcommands: 27 + 28 + * `knot server`: the main knot server process, typically run as a 29 + supervised service 30 + * `knot guard`: handles role-based access control for git over SSH 31 + (you'll never have to run this yourself) 32 + * `knot keys`: fetches SSH keys associated with your knot; we'll use 33 + this to generate the SSH `AuthorizedKeysCommand` 34 + 35 + ``` 36 + cd core 37 + export CGO_ENABLED=1 38 + go build -o knot ./cmd/knot 39 + ``` 40 + 41 + Next, move the `knot` binary to a location owned by `root` -- 42 + `/usr/local/bin/` is a good choice. Make sure the binary itself is also owned by `root`: 43 + 44 + ``` 45 + sudo mv knot /usr/local/bin/knot 46 + sudo chown root:root /usr/local/bin/knot 47 + ``` 48 + 49 + This is necessary because SSH `AuthorizedKeysCommand` requires [really 50 + specific permissions](https://stackoverflow.com/a/27638306). The 51 + `AuthorizedKeysCommand` specifies a command that is run by `sshd` to 52 + retrieve a user's public SSH keys dynamically for authentication. Let's 53 + set that up. 54 + 55 + ``` 56 + sudo tee /etc/ssh/sshd_config.d/authorized_keys_command.conf <<EOF 57 + Match User git 58 + AuthorizedKeysCommand /usr/local/bin/knot keys -o authorized-keys 59 + AuthorizedKeysCommandUser nobody 60 + EOF 61 + ``` 62 + 63 + Then, reload `sshd`: 64 + 65 + ``` 66 + sudo systemctl reload ssh 67 + ``` 68 + 69 + Next, create the `git` user. We'll use the `git` user's home directory 70 + to store repositories: 71 + 72 + ``` 73 + sudo adduser git 74 + ``` 75 + 76 + Create `/home/git/.knot.env` with the following, updating the values as 77 + necessary. The `KNOT_SERVER_OWNER` should be set to your 78 + DID, you can find your DID in the [Settings](https://tangled.sh/settings) page. 79 + 80 + ``` 81 + KNOT_REPO_SCAN_PATH=/home/git 82 + KNOT_SERVER_HOSTNAME=knot.example.com 83 + APPVIEW_ENDPOINT=https://tangled.sh 84 + KNOT_SERVER_OWNER=did:plc:foobar 85 + KNOT_SERVER_INTERNAL_LISTEN_ADDR=127.0.0.1:5444 86 + KNOT_SERVER_LISTEN_ADDR=127.0.0.1:5555 87 + ``` 88 + 89 + If you run a Linux distribution that uses systemd, you can use the provided 90 + service file to run the server. Copy 91 + [`knotserver.service`](/systemd/knotserver.service) 92 + to `/etc/systemd/system/`. Then, run: 93 + 94 + ``` 95 + systemctl enable knotserver 96 + systemctl start knotserver 97 + ``` 98 + 99 + The last step is to configure a reverse proxy like Nginx or Caddy to front your 100 + knot. Here's an example configuration for Nginx: 101 + 102 + ``` 103 + server { 104 + listen 80; 105 + listen [::]:80; 106 + server_name knot.example.com; 107 + 108 + location / { 109 + proxy_pass http://localhost:5555; 110 + proxy_set_header Host $host; 111 + proxy_set_header X-Real-IP $remote_addr; 112 + proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; 113 + proxy_set_header X-Forwarded-Proto $scheme; 114 + } 115 + 116 + # wss endpoint for git events 117 + location /events { 118 + proxy_set_header X-Forwarded-For $remote_addr; 119 + proxy_set_header Host $http_host; 120 + proxy_set_header Upgrade websocket; 121 + proxy_set_header Connection Upgrade; 122 + proxy_pass http://localhost:5555; 123 + } 124 + # additional config for SSL/TLS go here. 125 + } 126 + 127 + ``` 128 + 129 + Remember to use Let's Encrypt or similar to procure a certificate for your 130 + knot domain. 131 + 132 + You should now have a running knot server! You can finalize 133 + your registration by hitting the `verify` button on the 134 + [/settings/knots](https://tangled.org/settings/knots) page. This simply creates 135 + a record on your PDS to announce the existence of the knot. 136 + 137 + ### custom paths 138 + 139 + (This section applies to manual setup only. Docker users should edit the mounts 140 + in `docker-compose.yml` instead.) 141 + 142 + Right now, the database and repositories of your knot lives in `/home/git`. You 143 + can move these paths if you'd like to store them in another folder. Be careful 144 + when adjusting these paths: 145 + 146 + * Stop your knot when moving data (e.g. `systemctl stop knotserver`) to prevent 147 + any possible side effects. Remember to restart it once you're done. 148 + * Make backups before moving in case something goes wrong. 149 + * Make sure the `git` user can read and write from the new paths. 150 + 151 + #### database 152 + 153 + As an example, let's say the current database is at `/home/git/knotserver.db`, 154 + and we want to move it to `/home/git/database/knotserver.db`. 155 + 156 + Copy the current database to the new location. Make sure to copy the `.db-shm` 157 + and `.db-wal` files if they exist. 158 + 159 + ``` 160 + mkdir /home/git/database 161 + cp /home/git/knotserver.db* /home/git/database 162 + ``` 163 + 164 + In the environment (e.g. `/home/git/.knot.env`), set `KNOT_SERVER_DB_PATH` to 165 + the new file path (_not_ the directory): 166 + 167 + ``` 168 + KNOT_SERVER_DB_PATH=/home/git/database/knotserver.db 169 + ``` 170 + 171 + #### repositories 172 + 173 + As an example, let's say the repositories are currently in `/home/git`, and we 174 + want to move them into `/home/git/repositories`. 175 + 176 + Create the new folder, then move the existing repositories (if there are any): 177 + 178 + ``` 179 + mkdir /home/git/repositories 180 + # move all DIDs into the new folder; these will vary for you! 181 + mv /home/git/did:plc:wshs7t2adsemcrrd4snkeqli /home/git/repositories 182 + ``` 183 + 184 + In the environment (e.g. `/home/git/.knot.env`), update `KNOT_REPO_SCAN_PATH` 185 + to the new directory: 186 + 187 + ``` 188 + KNOT_REPO_SCAN_PATH=/home/git/repositories 189 + ``` 190 + 191 + Similarly, update your `sshd` `AuthorizedKeysCommand` to use the updated 192 + repository path: 193 + 194 + ``` 195 + sudo tee /etc/ssh/sshd_config.d/authorized_keys_command.conf <<EOF 196 + Match User git 197 + AuthorizedKeysCommand /usr/local/bin/knot keys -o authorized-keys -git-dir /home/git/repositories 198 + AuthorizedKeysCommandUser nobody 199 + EOF 200 + ``` 201 + 202 + Make sure to restart your SSH server! 203 + 204 + #### MOTD (message of the day) 205 + 206 + To configure the MOTD used ("Welcome to this knot!" by default), edit the 207 + `/home/git/motd` file: 208 + 209 + ``` 210 + printf "Hi from this knot!\n" > /home/git/motd 211 + ``` 212 + 213 + Note that you should add a newline at the end if setting a non-empty message 214 + since the knot won't do this for you.
+59
docs/migrations.md
··· 1 + # Migrations 2 + 3 + This document is laid out in reverse-chronological order. 4 + Newer migration guides are listed first, and older guides 5 + are further down the page. 6 + 7 + ## Upgrading from v1.8.x 8 + 9 + After v1.8.2, the HTTP API for knot and spindles have been 10 + deprecated and replaced with XRPC. Repositories on outdated 11 + knots will not be viewable from the appview. Upgrading is 12 + straightforward however. 13 + 14 + For knots: 15 + 16 + - Upgrade to latest tag (v1.9.0 or above) 17 + - Head to the [knot dashboard](https://tangled.org/settings/knots) and 18 + hit the "retry" button to verify your knot 19 + 20 + For spindles: 21 + 22 + - Upgrade to latest tag (v1.9.0 or above) 23 + - Head to the [spindle 24 + dashboard](https://tangled.org/settings/spindles) and hit the 25 + "retry" button to verify your spindle 26 + 27 + ## Upgrading from v1.7.x 28 + 29 + After v1.7.0, knot secrets have been deprecated. You no 30 + longer need a secret from the appview to run a knot. All 31 + authorized commands to knots are managed via [Inter-Service 32 + Authentication](https://atproto.com/specs/xrpc#inter-service-authentication-jwt). 33 + Knots will be read-only until upgraded. 34 + 35 + Upgrading is quite easy, in essence: 36 + 37 + - `KNOT_SERVER_SECRET` is no more, you can remove this 38 + environment variable entirely 39 + - `KNOT_SERVER_OWNER` is now required on boot, set this to 40 + your DID. You can find your DID in the 41 + [settings](https://tangled.org/settings) page. 42 + - Restart your knot once you have replaced the environment 43 + variable 44 + - Head to the [knot dashboard](https://tangled.org/settings/knots) and 45 + hit the "retry" button to verify your knot. This simply 46 + writes a `sh.tangled.knot` record to your PDS. 47 + 48 + If you use the nix module, simply bump the flake to the 49 + latest revision, and change your config block like so: 50 + 51 + ```diff 52 + services.tangled.knot = { 53 + enable = true; 54 + server = { 55 + - secretFile = /path/to/secret; 56 + + owner = "did:plc:foo"; 57 + }; 58 + }; 59 + ```
+25
docs/spindle/architecture.md
··· 1 + # spindle architecture 2 + 3 + Spindle is a small CI runner service. Here's a high level overview of how it operates: 4 + 5 + * listens for [`sh.tangled.spindle.member`](/lexicons/spindle/member.json) and 6 + [`sh.tangled.repo`](/lexicons/repo.json) records on the Jetstream. 7 + * when a new repo record comes through (typically when you add a spindle to a 8 + repo from the settings), spindle then resolves the underlying knot and 9 + subscribes to repo events (see: 10 + [`sh.tangled.pipeline`](/lexicons/pipeline.json)). 11 + * the spindle engine then handles execution of the pipeline, with results and 12 + logs beamed on the spindle event stream over wss 13 + 14 + ### the engine 15 + 16 + At present, the only supported backend is Docker (and Podman, if Docker 17 + compatibility is enabled, so that `/run/docker.sock` is created). Spindle 18 + executes each step in the pipeline in a fresh container, with state persisted 19 + across steps within the `/tangled/workspace` directory. 20 + 21 + The base image for the container is constructed on the fly using 22 + [Nixery](https://nixery.dev), which is handy for caching layers for frequently 23 + used packages. 24 + 25 + The pipeline manifest is [specified here](/docs/spindle/pipeline.md).
+52
docs/spindle/hosting.md
··· 1 + # spindle self-hosting guide 2 + 3 + ## prerequisites 4 + 5 + * Go 6 + * Docker (the only supported backend currently) 7 + 8 + ## configuration 9 + 10 + Spindle is configured using environment variables. The following environment variables are available: 11 + 12 + * `SPINDLE_SERVER_LISTEN_ADDR`: The address the server listens on (default: `"0.0.0.0:6555"`). 13 + * `SPINDLE_SERVER_DB_PATH`: The path to the SQLite database file (default: `"spindle.db"`). 14 + * `SPINDLE_SERVER_HOSTNAME`: The hostname of the server (required). 15 + * `SPINDLE_SERVER_JETSTREAM_ENDPOINT`: The endpoint of the Jetstream server (default: `"wss://jetstream1.us-west.bsky.network/subscribe"`). 16 + * `SPINDLE_SERVER_DEV`: A boolean indicating whether the server is running in development mode (default: `false`). 17 + * `SPINDLE_SERVER_OWNER`: The DID of the owner (required). 18 + * `SPINDLE_PIPELINES_NIXERY`: The Nixery URL (default: `"nixery.tangled.sh"`). 19 + * `SPINDLE_PIPELINES_WORKFLOW_TIMEOUT`: The default workflow timeout (default: `"5m"`). 20 + * `SPINDLE_PIPELINES_LOG_DIR`: The directory to store workflow logs (default: `"/var/log/spindle"`). 21 + 22 + ## running spindle 23 + 24 + 1. **Set the environment variables.** For example: 25 + 26 + ```shell 27 + export SPINDLE_SERVER_HOSTNAME="your-hostname" 28 + export SPINDLE_SERVER_OWNER="your-did" 29 + ``` 30 + 31 + 2. **Build the Spindle binary.** 32 + 33 + ```shell 34 + cd core 35 + go mod download 36 + go build -o cmd/spindle/spindle cmd/spindle/main.go 37 + ``` 38 + 39 + 3. **Create the log directory.** 40 + 41 + ```shell 42 + sudo mkdir -p /var/log/spindle 43 + sudo chown $USER:$USER -R /var/log/spindle 44 + ``` 45 + 46 + 4. **Run the Spindle binary.** 47 + 48 + ```shell 49 + ./cmd/spindle/spindle 50 + ``` 51 + 52 + Spindle will now start, connect to the Jetstream server, and begin processing pipelines.
+285
docs/spindle/openbao.md
··· 1 + # spindle secrets with openbao 2 + 3 + This document covers setting up Spindle to use OpenBao for secrets 4 + management via OpenBao Proxy instead of the default SQLite backend. 5 + 6 + ## overview 7 + 8 + Spindle now uses OpenBao Proxy for secrets management. The proxy handles 9 + authentication automatically using AppRole credentials, while Spindle 10 + connects to the local proxy instead of directly to the OpenBao server. 11 + 12 + This approach provides better security, automatic token renewal, and 13 + simplified application code. 14 + 15 + ## installation 16 + 17 + Install OpenBao from nixpkgs: 18 + 19 + ```bash 20 + nix shell nixpkgs#openbao # for a local server 21 + ``` 22 + 23 + ## setup 24 + 25 + The setup process can is documented for both local development and production. 26 + 27 + ### local development 28 + 29 + Start OpenBao in dev mode: 30 + 31 + ```bash 32 + bao server -dev -dev-root-token-id="root" -dev-listen-address=127.0.0.1:8201 33 + ``` 34 + 35 + This starts OpenBao on `http://localhost:8201` with a root token. 36 + 37 + Set up environment for bao CLI: 38 + 39 + ```bash 40 + export BAO_ADDR=http://localhost:8200 41 + export BAO_TOKEN=root 42 + ``` 43 + 44 + ### production 45 + 46 + You would typically use a systemd service with a configuration file. Refer to 47 + [@tangled.org/infra](https://tangled.org/@tangled.org/infra) for how this can be 48 + achieved using Nix. 49 + 50 + Then, initialize the bao server: 51 + ```bash 52 + bao operator init -key-shares=1 -key-threshold=1 53 + ``` 54 + 55 + This will print out an unseal key and a root key. Save them somewhere (like a password manager). Then unseal the vault to begin setting it up: 56 + ```bash 57 + bao operator unseal <unseal_key> 58 + ``` 59 + 60 + All steps below remain the same across both dev and production setups. 61 + 62 + ### configure openbao server 63 + 64 + Create the spindle KV mount: 65 + 66 + ```bash 67 + bao secrets enable -path=spindle -version=2 kv 68 + ``` 69 + 70 + Set up AppRole authentication and policy: 71 + 72 + Create a policy file `spindle-policy.hcl`: 73 + 74 + ```hcl 75 + # Full access to spindle KV v2 data 76 + path "spindle/data/*" { 77 + capabilities = ["create", "read", "update", "delete"] 78 + } 79 + 80 + # Access to metadata for listing and management 81 + path "spindle/metadata/*" { 82 + capabilities = ["list", "read", "delete", "update"] 83 + } 84 + 85 + # Allow listing at root level 86 + path "spindle/" { 87 + capabilities = ["list"] 88 + } 89 + 90 + # Required for connection testing and health checks 91 + path "auth/token/lookup-self" { 92 + capabilities = ["read"] 93 + } 94 + ``` 95 + 96 + Apply the policy and create an AppRole: 97 + 98 + ```bash 99 + bao policy write spindle-policy spindle-policy.hcl 100 + bao auth enable approle 101 + bao write auth/approle/role/spindle \ 102 + token_policies="spindle-policy" \ 103 + token_ttl=1h \ 104 + token_max_ttl=4h \ 105 + bind_secret_id=true \ 106 + secret_id_ttl=0 \ 107 + secret_id_num_uses=0 108 + ``` 109 + 110 + Get the credentials: 111 + 112 + ```bash 113 + # Get role ID (static) 114 + ROLE_ID=$(bao read -field=role_id auth/approle/role/spindle/role-id) 115 + 116 + # Generate secret ID 117 + SECRET_ID=$(bao write -f -field=secret_id auth/approle/role/spindle/secret-id) 118 + 119 + echo "Role ID: $ROLE_ID" 120 + echo "Secret ID: $SECRET_ID" 121 + ``` 122 + 123 + ### create proxy configuration 124 + 125 + Create the credential files: 126 + 127 + ```bash 128 + # Create directory for OpenBao files 129 + mkdir -p /tmp/openbao 130 + 131 + # Save credentials 132 + echo "$ROLE_ID" > /tmp/openbao/role-id 133 + echo "$SECRET_ID" > /tmp/openbao/secret-id 134 + chmod 600 /tmp/openbao/role-id /tmp/openbao/secret-id 135 + ``` 136 + 137 + Create a proxy configuration file `/tmp/openbao/proxy.hcl`: 138 + 139 + ```hcl 140 + # OpenBao server connection 141 + vault { 142 + address = "http://localhost:8200" 143 + } 144 + 145 + # Auto-Auth using AppRole 146 + auto_auth { 147 + method "approle" { 148 + mount_path = "auth/approle" 149 + config = { 150 + role_id_file_path = "/tmp/openbao/role-id" 151 + secret_id_file_path = "/tmp/openbao/secret-id" 152 + } 153 + } 154 + 155 + # Optional: write token to file for debugging 156 + sink "file" { 157 + config = { 158 + path = "/tmp/openbao/token" 159 + mode = 0640 160 + } 161 + } 162 + } 163 + 164 + # Proxy listener for Spindle 165 + listener "tcp" { 166 + address = "127.0.0.1:8201" 167 + tls_disable = true 168 + } 169 + 170 + # Enable API proxy with auto-auth token 171 + api_proxy { 172 + use_auto_auth_token = true 173 + } 174 + 175 + # Enable response caching 176 + cache { 177 + use_auto_auth_token = true 178 + } 179 + 180 + # Logging 181 + log_level = "info" 182 + ``` 183 + 184 + ### start the proxy 185 + 186 + Start OpenBao Proxy: 187 + 188 + ```bash 189 + bao proxy -config=/tmp/openbao/proxy.hcl 190 + ``` 191 + 192 + The proxy will authenticate with OpenBao and start listening on 193 + `127.0.0.1:8201`. 194 + 195 + ### configure spindle 196 + 197 + Set these environment variables for Spindle: 198 + 199 + ```bash 200 + export SPINDLE_SERVER_SECRETS_PROVIDER=openbao 201 + export SPINDLE_SERVER_SECRETS_OPENBAO_PROXY_ADDR=http://127.0.0.1:8201 202 + export SPINDLE_SERVER_SECRETS_OPENBAO_MOUNT=spindle 203 + ``` 204 + 205 + Start Spindle: 206 + 207 + Spindle will now connect to the local proxy, which handles all 208 + authentication automatically. 209 + 210 + ## production setup for proxy 211 + 212 + For production, you'll want to run the proxy as a service: 213 + 214 + Place your production configuration in `/etc/openbao/proxy.hcl` with 215 + proper TLS settings for the vault connection. 216 + 217 + ## verifying setup 218 + 219 + Test the proxy directly: 220 + 221 + ```bash 222 + # Check proxy health 223 + curl -H "X-Vault-Request: true" http://127.0.0.1:8201/v1/sys/health 224 + 225 + # Test token lookup through proxy 226 + curl -H "X-Vault-Request: true" http://127.0.0.1:8201/v1/auth/token/lookup-self 227 + ``` 228 + 229 + Test OpenBao operations through the server: 230 + 231 + ```bash 232 + # List all secrets 233 + bao kv list spindle/ 234 + 235 + # Add a test secret via Spindle API, then check it exists 236 + bao kv list spindle/repos/ 237 + 238 + # Get a specific secret 239 + bao kv get spindle/repos/your_repo_path/SECRET_NAME 240 + ``` 241 + 242 + ## how it works 243 + 244 + - Spindle connects to OpenBao Proxy on localhost (typically port 8200 or 8201) 245 + - The proxy authenticates with OpenBao using AppRole credentials 246 + - All Spindle requests go through the proxy, which injects authentication tokens 247 + - Secrets are stored at `spindle/repos/{sanitized_repo_path}/{secret_key}` 248 + - Repository paths like `did:plc:alice/myrepo` become `did_plc_alice_myrepo` 249 + - The proxy handles all token renewal automatically 250 + - Spindle no longer manages tokens or authentication directly 251 + 252 + ## troubleshooting 253 + 254 + **Connection refused**: Check that the OpenBao Proxy is running and 255 + listening on the configured address. 256 + 257 + **403 errors**: Verify the AppRole credentials are correct and the policy 258 + has the necessary permissions. 259 + 260 + **404 route errors**: The spindle KV mount probably doesn't exist - run 261 + the mount creation step again. 262 + 263 + **Proxy authentication failures**: Check the proxy logs and verify the 264 + role-id and secret-id files are readable and contain valid credentials. 265 + 266 + **Secret not found after writing**: This can indicate policy permission 267 + issues. Verify the policy includes both `spindle/data/*` and 268 + `spindle/metadata/*` paths with appropriate capabilities. 269 + 270 + Check proxy logs: 271 + 272 + ```bash 273 + # If running as systemd service 274 + journalctl -u openbao-proxy -f 275 + 276 + # If running directly, check the console output 277 + ``` 278 + 279 + Test AppRole authentication manually: 280 + 281 + ```bash 282 + bao write auth/approle/login \ 283 + role_id="$(cat /tmp/openbao/role-id)" \ 284 + secret_id="$(cat /tmp/openbao/secret-id)" 285 + ```
+183
docs/spindle/pipeline.md
··· 1 + # spindle pipelines 2 + 3 + Spindle workflows allow you to write CI/CD pipelines in a simple format. They're located in the `.tangled/workflows` directory at the root of your repository, and are defined using YAML. 4 + 5 + The fields are: 6 + 7 + - [Trigger](#trigger): A **required** field that defines when a workflow should be triggered. 8 + - [Engine](#engine): A **required** field that defines which engine a workflow should run on. 9 + - [Clone options](#clone-options): An **optional** field that defines how the repository should be cloned. 10 + - [Dependencies](#dependencies): An **optional** field that allows you to list dependencies you may need. 11 + - [Environment](#environment): An **optional** field that allows you to define environment variables. 12 + - [Steps](#steps): An **optional** field that allows you to define what steps should run in the workflow. 13 + 14 + ## Trigger 15 + 16 + The first thing to add to a workflow is the trigger, which defines when a workflow runs. This is defined using a `when` field, which takes in a list of conditions. Each condition has the following fields: 17 + 18 + - `event`: This is a **required** field that defines when your workflow should run. It's a list that can take one or more of the following values: 19 + - `push`: The workflow should run every time a commit is pushed to the repository. 20 + - `pull_request`: The workflow should run every time a pull request is made or updated. 21 + - `manual`: The workflow can be triggered manually. 22 + - `branch`: Defines which branches the workflow should run for. If used with the `push` event, commits to the branch(es) listed here will trigger the workflow. If used with the `pull_request` event, updates to pull requests targeting the branch(es) listed here will trigger the workflow. This field has no effect with the `manual` event. Supports glob patterns using `*` and `**` (e.g., `main`, `develop`, `release-*`). Either `branch` or `tag` (or both) must be specified for `push` events. 23 + - `tag`: Defines which tags the workflow should run for. Only used with the `push` event - when tags matching the pattern(s) listed here are pushed, the workflow will trigger. This field has no effect with `pull_request` or `manual` events. Supports glob patterns using `*` and `**` (e.g., `v*`, `v1.*`, `release-**`). Either `branch` or `tag` (or both) must be specified for `push` events. 24 + 25 + For example, if you'd like to define a workflow that runs when commits are pushed to the `main` and `develop` branches, or when pull requests that target the `main` branch are updated, or manually, you can do so with: 26 + 27 + ```yaml 28 + when: 29 + - event: ["push", "manual"] 30 + branch: ["main", "develop"] 31 + - event: ["pull_request"] 32 + branch: ["main"] 33 + ``` 34 + 35 + You can also trigger workflows on tag pushes. For instance, to run a deployment workflow when tags matching `v*` are pushed: 36 + 37 + ```yaml 38 + when: 39 + - event: ["push"] 40 + tag: ["v*"] 41 + ``` 42 + 43 + You can even combine branch and tag patterns in a single constraint (the workflow triggers if either matches): 44 + 45 + ```yaml 46 + when: 47 + - event: ["push"] 48 + branch: ["main", "release-*"] 49 + tag: ["v*", "stable"] 50 + ``` 51 + 52 + ## Engine 53 + 54 + Next is the engine on which the workflow should run, defined using the **required** `engine` field. The currently supported engines are: 55 + 56 + - `nixery`: This uses an instance of [Nixery](https://nixery.dev) to run steps, which allows you to add [dependencies](#dependencies) from [Nixpkgs](https://github.com/NixOS/nixpkgs). You can search for packages on https://search.nixos.org, and there's a pretty good chance the package(s) you're looking for will be there. 57 + 58 + Example: 59 + 60 + ```yaml 61 + engine: "nixery" 62 + ``` 63 + 64 + ## Clone options 65 + 66 + When a workflow starts, the first step is to clone the repository. You can customize this behavior using the **optional** `clone` field. It has the following fields: 67 + 68 + - `skip`: Setting this to `true` will skip cloning the repository. This can be useful if your workflow is doing something that doesn't require anything from the repository itself. This is `false` by default. 69 + - `depth`: This sets the number of commits, or the "clone depth", to fetch from the repository. For example, if you set this to 2, the last 2 commits will be fetched. By default, the depth is set to 1, meaning only the most recent commit will be fetched, which is the commit that triggered the workflow. 70 + - `submodules`: If you use [git submodules](https://git-scm.com/book/en/v2/Git-Tools-Submodules) in your repository, setting this field to `true` will recursively fetch all submodules. This is `false` by default. 71 + 72 + The default settings are: 73 + 74 + ```yaml 75 + clone: 76 + skip: false 77 + depth: 1 78 + submodules: false 79 + ``` 80 + 81 + ## Dependencies 82 + 83 + Usually when you're running a workflow, you'll need additional dependencies. The `dependencies` field lets you define which dependencies to get, and from where. It's a key-value map, with the key being the registry to fetch dependencies from, and the value being the list of dependencies to fetch. 84 + 85 + Say you want to fetch Node.js and Go from `nixpkgs`, and a package called `my_pkg` you've made from your own registry at your repository at `https://tangled.sh/@example.com/my_pkg`. You can define those dependencies like so: 86 + 87 + ```yaml 88 + dependencies: 89 + # nixpkgs 90 + nixpkgs: 91 + - nodejs 92 + - go 93 + # custom registry 94 + git+https://tangled.org/@example.com/my_pkg: 95 + - my_pkg 96 + ``` 97 + 98 + Now these dependencies are available to use in your workflow! 99 + 100 + ## Environment 101 + 102 + The `environment` field allows you define environment variables that will be available throughout the entire workflow. **Do not put secrets here, these environment variables are visible to anyone viewing the repository. You can add secrets for pipelines in your repository's settings.** 103 + 104 + Example: 105 + 106 + ```yaml 107 + environment: 108 + GOOS: "linux" 109 + GOARCH: "arm64" 110 + NODE_ENV: "production" 111 + MY_ENV_VAR: "MY_ENV_VALUE" 112 + ``` 113 + 114 + ## Steps 115 + 116 + The `steps` field allows you to define what steps should run in the workflow. It's a list of step objects, each with the following fields: 117 + 118 + - `name`: This field allows you to give your step a name. This name is visible in your workflow runs, and is used to describe what the step is doing. 119 + - `command`: This field allows you to define a command to run in that step. The step is run in a Bash shell, and the logs from the command will be visible in the pipelines page on the Tangled website. The [dependencies](#dependencies) you added will be available to use here. 120 + - `environment`: Similar to the global [environment](#environment) config, this **optional** field is a key-value map that allows you to set environment variables for the step. **Do not put secrets here, these environment variables are visible to anyone viewing the repository. You can add secrets for pipelines in your repository's settings.** 121 + 122 + Example: 123 + 124 + ```yaml 125 + steps: 126 + - name: "Build backend" 127 + command: "go build" 128 + environment: 129 + GOOS: "darwin" 130 + GOARCH: "arm64" 131 + - name: "Build frontend" 132 + command: "npm run build" 133 + environment: 134 + NODE_ENV: "production" 135 + ``` 136 + 137 + ## Complete workflow 138 + 139 + ```yaml 140 + # .tangled/workflows/build.yml 141 + 142 + when: 143 + - event: ["push", "manual"] 144 + branch: ["main", "develop"] 145 + - event: ["pull_request"] 146 + branch: ["main"] 147 + 148 + engine: "nixery" 149 + 150 + # using the default values 151 + clone: 152 + skip: false 153 + depth: 1 154 + submodules: false 155 + 156 + dependencies: 157 + # nixpkgs 158 + nixpkgs: 159 + - nodejs 160 + - go 161 + # custom registry 162 + git+https://tangled.org/@example.com/my_pkg: 163 + - my_pkg 164 + 165 + environment: 166 + GOOS: "linux" 167 + GOARCH: "arm64" 168 + NODE_ENV: "production" 169 + MY_ENV_VAR: "MY_ENV_VALUE" 170 + 171 + steps: 172 + - name: "Build backend" 173 + command: "go build" 174 + environment: 175 + GOOS: "darwin" 176 + GOARCH: "arm64" 177 + - name: "Build frontend" 178 + command: "npm run build" 179 + environment: 180 + NODE_ENV: "production" 181 + ``` 182 + 183 + If you want another example of a workflow, you can look at the one [Tangled uses to build the project](https://tangled.sh/@tangled.sh/core/blob/master/.tangled/workflows/build.yml).
-101
docs/styles.css
··· 1 - svg { 2 - width: 16px; 3 - height: 16px; 4 - } 5 - 6 - :root { 7 - --syntax-alert: #d20f39; 8 - --syntax-annotation: #fe640b; 9 - --syntax-attribute: #df8e1d; 10 - --syntax-basen: #40a02b; 11 - --syntax-builtin: #1e66f5; 12 - --syntax-controlflow: #8839ef; 13 - --syntax-char: #04a5e5; 14 - --syntax-constant: #fe640b; 15 - --syntax-comment: #9ca0b0; 16 - --syntax-commentvar: #7c7f93; 17 - --syntax-documentation: #9ca0b0; 18 - --syntax-datatype: #df8e1d; 19 - --syntax-decval: #40a02b; 20 - --syntax-error: #d20f39; 21 - --syntax-extension: #4c4f69; 22 - --syntax-float: #40a02b; 23 - --syntax-function: #1e66f5; 24 - --syntax-import: #40a02b; 25 - --syntax-information: #04a5e5; 26 - --syntax-keyword: #8839ef; 27 - --syntax-operator: #179299; 28 - --syntax-other: #8839ef; 29 - --syntax-preprocessor: #ea76cb; 30 - --syntax-specialchar: #04a5e5; 31 - --syntax-specialstring: #ea76cb; 32 - --syntax-string: #40a02b; 33 - --syntax-variable: #8839ef; 34 - --syntax-verbatimstring: #40a02b; 35 - --syntax-warning: #df8e1d; 36 - } 37 - 38 - @media (prefers-color-scheme: dark) { 39 - :root { 40 - --syntax-alert: #f38ba8; 41 - --syntax-annotation: #fab387; 42 - --syntax-attribute: #f9e2af; 43 - --syntax-basen: #a6e3a1; 44 - --syntax-builtin: #89b4fa; 45 - --syntax-controlflow: #cba6f7; 46 - --syntax-char: #89dceb; 47 - --syntax-constant: #fab387; 48 - --syntax-comment: #6c7086; 49 - --syntax-commentvar: #585b70; 50 - --syntax-documentation: #6c7086; 51 - --syntax-datatype: #f9e2af; 52 - --syntax-decval: #a6e3a1; 53 - --syntax-error: #f38ba8; 54 - --syntax-extension: #cdd6f4; 55 - --syntax-float: #a6e3a1; 56 - --syntax-function: #89b4fa; 57 - --syntax-import: #a6e3a1; 58 - --syntax-information: #89dceb; 59 - --syntax-keyword: #cba6f7; 60 - --syntax-operator: #94e2d5; 61 - --syntax-other: #cba6f7; 62 - --syntax-preprocessor: #f5c2e7; 63 - --syntax-specialchar: #89dceb; 64 - --syntax-specialstring: #f5c2e7; 65 - --syntax-string: #a6e3a1; 66 - --syntax-variable: #cba6f7; 67 - --syntax-verbatimstring: #a6e3a1; 68 - --syntax-warning: #f9e2af; 69 - } 70 - } 71 - 72 - /* pandoc syntax highlighting classes */ 73 - code span.al { color: var(--syntax-alert); font-weight: bold; } /* alert */ 74 - code span.an { color: var(--syntax-annotation); font-weight: bold; font-style: italic; } /* annotation */ 75 - code span.at { color: var(--syntax-attribute); } /* attribute */ 76 - code span.bn { color: var(--syntax-basen); } /* basen */ 77 - code span.bu { color: var(--syntax-builtin); } /* builtin */ 78 - code span.cf { color: var(--syntax-controlflow); font-weight: bold; } /* controlflow */ 79 - code span.ch { color: var(--syntax-char); } /* char */ 80 - code span.cn { color: var(--syntax-constant); } /* constant */ 81 - code span.co { color: var(--syntax-comment); font-style: italic; } /* comment */ 82 - code span.cv { color: var(--syntax-commentvar); font-weight: bold; font-style: italic; } /* commentvar */ 83 - code span.do { color: var(--syntax-documentation); font-style: italic; } /* documentation */ 84 - code span.dt { color: var(--syntax-datatype); } /* datatype */ 85 - code span.dv { color: var(--syntax-decval); } /* decval */ 86 - code span.er { color: var(--syntax-error); font-weight: bold; } /* error */ 87 - code span.ex { color: var(--syntax-extension); } /* extension */ 88 - code span.fl { color: var(--syntax-float); } /* float */ 89 - code span.fu { color: var(--syntax-function); } /* function */ 90 - code span.im { color: var(--syntax-import); font-weight: bold; } /* import */ 91 - code span.in { color: var(--syntax-information); font-weight: bold; font-style: italic; } /* information */ 92 - code span.kw { color: var(--syntax-keyword); font-weight: bold; } /* keyword */ 93 - code span.op { color: var(--syntax-operator); } /* operator */ 94 - code span.ot { color: var(--syntax-other); } /* other */ 95 - code span.pp { color: var(--syntax-preprocessor); } /* preprocessor */ 96 - code span.sc { color: var(--syntax-specialchar); } /* specialchar */ 97 - code span.ss { color: var(--syntax-specialstring); } /* specialstring */ 98 - code span.st { color: var(--syntax-string); } /* string */ 99 - code span.va { color: var(--syntax-variable); } /* variable */ 100 - code span.vs { color: var(--syntax-verbatimstring); } /* verbatimstring */ 101 - code span.wa { color: var(--syntax-warning); font-weight: bold; font-style: italic; } /* warning */
-117
docs/template.html
··· 1 - <!DOCTYPE html> 2 - <html xmlns="http://www.w3.org/1999/xhtml" lang="$lang$" xml:lang="$lang$"$if(dir)$ dir="$dir$"$endif$> 3 - <head> 4 - <meta charset="utf-8" /> 5 - <meta name="generator" content="pandoc" /> 6 - <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> 7 - $for(author-meta)$ 8 - <meta name="author" content="$author-meta$" /> 9 - $endfor$ 10 - 11 - $if(date-meta)$ 12 - <meta name="dcterms.date" content="$date-meta$" /> 13 - $endif$ 14 - 15 - $if(keywords)$ 16 - <meta name="keywords" content="$for(keywords)$$keywords$$sep$, $endfor$" /> 17 - $endif$ 18 - 19 - $if(description-meta)$ 20 - <meta name="description" content="$description-meta$" /> 21 - $endif$ 22 - 23 - <title>$pagetitle$ - Tangled docs</title> 24 - 25 - <style> 26 - $styles.css()$ 27 - </style> 28 - 29 - $for(css)$ 30 - <link rel="stylesheet" href="$css$" /> 31 - $endfor$ 32 - 33 - $for(header-includes)$ 34 - $header-includes$ 35 - $endfor$ 36 - 37 - <link rel="preload" href="/static/fonts/InterVariable.woff2" as="font" type="font/woff2" crossorigin /> 38 - 39 - </head> 40 - <body class="bg-white dark:bg-gray-900 min-h-screen flex flex-col min-h-screen"> 41 - $for(include-before)$ 42 - $include-before$ 43 - $endfor$ 44 - 45 - $if(toc)$ 46 - <!-- mobile topbar toc --> 47 - <details id="mobile-$idprefix$TOC" role="doc-toc" class="md:hidden bg-gray-50 dark:bg-gray-800 border-b border-gray-200 dark:border-gray-700 z-50 space-y-4 group px-6 py-4"> 48 - <summary class="cursor-pointer list-none text-sm font-semibold select-none flex gap-2 justify-between items-center dark:text-white"> 49 - $if(toc-title)$$toc-title$$else$Table of Contents$endif$ 50 - <span class="group-open:hidden inline">${ menu.svg() }</span> 51 - <span class="hidden group-open:inline">${ x.svg() }</span> 52 - </summary> 53 - ${ table-of-contents:toc.html() } 54 - </details> 55 - <!-- desktop sidebar toc --> 56 - <nav id="$idprefix$TOC" role="doc-toc" class="hidden md:block fixed left-0 top-0 w-80 h-screen bg-gray-50 dark:bg-gray-800 border-r border-gray-200 dark:border-gray-700 overflow-y-auto p-4 z-50"> 57 - $if(toc-title)$ 58 - <h2 id="$idprefix$toc-title" class="text-lg font-semibold mb-4 text-gray-900">$toc-title$</h2> 59 - $endif$ 60 - ${ table-of-contents:toc.html() } 61 - </nav> 62 - $endif$ 63 - 64 - <div class="$if(toc)$md:ml-80$endif$ flex-1 flex flex-col"> 65 - <main class="max-w-4xl w-full mx-auto p-6 flex-1"> 66 - $if(top)$ 67 - $-- only print title block if this is NOT the top page 68 - $else$ 69 - $if(title)$ 70 - <header id="title-block-header" class="mb-8 pb-8 border-b border-gray-200 dark:border-gray-700"> 71 - <h1 class="text-4xl font-bold mb-2 text-black dark:text-white">$title$</h1> 72 - $if(subtitle)$ 73 - <p class="text-xl text-gray-500 dark:text-gray-400 mb-2">$subtitle$</p> 74 - $endif$ 75 - $for(author)$ 76 - <p class="text-sm text-gray-500 dark:text-gray-400">$author$</p> 77 - $endfor$ 78 - $if(date)$ 79 - <p class="text-sm text-gray-500 dark:text-gray-400">Updated on $date$</p> 80 - $endif$ 81 - $if(abstract)$ 82 - <div class="mt-6 p-4 bg-gray-50 rounded-lg"> 83 - <div class="text-sm font-semibold text-gray-700 uppercase mb-2">$abstract-title$</div> 84 - <div class="text-gray-700">$abstract$</div> 85 - </div> 86 - $endif$ 87 - $endif$ 88 - </header> 89 - $endif$ 90 - <article class="prose dark:prose-invert max-w-none"> 91 - $body$ 92 - </article> 93 - </main> 94 - <nav id="sitenav" class="border-t border-gray-200 dark:border-gray-700 bg-gray-50 dark:bg-gray-800 "> 95 - <div class="max-w-4xl mx-auto px-8 py-4"> 96 - <div class="flex justify-between gap-4"> 97 - <span class="flex-1"> 98 - $if(previous.url)$ 99 - <span class="text-xs text-gray-500 dark:text-gray-400 uppercase block mb-1">Previous</span> 100 - <a href="$previous.url$" accesskey="p" rel="previous">$previous.title$</a> 101 - $endif$ 102 - </span> 103 - <span class="flex-1 text-right"> 104 - $if(next.url)$ 105 - <span class="text-xs text-gray-500 dark:text-gray-400 uppercase block mb-1">Next</span> 106 - <a href="$next.url$" accesskey="n" rel="next">$next.title$</a> 107 - $endif$ 108 - </span> 109 - </div> 110 - </div> 111 - </nav> 112 - </div> 113 - $for(include-after)$ 114 - $include-after$ 115 - $endfor$ 116 - </body> 117 - </html>
-4
docs/toc.html
··· 1 - <div class="[&_ul]:space-y-6 [&_ul]:pl-0 [&_ul]:font-bold [&_ul_ul]:pl-4 [&_ul_ul]:font-normal [&_ul_ul]:space-y-2 [&_li]:space-y-2"> 2 - $table-of-contents$ 3 - </div> 4 -
+2 -5
flake.nix
··· 88 88 inherit htmx-src htmx-ws-src lucide-src inter-fonts-src ibm-plex-mono-src actor-typeahead-src; 89 89 }; 90 90 appview = self.callPackage ./nix/pkgs/appview.nix {}; 91 - docs = self.callPackage ./nix/pkgs/docs.nix { 92 - inherit inter-fonts-src ibm-plex-mono-src lucide-src; 93 - }; 94 91 spindle = self.callPackage ./nix/pkgs/spindle.nix {}; 95 92 knot-unwrapped = self.callPackage ./nix/pkgs/knot-unwrapped.nix {}; 96 93 knot = self.callPackage ./nix/pkgs/knot.nix {}; 97 94 }); 98 95 in { 99 96 overlays.default = final: prev: { 100 - inherit (mkPackageSet final) lexgen goat sqlite-lib spindle knot-unwrapped knot appview docs; 97 + inherit (mkPackageSet final) lexgen goat sqlite-lib spindle knot-unwrapped knot appview; 101 98 }; 102 99 103 100 packages = forAllSystems (system: let ··· 106 103 staticPackages = mkPackageSet pkgs.pkgsStatic; 107 104 crossPackages = mkPackageSet pkgs.pkgsCross.gnu64.pkgsStatic; 108 105 in { 109 - inherit (packages) appview appview-static-files lexgen goat spindle knot knot-unwrapped sqlite-lib docs; 106 + inherit (packages) appview appview-static-files lexgen goat spindle knot knot-unwrapped sqlite-lib; 110 107 111 108 pkgsStatic-appview = staticPackages.appview; 112 109 pkgsStatic-knot = staticPackages.knot;
+1 -1
input.css
··· 162 162 } 163 163 164 164 .prose a.mention { 165 - @apply no-underline hover:underline font-bold; 165 + @apply no-underline hover:underline; 166 166 } 167 167 168 168 .prose li {
+103
knotserver/xrpc/list_repos.go
··· 1 + package xrpc 2 + 3 + import ( 4 + "net/http" 5 + "os" 6 + "path/filepath" 7 + "strings" 8 + 9 + securejoin "github.com/cyphar/filepath-securejoin" 10 + "tangled.org/core/api/tangled" 11 + "tangled.org/core/knotserver/git" 12 + xrpcerr "tangled.org/core/xrpc/errors" 13 + ) 14 + 15 + // ListRepos lists all users (DIDs) and their repositories by scanning the repository directory 16 + func (x *Xrpc) ListRepos(w http.ResponseWriter, r *http.Request) { 17 + scanPath := x.Config.Repo.ScanPath 18 + 19 + didEntries, err := os.ReadDir(scanPath) 20 + if err != nil { 21 + x.Logger.Error("failed to read scan path", "error", err, "path", scanPath) 22 + writeError(w, xrpcerr.GenericError(err), http.StatusInternalServerError) 23 + return 24 + } 25 + 26 + var users []*tangled.RepoListRepos_User 27 + 28 + for _, didEntry := range didEntries { 29 + if !didEntry.IsDir() { 30 + continue 31 + } 32 + 33 + did := didEntry.Name() 34 + 35 + // Validate DID format (basic check) 36 + if !strings.HasPrefix(did, "did:") { 37 + continue 38 + } 39 + 40 + didPath, err := securejoin.SecureJoin(scanPath, did) 41 + if err != nil { 42 + x.Logger.Warn("failed to join path for did", "did", did, "error", err) 43 + continue 44 + } 45 + 46 + // Read repositories for this DID 47 + repoEntries, err := os.ReadDir(didPath) 48 + if err != nil { 49 + x.Logger.Warn("failed to read did directory", "did", did, "error", err) 50 + continue 51 + } 52 + 53 + var repos []*tangled.RepoListRepos_RepoEntry 54 + 55 + for _, repoEntry := range repoEntries { 56 + if !repoEntry.IsDir() { 57 + continue 58 + } 59 + 60 + repoName := repoEntry.Name() 61 + 62 + // Check if it's a valid git repository 63 + repoPath, err := securejoin.SecureJoin(didPath, repoName) 64 + if err != nil { 65 + continue 66 + } 67 + 68 + repo, err := git.PlainOpen(repoPath) 69 + if err != nil { 70 + // Not a valid git repository, skip 71 + continue 72 + } 73 + 74 + // Get default branch 75 + defaultBranch := "master" 76 + branch, err := repo.FindMainBranch() 77 + if err == nil { 78 + defaultBranch = branch 79 + } 80 + 81 + repos = append(repos, &tangled.RepoListRepos_RepoEntry{ 82 + Name: repoName, 83 + Did: did, 84 + FullPath: filepath.Join(did, repoName), 85 + DefaultBranch: &defaultBranch, 86 + }) 87 + } 88 + 89 + // Only add user if they have repositories 90 + if len(repos) > 0 { 91 + users = append(users, &tangled.RepoListRepos_User{ 92 + Did: did, 93 + Repos: repos, 94 + }) 95 + } 96 + } 97 + 98 + response := tangled.RepoListRepos_Output{ 99 + Users: users, 100 + } 101 + 102 + writeJson(w, response) 103 + }
+1
knotserver/xrpc/xrpc.go
··· 73 73 74 74 // service query endpoints (no auth required) 75 75 r.Get("/"+tangled.OwnerNSID, x.Owner) 76 + r.Get("/"+tangled.RepoListReposNSID, x.ListRepos) 76 77 77 78 return r 78 79 }
+71
lexicons/repo/listRepos.json
··· 1 + { 2 + "lexicon": 1, 3 + "id": "sh.tangled.repo.listRepos", 4 + "defs": { 5 + "main": { 6 + "type": "query", 7 + "description": "Lists all users (DIDs) and their repositories", 8 + "parameters": { 9 + "type": "params", 10 + "properties": {} 11 + }, 12 + "output": { 13 + "encoding": "application/json", 14 + "schema": { 15 + "type": "object", 16 + "required": ["users"], 17 + "properties": { 18 + "users": { 19 + "type": "array", 20 + "items": { 21 + "type": "ref", 22 + "ref": "#user" 23 + } 24 + } 25 + } 26 + } 27 + } 28 + }, 29 + "user": { 30 + "type": "object", 31 + "required": ["did", "repos"], 32 + "properties": { 33 + "did": { 34 + "type": "string", 35 + "format": "did", 36 + "description": "DID of the user" 37 + }, 38 + "repos": { 39 + "type": "array", 40 + "items": { 41 + "type": "ref", 42 + "ref": "#repoEntry" 43 + } 44 + } 45 + } 46 + }, 47 + "repoEntry": { 48 + "type": "object", 49 + "required": ["name", "did", "fullPath"], 50 + "properties": { 51 + "name": { 52 + "type": "string", 53 + "description": "Repository name" 54 + }, 55 + "did": { 56 + "type": "string", 57 + "format": "did", 58 + "description": "DID of the repository owner" 59 + }, 60 + "fullPath": { 61 + "type": "string", 62 + "description": "Full path to the repository" 63 + }, 64 + "defaultBranch": { 65 + "type": "string", 66 + "description": "Default branch of the repository" 67 + } 68 + } 69 + } 70 + } 71 + }
-41
nix/pkgs/docs.nix
··· 1 - { 2 - pandoc, 3 - tailwindcss, 4 - runCommandLocal, 5 - inter-fonts-src, 6 - ibm-plex-mono-src, 7 - lucide-src, 8 - src, 9 - }: 10 - runCommandLocal "docs" {} '' 11 - mkdir -p working 12 - 13 - # copy templates, themes, styles, filters to working directory 14 - cp ${src}/docs/*.html working/ 15 - cp ${src}/docs/*.theme working/ 16 - cp ${src}/docs/*.css working/ 17 - 18 - # icons 19 - cp -rf ${lucide-src}/*.svg working/ 20 - 21 - # content 22 - ${pandoc}/bin/pandoc ${src}/docs/DOCS.md \ 23 - -o $out/ \ 24 - -t chunkedhtml \ 25 - --variable toc \ 26 - --toc-depth=2 \ 27 - --css=stylesheet.css \ 28 - --chunk-template="%i.html" \ 29 - --highlight-style=working/highlight.theme \ 30 - --template=working/template.html 31 - 32 - # fonts 33 - mkdir -p $out/static/fonts 34 - cp -f ${inter-fonts-src}/web/InterVariable*.woff2 $out/static/fonts/ 35 - cp -f ${inter-fonts-src}/web/InterDisplay*.woff2 $out/static/fonts/ 36 - cp -f ${inter-fonts-src}/InterVariable*.ttf $out/static/fonts/ 37 - cp -f ${ibm-plex-mono-src}/fonts/complete/woff2/IBMPlexMono*.woff2 $out/static/fonts/ 38 - 39 - # styles 40 - cd ${src} && ${tailwindcss}/bin/tailwindcss -i input.css -o $out/stylesheet.css 41 - ''
+1 -1
tailwind.config.js
··· 2 2 const colors = require("tailwindcss/colors"); 3 3 4 4 module.exports = { 5 - content: ["./appview/pages/templates/**/*.html", "./appview/pages/chroma.go", "./docs/*.html"], 5 + content: ["./appview/pages/templates/**/*.html", "./appview/pages/chroma.go"], 6 6 darkMode: "media", 7 7 theme: { 8 8 container: {

History

3 rounds 1 comment
sign up or login to add to the discussion
3 commits
expand
feat(knotserver): add list repo endpoint
use lexgen
chore: revert to main go.mod
expand 1 comment

we have some reworks around this planned, along with refining our knotserver APIs, so ill be closing this in favor of future changes. one of the things is that we will be backfilling repo info in knots using tap, and subsequently, all XRPC APIs can speak ATURIs instead of did/repo-name.

i really do appreciate the effort gone into this!

closed without merging
2 commits
expand
feat(knotserver): add list repo endpoint
use lexgen
expand 0 comments
2 commits
expand
feat(knotserver): add list repo endpoint
use lexgen
expand 0 comments