Compare commits

...

3 Commits

Author SHA1 Message Date
thomas db41ec6e93 pi settings 2026-03-19 15:20:50 +00:00
thomas c44420ce7c osc52 or whatever for ssh clipboard 2026-03-19 15:20:27 +00:00
thomas f74242ed02 jj rules 2026-03-19 15:20:15 +00:00
5 changed files with 31 additions and 28 deletions
+23 -21
View File
@@ -51,27 +51,29 @@ return {
-- end,
-- desc = "Git Blame Line",
-- },
{
"<leader>gf",
function()
Snacks.lazygit.log_file()
end,
desc = "Lazygit Current File History",
},
{
"<leader>lg",
function()
Snacks.lazygit()
end,
desc = "Lazygit",
},
{
"<leader>gl",
function()
Snacks.lazygit.log()
end,
desc = "Lazygit Log (cwd)",
},
--
-- Commented out LazyGit in favor of separated jj
-- {
-- "<leader>gf",
-- function()
-- Snacks.lazygit.log_file()
-- end,
-- desc = "Lazygit Current File History",
-- },
-- {
-- "<leader>lg",
-- function()
-- Snacks.lazygit()
-- end,
-- desc = "Lazygit",
-- },
-- {
-- "<leader>gl",
-- function()
-- Snacks.lazygit.log()
-- end,
-- desc = "Lazygit Log (cwd)",
-- },
{
"<leader>dn",
function()
+6
View File
@@ -89,6 +89,12 @@ vim.o.confirm = true
-- vim.o.winborder = "rounded"
-- Clipboard: prefer system clipboard, and over SSH use OSC52 so yanks can reach local clipboard
vim.opt.clipboard = "unnamedplus"
if vim.env.SSH_TTY then
vim.g.clipboard = "osc52"
end
-- Highlight text on yank
vim.api.nvim_create_autocmd("TextYankPost", {
callback = function()
+1 -1
View File
@@ -1,7 +1,7 @@
{
"activePack": "glados",
"volume": 1,
"muted": false,
"muted": true,
"enabledCategories": {
"session.start": true,
"task.acknowledge": true,
@@ -21,11 +21,6 @@
"file": "sounds/IKnowYoureThere.mp3",
"label": "I know you're there. I can feel you here.",
"sha256": "df3780607b7a480fd3968c8aae5e0a397ea956008a5c7a47fecb887a05d61622"
},
{
"file": "sounds/HelloImbecile.mp3",
"label": "Hello, imbecile!",
"sha256": "dd10461e79bb4b1319f436cef5f0541f18a9505638824a6e765b9f2824a3380f"
}
]
},
+1 -1
View File
@@ -1,5 +1,5 @@
{
"lastChangelogVersion": "0.57.1",
"lastChangelogVersion": "0.60.0",
"defaultProvider": "openai-codex",
"defaultModel": "gpt-5.3-codex",
"defaultThinkingLevel": "medium",