Jiri Stribny has released a free, online, and modern command-line handbook aimed at both beginners and experienced users. The handbook covers a wide range of topics from basic navigation and file manipulation to more advanced concepts like shell scripting, process management, and using the command line effectively with cloud services like AWS. It focuses on practical examples and aims to be a comprehensive resource, updated for the current computing landscape, including discussions of newer tools and best practices. The handbook encourages interactive learning through built-in exercises and code examples that readers can experiment with directly in their terminal.
Brush is a new shell written in Rust, aiming for full POSIX compatibility and improved Bash compatibility. It leverages Rust's performance and safety features to create a potentially faster and more robust alternative to existing shells. While still in early development, Brush already supports many common shell features, including pipelines, globbing, and redirections. The project aims to eventually provide a drop-in replacement for Bash, offering a modern shell experience with improved performance and security.
HN commenters generally express excitement about Brush, praising its Rust implementation for potential performance and safety improvements over Bash. Several discuss the challenges of full Bash compatibility, particularly regarding corner cases and the complexities of parsing. Some suggest focusing on a smaller, cleaner subset of Bash functionality rather than striving for complete parity. Others raise concerns about potential performance overhead from Rust, especially regarding system calls, and question whether the benefits outweigh the costs. A few users mention looking forward to trying Brush, while others highlight similar projects like Ion and Nushell as alternative Rust-based shells. The maintainability of a complex project like a shell written in Rust is also discussed, with some expressing concerns about the long-term feasibility.
Linux in Excel demonstrates running a basic Linux system within a Microsoft Excel spreadsheet. Leveraging VBA scripting and x86 emulation, the project allows users to interact with a simplified Linux environment, complete with a command line interface, directly within Excel. It emulates a small subset of Linux system calls, enabling basic commands like ls
, cat
, and file manipulation within the spreadsheet's cells. While highly constrained and not a practical Linux replacement, it serves as a fascinating proof-of-concept, showcasing the flexibility of both VBA and the underlying architecture of Excel.
Hacker News users expressed both amusement and skepticism towards running Linux in Excel. Several commenters questioned the practicality and performance of such a setup, with some suggesting it's more of a novelty than a useful tool. Others were impressed by the technical feat, appreciating the ingenuity and creativity involved. Some discussed the potential for misuse, particularly in bypassing corporate security measures. There was also debate on whether this qualified as truly "running Linux," with some arguing it was merely simulating a limited environment. A few pointed out the historical precedent of running Doom in unexpected places, placing this project in a similar category of playful hacking.
Frustrated with the complexity and performance overhead of dynamic CMS platforms like WordPress, the author developed BSSG, a static site generator written entirely in Bash. Driven by a desire for simplicity, speed, and portability, they transitioned their website from WordPress to this custom solution. BSSG utilizes Pandoc for Markdown conversion and a templating system based on heredocs, offering a lightweight and efficient approach to website generation. The author emphasizes the benefits of this minimalist setup, highlighting improved site speed, reduced attack surface, and easier maintenance. While acknowledging potential limitations in features compared to full-fledged CMS platforms, they champion BSSG as a viable alternative for those prioritizing speed and simplicity.
HN commenters generally praised the author's simple, pragmatic approach to static site generation, finding it refreshing compared to more complex solutions. Several appreciated the focus on Bash scripting for its accessibility and ease of understanding. Some questioned the long-term maintainability and scalability of a Bash-based generator, suggesting alternatives like Python or Go for more complex sites. Others offered specific improvements, such as using rsync
for deployment and incorporating a templating engine. A few pointed out potential vulnerabilities in the provided code examples, particularly regarding HTML escaping. The overall sentiment leaned towards appreciation for the author's ingenuity and the project's minimalist philosophy.
This presentation provides a deep dive into advanced Bash scripting techniques. It covers crucial topics like regular expressions for pattern matching, utilizing built-in commands for string manipulation and file processing, and leveraging external utilities like sed
and awk
for more complex operations. The guide emphasizes practical scripting skills, demonstrating how to control program flow with loops and conditional statements, handle signals and traps for robust script behavior, and effectively manage variables and functions for modular and reusable code. It also delves into input/output redirection, process management, and here documents, equipping users to write powerful and efficient shell scripts for automating various system administration tasks.
HN commenters generally praise the linked Bash scripting guide for its clarity and comprehensiveness, especially regarding lesser-known features and best practices. Several highlight the sections on quoting and variable expansion as particularly valuable for avoiding common pitfalls. Some suggest the guide, while older, remains relevant for intermediate/advanced users looking to solidify their understanding. A few users mention alternative resources or offer minor critiques, such as the guide's lack of coverage on newer Bash features or the density of information, but the overall sentiment is positive, viewing the PDF as a valuable resource for improving Bash scripting skills. The mention of set -u
(nounset) to catch undefined variables is brought up multiple times as a crucial takeaway.
The author details their method for installing and managing personal versions of software on Unix systems, emphasizing a clean, organized approach. They create a dedicated directory within their home folder (e.g., ~/software
) to house all personally installed programs. Within this directory, each program gets its own subdirectory, containing the source code, build artifacts, and the compiled binaries. Critically, they manage dependencies by either statically linking them or bundling them within the program's directory. Finally, they modify their shell's PATH
environment variable to prioritize these personal installations over system-wide versions, enabling easy access and preventing conflicts. This method allows for running multiple versions of the same software concurrently and simplifies upgrading or removing personally installed programs.
HN commenters largely appreciate the author's approach of compiling and managing personal software installations in their home directory, praising it as clean, organized, and a good way to avoid dependency conflicts or polluting system directories. Several suggest using tools like stow
or GNU Stow for simplified management of this setup, allowing easy enabling/disabling of different software versions. Some discuss alternatives like Nix, Guix, or containers, offering more robust isolation. Others caution against potential downsides like increased compile times and the need for careful dependency management, especially for libraries. A few commenters mention difficulties encountered with specific tools or libraries in this type of personalized setup.
Dockerfmt is a command-line tool that automatically formats Dockerfiles, improving their readability and consistency. It restructures instructions, normalizes keywords, and adjusts indentation to adhere to best practices. The tool aims to eliminate manual formatting efforts and promote a standardized style across Dockerfiles, ultimately making them easier to maintain and understand. Dockerfmt is written in Go and can be installed as a standalone binary or used as a library.
HN users generally praised dockerfmt
for addressing a real need for Dockerfile formatting consistency. Several commenters appreciated the project's simplicity and ease of use, particularly its integration with gofmt
. Some raised concerns, including the potential for unwanted changes to existing Dockerfiles during formatting and the limited scope of the current linting capabilities, wishing for more comprehensive Dockerfile analysis. A few suggested potential improvements, such as options to ignore certain lines or files and integration with pre-commit hooks. The project's reliance on regular expressions for parsing also sparked discussion, with some advocating for a more robust parsing approach using a proper grammar. Overall, the reception was positive, with many seeing dockerfmt
as a useful tool despite acknowledging its current limitations.
Pico.sh offers developers instant, SSH-accessible Linux containers, pre-configured with popular development tools and languages. These containers act as personal servers, allowing developers to run web apps, databases, and background tasks without complex server management. Pico emphasizes simplicity and speed, providing a web-based terminal for direct access, custom domains, and built-in tools like Git, Docker, and various programming language runtimes. They aim to streamline the development workflow by eliminating the need for local setup and providing a consistent environment accessible from anywhere.
HN commenters generally expressed interest in Pico.sh, praising its simplicity and potential for streamlining development workflows. Several users appreciated the focus on SSH, viewing it as a secure and familiar access method. Some questioned the pricing model's long-term viability and compared it to similar services like Fly.io and Railway. The reliance on Tailscale for networking was both lauded for its ease of use and questioned for its potential limitations. A few commenters expressed concern about vendor lock-in, while others saw the open-source nature of the platform as mitigating that risk. The project's early stage was acknowledged, with some anticipating future features and improvements.
This blog post demonstrates how to efficiently integrate Large Language Models (LLMs) into bash scripts for automating text-based tasks. It leverages the curl
command to send prompts to LLMs via API, specifically using OpenAI's API as an example. The author provides practical examples of formatting prompts with variables and processing the JSON responses to extract desired text output. This allows for dynamic prompt generation and seamless integration of LLM-generated content into existing shell workflows, opening possibilities for tasks like code generation, text summarization, and automated report creation directly within a familiar scripting environment.
Hacker News users generally found the concept of using LLMs in bash scripts intriguing but impractical. Several commenters highlighted potential issues like rate limiting, cost, and the inherent unreliability of LLMs for tasks that demand precision. One compelling argument was that relying on an LLM for simple string manipulation or data extraction in bash is overkill when more robust and predictable tools like sed
, awk
, or jq
already exist. The discussion also touched upon the security implications of sending potentially sensitive data to an external LLM API and the lack of reproducibility in scripts relying on probabilistic outputs. Some suggested alternative uses for LLMs within scripting, such as generating boilerplate code or documentation.
Paul Samuels advocates for using simple, project-specific shell scripts instead of complex build systems or task runners for small to medium-sized projects. He argues that shell scripts offer better transparency, debuggability, and control, while reducing cognitive overhead. They facilitate easier understanding of project dependencies and build processes, which ultimately contributes to better maintainability, especially for solo developers or small teams. By leveraging the shell's built-in features and readily available Unix tools, project scripts provide a lightweight yet powerful approach to managing common development tasks.
Hacker News users generally praised the simplicity and practicality of "Project Scripts." Several commenters appreciated the lightweight nature of the approach compared to more complex build systems or dedicated project management tools, highlighting the benefit of reduced cognitive overhead. Some suggested potential improvements like incorporating direnv or using a Makefile for more complex projects. A few users expressed skepticism, arguing that the proposed "Project Scripts" offered little beyond basic shell scripting and questioned the need for a dedicated term. Others found the idea valuable for its focus on explicitness and ease of sharing project setup within a team. The discussion also touched on related tools like Taskfile and justfile, comparing their features and complexity to the author's approach.
This blog post details how to automatically remove macOS-specific files (.DS_Store
and ._*
) from external drives upon ejection. The author uses a combination of AppleScript and a LaunchAgent to trigger a cleanup script whenever a volume is ejected. The script leverages dot_clean
to efficiently delete these often-annoying hidden files, preventing their proliferation on non-macOS systems. This automated approach replaces the need for manual cleanup and ensures a cleaner experience when sharing drives between different operating systems.
Commenters on Hacker News largely appreciated the simplicity and directness of the provided AppleScript solution for removing macOS-specific files from external drives upon ejection. Some highlighted the potential for data loss if used carelessly, especially with networked drives or if the script were modified to delete different files. Others offered alternative solutions, including using dot_clean
, incorporating the script into a Hazel rule, or employing a shell script with find
. The discussion also touched upon the annoyance factor of these files on other operating systems and the historical reasons for their existence, with some suggesting that their prevalence has diminished. A few commenters mentioned more robust solutions for syncing and backing up, which would obviate the need for such a script altogether.
"Do-nothing scripting" advocates for a gradual approach to automation. Instead of immediately trying to fully automate a complex task, you start by writing a script that simply performs the steps manually, echoing each command to the screen. This allows you to document the process precisely and identify potential issues without the risk of automated errors. As you gain confidence, you incrementally replace the manual execution of each command within the script with its automated equivalent. This iterative process minimizes disruption, allows for easy rollback, and makes the transition to full automation smoother and more manageable.
Hacker News users generally praised the "do-nothing scripting" approach as a valuable tool for understanding existing processes before automating them. Several commenters highlighted the benefit of using this technique to gain stakeholder buy-in and build trust, particularly when dealing with complex or mission-critical systems. Some shared similar experiences or suggested alternative methods like using strace
or dtrace
. One commenter suggested incorporating progressive logging to further refine the script's insights over time, while another cautioned against over-reliance on this approach, advocating for a move towards true automation once sufficient understanding is gained. Some skepticism was expressed regarding the practicality for highly interactive processes. Overall, the commentary reflects strong support for the core idea as a practical step toward thoughtful and effective automation.
Milo Fultz's blog post details a method for finding the oldest lines of code in a Git repository. The approach leverages git blame
combined with awk
and sort
to extract commit dates and line numbers. By sorting the output based on these dates, the script identifies and displays the oldest surviving lines, effectively pinpointing code that has remained unchanged since its initial introduction. This technique can be useful for understanding the evolution of a codebase, identifying potential legacy code, or simply satisfying curiosity about a project's history.
Hacker News users discussed various methods and tools for finding the oldest lines of code in a repository, expanding on the article's git blame
approach. Several commenters suggested using git log -L
for more precise tracking of specific lines or functions, highlighting its ability to handle code moves and rewrites. The practicality of such analysis was debated, with some arguing its usefulness for understanding legacy code and identifying potential refactoring targets, while others questioned its value beyond curiosity. Alternatives like git-quick-stats
and commercial tools like CodeScene were also mentioned for broader code history analysis, including visualizing code churn and developer contributions over time. The potential pitfalls of relying solely on line age were also brought up, emphasizing the importance of considering code quality and functionality regardless of its age.
Shunpo is a minimalist Bash tool designed to streamline directory navigation. It learns frequently visited directories and allows users to quickly jump to them using short, custom aliases. By storing these aliases and their corresponding paths in a simple text file, Shunpo avoids complex databases and remains lightweight and portable. It offers basic commands for adding, removing, listing, and navigating to saved locations, simplifying the process of moving between commonly accessed folders within the terminal.
Hacker News users discussed Shunpo's utility and potential drawbacks. Some found its core functionality—quickly jumping to frequently used directories—appealing, especially combined with tools like fzf. Others questioned its value proposition over existing solutions like autojump, z, or fasd, particularly given its reliance on find
. Concerns were raised about performance in large directory trees and the security implications of executing arbitrary commands generated from find
results. Some suggested improvements, including leveraging shell builtins for better performance and integrating more advanced selection mechanisms. The project's minimalism was both praised and criticized, with some appreciating its simplicity and others desiring more features like directory tracking or the ability to ignore certain paths.
Bunster is a tool that compiles Bash scripts into standalone, statically-linked executables. This allows for easy distribution and execution of Bash scripts without requiring a separate Bash installation on the target system. It achieves this by embedding a minimal Bash interpreter and necessary dependencies within the generated executable. This makes scripts more portable and user-friendly, especially for scenarios where installing dependencies or ensuring a specific Bash version is impractical.
Hacker News users discussed Bunster's novel approach to compiling Bash scripts, expressing interest in its potential while also raising concerns. Several questioned the practical benefits over existing solutions like shc
or containers, particularly regarding dependency management and debugging complexity. Some highlighted the inherent limitations of Bash as a scripting language compared to more robust alternatives for complex applications. Others appreciated the project's ingenuity and suggested potential use cases like simplifying distribution of simple scripts or bypassing system-level restrictions on scripting. The discussion also touched upon the performance implications of this compilation method and the challenges of handling Bash's dynamic nature. A few commenters expressed curiosity about the inner workings of the compilation process and its handling of external commands.
Tabby is a self-hosted AI coding assistant designed to enhance programming productivity. It offers code completion, generation, translation, explanation, and chat functionality, all within a secure local environment. By leveraging large language models like StarCoder and CodeLlama, Tabby provides powerful assistance without sharing code with external servers. It's designed to be easily installed and customized, offering both a desktop application and a VS Code extension. The project aims to be a flexible and private alternative to cloud-based AI coding tools.
Hacker News users discussed Tabby's potential, limitations, and privacy implications. Some praised its self-hostable nature as a key advantage over cloud-based alternatives like GitHub Copilot, emphasizing data security and cost savings. Others questioned its offline performance compared to online models and expressed skepticism about its ability to truly compete with more established tools. The practicality of self-hosting a large language model (LLM) for individual use was also debated, with some highlighting the resource requirements. Several commenters showed interest in using Tabby for exploring and learning about LLMs, while others were more focused on its potential as a practical coding assistant. Concerns about the computational costs and complexity of setup were common threads. There was also some discussion comparing Tabby to similar projects.
This project demonstrates a surprisingly functional 3D raycaster engine implemented entirely within a Bash script. By cleverly leveraging ASCII characters and terminal output manipulation, it renders a simple maze-like environment in pseudo-3D. The script calculates ray intersections with walls and represents distances with varying shades of characters, creating a surprisingly immersive experience given the limitations of the medium. While performance is understandably limited, it showcases the flexibility and unexpected capabilities of Bash beyond typical scripting tasks.
Hacker News users discuss the ingenuity and limitations of a bash raycaster. Several express admiration for the project's creativity, highlighting the unexpected capability of bash for such a task. Some commenters delve into the technical details, discussing the clever use of shell built-ins and the performance implications of using bash for computationally intensive tasks. Others point out that the "raycasting" is actually a 2.5D projection technique and not true raycasting. The novelty of the project and its demonstration of bash's flexibility are the main takeaways, though its practicality is questioned. Some users also shared links to similar projects in other unexpected languages.
Summary of Comments ( 16 )
https://news.ycombinator.com/item?id=44126612
HN commenters largely praised the Command Line Handbook for its modern approach, covering newer tools and techniques omitted from older resources. Several appreciated the inclusion of practical examples and the focus on interactive use. Some suggested additions, including coverage of specific tools like
jq
,fzf
, andripgrep
, more detail on shell scripting, and explanations of underlying concepts like the filesystem hierarchy. A few pointed out minor typos or formatting inconsistencies. The overall sentiment was highly positive, with many expressing their intent to use the handbook themselves or recommend it to others.The Hacker News post "Show HN: I wrote a modern Command Line Handbook" at https://news.ycombinator.com/item?id=44126612 generated a moderate amount of discussion, with several commenters expressing appreciation for the resource and offering suggestions for improvement.
One of the most compelling comments highlighted the importance of focusing on the conceptual underpinnings of the command line, rather than just presenting a collection of commands. This commenter advocated for explaining concepts like standard input/output/error, pipes, and redirection early on, emphasizing that understanding these fundamental concepts is crucial for effectively utilizing the command line. They argued that mastering these principles allows users to extrapolate and combine commands creatively, rather than just memorizing individual commands.
Another user suggested incorporating information about shell scripting, particularly using Bash or Zsh. They believed that introducing basic scripting could significantly enhance the handbook's value by empowering users to automate tasks and create more powerful workflows.
Several commenters praised the handbook's clear and concise writing style, finding it accessible and easy to understand. The modern focus, including coverage of tools like
fd
,ripgrep
, andfzf
, was also well-received.Some suggestions for improvement included adding more interactive elements, such as exercises or quizzes, to reinforce learning. Another commenter requested a section on common command-line utilities for tasks like managing archives, working with images, and performing system administration tasks.
One user expressed a desire for more platform-specific information, particularly for Windows users, recognizing that many command-line tools behave differently across operating systems. They pointed out the growing popularity of Windows Subsystem for Linux (WSL) and suggested including guidance on using it effectively.
Finally, a few comments pointed out minor typos and formatting issues, demonstrating the community's engagement with the project and their willingness to contribute to its improvement. Overall, the comments reflect a positive reception to the handbook and provide valuable feedback for its continued development.