The author details their process of creating a WebAssembly (Wasm) virtual machine (VM) written entirely in C. Driven by a desire for a lightweight, embeddable Wasm runtime for resource-constrained environments, they built the VM from scratch, implementing core features like the stack-based execution model, linear memory, and basic WebAssembly System Interface (WASI) support. The project focused on simplicity and understandability over performance, serving primarily as a learning exercise and a platform for experimentation with Wasm. The post walks through key aspects of the VM's design and implementation, including parsing the Wasm binary format, handling function calls, and managing memory. It also highlights the challenges faced and lessons learned during the development process.
The blog post details methods for eliminating left and mutual recursion in context-free grammars, crucial for parser construction. Left recursion, where a non-terminal derives itself as the leftmost symbol, is problematic for top-down parsers. The post demonstrates how to remove direct left recursion using factorization and substitution. It then explains how to handle indirect left recursion by ordering non-terminals and systematically applying the direct recursion removal technique. Finally, it addresses mutual recursion, where two or more non-terminals derive each other, converting it into direct left recursion, which can then be eliminated using the previously described methods. The post uses concrete examples to illustrate these transformations, making it easier to understand the process of converting a grammar into a parser-friendly form.
Hacker News users discussed the potential inefficiency of the presented left-recursion elimination algorithm, particularly its reliance on repeated string concatenation. They suggested alternative approaches using stacks or accumulating results in a list for better performance. Some commenters questioned the necessity of fully eliminating left recursion in all cases, pointing out that modern parsing techniques, like packrat parsing, can handle left-recursive grammars directly. The lack of formal proofs or performance comparisons with established methods was also noted. A few users discussed the benefits and drawbacks of different parsing libraries and techniques, including ANTLR and various parser combinator libraries.
plrust is a PostgreSQL extension that allows developers to write stored procedures and functions in Rust. It leverages the PostgreSQL procedural language handler framework and offers safe, performant execution within the database. By compiling Rust code into shared libraries, plrust provides direct access to PostgreSQL internals and avoids the overhead of external processes or interpreters. This allows developers to harness Rust's speed and safety for complex database tasks while integrating seamlessly with existing PostgreSQL infrastructure.
HN users discuss the complexities and potential benefits of writing PostgreSQL extensions in Rust. Several express interest in the project (plrust), citing Rust's performance advantages and memory safety as key motivators for moving away from C. Concerns are raised about the overhead of crossing the FFI boundary between Rust and PostgreSQL, and the potential difficulties in debugging. Some commenters suggest comparing plrust's performance to existing solutions like PL/pgSQL and C extensions, while others highlight the potential for improved developer experience and safety that Rust offers. The maintainability of generated Rust code from PostgreSQL queries is also questioned. Overall, the comments reflect cautious optimism about plrust's potential, tempered by a pragmatic awareness of the challenges involved in integrating Rust into the PostgreSQL ecosystem.
Astral is a new static type checker being developed for Python that aims to be faster and more ergonomic than existing options like MyPy. It leverages a new type inference algorithm designed for performance and boasts features like auto-completion, goto-definition, and an improved developer experience. The project is still early in development but claims significant speed improvements, with a goal of being at least 5x faster than MyPy on real-world codebases. Astral also intends to offer seamless integration with existing Python tooling and provide enhanced support for popular libraries like NumPy and Pandas.
Hacker News users discuss Astral's potential, drawing parallels to MyPy but with a focus on performance. Some express skepticism about static typing in Python, questioning its necessity and impact on the language's flexibility. Others are interested in Astral's approach to gradual typing and its ability to handle complex codebases. Performance improvements over MyPy are frequently mentioned as a key benefit. Several commenters inquire about specific features, such as handling metaclasses and integration with existing tools. Overall, there's a mix of cautious optimism and interest in seeing how Astral develops.
Preserves is a new data language designed for clarity and expressiveness, aiming to bridge the gap between simple configuration formats like JSON/YAML and full-fledged programming languages. It focuses on data transformation and manipulation with a concise syntax inspired by functional programming. Key features include immutability, a type system emphasizing structural types, built-in support for common data structures like maps and lists, and user-defined functions for more complex logic. The project aims to offer a powerful yet approachable tool for tasks ranging from simple configuration to data processing and analysis, especially where maintainability and readability are paramount.
Hacker News users discussed Preserves' potential, comparing it to tools like JSON, YAML, TOML, and edn. Some lauded its expressiveness, particularly its support for comments and arbitrary keys. Others questioned its practical value beyond configuration files, wondering about performance, tooling, and whether its added complexity justified the benefits over simpler formats. The lack of a formal specification was also a concern. Several commenters expressed interest in seeing real-world use cases and benchmarks to better assess Preserves' viability. Some saw potential for niche applications like game modding or creative coding, while others remained skeptical about its broad adoption. The discussion highlighted the trade-off between expressiveness and simplicity in data languages.
Go 1.24's revamped go
tool significantly streamlines dependency management and build processes. By embedding version information directly within the go.mod
file and leveraging a content-addressable file system (CAS), builds become more reproducible and efficient. This eliminates the need for separate go.sum
files and simplifies workflows, especially in environments with limited network access. The improved tooling allows developers to more easily vendor dependencies, create reproducible builds across different machines, and share builds efficiently, making it a major improvement for the Go ecosystem.
HN users largely agree that the go
tool improvements in 1.24 are significant and welcome. Several commenters highlight the improved dependency management as a major win, specifically the reduced verbosity and simplified workflow when adding, updating, or vending dependencies. Some express appreciation for the enhanced transparency, allowing developers to more easily understand the tool's actions. A few users note that the improvements bring Go's tooling closer to the experience offered by other languages like Rust's Cargo. There's also discussion around the specific benefits of lazy loading, minimal version selection (MVS), and the implications for package management within monorepos. While largely positive, some users mention lingering minor frustrations or express curiosity about further planned improvements.
The blog post explores building a composable SQL query builder in Haskell using the concept of functors. Instead of relying on string concatenation, which is prone to SQL injection vulnerabilities, it leverages Haskell's type system and the Functor
typeclass to represent SQL fragments as data structures. These fragments can then be safely combined and transformed using pure functions. The approach allows for building complex queries piece by piece, abstracting away the underlying SQL syntax and promoting code reusability. This results in a more type-safe, maintainable, and composable way to generate SQL queries compared to traditional string-based methods.
HN commenters generally appreciate the composability approach to SQL queries presented in the article, finding it cleaner and more maintainable than traditional string concatenation. Several highlight the similarity to functional programming concepts and appreciate the use of Python's type hinting. Some express concern about performance implications, particularly with nested queries, and suggest comparing it to ORMs. Others question the practicality for complex queries or the necessity for simpler ones. A few users mention existing libraries with similar functionality, like SQLAlchemy Core. The discussion also touches upon alternative approaches like using CTEs (Common Table Expressions) for composability and the potential benefits for testing and debugging.
The blog post "The Simplicity of Prolog" argues that Prolog's declarative nature makes it easier to learn and use than imperative languages for certain problem domains. It demonstrates this by building a simple genealogy program in Prolog, highlighting how its concise syntax and built-in search mechanism naturally express relationships and deduce facts. The author contrasts this with the iterative loops and explicit state management required in imperative languages, emphasizing how Prolog abstracts away these complexities. The post concludes that while Prolog may not be suitable for all tasks, its elegant approach to logic programming offers a powerful and efficient solution for problems involving knowledge representation and inference.
Hacker News users generally praised the article for its clear introduction to Prolog, with several noting its effectiveness in sparking their own interest in the language. Some pointed out Prolog's historical significance and its continued relevance in specific domains like AI and knowledge representation. A few users highlighted the contrast between Prolog's declarative approach and the more common imperative style of programming, emphasizing the shift in mindset required to effectively use it. Others shared personal anecdotes of their experiences with Prolog, both positive and negative, with some mentioning its limitations in performance-critical applications. A couple of comments also touched on the learning curve associated with Prolog and the challenges in debugging complex programs.
Mukul Rathi details his journey of creating a custom programming language, focusing on the compiler construction process. He explains the key stages involved, from lexing (converting source code into tokens) and parsing (creating an Abstract Syntax Tree) to code generation and optimization. Rathi uses his language, which he implements in OCaml, to illustrate these concepts, providing code examples and explanations of how each component works together to transform high-level code into executable machine instructions. He emphasizes the importance of understanding these foundational principles for anyone interested in building their own language or gaining a deeper appreciation for how programming languages function.
Hacker News users generally praised the article for its clarity and accessibility in explaining compiler construction. Several commenters appreciated the author's approach of building a complete, albeit simple, language instead of just a toy example. Some pointed out the project's similarity to the "Let's Build a Compiler" series, while others suggested alternative or supplementary resources like Crafting Interpreters and the LLVM tutorial. A few users discussed the tradeoffs between hand-written lexers/parsers and using parser generator tools, and the challenges of garbage collection implementation. One commenter shared their personal experience of writing a language and the surprising complexity of seemingly simple features.
Yasser is developing "Tilde," a new compiler infrastructure designed as a simpler, more modular alternative to LLVM. Frustrated with LLVM's complexity and monolithic nature, he's building Tilde with a focus on ease of use, extensibility, and better diagnostics. The project is in its early stages, currently capable of compiling a subset of C and targeting x86-64 Linux. Key differentiating features include a novel intermediate representation (IR) designed for efficient analysis and transformation, a pipeline architecture that facilitates experimentation and customization, and a commitment to clear documentation and a welcoming community. While performance isn't the primary focus initially, the long-term goal is to be competitive with LLVM.
Hacker News users discuss the author's approach to building a compiler, "Tilde," positioned as an LLVM alternative. Several commenters express skepticism about the project's practicality and scope, questioning the rationale behind reinventing LLVM, especially given its maturity and extensive community. Some doubt the performance claims and suggest benchmarks are needed. Others appreciate the author's ambition and the technical details shared, seeing value in exploring alternative compiler designs even if Tilde doesn't replace LLVM. A few users offer constructive feedback on specific aspects of the compiler's architecture and potential improvements. The overall sentiment leans towards cautious interest with a dose of pragmatism regarding the challenges of competing with an established project like LLVM.
The author argues that Go's context.Context
is overused and often misused as a dumping ground for arbitrary values, leading to unclear dependencies and difficult-to-test code. Instead of propagating values through Context
, they propose using explicit function parameters, promoting clearer code, better separation of concerns, and easier testability. They contend that using Context
primarily for cancellation and timeouts, its intended purpose, would streamline code and improve its maintainability.
HN commenters largely agree with the author's premise that context.Context
in Go is overused and often misused for dependency injection or as a dumping ground for miscellaneous values. Several suggest that structured concurrency, improved error handling, and better language features for cancellation and deadlines could alleviate the need for context
in many cases. Some argue that context
is still useful for request-scoped values, especially in server contexts, and shouldn't be entirely removed. A few commenters express concern about the practicality of removing context
given its widespread adoption and integration into the standard library. There is a strong desire for better alternatives, rather than simply discarding the existing mechanism without a replacement. Several commenters also mention the similarities between context
overuse in Go and similar issues with dependency injection frameworks in other languages.
The Hacker News post discusses whether any programming languages allow specifying package dependencies directly within import or include statements, rather than separately in a dedicated dependency management file. The original poster highlights the potential benefits of this approach, such as improved clarity and ease of understanding dependencies for individual files. They suggest a syntax where version numbers or constraints could be incorporated into the import statement itself. While no existing mainstream languages seem to offer this feature, some commenters mention related concepts like import maps in JavaScript and conditional imports in some languages. The core idea is to make dependency management more localized and transparent at the file level.
The Hacker News comments discuss the pros and cons of specifying package requirements directly within import statements. Several commenters appreciate the clarity and explicitness this would bring, as it makes dependencies immediately obvious and reduces the need for separate dependency management files. Others argue against it, citing potential drawbacks like redundancy, increased code verbosity, and difficulties managing complex dependency graphs. Some propose alternative solutions, like embedding version requirements in comments or using language-specific mechanisms for dependency specification. A few commenters mention existing languages or tools that offer similar functionality, such as Nix and Dhall, pointing to these as potential examples or inspiration for how such a system could work. The discussion also touches on the practical implications for tooling and build systems, with commenters considering the impact on IDE integration and compilation processes.
"Alligator Eggs" explores the surprising computational power hidden within a simple system of rewriting strings. Inspired by a children's puzzle involving moving colored eggs, the post demonstrates how a carefully designed set of rules for replacing egg sequences can emulate the functionality of a Turing Machine, a theoretical model capable of performing any computation. By encoding logic and data within the arrangement of the eggs, the system can execute arbitrary programs, effectively turning a seemingly trivial game into a universal computer. The post emphasizes the elegance and minimalism of this computational model, highlighting how complex behavior can emerge from simple, well-defined rules.
HN users generally praised the clarity and approachability of Bret Victor's explanation of lambda calculus, with several highlighting its effectiveness as an introductory resource even for those without a strong math background. Some discussed the challenges of teaching and visualizing these concepts, appreciating Victor's interactive approach. A few commenters delved into more technical nuances, comparing lambda calculus to combinatory logic and touching upon topics like currying and the SKI calculus. Others reminisced about learning from similar resources in the past and shared related links, demonstrating the article's enduring relevance. A recurring theme was the power of visual and interactive learning tools in making complex topics more accessible.
Justine Tunney's "Lambda Calculus in 383 Bytes" presents a remarkably small, self-hosting Lambda Calculus interpreter written in x86-64 assembly. It parses, evaluates, and prints lambda expressions, supporting variables, application, and abstraction using a custom encoding. Despite its tiny size, the interpreter implements a complete, albeit slow, evaluation strategy by translating lambda terms into De Bruijn indices and employing normal order reduction. The project showcases the minimal computational requirements of lambda calculus and the power of concise, low-level programming.
Hacker News users discuss the cleverness and efficiency of the 383-byte lambda calculus implementation, praising its conciseness and educational value. Some debate the practicality of such a minimal implementation, questioning its performance and highlighting the trade-offs made for size. Others delve into technical details, comparing it to other small language implementations and discussing optimization strategies. Several comments point out the significance of understanding lambda calculus fundamentals and appreciate the author's clear explanation and accompanying code. A few users express interest in exploring similar projects and adapting the code for different architectures. The overall sentiment is one of admiration for the technical feat and its potential as a learning tool.
This blog post explores a simplified variant of Generalized LR (GLR) parsing called "right-nulled" GLR. Instead of maintaining a graph-structured stack during parsing ambiguities, this technique uses a single stack and resolves conflicts by prioritizing reduce actions over shift actions. When a conflict occurs, the parser performs all possible reductions before attempting to shift. This approach sacrifices some of GLR's generality, as it cannot handle all types of grammars, but it significantly reduces the complexity and overhead associated with maintaining the graph-structured stack, leading to a faster and more memory-efficient parser. The post provides a conceptual overview, highlights the limitations compared to full GLR, and demonstrates the algorithm with a simple example.
Hacker News users discuss the practicality and efficiency of GLR parsing, particularly in comparison to other parsing techniques. Some commenters highlight its theoretical power and ability to handle ambiguous grammars, while acknowledging its potential performance overhead. Others question its suitability for real-world applications, suggesting that simpler methods like PEG or recursive descent parsers are often sufficient and more efficient. A few users mention specific use cases where GLR parsing shines, such as language servers and situations requiring robust error recovery. The overall sentiment leans towards appreciating GLR's theoretical elegance but expressing reservations about its widespread adoption due to perceived complexity and performance concerns. A recurring theme is the trade-off between parsing power and practical efficiency.
This paper introduces Crusade, a formally verified translation from a subset of C to safe Rust. Crusade targets a memory-safe dialect of C, excluding features like arbitrary pointer arithmetic and casts. It leverages the Coq proof assistant to formally verify the translation's correctness, ensuring that the generated Rust code behaves identically to the original C, modulo non-determinism inherent in C. This rigorous approach aims to facilitate safe integration of legacy C code into Rust projects without sacrificing confidence in memory safety, a critical aspect of modern systems programming. The translation handles a substantial subset of C, including structs, unions, and functions, and demonstrates its practical applicability by successfully converting real-world C libraries.
HN commenters discuss the challenges and nuances of formally verifying the C to Rust transpiler, Cracked. Some express skepticism about the practicality of fully verifying such a complex tool, citing the potential for errors in the formal proofs themselves and the inherent difficulty of capturing all undefined C behavior. Others question the performance impact of the generated Rust code. However, many commend the project's ambition and see it as a significant step towards safer systems programming. The discussion also touches upon the trade-offs between a fully verified transpiler and a more pragmatic approach focusing on common C patterns, with some suggesting that prioritizing practical safety improvements could be more beneficial in the short term. There's also interest in the project's handling of concurrency and the potential for integrating Cracked with existing Rust tooling.
This blog post explores the powerful concept of functions as the fundamental building blocks of computation, drawing insights from the book Structure and Interpretation of Computer Programs (SICP) and David Beazley's work. It illustrates how even seemingly complex structures like objects and classes can be represented and implemented using functions, emphasizing the elegance and flexibility of this approach. The author demonstrates building a simple object system solely with functions, highlighting closures for managing state and higher-order functions for method dispatch. This functional perspective provides a deeper understanding of object-oriented programming and showcases the unifying power of functions in expressing diverse programming paradigms. By breaking down familiar concepts into their functional essence, the post encourages a more fundamental and adaptable approach to software design.
Hacker News users discuss the transformative experience of learning Scheme and SICP, particularly under David Beazley's tutelage. Several commenters emphasize the power of Beazley's teaching style, highlighting his ability to simplify complex concepts and make them engaging. Some found the author's surprise at the functional paradigm's elegance noteworthy, with one suggesting that other languages like Python and Javascript offer similar functional capabilities, perhaps underappreciated by the author. Others debated the benefits and drawbacks of "pure" functional programming, its practicality in real-world projects, and the learning curve associated with Scheme. A few users also shared their own positive experiences with SICP and its impact on their understanding of computer science fundamentals. The overall sentiment reflects an appreciation for the article's insights and the enduring relevance of SICP in shaping programmers' perspectives.
This blog post explores using Go's strengths for web service development while leveraging Python's rich machine learning ecosystem. The author details a "sidecar" approach, where a Go web service communicates with a separate Python process responsible for ML tasks. This allows the Go service to handle routing, request processing, and other web-related functionalities, while the Python sidecar focuses solely on model inference. Communication between the two is achieved via gRPC, chosen for its performance and cross-language compatibility. The article walks through the process of setting up the gRPC connection, preparing a simple ML model in Python using scikit-learn, and implementing the corresponding Go service. This architectural pattern isolates the complexity of the ML component and allows for independent scaling and development of both the Go and Python parts of the application.
HN commenters discuss the practicality and performance implications of the Python sidecar approach for ML in Go. Some express skepticism about the added complexity and overhead, suggesting gRPC or REST might be overkill for simple tasks and questioning the performance benefits compared to pure Python or using GoML libraries directly. Others appreciate the author's exploration of different approaches and the detailed benchmarks provided. The discussion also touches on alternative solutions like using shared memory or embedding Python in Go, as well as the broader topic of language interoperability for ML tasks. A few comments mention specific Go ML libraries like gorgonia/tensor as potential alternatives to the sidecar approach. Overall, the consensus seems to be that while interesting, the sidecar approach may not be the most efficient solution in many cases, but could be valuable in specific circumstances where existing Go ML libraries are insufficient.
Summary of Comments ( 43 )
https://news.ycombinator.com/item?id=42918524
Hacker News users generally praised the author's clear writing style and the educational value of the post. Several commenters discussed the project's performance, noting that it's not optimized for speed and suggesting potential improvements like just-in-time compilation. Some shared their own experiences with WASM interpreters and related projects, including comparisons to other implementations and alternative approaches like using a stack machine. Others appreciated the detailed explanation of the parsing and execution process, finding it helpful for understanding WASM internals. A few users pointed out minor corrections or areas for potential enhancement in the code, demonstrating active engagement with the technical details.
The Hacker News post "I Wrote a WebAssembly VM in C" (https://news.ycombinator.com/item?id=42918524) generated a moderate amount of discussion, with several commenters engaging with the project and offering insights or related experiences.
A recurring theme was admiration for the author's undertaking, with several commenters acknowledging the complexity and difficulty of writing a Wasm VM. One commenter pointed out the educational value of such projects, emphasizing the deep understanding of Wasm's internals that one gains through implementation. They also noted that while Wasm is often perceived as a compilation target, understanding its runtime environment is equally crucial.
Another user shared a personal anecdote of a similar project, where they wrote a Wasm interpreter in Rust. They explained that their motivation stemmed from a need to run Wasm in a constrained embedded environment lacking a JIT compiler. This comment highlighted a practical use case for Wasm interpreters, contrasting with the more common JIT-based implementations.
A discussion unfolded about the performance characteristics of interpreted Wasm versus compiled Wasm. One commenter questioned the practical applicability of interpreters, speculating that their performance limitations might restrict their usefulness. Another user countered this by suggesting potential niche applications, such as debugging or educational purposes, where raw performance is less critical than other features like understandability and control. They also mentioned the possibility of using an interpreter as a fallback mechanism when JIT compilation is unavailable.
The author of the Wasm VM chimed in to address some of these questions. They clarified that the project was primarily an educational exercise, not intended for production use. They acknowledged the performance limitations of interpretation and confirmed they had no plans to add a JIT compiler. They also engaged with other commenters, discussing technical details of their implementation, such as the handling of garbage collection.
Finally, one comment drew a parallel between the author's project and the early days of Java, where interpreted execution was common before JIT compilation became prevalent. This comparison highlighted the potential evolution of Wasm runtimes, suggesting that interpreters might play a more significant role in the future, particularly in resource-constrained environments.