By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
AIModelKitAIModelKitAIModelKit
  • Home
  • News
    NewsShow More
    OpenAI Acquires AI Personal Finance Startup Hiro: What This Means for the Future
    OpenAI Acquires AI Personal Finance Startup Hiro: What This Means for the Future
    5 Min Read
    Microsoft Develops New OpenClaw-like AI Agent: What to Expect
    Microsoft Develops New OpenClaw-like AI Agent: What to Expect
    4 Min Read
    Microsoft Tests OpenClaw-Inspired AI Bots for Enhanced Copilot Functionality
    Microsoft Tests OpenClaw-Inspired AI Bots for Enhanced Copilot Functionality
    4 Min Read
    How Companies Are Expanding AI Adoption While Maintaining Control
    How Companies Are Expanding AI Adoption While Maintaining Control
    6 Min Read
    Explore the World’s Largest Orbital Compute Cluster Now Open for Business
    Explore the World’s Largest Orbital Compute Cluster Now Open for Business
    6 Min Read
  • Open-Source Models
    Open-Source ModelsShow More
    Pioneering the Future of Computer Use: Expanding Digital Frontiers
    Pioneering the Future of Computer Use: Expanding Digital Frontiers
    5 Min Read
    Protecting Cryptocurrency: How to Responsibly Disclose Quantum Vulnerabilities
    Protecting Cryptocurrency: How to Responsibly Disclose Quantum Vulnerabilities
    4 Min Read
    Boosting AI and XR Prototyping Efficiency with XR Blocks and Gemini
    Boosting AI and XR Prototyping Efficiency with XR Blocks and Gemini
    5 Min Read
    Transforming News Reports into Data Insights with Gemini: A Comprehensive Guide
    Transforming News Reports into Data Insights with Gemini: A Comprehensive Guide
    6 Min Read
    Enhancing Urban Safety: AI-Powered Flash Flood Forecasting Solutions for Cities
    Enhancing Urban Safety: AI-Powered Flash Flood Forecasting Solutions for Cities
    5 Min Read
  • Guides
    GuidesShow More
    Master Python Continuous Integration and Deployment with GitHub Actions: Take the Real Python Quiz
    Master Python Continuous Integration and Deployment with GitHub Actions: Take the Real Python Quiz
    3 Min Read
    Exploring the Role of Data Generalists: Why Range is More Important than Depth
    Exploring the Role of Data Generalists: Why Range is More Important than Depth
    6 Min Read
    Master Python Protocols: Take the Ultimate Quiz with Real Python
    Master Python Protocols: Take the Ultimate Quiz with Real Python
    4 Min Read
    Mastering Input and Output in Python: Quiz from Real Python
    Mastering Input and Output in Python: Quiz from Real Python
    3 Min Read
    Mastering Python Logging: Simplify Your Workflow with Loguru – A Real Python Guide
    Mastering Python Logging: Simplify Your Workflow with Loguru – A Real Python Guide
    4 Min Read
  • Tools
    ToolsShow More
    Safetensors Partners with PyTorch Foundation: Strengthening AI Development
    Safetensors Partners with PyTorch Foundation: Strengthening AI Development
    5 Min Read
    High Throughput Computer Use Agent: Understanding 12B for Optimal Performance
    High Throughput Computer Use Agent: Understanding 12B for Optimal Performance
    5 Min Read
    Introducing the First Comprehensive Healthcare Robotics Dataset and Essential Physical AI Models for Advancing Healthcare Robotics
    Introducing the First Comprehensive Healthcare Robotics Dataset and Essential Physical AI Models for Advancing Healthcare Robotics
    6 Min Read
    Creating Native Multimodal Agents with Qwen 3.5 VLM on NVIDIA GPU-Accelerated Endpoints
    Creating Native Multimodal Agents with Qwen 3.5 VLM on NVIDIA GPU-Accelerated Endpoints
    5 Min Read
    Discover SyGra Studio: Your Gateway to Exceptional Creative Solutions
    Discover SyGra Studio: Your Gateway to Exceptional Creative Solutions
    6 Min Read
  • Events
    EventsShow More
    Navigating the ESSER Cliff: Key Reasons Education Company Leaders are Attending the 2026 EdExec Summit
    Navigating the ESSER Cliff: Key Reasons Education Company Leaders are Attending the 2026 EdExec Summit
    6 Min Read
    Exploring National Robotics Week: Key Physical AI Research Breakthroughs and Essential Resources
    Exploring National Robotics Week: Key Physical AI Research Breakthroughs and Essential Resources
    5 Min Read
    Developing a Comprehensive Four-Part Professional Development Series on AI Education
    Developing a Comprehensive Four-Part Professional Development Series on AI Education
    6 Min Read
    NVIDIA and Thinking Machines Lab Forge Strategic Gigawatt-Scale Partnership for Long-Term Innovation
    NVIDIA and Thinking Machines Lab Forge Strategic Gigawatt-Scale Partnership for Long-Term Innovation
    5 Min Read
    ABB Robotics Utilizes NVIDIA Omniverse for Scalable Industrial-Grade Physical AI Solutions
    ABB Robotics Utilizes NVIDIA Omniverse for Scalable Industrial-Grade Physical AI Solutions
    5 Min Read
  • Ethics
    EthicsShow More
    Meta Faces Warning: Facial Recognition Glasses Could Empower Sexual Predators
    Meta Faces Warning: Facial Recognition Glasses Could Empower Sexual Predators
    5 Min Read
    How Increased Job Commodification Makes Your Role More Susceptible to AI: Insights from Online Freelancing
    How Increased Job Commodification Makes Your Role More Susceptible to AI: Insights from Online Freelancing
    6 Min Read
    Exclusive Jeff VanderMeer Story & Unreleased AI Models: The Download You Can’t Miss
    Exclusive Jeff VanderMeer Story & Unreleased AI Models: The Download You Can’t Miss
    5 Min Read
    Exploring Psychological Learning Paradigms: Their Impact on Shaping and Constraining Artificial Intelligence
    Exploring Psychological Learning Paradigms: Their Impact on Shaping and Constraining Artificial Intelligence
    4 Min Read
    Anthropic Faces Supply Chain Risk Limbo Amid Conflicting Legal Rulings
    Anthropic Faces Supply Chain Risk Limbo Amid Conflicting Legal Rulings
    6 Min Read
  • Comparisons
    ComparisonsShow More
    Enhancing Mission-Critical Small Language Models through Multi-Model Synthetic Training: Insights from Research 2509.13047
    Enhancing Mission-Critical Small Language Models through Multi-Model Synthetic Training: Insights from Research 2509.13047
    4 Min Read
    Google Launches Gemma 4: Emphasizing Local-First, On-Device AI Inference for Enhanced Performance
    Google Launches Gemma 4: Emphasizing Local-First, On-Device AI Inference for Enhanced Performance
    5 Min Read
    Overcoming Limitations of Discrete Neuronal Attribution in Neuroscience
    Overcoming Limitations of Discrete Neuronal Attribution in Neuroscience
    5 Min Read
    Optimizing Bandwidth for Cooperative Multi-Agent Reinforcement Learning: Variational Message Encoding Techniques
    Optimizing Bandwidth for Cooperative Multi-Agent Reinforcement Learning: Variational Message Encoding Techniques
    4 Min Read
    Anthropic Unveils Claude Mythos Preview Featuring Advanced Cybersecurity Features, Access Restricted for Public
    Anthropic Unveils Claude Mythos Preview Featuring Advanced Cybersecurity Features, Access Restricted for Public
    6 Min Read
Search
  • Privacy Policy
  • Terms of Service
  • Contact Us
  • FAQ / Help Center
  • Advertise With Us
  • Latest News
  • Model Comparisons
  • Tutorials & Guides
  • Open-Source Tools
  • Community Events
© 2025 AI Model Kit. All Rights Reserved.
Reading: Explore Our Open Source Build System: Streamline Your Development Process
Share
Notification Show More
Font ResizerAa
AIModelKitAIModelKit
Font ResizerAa
  • 🏠
  • 🚀
  • 📰
  • 💡
  • 📚
  • ⭐
Search
  • Home
  • News
  • Models
  • Guides
  • Tools
  • Ethics
  • Events
  • Comparisons
Follow US
  • Latest News
  • Model Comparisons
  • Tutorials & Guides
  • Open-Source Tools
  • Community Events
© 2025 AI Model Kit. All Rights Reserved.
AIModelKit > Comparisons > Explore Our Open Source Build System: Streamline Your Development Process
Comparisons

Explore Our Open Source Build System: Streamline Your Development Process

aimodelkit
Last updated: April 12, 2025 8:07 am
aimodelkit
Share
Explore Our Open Source Build System: Streamline Your Development Process
SHARE

Write a detailed, engaging, and SEO-friendly article about

Contents
  • Why rebuild Buck?
  • The user view
  • The rule author view
    • OCaml dependencies
    • C++ link dependencies
  • Try Buck2 now
  • Buck2, our new open source, large-scale build system, is now available on GitHub.
  • Buck2 is an extensible and performant build system written in Rust and designed to make your build experience faster and more efficient. 
  • In our internal tests at Meta, we observed that Buck2 completed builds 2x as fast as Buck1.

Buck2, Meta’s open source large-scale build system, is now publicly available via the Buck2 website and the Buck2 GitHub repository. While it shares some commonalities with other build systems (like Buck1 and Bazel), Buck2 is a from-scratch rewrite. Buck2 features a complete separation of the core and language-specific rules, with increased parallelism, integration with remote execution and virtual file systems, and a redesigned console output. All of these changes are aimed at helping engineers and developers spend less time waiting, and more time iterating on their code.

Thousands of developers at Meta are already using Buck2 and performing millions of builds per day, with builds completing twice as fast as with Buck1. Our own internal analysis has shown that engineers were able to produce meaningfully more code when their builds were executed by Buck2, and we hope the wider industry will also see benefits.

Why rebuild Buck?

Build systems stand between a programmer and running their code, so anything we can do to make the experience quicker or more productive directly impacts how effective a developer can be. The goal of Buck2 was to keep what we liked about Buck1 (the core concepts and workflows), draw inspiration from innovations after Buck1 (including Bazel, Adapton, and Shake), and focus on speed and enabling new experiences.

Buck2’s design is based on the following principles:

  • The core build system has no knowledge of any language-specific rules. Having the rules separated from the core means that the rules are easier to change and understand. The core of Buck2 is written in Rust, and its language rules (like how to build C++) are written in Starlark. This separation is in contrast to Buck1 (where all rules are written in the core) and Bazel (where C++/Java are written in the core).
  • The build system is powered by a single incremental dependency graph, avoiding any phases (in contrast to Buck1 or Bazel). This decision eliminates many types of bugs and increases parallelism.
  • The rules API is designed to contain advanced features for performance, along with dynamic (or monadic) dependency features for expressibility. At the same time, these features are carefully restricted to ensure other properties (for example, fast queries or hermeticity) are not harmed.
  • The open source release is almost identical to our internal version. The only pieces swapped out are the toolchains (which point at the internal copies of our compilers) and remote execution (which points at our internal servers) — both have open source alternatives supplied. We are also releasing all the rules exactly as they are used internally. Furthermore, we have separated some of the logical components into separate crates (e.g. Starlark, Superconsole, Allocative, Gazebo) so that they can be used outside Buck2.
  • Buck2 is written to integrate with remote execution, with the ability to run actions on remote machines. We use the same API as Bazel, and have been testing remote execution with Buildbarn and EngFlow. While not required (and not really expected for people starting out with the open source version), we are able to efficiently compute recursive digests and send them to remote execution efficiently.
  • Buck2 is written to integrate with virtual file systems, where the entire repository is not all checked out, but fetched on demand as the files are accessed. In particular, we support Sapling-based file systems. To integrate well, we watch for file notifications (with Watchman) and request both files and file-digests without direct file operations. The benefit is that we can make virtual file systems as fast as a full checkout, but with the benefits of much faster checkout and much lower disk usage.

The key takeaway from all these improvements is that we have designed Buck2 to be fast. In real world usage, depending on the build, Buck2 is significantly faster than Buck1. If there are no source code changes, Buck2 is almost instant on subsequent builds. If there is a lot of work to do, Buck2 starts executing faster and has greater parallelism. This increase in speed is both a consequence of many of the factors above, but also care and attention.

The user view

For end users, Buck2 works mostly the same as Buck1 (which, to a first approximation, is fairly similar to Bazel). A user defines targets in a BUCK file:

rust_binary(
    name = “my_binary”,
    srcs = [“main.rs”],
    deps = [“:my_library”],
)

A user can then build with buck2 build //:my_binary. The value main.rs is a source file, and :my_library is a dependency defined in the same BUCK file. It’s worth noting that Buck2 is mostly compatible with the BUCK files of Buck1. 

As well as the increase in speed, there are two major additional user-visible differences compared to Buck1.

First, the console output has been redesigned on top of the Superconsole library, which we specifically developed for Buck2. The console shows a few more details and feels a lot nicer to use:

Second, there is a persistent daemon that maintains a single dependency graph. When you change a BUCK file, a dependency, or a source file, we invalidate the appropriate things on the dependency graph, then request the output artifacts per the command line. In Buck1 there are multiple distinct dependency graphs, which result in phases like target graph construction, action graph construction, and then action graph execution. There are also some operations that aren’t performed on the graph. If certain things change in Buck1, then entire graphs are thrown away, rather than the minimum pieces being invalidated. With a single dependency graph, Buck2 is simpler, avoids more redundant work, and avoids explicit phases. Everything on the dependency graph has a key (how it is identified) and a value, along with a function to compute the value from the key and other related keys (following the model in the paper, “Build Systems a la Carte”).

The rule author view

While the user model follows Buck1 very closely, the approach for rules is completely different. There are lots of rules in Buck, for example rust_binary used above. While a rule in Buck1 was a Java class, baked into Buck1, a rule in Buck2 is entirely decoupled. Buck2 also ships with a “prelude” of rules that implement most of the Buck1 rules. 

Buck1 rules were tuned over time, had lots of performance optimizations and powerful features like graph traversal, but those rules were also expected to obey a lot of complex invariants—sometimes breaking those rules. For Buck2, the rule API is entirely in Starlark, which forced us to abstract those features as generically reusable APIs, aiming to make them safe, expressive, and performant—a tricky balance. We’ll touch on two such examples.

OCaml dependencies

The dependency structure of the OCaml library is hard to express in Buck1. An OCaml library consists of a number of OCaml files. These must be compiled in dependency order—so if A.ml uses B.ml, you must compile B.ml first. Bazel requires the dependency of A.ml on B.ml to be written explicitly in the BUCK file. Buck1 and Buck2 both leave that internal dependency implicit and run the tool ocamldep to infer it, which requires less maintenance as the structure changes.  What Buck1 did is run ocamldep just after parsing the BUCK file, which wasn’t really allowed, and it didn’t track dependencies, so if you changed the imports too much Buck1 gave spurious compilation failures. With Buck2, we can use the new primitive dynamic_output, which lets you run a command, read the output of the file, then wire up the rest of the graph—putting in the correct dependencies between the .ml files automatically.

C++ link dependencies

Consider the C++ linking model: To produce a library, you usually need to link together its build output, along with the transitive closure of the build output of its dependencies. If you simply duplicate the set of dependencies at each layer as you move up the graph, you end up with O(n2) memory usage. In Buck1, there was custom code in many rules to capture this pattern, relying on the ability to share Java values in memory and for the dependencies to be represented in place within the rule structure (as there was no reified dependency graph). In Buck2, there are much stronger abstraction boundaries, so such reuse has to be made more explicit. Therefore, we introduced transitive-sets (tsets) to capture this pattern of sets representing a transitive closure. By making tsets more abstract, we were also able to wire the tset directly into the underlying dependency graph, meaning this representation is efficient in both memory and computation time.

Try Buck2 now

We’re keen for people to give Buck2 a try, and we would be happy to hear any feedback (GitHub issues are the best way). We expect Buck2 will be most interesting to moderately sized multi-language projects. Visit the Buck2 getting started page for more information.

. Structure the content with clear, distinct paragraphs, each covering a specific aspect of the topic. Use a natural, human, and conversational tone while ensuring the article remains informative, well-organized, and easy to read. Integrate relevant keywords naturally, avoid repetition, and do not include a conclusion. Focus entirely on delivering valuable and well-structured insights.

EditTrack: Uncovering and Attributing AI-Enhanced Image Editing Techniques
Enhancing Downhole Depth Sensing and Field Validation with Data-Augmented Deep Learning Techniques
CoPE-VideoLM: Optimizing Codec Primitives for Improved Efficiency in Video Language Models
Effects of Evaluation on Language Models: Understanding Their Impact
Enhance SGLang Inference with Native NVIDIA Model Optimizer Integration for Streamlined Quantization and Deployment

Sign Up For Daily Newsletter

Get AI news first! Join our newsletter for fresh updates on open-source models.

By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Copy Link Print
Previous Article Assessing the Advancement of Large Language Models in Scientific Problem-Solving Assessing the Advancement of Large Language Models in Scientific Problem-Solving
Next Article Explore Arabic Instruction Following, AraGen Updates, and Additional Enhancements Explore Arabic Instruction Following, AraGen Updates, and Additional Enhancements

Stay Connected

XFollow
PinterestPin
TelegramFollow
LinkedInFollow

							banner							
							banner
Explore Top AI Tools Instantly
Discover, compare, and choose the best AI tools in one place. Easy search, real-time updates, and expert-picked solutions.
Browse AI Tools

Latest News

Enhancing Mission-Critical Small Language Models through Multi-Model Synthetic Training: Insights from Research 2509.13047
Enhancing Mission-Critical Small Language Models through Multi-Model Synthetic Training: Insights from Research 2509.13047
Comparisons
OpenAI Acquires AI Personal Finance Startup Hiro: What This Means for the Future
OpenAI Acquires AI Personal Finance Startup Hiro: What This Means for the Future
News
Google Launches Gemma 4: Emphasizing Local-First, On-Device AI Inference for Enhanced Performance
Google Launches Gemma 4: Emphasizing Local-First, On-Device AI Inference for Enhanced Performance
Comparisons
Master Python Continuous Integration and Deployment with GitHub Actions: Take the Real Python Quiz
Master Python Continuous Integration and Deployment with GitHub Actions: Take the Real Python Quiz
Guides
//

Leading global tech insights for 20M+ innovators

Quick Link

  • Latest News
  • Model Comparisons
  • Tutorials & Guides
  • Open-Source Tools
  • Community Events

Support

  • Privacy Policy
  • Terms of Service
  • Contact Us
  • FAQ / Help Center
  • Advertise With Us

Sign Up for Our Newsletter

Get AI news first! Join our newsletter for fresh updates on open-source models.

AIModelKitAIModelKit
Follow US
© 2025 AI Model Kit. All Rights Reserved.
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?