# accumulo-fluo-uno **Repository Path**: mirrors_apache/accumulo-fluo-uno ## Basic Information - **Project Name**: accumulo-fluo-uno - **Description**: Apache Fluo Uno - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-07-15 - **Last Updated**: 2025-10-12 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README ![Uno][logo] --- [![ShellCheck][ci_img]][ci_link] [![Apache License][li]][ll] **Uno automates setting up [Apache Accumulo][accumulo] or [Apache Fluo][fluo] (and their dependencies) on a single machine.** Uno makes it easy for a developer to experiment with Accumulo or Fluo in a realistic environment. Uno is designed for developers who need to frequently upgrade and test their code, and do not care about preserving data. While Uno makes it easy to setup a dev stack running Fluo or Accumulo, it also makes it easy to clear your data and setup your dev stack again. To avoid inadvertent data loss, Uno should not be used in production. Checkout [Muchos] for setting up Accumulo or Fluo on multiple machines. ## Requirements Uno requires the following software to be installed on your machine. * Java - JDK 11 is required for running Fluo (with Accumulo 2.1). * wget - Needed for `fetch` command to download tarballs. * Maven - Only needed if `fetch` command builds tarball from local repo. You should also be able to [ssh to localhost without a passphrase][ssh-docs]. The following instructions can help you setup these requirements in your environment : * [Ubuntu 16.04](/docs/ubuntu1604.md) * [Ubuntu 18.04](/docs/ubuntu1804.md) * [CentOS 7](/docs/centos7.md) ## Quickstart The following commands will get you up and running with an Accumulo instance if you have satisfied the requirements mentioned above. Replace `accumulo` with `fluo` to setup a Fluo instance. ```bash git clone https://github.com/apache/fluo-uno.git cd fluo-uno ./bin/uno fetch accumulo # Fetches binary tarballs of Accumulo and its dependencies ./bin/uno setup accumulo # Sets up Accumulo and its dependencies (Hadoop & ZooKeeper) source <(./bin/uno env) # Bash-specific command that sets up current shell ``` Accumulo is now ready to use. Verify your installation by checking the [Accumulo Monitor](http://localhost:9995/) and [Hadoop NameNode](http://localhost:9870/) status pages. Note that the Accumulo shell can be accessed in one of two ways. The easiest is method is to use the `uno` command. ``` ./bin/uno ashell ``` You can also access the shell directly. The Accumulo installation is initialized using the username `root` and password `secret` (set in the `uno.conf` file). Therefore, the shell can be accessed directly using: ``` accumulo shell -u root -p secret ``` Starting with Accumulo 2.1, a Jshell session can also be used. ``` ./bin/uno jshell ``` When you're all done testing out Accumulo you can clean up: ``` ./bin/uno wipe ``` For a more complete understanding of Uno, please continue reading. ## Installation First, clone the Uno repo on a local disk with enough space to run Hadoop, Accumulo, etc: git clone https://github.com/apache/fluo-uno.git The `uno` command uses `conf/uno.conf` for its default configuration which should be sufficient for most users. Optionally, you can customize this configuration by modifying the `uno.conf` file for your environment. Inside this script the variable `UNO_HOME` defaults to the root of the Uno repository. ```bash vim conf/uno.conf ``` If you would like to avoid modifying `uno.conf` because it is managed by git, there is a second way to configure uno. If `conf/uno-local.conf` exists then it is used instead of `uno.conf`. After pulling the latest changes to Uno, a tool like meld can be used to compare `uno.conf` and `uno-local.conf`. ```bash cp conf/uno.conf conf/uno-local.conf vim conf/uno-local.conf ``` All commands are run using the `uno` script in `bin/`. Uno has a command that helps you configure your shell so that you can run commands from any directory and easily set common environment variables in your shell for Uno, Hadoop, ZooKeeper, Fluo, and Spark. Run the following command to print this shell configuration. You can also add `--paths` or `--vars` to the command below to limit output to PATH or environment variable configuration: uno env You can either copy and paste this output into your shell or add the following (with a correct path) to your ~/.bashrc automatically configure every new shell. ```bash source <(/path/to/uno/bin/uno env) ``` With `uno` script set up, you can now use it to download, configure, and run Fluo's dependencies. ## Fetch command The `uno fetch ` command fetches the tarballs of a component and its dependencies for later use by the `setup` command. By default, the `fetch` command downloads tarballs but you can configure it to build Fluo or Accumulo from a local git repo by setting `FLUO_REPO` or `ACCUMULO_REPO` in `uno.conf`. Run `uno fetch` to see a list of possible components. After the `fetch` command is run for the first time, it only needs to run again if you want to upgrade components and need to download/build the latest version. ## Setup command The `uno setup` command combines `uno install` and `uno run` into one command. It will install the downloaded tarballs to the directory set by `$INSTALL` in your `uno.conf` and run you local development cluster. The command can be run in several different ways: 1. Sets up Apache Accumulo and its dependencies of Hadoop, ZooKeeper. This starts all processes and will wipe Accumulo/Hadoop if this command was run previously. uno setup accumulo 2. Sets up Apache Fluo along with Accumulo (and its dependencies). This command will wipe your cluster. While Fluo is set up, it does not start any Fluo applications. uno setup fluo 3. For Fluo & Accumulo, you can setup the software again without wiping/setting up their underlying dependencies. You can upgrade Accumulo or Fluo by running `uno fetch` before running this command. uno setup fluo --no-deps uno setup accumulo --no-deps You can confirm that everything started by checking the monitoring pages below: * [Hadoop NameNode](http://localhost:9870/) * [Hadoop ResourceManager](http://localhost:8088/) * [Accumulo Monitor](http://localhost:9995/) If you run some tests and then want a fresh cluster, run the `setup` command again which will kill all running processes, clear any data and logs, and restart your cluster. ## Plugins Uno is focused on running Accumulo & Fluo. Optional features and service can be run using plugins. These plugins can optionally execute after the `install` or `run` commands. They are configured by setting `POST_INSTALL_PLUGINS` and `POST_RUN_PLUGINS` in `uno.conf`. ### Post install plugins These plugins can optionally execute after the `install` command for Accumulo and Fluo: * `accumulo-encryption` - Turns on Accumulo encryption * `influx-metrics` - Install and run metrics service using InfluxDB & Grafana * [Grafana](http://localhost:3000/) * [InfluxDB Admin](http://localhost:8083/) ### Post run plugins These plugins can optionally execute after the `run` command for Accumulo and Fluo: * `spark` - Install Apache Spark and start Spark's History server * [Spark HistoryServer](http://localhost:18080/) * `accumulo-proxy` - Starts an [Accumulo Proxy] which enables Accumulo clients in other languages. ## Wipe command The `uno wipe` command will kill all running processes for your local development cluster and clear all the data and logs. It does *not* delete the binary tarballs downloaded by the `fetch` command so you can use `setup` directly again in the future. If you need to reclaim the space used by the binary tarballs you'll have to manually delete them. ## Running Apache Fluo applications Before running an Apache Fluo application, it is recommended that you configure your shell using `uno env`. If this is done, many Fluo example applications (such as [Webindex] and [Phrasecount]) can be run by simply cloning their repo and executing their start scripts (which will use environment variables set in your shell by `uno env`). If you want to create your own Fluo application, you should mimic the scripts of example Fluo applications or follow the instructions starting at the [Configure a Fluo application][configure] section of the Fluo install instructions. These instructions will guide you through the process of configuring, initializing, and starting your application. [Accumulo Proxy]: https://github.com/apache/accumulo-proxy [Muchos]: https://github.com/apache/fluo-muchos [Phrasecount]: https://github.com/apache/fluo-examples/tree/main/phrasecount [Webindex]: https://github.com/apache/fluo-examples/tree/main/webindex [accumulo]: https://accumulo.apache.org/ [ci_img]: https://github.com/apache/fluo-uno/workflows/ShellCheck/badge.svg [ci_link]: https://github.com/apache/fluo-uno/actions [configure]: https://github.com/apache/fluo/blob/main/docs/install.md#configure-a-fluo-application [fluo]: https://fluo.apache.org/ [hadoop]: https://hadoop.apache.org/ [li]: https://img.shields.io/badge/license-Apache%202.0-blue.svg [ll]: https://github.com/apache/fluo-uno/blob/main/LICENSE [logo]: contrib/uno-logo.png [mirrors]: https://www.apache.org/dyn/closer.cgi [ssh-docs]: https://hadoop.apache.org/docs/r3.3.0/hadoop-project-dist/hadoop-common/SingleCluster.html#Setup_passphraseless_ssh [zookeeper]: https://zookeeper.apache.org/