Download the worker script using a curl command or download the files through github.
Ensure you have the following installed and running on your system:
Download this repository, extract the folder, and drag and drop infera-windows.exe
to your desktop.
Double click on the infera-windows.exe
icon on your desktop and the Infera node will boot up.
Please ensure that Ollama is running or infera-windows.exe
will not boot up.
Open up Terminal and copy and paste the following commands into your command line and press enter to install the Infera Node.
Note: The current Linux build is only for Linux + Intel chips, AMD chips will be added shortly
curl -O https://www.infera.org/scripts/infera-apple-m.sh
chmod +x ./infera-apple-m.sh
./infera-apple-m.sh
echo "alias init-infera='~/infera'" >> ~/.zshrc && source ~/.zshrc
curl -O https://www.infera.org/scripts/infera-apple-intel.sh
chmod +x ./infera-apple-intel.sh
./infera-apple-intel.sh
echo "alias init-infera='~/infera'" >> ~/.zshrc && source ~/.zshrc
curl -O https://www.infera.org/scripts/infera-linux-intel.sh
chmod +x ./infera-linux-intel.sh
./infera-linux-intel.sh
To start the node, enter the following to your terminal and the node will automatically start
init-infera
To update Infera on your computer, enter the following command into your Terminal to remove and replace the depricated version of Infera.
rm -rf ~/infera
curl -O https://www.infera.org/scripts/infera-apple-m.sh
chmod +x ./infera-apple-m.sh
./infera-apple-m.sh
rm -rf ~/infera
curl -O https://www.infera.org/scripts/infera-apple-intel.sh
chmod +x ./infera-apple-intel.sh
./infera-apple-intel.sh
rm -rf ~/infera
curl -O https://www.infera.org/scripts/infera-linux-intel.sh
chmod +x ./infera-linux-intel.sh
./infera-linux-intel.sh
To delete infera from your computer, enter the following command into your Terminal.
rm -rf ~/infera
While the node is running, you can install llama3:latest
with CLI using the following terminal
command in a second terminal window.
curl -X POST http://localhost:11025/install_new_model \
-H "Content-Type: application/json" \
-d '{"install_model_name": "llama3:latest"}'
The node should now have llama3:latest
installed and your node is now fully operational on the network.
llama3:latest
can also be replace with the following models
gemma:latest
gemma2:latest
dolphin-mistral:latest
mistral:latest
llama3:latest
llama3.1:latest
llama2-uncensored:latest
Users can also user their browser and go to localhost:11025/docs
to view the node API and see node stats and information.
Models can also be managed under the node API without having to use a terminal curl request.
Infera offers a Google Extension for users who wish to use a visual interface for accessing their hosted node.