🚀 How to onboard Render compute nodes into TensorOpera?
Introduction
As announced earlier (https://github.com/rendernetwork/RNPs/blob/main/RNP-007.md), the collaboration between TensorOpera and Render aims at bringing Generative AI workloads to Render’s compute network. In particular, by integrating Render’s community of GPU owners into TensorOpera decentralized cloud, GenAI developers will now be offered new compute resources to tap into on TensorOpera GenAI platform.
We are now thrilled to move forward and complete phase 1 of this partnership by onboarding Render compute nodes into TensorOpera.
The process is very simple and each GPU provider from the Render community needs to only follow two steps as described below!
Onboarding instructions (requiring execution of only 2 commands)
The following two commands are to be executed from within the compute nodes which you would like to bind to the TensorOpera platform
1. Execute the first command to install fedml and related libraries​
sudo curl -sSf https://doc.fedml.ai/render/shell/install.sh | bash && exec $SHELL
What does the terminal output of successful execution of this step look like?
2. Execute the second command to verify installation setup, and to bind your node to TensorOpera Platform​
During this step, you'll need to input your render auth token. Your node's earnings are linked to this token, so ensuring the correct one is entered is essential; otherwise, distributing earnings accurately may not be possible.
sudo curl -sSf https://doc.fedml.ai/render/shell/setup.sh | bash
What does the terminal output of successful execution of this step look like?
You should see output like below in your node terminal​
Frequently Asked Questions?
What if my node binding failed?
Kindly refer to this documentation to ensure that your node possesses the necessary environment prerequisites: Node prerequisites for binding to FEDML Platform
How to make sure my node is successfully binded to the platform?
Verify the installation of fedml environment on your GPU server:​
sudo wget -q https://doc.fedml.ai/shell/verify_installation.sh && sudo chmod +x verify_installation.sh && bash verify_installation.sh
The output should look like below:​
âś” Miniconda is installed.
âś” fedml is installed in the fedml conda environment.
âś” Docker is installed.
âś” Redis is installed.
âś” NVIDIA Container Toolkit is installed.
âś” All components installed successfully.
If any of the above components failed to install, please execute the following command to do a hard clean of fedml environment and re-try the process from the beginning:
fedml logout; sudo pkill -9 python; sudo rm -rf ~/.fedml; redis-cli flushall; pidof python | xargs kill -9
Verify the node is successfully binded to platform:​
ps aux | grep fedml | wc -l
❌ If the output of above command is anything <10, then that means the node was not binded to the platform:
âś… Otherwise, it means node was successfully binded to the platform: