- Darryl Mercado
- Pranav Chiploonkar
- Danny Fraijo
- Isabella Ducey
Date: March 23, 2025
To create a new way to interpret sensor data from Biosphere 2 using locally run LLMs with secure SSL encryption.
Biosphere 2, a research facility unlike any other in the world, stands as a testament to our quest for understanding Earth's intricate and intelligent ecosystems. Inspired by Buckminster Fuller's "Spaceship Earth", it is the world’s largest controlled environment dedicated to unraveling the complexities of our planet. Now, by harnessing the awesome potential of Artificial Intelligence and the ingenuity of young engineers drawn to this university, we are poised to elevate our comprehension of these living systems to unprecedented, super-intelligent levels. This exciting collaboration marks a pathway into the future, where together we will deepen our knowledge of Earth's stewardship and even explore the possibilities of off-planet living.
B2Twin is the brand name for a new AI / XR system development strategy from Biosphere 2 and AI Core. Your goal as a hacker is to innovate ways to create a digital twin of Biosphere 2 that will help scientists restore degraded environments on earth and prepare for space travel.
From AI Core + Biosphere 2 (2025).
- Recognizing hardware constraints, we setup 3 LLMs running locally on moderately powerful laptops.
- We used the Biosphere 2 data provided to create a knowledge base for the LLMs.
- We used a well-supported open source web server for a familiar user-friendly platform (reminiscent of OpenAI's ChatGPT web-interface).
- We used SSL encryption to ensure secure communication between the LLMs and the web server.
- We intend for our platform to allow distributed computing on multiple devices, with robust Progressive Web App (PWA) support.
- Ollama Setup: Configured Ollama with the NVIDIA® llama3-chatqa 8B parameter model across multiple operating systems (Linux, Windows, macOS) to ensure cross-platform compatibility.
Note
This step is designed to run on Linux systems.
curl -fsSL https://ollama.com/install.sh | sh
ollama pull llama3-chatqa
- OpenWebUI Deployment: Established OpenWebUI within a Dockerized environment, streamlining container orchestration and deployment.
Note
This step requires an NVIDIA® GPU with CUDA® support. Refer to Open WebUI's documentation for other configurations.
docker run -d --gpus all --runtime=nvidia \
--network=host \
--restart unless-stopped \
-v open-webui:/app/backend/data \
-e OLLAMA_BASE_URL=http://127.0.0.1:11434 \
--name open-webui \
ghcr.io/open-webui/open-webui:cuda
- Tailscale Configuration: Implemented Tailscale on Linux, Windows, and macOS to facilitate secure, peer-to-peer WireGuard® connectivity.
- SSH Server Integration: Deployed SSH servers on devices hosting LLMs to enable robust remote workflows.
- Miniconda Environment: Implemented a Miniconda environment for agile, volatile testing of Python code.
- Data Parsing and Organization: Manually parsed and organized Biosphere 2 sensor CSV data, categorizing records by habitat to create a coherent knowledge base.
- Primary Relay Selection: Designated a primary relay/web server device, while configuring ancillary devices to open ports for remote access to Ollama.
- OpenWebUI Customization: Adapted OpenWebUI to our specific use case by configuring user accounts, models, and SSL encryption parameters.
- Custom Model Development: Developed habitat-specific models to curate and structure the knowledge base effectively.
- Quality-of-Life Enhancements: Integrated OpenAI Whisper SST, ElevenLabs TTS, and DuckDuckGo web search capabilities to augment user experience.
- Proof-of-Concept Code Execution: Implemented a preliminary code execution environment for data plotting and handling, demonstrating functional feasibility.
Through the integration of these diverse technologies, we achieved a robust, distributed system capable of processing and interpreting sensor data from Biosphere 2 with notable efficiency.
While the current implementation serves as a proof-of-concept, several avenues for further enhancement have been identified:
- Scalability and Throughput: Expand the distributed computing framework to support increased sensor data volumes and higher concurrent processing demands.
- Model Refinement: Optimize and fine-tune the habitat-specific models to bolster predictive accuracy and response times.
- Enhanced Automation: Develop automated data ingestion pipelines to streamline the processing of real-time sensor data, reducing manual intervention.
- Better Code Execution: Allow the LLMs to access a Jupyter Server with data access, allowing for more robust and complex data analysis using Python.
We extend our sincere gratitude to the following:
- Biosphere 2 Team: For providing the indispensable sensor data and hackathon challenge that underpins this project.
- Tool Developers: The communities behind Ollama, OpenWebUI, and Tailscale for their invaluable contributions and extensive toolsets.
- Research Community: For inspiring the integration of AI into environmental monitoring and digital twin simulations.

