Deploy AI capabilities directly to your device fleet with locally running foundation models through Ollama. Your operations continue uninterrupted with local intelligence even when cloud connectivity is limited or unavailable.
Overview
This Guidance demonstrates how to leverage AWS IoT Greengrass to deploy StrandsAgents with local Small Language Models (SLMs) at the edge, enabling robust agentic operations across distributed device fleets. It helps organizations achieve critical operational requirements including low-latency processing, offline capabilities, and enhanced data confidentiality. The solution is particularly valuable for industries with stringent edge computing needs, such as robotics, automotive, oil and gas, and smart home automation. By combining cloud support with edge processing through Ollama inference engine, businesses can maintain reliable operations even with intermittent connectivity while ensuring sensitive data remains secure and local.
Benefits
Enable intelligent edge operations
Streamline fleet-wide deployments
Manage AI model distribution to thousands of edge devices through a centralized workflow using AWS IoT Greengrass and S3. Focus on developing intelligent applications while AWS services handle secure model deployment and device management.
Access comprehensive data insights
Empower edge devices to reason across multiple data sources including documentation and OPC-UA industrial systems. The orchestration capabilities of Amazon Strands Agents SDK intelligently coordinate specialized agents to deliver contextual responses to complex queries.
How it works
These technical details feature an architecture diagram to illustrate how to effectively use this solution. The architecture diagram shows the key components and their interactions, providing an overview of the architecture's structure and functionality step-by-step.
Step 1
Deploy with confidence
Everything you need to launch this Guidance in your account is right here.
Let's make it happen
Ready to deploy? Review the sample code on GitHub for detailed deployment instructions to deploy as-is or customize to fit your needs.