Implement complete S2I deployment script - fully self-service

- Create deploy-s2i-complete.sh with all required resources
- Automatically applies RBAC, ConfigMap, S2I build, Service, Route
- Configures ServiceAccount, resource limits, and replicas
- Single command deployment - no additional steps needed
- Fix service routing to use correct service created by oc new-app
- Update README.md to highlight complete S2I option
- Ensure application is fully functional after deployment
This commit is contained in:
2025-10-04 09:33:01 -03:00
parent d79768d00b
commit f80b488949
2 changed files with 248 additions and 10 deletions

View File

@@ -31,7 +31,7 @@ A comprehensive tool for analyzing user workloads and resource usage in OpenShif
### 🚀 Quick Deploy (Recommended)
#### Option 1: Source-to-Image (S2I) - Fastest
#### Option 1: Source-to-Image (S2I) - Fastest & Complete
```bash
# 1. Clone the repository
git clone https://github.com/andersonid/openshift-resource-governance.git
@@ -40,8 +40,8 @@ cd openshift-resource-governance
# 2. Login to OpenShift
oc login <cluster-url>
# 3. Deploy using S2I (automatic build from Git)
./scripts/deploy-s2i.sh
# 3. Deploy using S2I (complete deployment with all resources)
./scripts/deploy-s2i-complete.sh
```
#### Option 2: Container Build (Traditional)
@@ -312,7 +312,7 @@ python -m uvicorn app.main:app --reload --host 0.0.0.0 --port 8080
./setup.sh # Initial environment setup
./scripts/build-and-push.sh # Build and push to Quay.io
./scripts/deploy-complete.sh # Complete OpenShift deployment (Container Build)
./scripts/deploy-s2i.sh # S2I deployment (Source-to-Image)
./scripts/deploy-s2i-complete.sh # Complete S2I deployment (Source-to-Image + All Resources)
./scripts/undeploy-complete.sh # Complete application removal
```
@@ -336,14 +336,19 @@ ORU Analyzer now supports **Source-to-Image (S2I)** deployment as an alternative
| **Registry Dependency** | ❌ None | ✅ Quay.io |
| **Build Control** | 🔒 Limited | 🎛️ Full Control |
### S2I Quick Start
### S2I Quick Start (Complete & Self-Service)
```bash
# Deploy using S2I
./scripts/deploy-s2i.sh
# Deploy using S2I with ALL resources automatically
./scripts/deploy-s2i-complete.sh
# Or use oc new-app
oc new-app python:3.11~https://github.com/andersonid/openshift-resource-governance.git \
--name=oru-analyzer --env=PYTHON_VERSION=3.11
# This single command creates:
# - Namespace
# - RBAC (ServiceAccount, ClusterRole, ClusterRoleBinding)
# - ConfigMap with all configurations
# - S2I Build and Deployment
# - Service and Route
# - Resource limits and requests
# - No additional commands needed!
```
For detailed S2I deployment information, see the S2I section above.

233
scripts/deploy-s2i-complete.sh Executable file
View File

@@ -0,0 +1,233 @@
#!/bin/bash
# ORU Analyzer - Complete S2I Deployment Script
# This script deploys the application with ALL required resources automatically
# No additional commands needed - completely self-service
set -e
echo "🚀 ORU Analyzer - Complete S2I Deployment"
echo "=========================================="
echo "📦 This will deploy the application with ALL required resources"
echo " - RBAC (ServiceAccount, ClusterRole, ClusterRoleBinding)"
echo " - ConfigMap with all configurations"
echo " - S2I Build and Deployment"
echo " - Service and Route"
echo " - Resource limits and requests"
echo ""
# Colors for output
RED='\033[0;31m'
GREEN='\033[0;32m'
YELLOW='\033[1;33m'
BLUE='\033[0;34m'
NC='\033[0m' # No Color
# Function to print colored output
print_status() {
echo -e "${BLUE}[INFO]${NC} $1"
}
print_success() {
echo -e "${GREEN}[SUCCESS]${NC} $1"
}
print_warning() {
echo -e "${YELLOW}[WARNING]${NC} $1"
}
print_error() {
echo -e "${RED}[ERROR]${NC} $1"
}
# Check if oc is available
if ! command -v oc &> /dev/null; then
print_error "OpenShift CLI (oc) is not installed or not in PATH"
exit 1
fi
# Check if user is logged in
if ! oc whoami &> /dev/null; then
print_error "Not logged in to OpenShift. Please run 'oc login' first"
exit 1
fi
print_success "OpenShift CLI is available and user is logged in"
# Get current directory (should be project root)
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
PROJECT_ROOT="$(dirname "$SCRIPT_DIR")"
K8S_DIR="$PROJECT_ROOT/k8s"
print_status "Project root: $PROJECT_ROOT"
print_status "K8s manifests: $K8S_DIR"
# Check if k8s directory exists
if [ ! -d "$K8S_DIR" ]; then
print_error "K8s directory not found: $K8S_DIR"
exit 1
fi
# Check if required manifest files exist
REQUIRED_FILES=("rbac.yaml" "configmap.yaml" "service.yaml" "route.yaml")
for file in "${REQUIRED_FILES[@]}"; do
if [ ! -f "$K8S_DIR/$file" ]; then
print_error "Required manifest file not found: $K8S_DIR/$file"
exit 1
fi
done
print_success "All required manifest files found"
# Step 1: Create namespace
print_status "Step 1: Creating namespace..."
if oc get namespace resource-governance &> /dev/null; then
print_warning "Namespace 'resource-governance' already exists"
else
oc new-project resource-governance
print_success "Namespace 'resource-governance' created"
fi
# Step 2: Apply RBAC
print_status "Step 2: Applying RBAC (ServiceAccount, ClusterRole, ClusterRoleBinding)..."
oc apply -f "$K8S_DIR/rbac.yaml"
print_success "RBAC applied successfully"
# Step 3: Apply ConfigMap
print_status "Step 3: Applying ConfigMap with application configurations..."
oc apply -f "$K8S_DIR/configmap.yaml"
print_success "ConfigMap applied successfully"
# Step 4: Deploy S2I application
print_status "Step 4: Deploying application using S2I..."
print_status " - Using Python 3.12 UBI9 base image"
print_status " - Building from GitHub repository"
print_status " - Configuring with ServiceAccount and ConfigMap"
# Deploy using S2I with proper configuration
oc new-app python:3.12-ubi9~https://github.com/andersonid/openshift-resource-governance.git \
--name=oru-analyzer \
--namespace=resource-governance \
--env=PYTHON_VERSION=3.12 \
--env=APP_ROOT=/app \
--env=HOST=0.0.0.0 \
--env=PORT=8080 \
--env=WORKERS=1
print_success "S2I application deployed"
# Step 5: Configure ServiceAccount and ConfigMap
print_status "Step 5: Configuring ServiceAccount and ConfigMap..."
oc patch deployment/oru-analyzer -p '{
"spec": {
"template": {
"spec": {
"serviceAccountName": "resource-governance-sa"
}
}
}
}' -n resource-governance
# Mount ConfigMap as environment variables
oc set env deployment/oru-analyzer --from=configmap/resource-governance-config -n resource-governance
print_success "ServiceAccount and ConfigMap configured"
# Step 6: Configure replicas
print_status "Step 6: Configuring replicas..."
oc scale deployment/oru-analyzer --replicas=1 -n resource-governance
print_success "Replicas configured (1 replica)"
# Step 7: Configure resources (CPU/Memory)
print_status "Step 7: Configuring resource requests and limits..."
oc patch deployment/oru-analyzer -p '{
"spec": {
"template": {
"spec": {
"containers": [{
"name": "oru-analyzer",
"resources": {
"requests": {
"cpu": "50m",
"memory": "64Mi"
},
"limits": {
"cpu": "200m",
"memory": "256Mi"
}
}
}]
}
}
}
}' -n resource-governance
print_success "Resource limits configured (CPU: 50m-200m, Memory: 64Mi-256Mi)"
# Step 8: Wait for deployment to be ready
print_status "Step 8: Waiting for deployment to be ready..."
oc rollout status deployment/oru-analyzer -n resource-governance --timeout=300s
print_success "Deployment is ready"
# Step 9: Configure Service (use the one created by oc new-app)
print_status "Step 9: Configuring Service..."
# The service is already created by oc new-app, just need to expose it
oc expose service/oru-analyzer
print_success "Service configured and exposed"
# Step 10: Get Route URL (created by oc expose)
print_status "Step 10: Getting Route URL..."
# The route is created by oc expose, no need to apply custom route
print_success "Route created by oc expose"
# Step 11: Get application URL
print_status "Step 11: Getting application URL..."
ROUTE_URL=$(oc get route oru-analyzer -o jsonpath='{.spec.host}' -n resource-governance 2>/dev/null)
if [ -z "$ROUTE_URL" ]; then
print_warning "Could not get route URL automatically"
print_status "You can get the URL manually with: oc get route -n resource-governance"
else
print_success "Application URL: https://$ROUTE_URL"
fi
# Step 12: Verify deployment
print_status "Step 12: Verifying deployment..."
print_status "Checking pod status..."
oc get pods -n resource-governance
print_status "Checking service status..."
oc get svc -n resource-governance
print_status "Checking route status..."
oc get route -n resource-governance
# Final status
echo ""
echo "🎉 DEPLOYMENT COMPLETED SUCCESSFULLY!"
echo "====================================="
echo "✅ All resources deployed:"
echo " - Namespace: resource-governance"
echo " - RBAC: ServiceAccount, ClusterRole, ClusterRoleBinding"
echo " - ConfigMap: resource-governance-config"
echo " - S2I Build: oru-analyzer"
echo " - Deployment: oru-analyzer"
echo " - Service: resource-governance-service"
echo " - Route: resource-governance-route"
echo ""
echo "🌐 Application Access:"
if [ -n "$ROUTE_URL" ]; then
echo " URL: https://$ROUTE_URL"
echo " Health: https://$ROUTE_URL/health"
echo " API: https://$ROUTE_URL/api/v1/cluster/status"
else
echo " Get URL: oc get route -n resource-governance"
fi
echo ""
echo "🔧 Management Commands:"
echo " View logs: oc logs -f deployment/oru-analyzer -n resource-governance"
echo " Check status: oc get all -n resource-governance"
echo " Restart: oc rollout restart deployment/oru-analyzer -n resource-governance"
echo ""
echo "📚 The application is now fully functional and self-service!"
echo " No additional configuration needed."