As machine learning (ML) models become increasingly deployed through cloud infrastructures, the confidentiality of user data during inference poses a significant security challenge. Homomorphic Encryption (HE) has emerged as a compelling cryptographic technique that enables computation on encrypted data, allowing predictions to be generated without decrypting sensitive inputs. However, the integration of HE within large scale cloud native pipelines remains constrained by high computational overhead, orchestration complexity, and model compatibility issues. This paper presents a systematic framework for the design and optimization of cloud native homomorphic encryption workflows that support privacy-preserving ML inference. The proposed architecture integrates containerized HE modules with Kubernetes-based orchestration, enabling elastic scaling and parallel encrypted computation across distributed environments. Furthermore, optimization strategies including ciphertext packing, polynomial modulus adjustment, and operator fusion are employed to minimize latency and resource consumption while preserving cryptographic integrity. Experimental results demonstrate that the proposed system achieves up to 3.2times inference acceleration and 40% reduction in memory utilization compared to conventional HE pipelines. These findings illustrate a practical pathway for deploying secure ML-as-a-Service (MLaaS) systems that guarantee data confidentiality under zero-trust cloud conditions.
翻译:随着机器学习模型日益通过云基础设施进行部署,推理过程中用户数据的机密性成为一项重要的安全挑战。同态加密作为一种极具潜力的密码学技术,能够在加密数据上直接进行计算,从而无需解密敏感输入即可生成预测结果。然而,将同态加密集成到大规模云原生流水线中仍面临计算开销高、编排复杂和模型兼容性等限制。本文提出了一种系统性的云原生同态加密工作流设计与优化框架,以支持隐私保护的机器学习推理。该架构将容器化的同态加密模块与基于Kubernetes的编排系统相结合,实现了分布式环境下的弹性伸缩与并行加密计算。此外,通过采用密文打包、多项式模数调整和算子融合等优化策略,在保证密码学完整性的同时,显著降低了延迟与资源消耗。实验结果表明,与传统同态加密流水线相比,所提系统实现了最高3.2倍的推理加速和40%的内存使用率降低。这些发现为部署安全的机器学习即服务系统提供了可行路径,确保在零信任云环境下实现数据机密性保障。