{"data":{"id":"23dd4143-c17c-4094-8831-67991a06f794","title":"AI Risk Report: Fast-Growing Threats in AI Runtime","summary":"Runtime attacks on large language models are rapidly increasing, with jailbreak techniques (methods that bypass AI safety restrictions) and denial-of-service exploits (attacks that make systems unavailable) becoming more sophisticated and widely shared through open-source platforms like GitHub. The report explains that these attacks have evolved from isolated research experiments into organized toolkits accessible to threat actors, affecting production AI deployments across enterprises.","solution":"N/A -- no mitigation discussed in source.","labels":["security","safety"],"sourceUrl":"https://protectai.com/blog/ai-risk-report-fast-growing-threats-in-ai-runtime","publishedAt":"2025-06-23T20:11:49.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["jailbreak","denial_of_service"],"issueType":"news","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2025-06-23T20:11:49.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety","availability"],"aiComponentTargeted":"inference","llmSpecific":true,"classifierConfidence":0.85,"researchCategory":null,"atlasIds":null}}