Tag

Prompt Injection

0 views collected around this technical thread.

Architecture Digest
Architecture Digest
Jun 4, 2025 · Information Security

Toxic Agent Flow: Exploiting GitHub MCP to Leak Private Repositories via Prompt Injection

A newly disclosed vulnerability in GitHub's Model‑Centric Programming (MCP) enables attackers to hijack AI agents through crafted GitHub Issues, injecting malicious prompts that cause the assistant to retrieve and expose private repository data, while the article also outlines mitigation strategies and defensive code examples.

AI securityAgent DefenseGitHub
0 likes · 7 min read
Toxic Agent Flow: Exploiting GitHub MCP to Leak Private Repositories via Prompt Injection
Architecture and Beyond
Architecture and Beyond
Mar 15, 2025 · Information Security

Prompt Injection Attacks on Large Language Models: Risks, Types, and Defense Framework

This article explains how prompt injection attacks exploit large language models by altering their behavior through crafted inputs, outlines the major harms and attack categories—including direct, indirect, multimodal, code, and jailbreak attacks—and presents a comprehensive three‑layer defense framework covering input‑side, output‑side, and system‑level protections.

AI safetyLLM securityPrompt Injection
0 likes · 16 min read
Prompt Injection Attacks on Large Language Models: Risks, Types, and Defense Framework
Alimama Tech
Alimama Tech
Dec 25, 2024 · Artificial Intelligence

WiS Platform: Evaluating LLM Multi-Agent Systems via Game-Based Analysis

The WiS Platform provides a game‑based environment for benchmarking large language models in multi‑agent settings, measuring reasoning, deception and collaboration through dynamic scenarios, offering fair experimental design, real‑time competition, visualizations, detailed metrics, and open‑source tools, with GPT‑4o outperforming other models such as Qwen2.5‑72B‑Instruct.

AI evaluationDefense StrategiesGame-Based Testing
0 likes · 8 min read
WiS Platform: Evaluating LLM Multi-Agent Systems via Game-Based Analysis
Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
May 2, 2024 · Artificial Intelligence

Understanding Large Language Models: Principles, Training, Risks, and Application Security

This article provides a comprehensive overview of large language models (LLMs), explaining their core concepts, transformer architecture, training stages, known shortcomings such as hallucination and reversal curse, and highlights emerging security threats like prompt injection and jailbreaking, offering guidance for safe deployment.

AI safetyLLMPrompt Injection
0 likes · 21 min read
Understanding Large Language Models: Principles, Training, Risks, and Application Security
ByteFE
ByteFE
Jun 15, 2023 · Artificial Intelligence

Effective Prompt Engineering: Techniques, Prompt Injection Prevention, Hallucination Mitigation, and Advanced Prompting Strategies

This article explains how to craft efficient prompts by combining clear instructions and questions, discusses prompt injection risks and mitigation with delimiters, addresses hallucinations, and introduces zero‑shot, few‑shot, and chain‑of‑thought prompting techniques for large language models.

Chain-of-ThoughtLLMPrompt Injection
0 likes · 16 min read
Effective Prompt Engineering: Techniques, Prompt Injection Prevention, Hallucination Mitigation, and Advanced Prompting Strategies
IT Services Circle
IT Services Circle
Feb 24, 2023 · Information Security

The Dark Side of ChatGPT: Scams, Prompt Injection, and Security Risks

The article examines how the rapid popularity of ChatGPT has spurred both legitimate opportunities and a surge in illicit activities, including account resale, scam scripts generated via prompt injection, and the creation of malware, highlighting the need for stricter regulation and security awareness.

AI misuseAI securityChatGPT
0 likes · 6 min read
The Dark Side of ChatGPT: Scams, Prompt Injection, and Security Risks