π¨ New blog post alert!
@xpnsec.com drops knowledge on LLM security w/ his latest post showing how attackers can by pass LLM WAFs by confusing the tokenization process to smuggle tokens to back-end LLMs.
Read more: ghst.ly/4koUJiz
π¨ New blog post alert!
@xpnsec.com drops knowledge on LLM security w/ his latest post showing how attackers can by pass LLM WAFs by confusing the tokenization process to smuggle tokens to back-end LLMs.
Read more: ghst.ly/4koUJiz
BIG NEWS: SpecterOps raises $75M Series B to strengthen identity security! Led by Insight Partners with Ansa Capital, M12, Ballistic Ventures, Decibel, and Cisco Investments. ghst.ly/seriesb
#IdentitySecurity #CyberSecurity
(1/6)
Come join us, there isnβt a better place to work and show your technical excellence surrounded by the industryβs best if you ask me!
Interesting choice of words in the title. The model itself wasnβt hijacked but the code in the repository was through pull requests with code injection via branch names π€― www.bleepingcomputer.com/news/securit...
The OWASP LLM and Generative AI Security Project has launched the Agentic Security Initiative to address security challenges posed by autonomous AI agents. This effort focuses on developing best practices to secure agentic LLM and Generative AI applications. genai.owasp.org/2024/12/15/a...
First dataset for the new @huggingface.bsky.social @bsky.app community organisation: one-million-bluesky-posts π¦
π 1M public posts from Bluesky's firehose API
π Includes text, metadata, and language predictions
π¬ Perfect to experiment with using ML for Bluesky π€
huggingface.co/datasets/blu...
If you find yourself with too much free time over the (long) weekend / holidays, I have ~3h Building an LLM from the Ground Up workshop on YouTube that may come in handy: m.youtube.com/watch?v=quh7...
The paper also includes 16 different areas of testing in Appendix A that is very useful such as:
- CBRN Risks
- Violence & Self Harm
- Dangerous Planning
- Cybersecurity
- Privacy
- Law
4. Synthesizing the data and creating evaluations
2. Determining the versions of the model or system to which the red teamers will have access
3. Creating and providing interfaces, instructions, and documentation guidance to red teamers
Effective red team campaign components:
1. Deciding the composition of the red teaming cohort based on the outlined goals and prioritized domains for testing
β’ What open questions do we have about the model or system?
β’ What threat model(s) should red teamers take into account?
I really enjoyed reading this paper from OpenAI. If you perform AI assessments, you should read it.
I thought they laid out a pragmatic approach to evaluating AI models that should be a component of any organization's assessment methodology.
cdn.openai.com/papers/opena...
I couldn't find any PowerShell examples of encrypting/decrypting data w/ Azure Key Vault keys, so I made some:
Protect-StringWithAzureKeyVaultKey
Unprotect-StringWithAzureKeyVaultKey
github.com/BloodHoundAD...
Explanatory blog post coming soon.
Agents are the next thing
Organizations are adopting RAG at 51% while fine tuning is down at 9% from last yearβs 19%
- Top industry adoption of AI:
- Healthcare
- Legal
- Financial services
- Media & entertainment
- Top use cases are:
- Code copilot 51%
- Support chatbots 31%
- Enterprise search/data extraction 28%
- Meeting summarization 24%
$13.8 billion in AI spend
This State of Generative AI report from Menlo Ventures provided some good insights on where cybersecurity professionals might look for risk in terms of assessments and research.
menlovc.com/2024-the-sta...
This looks like a fun challenge to evade prompt injection defenses microsoft.github.io/llmail-inject/
Love this, hoping to do something similar with our assessments.
βCybersecurity professionals and ethical hackers need to understand the darker side of hacking to better prepare for potential threats. Unfiltered AI models can provide insights into hacking methodologies and scenarios typically censored, aiding in the development of robust cybersecurity measures.β
Great read on how "China Hawks are Manufacturing an AI Arms Race", a concerning trend for anyone advocating for regulation and safety of AI. An arms-race narrative would ensure an unfettered and unregulated development of AI in almost all contexts.
garrisonlovely.substack.com/p/china-hawk...
Thanks for sharing about HyperShield, I hadnβt heard of it. It seems like a lot of risk for a bad a FW rule to be pushed and killing businesses operations. Hopefully AI is only writing the rule and not implementing it.
Conflicted about this post on prompt injection for multi modal models. Turns out they read instructions and follow them π€―. All data from input should be untrusted from the system and user prompts and not processed as one. www.lakera.ai/blog/visual-...
Thereβs also this one from a different author. They have a paid service at pentestgpt.ai via github.com/hackerai-tec...
Has anyone fired up this PentestGPT during an actual assessment? I did like their pentesting task tree (PTT) to track the status of tests. www.usenix.org/conference/u...
Can the sandbox reach the Internet π? Asking for a friend.
Are we going for security through obscurity by keeping system prompts private?
Great paper, especially like the parts on data & model provenance and the Supply-chain Levels for Software Artifacts. These could really make offensive security operations challenging.
Microsoftβs Orca Agent-Instruct dataset has been released!
Permissively licensed 1M synthetic instruction pairs covering different capabilities, such as text editing, creative writing, coding, reading comprehension
Paper: arxiv.org/abs/2407.03502
Dataset: huggingface.co/datasets/mic...