#mlsecops search results

#reversing #MLSecOps #Cyber_Education "Reverse Engineering GPT", 2024. github.com/mytechnotalent… // Drawing inspiration from Andrej Karpathy’s iconic lecture, "Let’s Build GPT: From Scratch, in Code, Spelled Out", this project takes you on an immersive journey into the inner…


#AIOps #MLSecOps #Offensive_security #Red_Team_Tactics "AutoBackdoor: Automating Backdoor Attacks via LLMAgents", Nov. 2025. ]-> Code, datasets, and experimental configurations - github.com/bboylyg/Backdo… // AutoBackdoor - general framework for automating backdoor injection,…


📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025! #pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs

vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs
vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs
vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs

#DFIR #AIOps #MLSecOps #RAG_Security AI Incident Response Framework, V1.0 github.com/cosai-oasis/ws… // This guides defenders on proactively minimizing the impact of AI system exploitation. It details how to maintain auditability, resiliency, and rapid recovery even when a system…


#AIOps #MLSecOps "WAInjectBench: Benchmarking Prompt Injection Detections for Web Agents", 2025. ]-> Comprehensive benchmark for prompt injection detection in web agents - github.com/Norrrrrrr-lyn/… // we presenting the first comprehensive benchmark study on detecting prompt…


#MLSecOps MCP Tool Poisoning Attacks invariantlabs.ai/blog/mcp-secur… ]-> MCP Tool Poisoning Experiments ]-> WhatsApp MCP Exploited: Exfiltrating your message history via MCP

HackingTeam777's tweet image. #MLSecOps
MCP Tool Poisoning Attacks
invariantlabs.ai/blog/mcp-secur…
]-> MCP Tool Poisoning Experiments
]-> WhatsApp MCP Exploited: Exfiltrating your message history via MCP

Our own Dr. Mehrin Kiani is speaking this Thursday - there's still time to register for this free online event! 🗓️Date: December 7, 2023 ⏰Time: 6:00 PM - 7:30 PM EST 📌Location: Online (Registration Link: bit.ly/3NdMJCs) #ProtectAI #MLSecOps Photo cred: Tina Aprile, CMP

ProtectAICorp's tweet image. Our own Dr. Mehrin Kiani is speaking this Thursday - there's still time to register for this free online event!
🗓️Date: December 7, 2023
⏰Time: 6:00 PM - 7:30 PM EST
📌Location: Online (Registration Link: bit.ly/3NdMJCs)
#ProtectAI #MLSecOps
Photo cred: Tina Aprile, CMP

#MLSecOps "CrossGuard: Safeguarding MLLMs against Joint-Modal Implicit Malicious Attacks", Oct. 2025. ]-> github.com/AI45Lab/MLLMGu… // We propose ImpForge, an automated red-teaming pipeline that leverages reinforcement learning with tailored reward modules to generate diverse…

github.com

GitHub - AI45Lab/MLLMGuard

Contribute to AI45Lab/MLLMGuard development by creating an account on GitHub.


#CogSec #MLSecOps Inside OpenAI Sora 2 - Uncovering System Prompts Driving Multi-Modal LLMs mindgard.ai/resources/open… // By chaining cross-modal prompts and clever framing, researchers surfaced hidden instructions from OpenAI’s video generator


Drop by booth #2169 at #BHUSA today for hands-on demos, insightful lightning talks, and community engagement. Engage with us and learn about securing your AI, and take home cool swag you'll actually want to use! #MLSecOps #AISecurity

ProtectAICorp's tweet image. Drop by booth #2169 at #BHUSA today for hands-on demos, insightful lightning talks, and community engagement.

Engage with us and learn about securing your AI, and take home cool swag you'll actually want to use!

#MLSecOps #AISecurity

Ok. Tested my knowledge:) #MLSecOps Foundations. What a great way for @ProtectAICorp to introduce their #AIDSPM set of tools and capabilities. It was fun! I'm going to explore now what others are doing in this space.

kerendg's tweet image. Ok. Tested my knowledge:) #MLSecOps Foundations. What a great way for @ProtectAICorp to introduce their #AIDSPM set of tools and capabilities. It was fun! I'm going to explore now what others are doing in this space.

#tools #AIOps #MLSecOps Same Model, Different Hat: Bypassing OpenAI Guardrails hiddenlayer.com/innovation-hub… ]-> tools to block/detect potentially harmful model behavior - github.com/openai/openai-… // OpenAI’s Guardrails framework is a thoughtful attempt to provide developers with…


🔐 Protect AI's MLSecOps course is your ticket to AI security mastery. Secure your AI future today! #AISecurity #MLSecOps aientrepreneurs.standout.digital/p/world-leader…

EntrepreneursAI's tweet image. 🔐 Protect AI's MLSecOps course is your ticket to AI security mastery. Secure your AI future today! #AISecurity #MLSecOps

aientrepreneurs.standout.digital/p/world-leader…

#tools #MLSecOps "AMULET: a Library for Assessing Interactions Among ML Defenses and Risks", 2025. ]-> Python ML package to evaluate the susceptibility of different risks to security, privacy, and fairness - github.com/ssg-research/a… // In addition to modules for risks, AMULET…


This week, Justin and Jack are talking #AI with one of the #security industry’s most well-known experts and influencers, Diana Kelley of Protect AI. Come hear what’s new in #MLSecOps and high-integrity AI, and some well-informed predictions for the future. hubs.la/Q01_-4Vy0


Just launched: A whitepaper from the AI/ML Security Working Group 🔐 Visualizing Secure MLOps (#MLSecOps) Read the blog by @DellTech’s Sarah Evans & @Ericsson’s Andrey Shorov + download the full guide: 🔗 hubs.la/Q03Bkym60

openssf's tweet image. Just launched: A whitepaper from the AI/ML Security Working Group
🔐 Visualizing Secure MLOps (#MLSecOps)
Read the blog by @DellTech’s Sarah Evans & @Ericsson’s Andrey Shorov + download the full guide:
🔗 hubs.la/Q03Bkym60

#AIOps #MLSecOps #Offensive_security #Red_Team_Tactics "AutoBackdoor: Automating Backdoor Attacks via LLMAgents", Nov. 2025. ]-> Code, datasets, and experimental configurations - github.com/bboylyg/Backdo… // AutoBackdoor - general framework for automating backdoor injection,…


#MLSecOps #Whitepaper "Automating Generative AI Guidelines: Reducing Prompt Injection Risk with "Shift-Left" MITRE ATLAS Mitigation Testing", Sept. 2025. ]-> LLMSecOps Research (Repo) - github.com/lightbroker/ll… // Automated testing during the build stage of the AI engineering…


#AIOps #MLSecOps #RAG_Security #Offensive_security AI pentest scoping playbook devansh.bearblog.dev/ai-pentest-sco… // Scoping AI security engagements is harder than traditional pentests because the attack surface is larger, the risks are novel, and the methodologies are still maturing


#CogSec #MLSecOps Inside OpenAI Sora 2 - Uncovering System Prompts Driving Multi-Modal LLMs mindgard.ai/resources/open… // By chaining cross-modal prompts and clever framing, researchers surfaced hidden instructions from OpenAI’s video generator


#reversing #MLSecOps #Cyber_Education "Reverse Engineering GPT", 2024. github.com/mytechnotalent… // Drawing inspiration from Andrej Karpathy’s iconic lecture, "Let’s Build GPT: From Scratch, in Code, Spelled Out", this project takes you on an immersive journey into the inner…


#SCA #MLSecOps "Whisper Leak: a side-channel attack on Large Language Models", Nov. 2025. ]-> github.com/yo-yo-yo-jbo/w… // Whisper Leak - side-channel attack that infers user prompt topics from encrypted LLM traffic by analyzing packet size and timing patterns in streaming…


#DFIR #AIOps #MLSecOps #RAG_Security AI Incident Response Framework, V1.0 github.com/cosai-oasis/ws… // This guides defenders on proactively minimizing the impact of AI system exploitation. It details how to maintain auditability, resiliency, and rapid recovery even when a system…


#AIOps #MLSecOps "Beyond the Protocol: Unveiling Attack Vectors in the Model Context Protocol (MCP) Ecosystem", 2025. ]-> Repo (MCP-Artifact) - github.com/MCP-Security/M… // In this paper, we present the first end-to-end empirical evaluation of attack vectors targeting the MCP…


#MLSecOps "CrossGuard: Safeguarding MLLMs against Joint-Modal Implicit Malicious Attacks", Oct. 2025. ]-> github.com/AI45Lab/MLLMGu… // We propose ImpForge, an automated red-teaming pipeline that leverages reinforcement learning with tailored reward modules to generate diverse…

github.com

GitHub - AI45Lab/MLLMGuard

Contribute to AI45Lab/MLLMGuard development by creating an account on GitHub.


📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025! #pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs

vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs
vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs
vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs

#AIOps #MLSecOps LOLMIL: Living Off the Land Models and Inference Libraries dreadnode.io/blog/lolmil-li… // the experiment proved that autonomous malware operating without any external infrastructure is not only possible but fairly straightforward to implement. For now, this…


#tools #AIOps #MLSecOps Same Model, Different Hat: Bypassing OpenAI Guardrails hiddenlayer.com/innovation-hub… ]-> tools to block/detect potentially harmful model behavior - github.com/openai/openai-… // OpenAI’s Guardrails framework is a thoughtful attempt to provide developers with…


#MLSecOps "Fewer Weights, More Problems: A Practical Attack on LLM Pruning", 2025. ]-> Repo + threat model - github.com/eth-sri/llm-pr… // Model pruning, i.e., removing a subset of model weights, has become a prominent approach to reducing the memory footprint of LLMs during…


#AIOps #MLSecOps #Threat_Modelling "Code Agent can be an End-to-end System Hacker: Benchmarking Real-world Threats of Computer-use Agent", Oct. 2025. ]-> Dataset - huggingface.co/datasets/MomoU… ]-> Code - github.com/EddyLuo1232/VR… // We propose AdvCUA, the first benchmark aligned with…


#AIOps #MLSecOps "WAInjectBench: Benchmarking Prompt Injection Detections for Web Agents", 2025. ]-> Comprehensive benchmark for prompt injection detection in web agents - github.com/Norrrrrrr-lyn/… // we presenting the first comprehensive benchmark study on detecting prompt…


#AIOps #MLSecOps "On the Security of Tool-Invocation Prompts for LLM-Based Agentic Systems: An Empirical Risk Assessment", 2025. ]-> tipexploit.github.io ]-> Repo - github.com/TIPExploit/TIP… // TIPs are critical yet vulnerable components of LLM-based agentic systems. Our…


#MLSecOps 1. From MCP to Shell: How MCP Authentication Flaws Enable RCE in Claude Code, Gemini CLI, and More verialabs.com/blog/from-mcp-… 2. REGEXSS: Overly-greedy regex replacements can break HTML sanitisation and lead to XSS sec.stealthcopter.com/regexss


#Research #MLSecOps "A Survey on Data Security in Large Language Models", 2025. // This survey offers a comprehensive overview of the main data security risks facing LLMs and reviews current defense strategies, including adversarial training, RLHF, and data augmentation See…

github.com

GitHub - wearetyomsmnv/Awesome-LLMSecOps: LLM | Security | Operations in one github repo with good...

LLM | Security | Operations in one github repo with good links and pictures. - wearetyomsmnv/Awesome-LLMSecOps


#tools #MLSecOps "AMULET: a Library for Assessing Interactions Among ML Defenses and Risks", 2025. ]-> Python ML package to evaluate the susceptibility of different risks to security, privacy, and fairness - github.com/ssg-research/a… // In addition to modules for risks, AMULET…


#tools #AIOps #MLSecOps #Malware_analysis The Risks of Code Assistant LLMs: Harmful Content, Misuse and Deception unit42.paloaltonetworks.com/code-assistant… ]-> Picklescan - Security scanner detecting Python Pickle files performing suspicious actions - github.com/mmaitre314/pic… ]-> Model-unpickler -…


📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025! #pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs

vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs
vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs
vkemerlis's tweet image. 📢 Last week, @__wunused__ presented our work on secure deserialization of pickle-based Machine Learning (ML) models at @acm_ccs 2025!

#pickleball #mlsec #mlsecops #acm_ccs #brownssl #browncs

#MLSecOps MCP Tool Poisoning Attacks invariantlabs.ai/blog/mcp-secur… ]-> MCP Tool Poisoning Experiments ]-> WhatsApp MCP Exploited: Exfiltrating your message history via MCP

HackingTeam777's tweet image. #MLSecOps
MCP Tool Poisoning Attacks
invariantlabs.ai/blog/mcp-secur…
]-> MCP Tool Poisoning Experiments
]-> WhatsApp MCP Exploited: Exfiltrating your message history via MCP

Exciting news: We've united with @ProtectAICorp to elevate AI/ML security! Follow us down the rabbit hole to learn more. #bugbounty #mlsecops

huntr_ai's tweet image. Exciting news: We've united with @ProtectAICorp to elevate AI/ML security! Follow us down the rabbit hole to learn more.
#bugbounty #mlsecops

Traditional security can’t protect your machine learning pipeline. From data poisoning to model inversion, ML systems face unique threats. Our new whitepaper introduces MLSecOps & outlines how to build a secure, compliant ML lifecycle: bit.ly/4kFRfHz #MLSecOps #MLOps #AI

cloudsa's tweet image. Traditional security can’t protect your machine learning pipeline. From data poisoning to model inversion, ML systems face unique threats. Our new whitepaper introduces MLSecOps & outlines how to build a secure, compliant ML lifecycle: bit.ly/4kFRfHz
#MLSecOps #MLOps #AI

🎉 Today, Protect AI announced new members to its executive team. This strategic move will enhance the company's ability to meet the increasing demands for AI/ML security technologies, and expand reach. Read the full PR here- bit.ly/3MTc0Rv #protectai #mlsecops

ProtectAICorp's tweet image. 🎉 Today, Protect AI announced new members to its executive team. This strategic move will enhance the company's ability to meet the increasing demands for AI/ML security technologies, and expand reach.

Read the full PR here- bit.ly/3MTc0Rv

#protectai #mlsecops

Protect AI is Hiring! Check out our Careers Page, and learn more about one of our own- Josh Miles. bit.ly/3YlKWP1 #protectai #mlsecops #devsecops #datascience #cybersecurity #machinelearning #hiring #techjobs

ProtectAICorp's tweet image. Protect AI is Hiring! Check out our Careers Page, and learn more about one of our own- Josh Miles. bit.ly/3YlKWP1 #protectai #mlsecops #devsecops #datascience #cybersecurity #machinelearning #hiring #techjobs

😎 garak: ... into the AI Red Teaming Rabbit Hole 💥So, you want to go down the AI Red Teaming rabbit hole? 📷 Let's check out how you can leverage garak for AI Red Teaming 🧵 #genaisecurity #mlsecops #aisecurity #llmsecurity #threatdetection #redteam #airedteaming

run2obtain's tweet image. 😎 garak: ... into the AI Red Teaming Rabbit Hole 

💥So, you want to go down the AI Red Teaming rabbit hole? 📷 Let's check out how you can leverage garak for AI Red Teaming  🧵

 #genaisecurity #mlsecops #aisecurity  #llmsecurity #threatdetection #redteam #airedteaming

Exciting news from Space ISAC's AI/ML Community! 🎉 Our MLSecOps white paper is coming soon! 🚀📄 Crafted by experts from @AerospaceCorp, @RS21, and @LockheedMartin, it explores AI/ML in Space, #MLSecOps origins, and security vulnerabilities. Stay tuned! #AI #ML #SpaceSecurity

SpaceISAC's tweet image. Exciting news from Space ISAC's AI/ML Community! 🎉 

Our MLSecOps white paper is coming soon! 🚀📄 Crafted by experts from @AerospaceCorp, @RS21, and  @LockheedMartin, it explores AI/ML in Space, #MLSecOps origins, and security vulnerabilities. Stay tuned! #AI #ML #SpaceSecurity

Protect AI's CEO and Founder, @ianrswanson is at Reuters MOMENTUM discussing AI regulatory challenges, considerations, & mitigations for AI compliance needs. Book a demo today 👉 bit.ly/431SPuB #protectai #mlsecops #ai #reutersmomentum

ProtectAICorp's tweet image. Protect AI's CEO and Founder, @ianrswanson is at Reuters MOMENTUM discussing AI regulatory challenges, considerations, & mitigations for AI compliance needs. 

Book a demo today 👉 bit.ly/431SPuB 

#protectai #mlsecops #ai #reutersmomentum

Our own Dr. Mehrin Kiani is speaking this Thursday - there's still time to register for this free online event! 🗓️Date: December 7, 2023 ⏰Time: 6:00 PM - 7:30 PM EST 📌Location: Online (Registration Link: bit.ly/3NdMJCs) #ProtectAI #MLSecOps Photo cred: Tina Aprile, CMP

ProtectAICorp's tweet image. Our own Dr. Mehrin Kiani is speaking this Thursday - there's still time to register for this free online event!
🗓️Date: December 7, 2023
⏰Time: 6:00 PM - 7:30 PM EST
📌Location: Online (Registration Link: bit.ly/3NdMJCs)
#ProtectAI #MLSecOps
Photo cred: Tina Aprile, CMP

the canonical MLOps CI/CD/CT pipeline allows organizations to automate their machine learning, and boost production impacts over ad hoc development. but it also introduces novel security challenges in workflows, supply chains, and other dependencies. #MLSecOps

disesdi's tweet image. the canonical MLOps CI/CD/CT pipeline allows organizations to automate their machine learning, and boost production impacts over ad hoc development. but it also introduces novel security challenges in workflows, supply chains, and other dependencies. #MLSecOps

🚨AI Skills Gap: Demand for ML and LLM SecOps Experts Soars!🛡️ Hey everyone, Bob here! Companies are hyping up AI in all their products, but guess what? This opens up new security holes! 🔓 Organizations are now on the hunt for professionals skilled in #MLSecOps and #LLMSecOps

ElusivePrivacy's tweet image. 🚨AI Skills Gap: Demand for ML and LLM SecOps Experts Soars!🛡️

Hey everyone, Bob here! Companies are hyping up AI in all their products, but guess what? This opens up new security holes! 🔓 Organizations are now on the hunt for professionals skilled in #MLSecOps and #LLMSecOps…

❇️ Enterprise-Grade Security for Model Context Protocol ❇️ Model Context Protocol is a critical aspect of AI systems due to the central role it plays in standardizing how AI models interact with the world around them. 🧵 #genaisecurity #mlsecops #aisecuritty #aiagents

run2obtain's tweet image. ❇️ Enterprise-Grade Security for Model Context Protocol  ❇️

Model Context Protocol is a critical aspect of AI systems due to the central role it plays in standardizing how AI models interact with the world around them. 🧵

#genaisecurity #mlsecops #aisecuritty #aiagents

Just launched: A whitepaper from the AI/ML Security Working Group 🔐 Visualizing Secure MLOps (#MLSecOps) Read the blog by @DellTech’s Sarah Evans & @Ericsson’s Andrey Shorov + download the full guide: 🔗 hubs.la/Q03Bkym60

openssf's tweet image. Just launched: A whitepaper from the AI/ML Security Working Group
🔐 Visualizing Secure MLOps (#MLSecOps)
Read the blog by @DellTech’s Sarah Evans & @Ericsson’s Andrey Shorov + download the full guide:
🔗 hubs.la/Q03Bkym60

We're now growing our #MLOps offerings with @Qwak_ai! Join us on July 22 / 23 / 24 for our webinar to learn how this unification delivers 1 platform for DevSecOps & #MLSecOps. Register here: jfrog.co/3VX9dwm

jfrog's tweet image. We're now growing our #MLOps offerings with @Qwak_ai! 

Join us on July 22 / 23 / 24 for our webinar to learn how this unification delivers 1 platform for DevSecOps & #MLSecOps.

Register here: jfrog.co/3VX9dwm

🚀Announcement! We're gearing up for @finjsio Toronto on Sep. 20, 2023! Hear our President and co-founder, @DaryanD13, on the panel below, and visit our booth for demos of our intuitive and popular solutions for #MLSecOps! 📷 Register here: hubs.li/Q021Y1MV0 #FinTech #LLM

ProtectAICorp's tweet image. 🚀Announcement! We're gearing up for
@finjsio Toronto on Sep. 20, 2023! Hear our President and co-founder, @DaryanD13, on the panel below, and visit our booth for demos of our intuitive and popular solutions for #MLSecOps! 📷 
Register here: hubs.li/Q021Y1MV0 #FinTech #LLM

Explore the Intersection of ML, Security, and SecOps at SECtember.ai! Dive deep into secure ML pipelines, model governance, and continuous monitoring. bit.ly/3y0adXt #SECtemberAI #CSAI #MLSecOps

cloudsa's tweet image. Explore the Intersection of ML, Security, and SecOps at SECtember.ai! Dive deep into secure ML pipelines, model governance, and continuous monitoring.

bit.ly/3y0adXt

#SECtemberAI #CSAI #MLSecOps

Loading...

Something went wrong.


Something went wrong.


United States Trends