Intel

company
Verified

AI & ML interests

None defined yet.

Recent Activity

n1ck-guo  updated a dataset about 9 hours ago
Intel/dynamic_model_information
PhillipHoward  published a dataset 1 day ago
Intel/Uncovering_LVLM_Bias
kylemin  updated a Space 2 days ago
Intel/UnlearnDiffAtk-Benchmark
View all activity

Articles

Intel's activity

ek-id 
in Intel/polite-guard 8 days ago

Add ONNX file of this model

#2 opened 8 days ago by
ek-id
bconsolvo 
updated a Space 16 days ago
daniel-de-leon 
posted an update 4 months ago
view post
Post
2406
As the rapid adoption of chat bots and QandA models continues, so do the concerns for their reliability and safety. In response to this, many state-of-the-art models are being tuned to act as Safety Guardrails to protect against malicious usage and avoid undesired, harmful output. I published a Hugging Face blog introducing a simple, proof-of-concept, RoBERTa-based LLM that my team and I finetuned to detect toxic prompt inputs into chat-style LLMs. The article explores some of the tradeoffs of fine-tuning larger decoder vs. smaller encoder models and asks the question if "simpler is better" in the arena of toxic prompt detection.

🔗 to blog: https://huggingface.co/blog/daniel-de-leon/toxic-prompt-roberta
🔗 to model: Intel/toxic-prompt-roberta
🔗 to OPEA microservice: https://github.com/opea-project/GenAIComps/tree/main/comps/guardrails/toxicity_detection

A huge thank you to my colleagues that helped contribute: @qgao007 , @mitalipo , @ashahba and Fahim Mohammad