Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
Juntao Dai's picture
1 4 8

Juntao Dai

calico-1226
panjinhao0320's profile picture Gaie's profile picture 21world's profile picture
·
  • calico-1226

AI & ML interests

RLHF

Recent Activity

upvoted a collection 4 days ago
AgentDoG
updated a model 6 months ago
calico-1226/cac_aligner
published a model 6 months ago
calico-1226/cac_aligner
View all activity

Organizations

OmniSafeAI's profile picture PKU-Alignment's profile picture

upvoted a collection 4 days ago

AgentDoG

Collection
A Diagnostic Guardrail Framework for AI Agent Safety and Security • 11 items • Updated 3 days ago • 82
upvoted a collection over 1 year ago

SafeSora

Collection
Towards Safety Alignment of Text2Video Generation • 4 items • Updated Aug 15, 2024 • 2
upvoted 2 papers over 2 years ago

Safe RLHF: Safe Reinforcement Learning from Human Feedback

Paper • 2310.12773 • Published Oct 19, 2023 • 28

BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset

Paper • 2307.04657 • Published Jul 10, 2023 • 6
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs