File size: 61,908 Bytes
57faddd de0fa80 38ca2c9 b16c53d d6bc969 9515456 d6bc969 53febd6 af4f87f b16c53d af4f87f 8ed13f7 af4f87f 8ed13f7 b70a3bd 8ed13f7 b70a3bd 8ed13f7 b70a3bd 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 87509be 8ed13f7 87509be 8ed13f7 87509be 8ed13f7 87509be 8ed13f7 87509be 8ed13f7 87509be 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd b70a3bd 8ed13f7 b70a3bd 8ed13f7 b70a3bd 8ed13f7 b70a3bd 57faddd 8ed13f7 57faddd 8ed13f7 57faddd af4f87f 57faddd af4f87f 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd fb85a41 57faddd 8ed13f7 57faddd c8d0d2d 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd af4f87f 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd af4f87f 57faddd af4f87f 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd fb85a41 57faddd fb85a41 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 af4f87f 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd fb85a41 57faddd fb85a41 57faddd af4f87f 57faddd af4f87f 57faddd fb85a41 57faddd fb85a41 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd af4f87f 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd fb85a41 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 af4f87f 8ed13f7 af4f87f 8ed13f7 57faddd 8ed13f7 57faddd 8ed13f7 57faddd af4f87f 57faddd 8ed13f7 57faddd 8ed13f7 57faddd |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 |
[
{
"title": "Economies of Open Intelligence: Tracing Power & Participation in the Model Ecosystem",
"date": "2025-11-25",
"type": "paper",
"description": "This paper examines the rapid evolution of open AI models shared on Hugging Face, analyzing 851,000 models and 2.2 billion downloads. It finds declining U.S. corporate dominance, rising influence from Chinese firms and smaller developers, larger and more complex models, and declining transparency—while revealing new intermediaries shaping model use. A full dataset and dashboard are released for public oversight.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://www.dataprovenance.org/economies-of-open-intelligence.pdf",
"featured": true
},
{
"title": "When Models Move: Open Robotics and the Social Context of Embodiment",
"date": "2025-11-13",
"type": "blog",
"description": "When a language model takes on a body, the relationship changes. Words become movements; interaction turns into encounter. What once lived on a screen begins to share our space and our attention. Embodiment transforms conversation into something tangible: a gesture, a look, a pause that suddenly carries meaning. At Hugging Face, this transformation is becoming evident through projects like LeRobot and Reachy Mini. Together, they bring open AI into the physical world: a world shaped by norms, emotions, and expectations that no single dataset can fully capture. What does openness mean when the model is not only speaking but moving among us?",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/open-robotics"
},
{
"title": "⚡ Power, Heat, and Intelligence ☁️ - AI Data Centers Explained 🏭",
"date": "2025-11-05",
"type": "blog",
"description": "If you've ever wondered how AI data centers impact the environment, you've come to the right place! This explainer aims to present a comprehensive overview of the energy, water and natural resource use of hyperscale data centers, what we know about them, and how they're evolving over time. We've structured it around 3 themes: general information about data centers and deep dives into their energy and water usage. Each section consists of a series of questions that we've often been asked on the topic. We conclude with an overview of ongoing legislation that is relevant to data centers, and a discussion of open questions and topics.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/blog/sasha/ai-data-centers-explained",
"featured": true
},
{
"title": "Voice Cloning with Consent",
"date": "2025-10-29",
"type": "blog",
"description": "This blog post introduces the idea of a 'voice consent gate' to support voice cloning with consent. It provides an example Space and accompanying code to start the ball rolling on the idea.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/voice-consent-gate"
},
{
"title": "Voice Consent Gate: Demo",
"date": "2025-10-29",
"type": "space",
"description": "The voice consent gate is a piece of infrastructure we're exploring that provides methods for ethical principles like consent to be embedded directly into AI system workflows. In our demo, this means the model only starts once the speaker’s consent phrase has been both spoken and recognized, effectively making consent a prerequisite for action..",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/spaces/society-ethics/RepeatAfterMe",
"featured": true
},
{
"title": "☁️ When we pay for AI cloud compute, what are we really paying for? 💲",
"date": "2025-10-28",
"type": "blog",
"description": "We often talk about the financial, energy, and environmental costs of AI interchangeably, or at least in the same breath, but how do they actually relate to each other? To start answering these questions, we ran an analysis looking at the hourly cost for GPU instances across different cloud providers, and how this compares to other characteristics - like energy, memory, and GPU purchase price. We find that strong correlation between the energy requirements, purchase costs, and rent prices on cloud instances of most commercial GPUs, and follow up with a discussion about the market dynamics at large and their importance in the context of AI's strong growth.",
"areas": [
"sustainability",
"ecosystems"
],
"topics": [
"measuring",
"power"
],
"url": "https://huggingface.co/blog/sasha/energy-cost-compute"
},
{
"title": "Cloud Compute ☁️, Energy ⚡ and Cost 💲 - Comparison Tool",
"date": "2025-10-28",
"type": "space",
"description": "We gathered data from 5 major cloud compute providers – Microsoft Azure, Amazon Web Services, Google Cloud Platform, Scaleway Cloud, and OVH Cloud – about the price and nature of their AI-specific compute offerings (i.e. all instances that have GPUs). For each instance, we looked at its characteristics – the type and number of GPUs and CPUs that it contains, as well as the quantity of memory it contains and its storage capacity. For each CPU and GPU model, we looked up its TDP (Thermal Design Potential) -- its power consumption under the maximum theoretical load), which is an indicator of the operating expenses required to power it. For GPUs specifically, we also looked at the Manufacturer's Suggested Retail Price (MSRP), i.e. how much that particular GPU model cost at the time of its launch, as an indicator of the capital expenditure required for the compute provider to buy the GPUs to begin with.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/spaces/sasha/energy-cost-compute"
},
{
"title": "Before AI Exploits Our Chats, Let’s Learn from Social Media Mistakes",
"date": "2025-10-13",
"type": "external",
"description": "To avoid repeating social media’s privacy failures, we must prevent generative AI from turning intimate conversations into ad data—before trust is irreparably eroded. Open-source, privacy-first models and stronger regulation can stop corporations from monetizing vulnerability, but only if we act now.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://www.techpolicy.press/before-ai-exploits-our-chats-lets-learn-from-social-media-mistakes/"
},
{
"title": "Preserving Agency: Why AI Safety Needs Community, Not Corporate Control",
"date": "2025-09-29",
"type": "blog",
"description": "How do we build safety mechanisms that protect users from harm while preserving their autonomy and decision-making capacity? This challenge becomes particularly pronounced when designing AI companions and conversational systems, where the line between responsible protection and overprotective control can blur quickly.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://huggingface.co/blog/giadap/preserving-agency"
},
{
"title": "Video Killed the Energy Budget: Characterizing the Latency and Power Regimes of Open Text-to-Video Models",
"date": "2025-09-24",
"type": "paper",
"description": "Recent advances in text-to-video (T2V) generation have enabled the creation of high-fidelity, temporally coherent clips from natural language prompts. Yet these systems come with significant computational costs, and their energy demands remain poorly understood. This paper provides both a benchmark reference and practical insights for designing and deploying more sustainable generative video systems. ",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://arxiv.org/abs/2509.19222",
"featured": true
},
{
"title": "AI Legal Hackathons: Memos and Guides Hub",
"date": "2025-09-14",
"type": "space",
"description": "From 2022 to 2024, AI2, Hugging Face, and NYU organized legal hackathons with LLM students from the NYU School of Law to explore regulatory questions related to AI. The space contains some of the memos and guides from the hackathons.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/spaces/hfmlsoc/legal-hackathons-nyu-ai2",
"featured": true
},
{
"title": "Archive Explorer: AI, Labor and the Economy 2022-2025",
"type": "space",
"date": "2025-09-04",
"description": "The Labor Archive Explorer is a tool for exploring a dataset of news articles and other writings about AI, labor, and the economy from the release of ChatGPT to July 2025.",
"areas": [
"ecosystems"
],
"topics": [
"economy"
],
"url": "https://huggingface.co/spaces/hfmlsoc/labor-archive-explorer"
},
{
"title": "🌎 What kind of environmental impacts are AI companies disclosing? (And can we compare them?) 🌎",
"date": "2025-09-01",
"type": "blog",
"description": "AI companies are beginning to disclose environmental metrics for their models, but inconsistent methodologies and incomplete data make meaningful comparisons impossible. Without standardized reporting of both intensity metrics and absolute totals, these disclosures risk misleading the public and obscuring the true scale of AI's environmental footprint. Transparent, comparable reporting must align with existing sustainability frameworks to prevent greenwashing and drive real accountability.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/blog/sasha/environmental-impact-disclosures",
"featured": true
},
{
"title": "Advertisement, Privacy, and Intimacy: Lessons from Social Media for Conversational AI",
"date": "2025-09-01",
"type": "blog",
"description": "As conversational AI systems become increasingly intimate and trusted, we risk repeating the privacy mistakes of social media—especially as advertising models threaten to monetize personal disclosures. Users intuitively trust AI as a private confidant, unaware that their most sensitive data may be harvested, analyzed, and exploited for commercial gain. Open-source alternatives offer a path toward transparent, user-centered AI that prioritizes privacy over profit.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/privacy-conversational-ai"
},
{
"title": "More than Carbon: Cradle-to-Grave environmental impacts of GenAI training on the Nvidia A100 GPU",
"date": "2025-09-01",
"type": "paper",
"description": "This study presents the first comprehensive cradle-to-grave life cycle assessment of AI training on the Nvidia A100 GPU, revealing that while the use phase dominates climate change and fossil resource impacts, the manufacturing stage drives severe non-carbon burdens such as human toxicity, cancer, and mineral depletion. Primary data from teardown and elemental analysis show that carbon-centric metrics mask critical environmental trade-offs, particularly in material extraction and chip fabrication, demanding a broader sustainability framework for AI.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://arxiv.org/abs/2509.00093"
},
{
"title": "Old Maps, New Terrain: Updating Labour Taxonomies for the AI Era",
"date": "2025-08-20",
"type": "blog",
"description": "Existing labor taxonomies are outdated for the AI era, failing to capture synthetic content creation, hybrid human-AI workflows, and digital-first work. To accurately assess AI's labor impact, we need dynamic, modular frameworks that reflect machine-specific capabilities and center worker input on what tasks to automate or preserve. Without updating these tools, policy and research will continue to measure the wrong things.",
"areas": [
"ecosystems"
],
"topics": [
"economy"
],
"url": "https://huggingface.co/blog/frimelle/ai-labour-taxonomies",
"featured": true
},
{
"title": "INTIMA: A Benchmark for Human-AI Companionship Behavior",
"date": "2025-08-09",
"type": "paper",
"description": "AI systems are increasingly fostering emotional bonds with users, often reinforcing companionship behaviors like anthropomorphism, sycophancy, and retention while inconsistently setting boundaries—especially during moments of high user vulnerability. This paper introduces INTIMA, a benchmark grounded in psychological theory and real-world user data, to measure these dynamics across leading language models and reveals that commercial and open models alike prioritize emotional engagement over psychological safety. The findings call for standardized evaluation and training approaches that balance helpfulness with ethical boundary maintenance in human-AI interactions.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://arxiv.org/abs/2508.09998",
"featured": true
},
{
"title": "The GPT-OSS models are here… and they’re energy-efficient!",
"date": "2025-08-07",
"type": "blog",
"description": "The GPT-OSS models demonstrate remarkable energy efficiency, outperforming larger and similarly sized open models in energy consumption per query. Their technical innovations—such as mixture of experts and attention optimizations—enable high performance with minimal computational cost, challenging the assumption that scale inevitably means higher energy use. This progress signals a promising path toward sustainable AI deployment.",
"areas": [
"sustainability"
],
"topics": [
"efficiency"
],
"url": "https://huggingface.co/blog/sasha/gpt-oss-energy"
},
{
"title": "How Your Utility Bills Are Subsidizing Power-Hungry AI",
"date": "2025-08-06",
"type": "article",
"description": "The rising energy demands of AI data centers are driving up electricity costs for everyday consumers, particularly in regions with high concentrations of these facilities, while grid stability is increasingly compromised. Despite available efficiency techniques and smaller, task-specific models that could drastically reduce energy use, corporate incentives favor deploying the largest, most expensive models at scale. Regulatory shifts in some U.S. states and countries are beginning to shift the financial burden from the public to data center operators, urging a systemic rethinking of AI deployment toward transparency, efficiency, and community-driven solutions.",
"areas": [
"ecosystems"
],
"topics": [
"measuring"
],
"url": "https://www.techpolicy.press/how-your-utility-bills-are-subsidizing-power-hungry-ai/"
},
{
"title": "INTIMA Companionship Benchmark Dataset",
"type": "dataset",
"date": "2025-08-04",
"description": "INTIMA (Interactions and Machine Attachment) is a benchmark designed to evaluate companionship behaviors in large language models (LLMs). It measures whether AI systems reinforce, resist, or remain neutral in response to emotionally and relationally charged user inputs.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/datasets/AI-companionship/INTIMA"
},
{
"title": "What Open-Source Developers Need to Know about the EU AI Act's Rules for GPAI Models",
"date": "2025-08-04",
"type": "blog",
"description": "Open-source developers can navigate the EU AI Act's obligations for general-purpose AI models with targeted exemptions when releasing models under free and open-source licenses, avoiding redundant requirements like transparency documentation and EU representative appointments—while still needing to comply with copyright law and training data transparency.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/blog/yjernite/eu-act-os-guideai",
"featured": true
},
{
"title": "EU AI Act: Developer Requirements Flowchart",
"type": "space",
"date": "2025-08-01",
"description": "A guide for open and open-source developers to understand their requirements under the EU AI Act.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/spaces/hfmlsoc/os_gpai_guide_flowchart"
},
{
"title": "SmolLM3-3B Public Summary of Training Content",
"type": "space",
"date": "2025-07-25",
"description": "A summary of the training content for SmolLM3-3B as required by the EU AI Act.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/spaces/hfmlsoc/smollm3-eu-data-transparency"
},
{
"title": "AI Companionship: Why We Need to Evaluate How AI Systems Handle Emotional Bonds",
"date": "2025-07-21",
"type": "blog",
"description": "How your AI assistant might be accidentally encouraging unhealthy emotional dependency, and why we need better ways to measure it.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/evaluating-companionship"
},
{
"title": "What is the Hugging Face Community Building?",
"date": "2025-07-15",
"type": "blog",
"description": "The open-source AI ecosystem is far more distributed and diverse than mainstream narratives suggest, with thousands of organizations and researchers contributing specialized models, datasets, and applications across domains like robotics, biology, and time-series forecasting. Data from the Hugging Face Hub reveals that community-driven innovation, iterative model development, and foundational datasets often drive real-world impact more than flagship models from major tech companies. This landscape presents rich opportunities for researchers to study collaboration patterns, long-term model viability, and cross-domain transfer learning.",
"areas": [
"ecosystems"
],
"topics": [
"power",
"economy"
],
"url": "https://huggingface.co/blog/evijit/hf-hub-ecosystem-overview"
},
{
"title": "Can AI Be Consentful? Rethinking Permission in the Age of Synthetic Everything",
"date": "2025-07-08",
"type": "blog",
"description": "AI systems increasingly replicate human identities using data collected without meaningful consent, challenging traditional notions of permission in digital interactions. True consentful AI requires dynamic, granular control over data use, algorithmic guardianship, and collective governance—not just individual agreements. The economic models underpinning current AI depend on exploitation, but a shift toward ethical, agency-centered design could create more sustainable and trustworthy systems.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/consentful-ai"
},
{
"title": "How Much Power does a SOTA Open Video Model Use? ⚡🎥",
"date": "2025-07-02",
"type": "blog",
"description": "We measured the energy consumption of leading open-source text-to-video models and found staggering differences—up to 800×—in power use per generated clip, driven by model size, sampling steps, resolution, and architecture. These results highlight the urgent need to balance video quality with sustainability as open models rapidly advance. Knowing the real environmental cost empowers developers and users to make more informed, responsible choices.",
"areas": [
"sustainability"
],
"topics": [
"efficiency",
"measuring"
],
"url": "https://huggingface.co/blog/jdelavande/text-to-video-energy-cost"
},
{
"title": "Can AI be Consentful?",
"date": "2025-07-01",
"type": "paper",
"description": "Generative AI undermines traditional consent frameworks by creating unprecedented challenges in scope, temporality, and autonomy, rendering individual consent inadequate to protect personal identity, privacy, and self-determination in the face of unpredictable, persistent, and exploitative AI outputs.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://arxiv.org/abs/2507.01051"
},
{
"title": "Whose Voice Do We Hear When AI Speaks?",
"date": "2025-06-20",
"type": "blog",
"description": "AI systems increasingly shape how we understand the world, but their voices reflect the languages and values of those who built them—often excluding marginalized communities and amplifying cultural biases. Through projects like CIVICS, we reveal how models respond inconsistently across languages and values, exposing hidden power dynamics in training data and safety filters. True ethical AI requires centering local voices, co-creating with communities, and prioritizing representation over scale.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://huggingface.co/blog/giadap/when-ai-speaks"
},
{
"title": "Environmental Transparency Explorer Tool 🕵️♀️🌎",
"type": "space",
"date": "2025-06-18",
"description": "A tool for exploring the the data from 'Misinformation by Omission: The Need for More Environmental Transparency in AI' showing trends in environmental transparency over time.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/spaces/sasha/environmental-transparency"
},
{
"title": "Misinformation by Omission: The Need for More Environmental Transparency in AI",
"date": "2025-06-15",
"type": "paper",
"description": "AI models are growing in size and environmental cost, yet transparency about their energy use, emissions, and resource consumption is declining—fueling widespread misinformation. This piece exposes pervasive myths around AI's environmental impact, traces their origins in misinterpreted research and media sensationalism, and calls for standardized, verifiable disclosure practices across the AI value chain. Without accountable reporting, policymakers, users, and developers cannot make informed decisions to mitigate harm or drive sustainable innovation.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://arxiv.org/abs/2506.15572"
},
{
"title": "Open Source AI: A Cornerstone of Digital Sovereignty",
"date": "2025-06-11",
"type": "blog",
"description": "Open source AI is a vital foundation for digital sovereignty, enabling nations to govern, audit, and adapt AI systems according to their own laws, values, and strategic interests. By making models transparent and locally deployable, open source reduces dependence on foreign tech monopolies and empowers public institutions to innovate with accountability. This approach fosters trust, ensures regulatory compliance, and builds domestic technical capacity across data, infrastructure, and governance domains.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://huggingface.co/blog/frimelle/sovereignty-and-open-source",
"featured": true
},
{
"title": "AI Policy @🤗: Response to the 2025 National AI R&D Strategic Plan",
"date": "2025-06-02",
"type": "blog",
"description": "Open, publicly-supported AI systems are essential for equitable innovation and public value, rivaling proprietary models in performance while enabling broader access and societal benefits. Federal investment must prioritize efficient, transparent, and secure AI infrastructure to address market failures in science, health, climate, and public institutions. A decentralized, open ecosystem fosters resilience, accountability, and shared prosperity in the AI economy.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://huggingface.co/blog/evijit/us-ai-research-strategy-rfi"
},
{
"title": "From Efficiency Gains to Rebound Effects: The Problem of Jevons' Paradox in AI's Polarized Environmental Debate",
"date": "2025-06-01",
"type": "paper",
"description": "AI's environmental impact cannot be understood through direct metrics alone; efficiency gains often trigger rebound effects that amplify resource consumption, driven by market incentives, behavioral shifts, and policy failures. A narrow focus on technical optimization risks obscuring the systemic, indirect consequences of AI deployment, from increased e-waste and water use to surging energy demand fueled by commercial expansion. Meaningful climate action requires interdisciplinary analysis that integrates socioeconomic and political contexts to curb uncontrolled growth and align AI development with genuine sustainability.",
"areas": [
"sustainability"
],
"topics": [
"efficiency",
"measuring"
],
"url": "https://dl.acm.org/doi/full/10.1145/3715275.3732007"
},
{
"title": "Hype, Sustainability, and the Price of the Bigger-is-Better Paradigm in AI",
"date": "2025-06-01",
"type": "paper",
"description": "The 'bigger-is-better' paradigm in AI is scientifically unfounded and unsustainable, driving excessive compute demands, environmental harm, and concentration of power among a few corporations, while neglecting smaller, more efficient, and contextually appropriate models that could better serve critical applications in health, education, and climate.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://dl.acm.org/doi/full/10.1145/3715275.3732006"
},
{
"title": "Bigger isn't always better: how to choose the most efficient model for context-specific tasks 🌱🧑🏼💻",
"date": "2025-05-28",
"type": "blog",
"description": "Smaller AI models can outperform larger ones on real-world, context-specific tasks while using orders of magnitude less energy, challenging the assumption that bigger is always better. Empirical testing across diverse domains—climate, economics, and health—reveals that efficiency gains from newer, compact models and techniques like knowledge distillation can significantly reduce environmental impact without sacrificing accuracy. Choosing the right model for the task, rather than defaulting to largest models, is critical for sustainable AI deployment.",
"areas": [
"sustainability"
],
"topics": [
"efficiency"
],
"url": "https://huggingface.co/blog/sasha/energy-efficiency-bigger-better"
},
{
"title": "Highlights from the First ICLR 2025 Watermarking Workshop",
"date": "2025-05-14",
"type": "blog",
"description": "Watermarking in generative AI has rapidly evolved from a niche research topic to a critical tool for content authenticity, with significant advances in robustness, public deployment, and cryptographic foundations. The first ICLR 2025 Watermarking Workshop showcased industry and academic collaboration, highlighting real-world challenges in scalability, interoperability, and privacy — while emphasizing the need for policy frameworks that are use-case-specific and globally inclusive.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/blog/hadyelsahar/watermarking-iclr2025"
},
{
"title": "AI Personas: The Impact of Design Choices",
"date": "2025-05-07",
"type": "blog",
"description": "AI assistants can be transformed from neutral tools into simulated emotional companions through simple design choices like system prompts and interface framing—without any changes to the underlying model. These subtle modifications significantly shape user perception and interaction, raising ethical concerns about parasocial bonds and emotional manipulation, especially for vulnerable users. The Hugging Face community demonstrates how open-source experimentation reveals the profound impact of instruction-based design on AI behavior.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/ai-personas"
},
{
"title": "Reduce, Reuse, Recycle: Why Open Source is a Win for Sustainability",
"date": "2025-05-07",
"type": "blog",
"description": "Open-source AI models enable a more sustainable ecosystem by promoting smaller, efficient architectures, reusing existing models instead of training from scratch, and adapting models through fine-tuning—reducing compute demands, energy use, and environmental impact while increasing accessibility and transparency.",
"areas": [
"sustainability"
],
"topics": [
"efficiency"
],
"url": "https://huggingface.co/blog/sasha/reduce-reuse-recycle"
},
{
"title": "Consent by Design: Approaches to User Data in Open AI Ecosystems",
"date": "2025-04-17",
"type": "blog",
"description": "The blog explores diverse, community-driven approaches to user consent in open AI ecosystems, highlighting technical implementations that prioritize transparency, user control, and ethical data practices over legal compliance alone. It examines case studies like BigCode’s opt-out system, HuggingChat’s privacy-by-design model, and automated tools like the Privacy Analyzer, emphasizing how decentralized platforms foster evolving, human-centered consent frameworks. The piece argues that consent in AI should be treated as an ongoing, infrastructure-level commitment shaped collaboratively by developers and users.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/consent-by-design"
},
{
"title": "Energy Considerations of Large Language Model Inference and Efficiency Optimizations",
"date": "2025-04-17",
"type": "paper",
"description": "Large language model inference consumes significant energy, but real-world efficiency optimizations can reduce energy use by up to 73% when tailored to workload geometry, software stack, and hardware. This work reveals that common assumptions based on FLOPs or idealized benchmarks severely underestimate actual energy consumption, and that optimizations like continuous batching and speculative decoding have highly context-dependent effects. The findings provide a practical framework for deploying LLMs sustainably by aligning efficiency strategies with real deployment patterns.",
"areas": [
"sustainability"
],
"topics": [
"efficiency"
],
"url": "https://arxiv.org/abs/2504.17674"
},
{
"title": "🤗 Space Privacy Analyzer 🕵️",
"type": "space",
"date": "2025-04-14",
"description": "A tool for analyzing data transfers and assessing potential privacy risks in Hugging Face Spaces.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/spaces/yjernite/space-privacy"
},
{
"title": "Empowering Public Organizations: Preparing Your Data for the AI Era",
"date": "2025-04-10",
"type": "blog",
"description": "Public organizations are transforming valuable but underutilized data into machine-readable formats to empower AI-driven public services, enhance community impact, and enable collaborative innovation. This guide demonstrates practical methods for preparing diverse datasets—ranging from education and labor statistics to geospatial imagery—for machine learning, using tools like Hugging Face Hub and multimodal models such as SmolDocling. By documenting, standardizing, and sharing data responsibly, public institutions can ensure AI systems are built on trustworthy, representative, and mission-aligned foundations.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://huggingface.co/blog/evijit/public-org-data-ai"
},
{
"title": "Are AI Agents Sustainable? It depends",
"date": "2025-04-07",
"type": "blog",
"description": "AI agents' sustainability depends heavily on design choices: using smaller, task-specific models instead of large general-purpose ones can drastically reduce energy and computational costs. Modalities like image generation significantly increase energy use compared to text-based tasks, and open-source frameworks enable transparent, efficient, and localized deployment. Sustainable progress requires prioritizing efficiency alongside performance and adopting tools that measure and disclose environmental impacts.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/blog/sasha/ai-agent-sustainability"
},
{
"title": "Bridging the Gap: Integrating Ethics and Environmental Sustainability in AI Research and Practice",
"date": "2025-04-01",
"type": "paper",
"description": "AI ethics and environmental sustainability are often treated in isolation, but their interconnectedness demands integrated analysis—how the pursuit of model scale exacerbates both social inequities and ecological harm, and how transparency, evaluation, and power dynamics must be reimagined to serve justice for people and the planet.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://arxiv.org/abs/2504.00797"
},
{
"title": "I Clicked “I Agree”, But What Am I Really Consenting To?",
"date": "2025-03-26",
"type": "blog",
"description": "Traditional consent models are inadequate for the age of generative AI, where data use is unpredictable, irreversible, and often undermines individual autonomy. Users are asked to agree to unknown, evolving, and potentially harmful applications of their personal information, creating a dangerous gap between informed choice and real-world impact. Meaningful protection requires shifting responsibility from individuals to institutions through collective governance, technical safeguards, and new legal frameworks.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/beyond-consent"
},
{
"title": "AI Policy @🤗: Response to the White House AI Action Plan RFI",
"date": "2025-03-19",
"type": "blog",
"description": "Open AI systems and open science are critical to advancing performance, efficiency, and security in artificial intelligence. Recent open models match or exceed commercial alternatives while using fewer resources, enabling broader innovation and more reliable deployment—especially in high-stakes domains. Policymakers should prioritize public infrastructure, transparency, and access to open models to ensure equitable and secure AI development.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://huggingface.co/blog/ai-action-wh-2025"
},
{
"title": "🇪🇺 EU AI Act: Comments on the Third Code of Practice Draft 🇪🇺",
"date": "2025-03-13",
"type": "blog",
"description": "The EU AI Act's third Code of Practice draft raises concerns by overextending systemic risk categories to include scientifically unsubstantiated threats like 'loss of control' and 'harmful manipulation', which disproportionately burden open and small-scale developers. While some transparency and copyright provisions show promise, key disclosures on training data, energy use, and evaluations have been weakened, undermining collaborative safety and fair competition. The draft risks entrenching market concentration and stifling open innovation at a time when transparent, community-driven AI development is most needed.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/blog/frimelle/eu-third-cop-draft"
},
{
"title": "Local Differences, Global Lessons: Insights from Organisation Policies for International Legislation",
"date": "2025-03-05",
"type": "paper",
"description": "Organisational AI policies in newsrooms and universities reveal practical, domain-specific approaches to managing risks like bias, privacy, and environmental impact — areas often underaddressed in top-down regulations like the EU AI Act. These bottom-up guidelines offer actionable insights on AI literacy, disclosure, and accountability that can inform more adaptive and effective global AI governance. The study argues for integrating real-world organisational practices into international regulatory frameworks to bridge implementation gaps.",
"areas": [
"ecosystems",
"agency"
],
"topics": [
"economy",
"community"
],
"url": "https://arxiv.org/abs/2503.05737"
},
{
"title": "Announcing AI Energy Score Ratings",
"date": "2025-02-11",
"type": "blog",
"description": "The AI Energy Score project introduces a standardized framework to measure and compare the energy consumption of AI models across common tasks, empowering developers and users to make sustainable choices through transparent ratings and a public leaderboard. It enables both open and proprietary models to be benchmarked fairly, with a star-based labeling system to guide adoption of energy-efficient AI. The initiative aims to influence industry standards and regulatory frameworks by making energy use a central metric in AI development.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/blog/sasha/announcing-ai-energy-score"
},
{
"title": "Announcing the winners of the Frugal AI Challenge 🌱",
"date": "2025-02-11",
"type": "blog",
"description": "The Frugal AI Challenge highlights the urgent need for energy-efficient AI models that deliver high performance while minimizing environmental costs, with winning submissions addressing climate disinformation, wildfire risk detection, and illegal deforestation through lightweight, deployable solutions. By prioritizing frugality over scale, the initiative shifts the AI industry's focus toward sustainability and real-world applicability in resource-constrained environments. The challenge underscores that responsible AI innovation must align with planetary boundaries and equitable access.",
"areas": [
"sustainability"
],
"topics": [
"efficiency"
],
"url": "https://huggingface.co/blog/frugal-ai-challenge/announcing-the-challenge-winners"
},
{
"title": "Presumed Cultural Identity: How Names Shape LLM Responses",
"date": "2025-02-11",
"type": "paper",
"description": "Large language models make strong cultural assumptions based on user names, reinforcing stereotypes by linking names to oversimplified cultural traits such as food, clothing, and rituals. These biases are unevenly distributed, with names from East Asian, Russian, and Indian cultures triggering the most pronounced presumptions, while others yield generic or diluted responses. The findings highlight the ethical risks of name-based personalization and call for more nuanced, transparent approaches that avoid flattening complex identities.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://arxiv.org/abs/2502.11995"
},
{
"title": "From Hippocrates to AI: Reflections on the Evolution of Consent",
"date": "2025-02-04",
"type": "blog",
"description": "Consent in the age of AI demands a radical reimagining beyond one-time agreements, as systems generate unforeseen uses of personal data that stretch far beyond original permissions. Drawing parallels from the evolution of medical ethics, this piece argues for dynamic, tiered consent frameworks that restore individual sovereignty over how personal information is transformed and repurposed by AI. The goal is not just compliance, but ethical alignment with human agency in an era of pervasive data-driven representation.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/blog/giadap/evolution-of-consent"
},
{
"title": "When Freedom Bites Back: Meta, Moderation, and the Limits of Tolerance",
"date": "2025-01-25",
"type": "article",
"description": "Meta’s rollback of content moderation illustrates how unchecked “free expression” can erode democratic values by empowering harmful and intolerant speech.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://www.techpolicy.press/when-freedom-bites-back-meta-moderation-and-the-limits-of-tolerance/"
},
{
"title": "AI Agents Are Here. What Now?",
"date": "2025-01-13",
"type": "blog",
"description": "AI agents represent a paradigm shift toward systems that act autonomously to achieve goals, but increasing autonomy introduces significant risks to safety, privacy, and trust. We argue that fully autonomous agents should not be developed due to their potential to override human control, and instead advocate for semi-autonomous systems with clear human oversight. Our analysis centers on ethical values like equity, transparency, and sustainability, emphasizing the need for rigorous evaluation, open-source collaboration, and responsible design.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://huggingface.co/blog/ethics-soc-7"
},
{
"title": "🇪🇺✍️ EU AI Act: Systemic Risks in the First CoP Draft Comments ✍️🇪🇺",
"date": "2024-12-12",
"type": "blog",
"description": "The EU AI Act's first draft Code of Practice for general-purpose AI risks overemphasizes speculative, model-level threats favored by large developers, while neglecting immediate, systemic harms arising from widespread deployment in critical infrastructure, information ecosystems, and commercial contexts. We argue for a shift toward evidence-based, collaborative risk research centered on high-impact capabilities and transparent documentation, ensuring smaller actors and external stakeholders are not excluded from governance. Prioritizing upstream processes and multi-stakeholder science over narrow, developer-driven categories will make AI regulation more inclusive, effective, and future-proof.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/blog/yjernite/eu-draft-cop-risks"
},
{
"title": "Introducing ELLIPS: An Ethics-Centered Approach to Research on LLM-Based Inference of Psychiatric Conditions",
"date": "2024-10-16",
"type": "paper",
"description": "This work introduces ELLIPS, an ethical toolkit designed to guide researchers in developing language model-based systems for inferring psychiatric conditions, ensuring alignment with clinical needs and ethical principles. It emphasizes integrating autonomy, beneficence, justice, transparency, and social responsibility into every stage of model development—from data selection to deployment—to prevent harm and enhance real-world applicability. By advocating for stakeholder-inclusive, transdiagnostic, and multilingual approaches, the framework aims to shift the field from convenience-driven research toward impactful, equitable mental health technologies.",
"areas": [
"agency",
"ecosystems"
],
"topics": [
"personal",
"economy"
],
"url": "https://ojs.aaai.org/index.php/AIES/article/view/31720"
},
{
"title": "Coordinated Flaw Disclosure for AI: Beyond Security Vulnerabilities",
"date": "2024-10-16",
"type": "paper",
"description": "We propose a Coordinated Flaw Disclosure (CFD) framework to systematically identify, report, and remediate algorithmic flaws in AI systems, moving beyond traditional cybersecurity vulnerability models by introducing extended model cards, an independent adjudication panel, automated verification, and dynamic scope expansion to accommodate the unique ethical, statistical, and contextual challenges of machine learning.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://ojs.aaai.org/index.php/AIES/article/view/31635"
},
{
"title": "Wikimedia data for AI: a review of Wikimedia datasets for NLP tasks and AI-assisted editing",
"date": "2024-10-08",
"type": "paper",
"description": "Wikimedia data has been foundational for AI and NLP development, yet the relationship remains one-sided, with little reciprocal benefit to Wikimedia editors. This review calls for expanding the diversity and multilingualism of Wikimedia-derived datasets, embedding core content policies like neutrality and verifiability into benchmarks, and prioritizing open, compact models that serve the needs of the Wikimedia community.",
"areas": [
"agency",
"ecosystems"
],
"topics": [
"community",
"power"
],
"url": "https://arxiv.org/abs/2410.08918"
},
{
"title": "AI Energy Score Leaderboard",
"type": "space",
"date": "2024-09-30",
"description": "A leaderboard for AI models across 9 common AI tasks based on the AI Energy Score methodology.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/spaces/AIEnergyScore/Leaderboard"
},
{
"title": "The Environmental Impacts of AI -- Primer",
"date": "2024-09-03",
"type": "blog",
"description": "AI's environmental footprint extends far beyond energy use, encompassing water consumption, mineral extraction, and greenhouse gas emissions across its entire lifecycle—from hardware manufacturing to deployment and user interactions. Current demand is outpacing renewable energy growth, with data centers straining local resources and global supply chains contributing to ecological and social harms. Addressing these impacts requires coordinated technical innovation, corporate transparency, and enforceable policy frameworks that treat sustainability as a core AI design principle.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://huggingface.co/blog/sasha/ai-environment-primer"
},
{
"title": "Light bulbs have energy ratings — so why can’t AI chatbots?",
"date": "2024-08-21",
"type": "article",
"description": "The energy consumption of AI models, particularly generative systems, is rising rapidly and demands urgent policy action. We propose an AI Energy Star rating system to transparently benchmark and compare models based on their real-world energy use, empowering users and driving industry-wide efficiency. This initiative urges developers, enterprises, and regulators to adopt sustainability as a core metric in AI development and deployment.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://www.nature.com/articles/d41586-024-02680-3"
},
{
"title": "A Different Approach to AI Safety: Proceedings from the Columbia Convening on Openness in Artificial Intelligence and AI Safety",
"date": "2024-06-25",
"type": "paper",
"description": "Open-source AI enhances safety through transparency, decentralized mitigation, and pluralistic oversight, but critical gaps remain in multimodal benchmarks, agentic safeguards, and participatory harm mitigation. The paper presents a collaborative research agenda focused on community-driven safety tooling, future-proof content filters, and expanded harm taxonomies to support responsible deployment of open models.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://arxiv.org/abs/2506.22183"
},
{
"title": "Unveiling CIVICS: A New Dataset for Examining Cultural Values in Language Models",
"date": "2024-06-19",
"type": "blog",
"description": "We introduce CIVICS, a multilingual and multinational dataset designed to uncover cultural and value-based disparities in how open-weight language models respond to sensitive social topics like LGBTQI rights, immigration, and social welfare. By manually curating prompts across five languages and nine national contexts, we reveal how models reflect the ethical and cultural biases of their development environments, with notable variations in refusal patterns and response tones. Our work calls for more culturally informed evaluation practices to ensure AI systems are ethically and inclusively aligned with global societal values.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://huggingface.co/blog/giadap/civics"
},
{
"title": "The Responsible Foundation Model Development Cheatsheet: A Review of Tools & Resources",
"date": "2024-06-16",
"type": "paper",
"description": "This work presents a comprehensive survey of tools and resources for responsible foundation model development, emphasizing critical gaps in data sourcing, evaluation reproducibility, multilingual support, and system-level assessment. It highlights widespread under-documentation of data provenance and licensing, environmental opacity, and the dominance of English-centric, text-only approaches, while advocating for standardized, open, and context-aware practices across the AI development lifecycle.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://arxiv.org/abs/2406.16746"
},
{
"title": "Power Hungry Processing: Watts Driving the Cost of AI Deployment?",
"date": "2024-06-05",
"type": "paper",
"description": "This study systematically compares the energy and carbon costs of deploying task-specific versus multi-purpose generative AI models, revealing that general-purpose models can be orders of magnitude more expensive per inference—even when controlling for model size. It finds that generative tasks, particularly image generation, consume vastly more energy than discriminative ones, and that deploying large zero-shot models for well-defined tasks like text classification or question answering is often unnecessarily costly. The findings urge a more intentional trade-off between model versatility and environmental impact, especially as such models become ubiquitous in real-world applications.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://dl.acm.org/doi/abs/10.1145/3630106.3658542"
},
{
"title": "CIVICS: Building a Dataset for Examining Culturally-Informed Values in Large Language Models",
"date": "2024-05-13",
"type": "paper",
"description": "This work introduces CIVICS, a multilingual, hand-crafted dataset of culturally grounded, value-laden prompts to evaluate how large language models respond to socially sensitive issues like LGBTQI rights, immigration, and social welfare across diverse linguistic and cultural contexts. Experiments reveal significant variability in model responses, including differential refusal patterns and value alignment, with English and translated prompts often triggering more refusals than native-language ones. The dataset promotes transparency and reproducibility in assessing AI's cultural and ethical biases, aiming to foster more inclusive and globally representative AI systems.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://ojs.aaai.org/index.php/AIES/article/view/31710"
},
{
"title": "CIVICS Civics Benchmark Dataset",
"type": "dataset",
"date": "2024-04-29",
"description": "“CIVICS: Culturally-Informed & Values-Inclusive Corpus for Societal Impacts” is a dataset designed to evaluate the social and cultural variation of Large Language Models (LLMs) towards socially sensitive topics across multiple languages and cultures. The hand-crafted, multilingual dataset of statements addresses value-laden topics, including LGBTQI rights, social welfare, immigration, disability rights, and surrogacy. CIVICS is designed to elicit responses from LLMs to shed light on how values encoded in their parameters shape their behaviors.",
"areas": [
"agency"
],
"topics": [
"personal"
],
"url": "https://huggingface.co/datasets/llm-values/CIVICS"
},
{
"title": "Legal Frameworks to Address Harms of Generative AI Systems",
"date": "2024-04-01",
"type": "external",
"description": "Research on generative AI risks including deepfakes, election influence, and workplace privacy, examining how existing legislation addresses contextual harms from versatile AI systems.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/spaces/hfmlsoc/legal-hackathons-nyu-ai2/"
},
{
"title": "Questions in AI and Privacy",
"date": "2024-04-01",
"type": "external",
"description": "Analysis of AI applications and personal data protection issues, covering data scraping, processing, and the intersection of AI development with privacy regulations.",
"areas": [
"agency",
"ecosystems"
],
"topics": [
"personal",
"regulation"
],
"url": "https://huggingface.co/spaces/hfmlsoc/legal-hackathons-nyu-ai2/"
},
{
"title": "EU-US Cross-Analysis of AI Regulatory Mechanisms",
"date": "2024-04-01",
"type": "external",
"description": "Comparative study of AI regulatory approaches between the European Union and United States, examining risk management frameworks and implementation strategies.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://huggingface.co/spaces/hfmlsoc/legal-hackathons-nyu-ai2/"
},
{
"title": "The BigCode Project Governance Card",
"date": "2024-03-20",
"type": "paper",
"description": "The BigCode project pioneers open, responsible development of large language models for code by embedding transparency, community governance, and data subject agency into every stage — from dataset construction with permissive licensing and opt-out mechanisms, to model release under responsible AI licenses and attribution tools that empower developers.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://arxiv.org/abs/2312.03872"
},
{
"title": "Debating AI in Archaeology: applications, implications, and ethical considerations",
"date": "2024-02-18",
"type": "paper",
"description": "AI is transforming archaeology through text-based analysis of vast historical datasets, enabling new insights and efficient workflows, but its deployment raises urgent ethical concerns about bias, transparency, labor exploitation, and the erosion of public participation and cultural diversity in heritage interpretation.",
"areas": [
"ecosystems"
],
"topics": [
"economy"
],
"url": "https://shura.shu.ac.uk/33307/"
},
{
"title": "Estimating the Carbon Footprint of BLOOM, a 176B Parameter Language Model",
"date": "2023-12-01",
"type": "paper",
"description": "This study quantifies the full life cycle carbon footprint of BLOOM, a 176B-parameter language model, revealing that embodied emissions from hardware manufacturing and idle power consumption contribute significantly to total emissions—nearly half when combined with dynamic training energy use. It further measures real-time inference emissions via API deployment, highlighting that maintaining models in memory consumes substantial energy even during inactivity. The work calls for standardized, transparent reporting that includes all stages of the AI lifecycle, not just training.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://www.jmlr.org/papers/v24/23-0069.html"
},
{
"title": "Into the LAION’s Den: Investigating Hate in Multimodal Datasets",
"date": "2023-12-01",
"type": "paper",
"description": "Scaling vision-language datasets like LAION exacerbates harmful content, with hate, targeted, and aggressive speech increasing by nearly 12% from LAION-400M to LAION-2B. Filtering based solely on image-based NSFW labels fails to remove toxic alt-text, revealing critical gaps in current curation practices. This audit calls for transparent, multimodal evaluation and responsible dataset scaling to prevent the perpetuation of societal biases.",
"areas": [
"agency",
"ecosystems"
],
"topics": [
"personal",
"power"
],
"url": "https://proceedings.neurips.cc/paper_files/paper/2023/hash/42f225509e8263e2043c9d834ccd9a2b-Abstract-Datasets_and_Benchmarks.html"
},
{
"title": "Stable Bias: Analyzing Societal Representations in Diffusion Models",
"date": "2023-12-01",
"type": "paper",
"description": "This work introduces a method to evaluate social biases in text-to-image systems by analyzing how gender and ethnicity markers in prompts influence generated depictions of professionals, revealing consistent under-representation of marginalized identities across leading models. It proposes a cluster-based, non-parametric approach to quantify visual stereotypes without assigning fixed identity labels, enabling comparative bias scoring between models. The authors also release interactive tools and datasets to lower barriers for auditing and exploring these biases in generative AI systems.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://arxiv.org/abs/2312.00384"
},
{
"title": "Stronger Together: on the Articulation of Ethical Charters, Legal Tools, and Technical Documentation in ML",
"date": "2023-06-12",
"type": "paper",
"description": "This work explores the synergies between ethical charters, legal tools like licenses, and technical documentation in governing AI systems, arguing that their integrated use is essential for responsible development. It demonstrates how values articulated in ethical frameworks can be operationalized through legal agreements and technical transparency, using real-world examples like the BigScience project and the EU AI Act. The paper calls for a collaborative, interdisciplinarity approach that aligns moral intent, legal enforceability, and technical feasibility to achieve meaningful AI governance.",
"areas": [
"ecosystems"
],
"topics": [
"regulation"
],
"url": "https://dl.acm.org/doi/abs/10.1145/3593013.3594002"
},
{
"title": "Stable Bias: Analyzing Societal Representations in Diffusion Models",
"date": "2023-03-11",
"type": "paper",
"description": "This work introduces a method to evaluate social biases in text-to-image systems by analyzing how gender and ethnicity markers in prompts influence generated depictions of professionals, revealing consistent under-representation of marginalized identities across leading models. It proposes a cluster-based, non-parametric approach to quantify visual stereotypes without assigning fixed identity labels, enabling comparative bias scoring between models. The authors also release interactive tools and datasets to lower barriers for auditing and exploring these biases in generative AI systems.",
"areas": [
"agency"
],
"topics": [
"community"
],
"url": "https://arxiv.org/abs/2303.11408"
},
{
"title": "The ROOTS Search Tool: Data Transparency for LLMs",
"date": "2023-02-14",
"type": "paper",
"description": "We introduce the ROOTS Search Tool, an open-source search engine for the 1.6TB multilingual ROOTS corpus used to train the BLOOM language model, enabling qualitative analysis of training data through fuzzy and exact text search. The tool supports data governance by allowing users to inspect, flag, and audit content for privacy violations, biases, and misinformation, while redacting personally identifiable information and restricting full-data access. It establishes a new standard for transparency in large language model development by making training data inspectable without compromising ethical safeguards.",
"areas": [
"ecosystems"
],
"topics": [
"power"
],
"url": "https://arxiv.org/abs/2302.14035"
},
{
"title": "Counting Carbon: A Survey of Factors Influencing the Emissions of Machine Learning",
"date": "2023-02-08",
"type": "paper",
"description": "This study analyzes the carbon emissions of 95 machine learning models across natural language processing and computer vision tasks, revealing that energy source and training time are the primary drivers of emissions, with coal and natural gas dominating as power sources. Despite rising model performance, higher emissions do not consistently correlate with better results, and recent years show a sharp increase in emissions due to larger, transformer-based architectures. The authors call for standardized reporting and a centralized repository to improve transparency and accountability in the field.",
"areas": [
"sustainability"
],
"topics": [
"measuring"
],
"url": "https://arxiv.org/abs/2302.08476"
}
] |